2025-12-04T09:32:17.1581898Z Current runner version: '2.330.0' 2025-12-04T09:32:17.1589425Z Runner name: 'i-092818c7270e5db43' 2025-12-04T09:32:17.1590340Z Runner group name: 'default' 2025-12-04T09:32:17.1591415Z Machine name: 'ip-10-0-50-234' 2025-12-04T09:32:17.1594639Z ##[group]GITHUB_TOKEN Permissions 2025-12-04T09:32:17.1597764Z Contents: read 2025-12-04T09:32:17.1598469Z Metadata: read 2025-12-04T09:32:17.1599080Z ##[endgroup] 2025-12-04T09:32:17.1601641Z Secret source: Actions 2025-12-04T09:32:17.1602480Z Prepare workflow directory 2025-12-04T09:32:17.2182302Z Prepare all required actions 2025-12-04T09:32:17.2228563Z Getting action download info 2025-12-04T09:32:17.6418098Z Download action repository 'pytorch/test-infra@main' (SHA:39aa74d619174326f4e2fb0e216151c2f29d9ffd) 2025-12-04T09:32:19.8796869Z Download action repository 'pytorch/pytorch@main' (SHA:7716da9fb23f27a65b41f9f016a2afadf281c18f) 2025-12-04T09:32:34.5847459Z Download action repository 'actions/setup-python@a26af69be951a213d495a4c3e4e4022e16d87065' (SHA:a26af69be951a213d495a4c3e4e4022e16d87065) 2025-12-04T09:32:34.9422592Z Download action repository 'aws-actions/configure-aws-credentials@ececac1a45f3b08a01d2dd070d28d111c5fe6722' (SHA:ececac1a45f3b08a01d2dd070d28d111c5fe6722) 2025-12-04T09:32:35.1518272Z Download action repository 'aws-actions/amazon-ecr-login@062b18b96a7aff071d4dc91bc00c4c1a7945b076' (SHA:062b18b96a7aff071d4dc91bc00c4c1a7945b076) 2025-12-04T09:32:35.3287333Z Download action repository 'seemethere/download-artifact-s3@1da556a7aa0a088e3153970611f6c432d58e80e6' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-12-04T09:32:35.6392740Z Download action repository 'seemethere/upload-artifact-s3@baba72d0712b404f646cebe0730933554ebce96a' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-12-04T09:32:35.9429427Z Getting action download info 2025-12-04T09:32:36.0504991Z Download action repository 'actions/checkout@v4' (SHA:34e114876b0b11c390a56381ad16ebd13914f8d5) 2025-12-04T09:32:36.3285784Z Getting action download info 2025-12-04T09:32:36.4507088Z Download action repository 'nick-fields/retry@v3.0.0' (SHA:7152eba30c6575329ac0576536151aca5a72780e) 2025-12-04T09:32:36.6615908Z Getting action download info 2025-12-04T09:32:36.7857686Z Download action repository 'nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482' (SHA:3e91a01664abd3c5cd539100d10d33b9c5b68482) 2025-12-04T09:32:37.0760967Z Getting action download info 2025-12-04T09:32:37.2531012Z Uses: pytorch/pytorch/.github/workflows/_linux-test.yml@refs/heads/main (ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32) 2025-12-04T09:32:37.2535454Z ##[group] Inputs 2025-12-04T09:32:37.2535900Z build-environment: linux-jammy-cuda12.4-py3.10-gcc11 2025-12-04T09:32:37.2543492Z test-matrix: {"include": [{"config": "legacy_nvidia_driver", "shard": 1, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 1, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 2, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 2, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 3, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 3, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 4, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 4, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 5, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 5, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}]} 2025-12-04T09:32:37.2551691Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:32:37.2552717Z sync-tag: 2025-12-04T09:32:37.2553677Z timeout-minutes: 240 2025-12-04T09:32:37.2553988Z use-gha: 2025-12-04T09:32:37.2554243Z dashboard-tag: 2025-12-04T09:32:37.2554515Z s3-bucket: gha-artifacts 2025-12-04T09:32:37.2554836Z aws-role-to-assume: 2025-12-04T09:32:37.2555497Z disable-monitor: false 2025-12-04T09:32:37.2555854Z monitor-log-interval: 5 2025-12-04T09:32:37.2556227Z monitor-data-collect-interval: 1 2025-12-04T09:32:37.2556615Z ##[endgroup] 2025-12-04T09:32:37.2557379Z Complete job name: linux-jammy-cuda12.4-py3.10-gcc11 / test (legacy_nvidia_driver, 3, 5, linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check, unstable) 2025-12-04T09:32:37.3135221Z A job started hook has been configured by the self-hosted runner administrator 2025-12-04T09:32:37.3249934Z ##[group]Run '/home/ec2-user/runner-scripts/before_job.sh' 2025-12-04T09:32:37.3260682Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:32:37.3261446Z ##[endgroup] 2025-12-04T09:32:38.7535987Z Runner Type: linux.g4dn.4xlarge.nvidia.gpu 2025-12-04T09:32:38.7536554Z Instance Type: g4dn.4xlarge 2025-12-04T09:32:38.7536980Z AMI Name: unknown 2025-12-04T09:32:38.7577123Z AMI ID: ami-08982f1c5bf93d976 2025-12-04T09:32:44.7759728Z ##[group]Run pytorch/test-infra/.github/actions/setup-ssh@main 2025-12-04T09:32:44.7760249Z with: 2025-12-04T09:32:44.7760883Z github-secret: *** 2025-12-04T09:32:44.7761734Z instructions: All testing is done inside the container, to start an interactive session run: docker exec -it $(docker container ps --format '{{.ID}}') bash 2025-12-04T09:32:44.7762680Z activate-with-label: false 2025-12-04T09:32:44.7763010Z label: with-ssh 2025-12-04T09:32:44.7763284Z remove-existing-keys: true 2025-12-04T09:32:44.7763611Z fail-silently: true 2025-12-04T09:32:44.7763887Z env: 2025-12-04T09:32:44.7764119Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:32:44.7764425Z ##[endgroup] 2025-12-04T09:32:44.9263231Z Please see https://github.com/pytorch/pytorch/wiki/Debugging-using-with-ssh-for-Github-Actions for more info. 2025-12-04T09:32:44.9264900Z Not on pull request and ciflow reference could not be extracted, skipping adding ssh keys 2025-12-04T09:32:44.9621259Z ##[group]Run pytorch/pytorch/.github/actions/checkout-pytorch@main 2025-12-04T09:32:44.9621777Z with: 2025-12-04T09:32:44.9622030Z no-sudo: true 2025-12-04T09:32:44.9622300Z submodules: recursive 2025-12-04T09:32:44.9622604Z fetch-depth: 0 2025-12-04T09:32:44.9622885Z env: 2025-12-04T09:32:44.9623117Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:32:44.9623414Z ##[endgroup] 2025-12-04T09:32:44.9706925Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:32:44.9708065Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:32:44.9718624Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:32:44.9719083Z env: 2025-12-04T09:32:44.9719342Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:32:44.9719686Z ##[endgroup] 2025-12-04T09:32:44.9810215Z ##[group]Run # Use all available CPUs for fetching 2025-12-04T09:32:44.9810725Z # Use all available CPUs for fetching 2025-12-04T09:32:44.9811119Z cd "${GITHUB_WORKSPACE}" 2025-12-04T09:32:44.9811512Z git config --global fetch.parallel 0 2025-12-04T09:32:44.9812148Z git config --global submodule.fetchJobs 0 2025-12-04T09:32:44.9812547Z  2025-12-04T09:32:44.9812959Z # Clean workspace. The default checkout action should also do this, but 2025-12-04T09:32:44.9813524Z # do it here as well just in case 2025-12-04T09:32:44.9813929Z if [[ -d .git ]]; then 2025-12-04T09:32:44.9814260Z  if [ -z "${NO_SUDO}" ]; then 2025-12-04T09:32:44.9814630Z  sudo git clean -ffdx 2025-12-04T09:32:44.9814961Z  else 2025-12-04T09:32:44.9815224Z  git clean -ffdx 2025-12-04T09:32:44.9815534Z  fi 2025-12-04T09:32:44.9815782Z fi 2025-12-04T09:32:44.9822323Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:32:44.9822774Z env: 2025-12-04T09:32:44.9823106Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:32:44.9823449Z NO_SUDO: true 2025-12-04T09:32:44.9823701Z ##[endgroup] 2025-12-04T09:32:44.9951738Z ##[group]Run actions/checkout@v4 2025-12-04T09:32:44.9952108Z with: 2025-12-04T09:32:44.9952401Z ref: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:32:44.9952784Z fetch-depth: 0 2025-12-04T09:32:44.9953059Z submodules: recursive 2025-12-04T09:32:44.9953357Z show-progress: false 2025-12-04T09:32:44.9953665Z repository: pytorch/pytorch 2025-12-04T09:32:44.9954129Z token: *** 2025-12-04T09:32:44.9954385Z ssh-strict: true 2025-12-04T09:32:44.9954659Z ssh-user: git 2025-12-04T09:32:44.9954922Z persist-credentials: true 2025-12-04T09:32:44.9955260Z clean: true 2025-12-04T09:32:44.9955559Z sparse-checkout-cone-mode: true 2025-12-04T09:32:44.9955896Z fetch-tags: false 2025-12-04T09:32:44.9956163Z lfs: false 2025-12-04T09:32:44.9956427Z set-safe-directory: true 2025-12-04T09:32:44.9956733Z env: 2025-12-04T09:32:44.9956981Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:32:44.9957280Z ##[endgroup] 2025-12-04T09:32:45.1240415Z Syncing repository: pytorch/pytorch 2025-12-04T09:32:45.1241994Z ##[group]Getting Git version info 2025-12-04T09:32:45.1242571Z Working directory is '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-12-04T09:32:45.1243369Z [command]/usr/bin/git version 2025-12-04T09:32:45.1424753Z git version 2.50.1 2025-12-04T09:32:45.1453775Z ##[endgroup] 2025-12-04T09:32:45.1464634Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/97023a8e-949f-412a-9298-5e913846cf84/.gitconfig' 2025-12-04T09:32:45.1482760Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/97023a8e-949f-412a-9298-5e913846cf84' before making global git config changes 2025-12-04T09:32:45.1483964Z Adding repository directory to the temporary git global config as a safe directory 2025-12-04T09:32:45.1488027Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:32:45.1532328Z Deleting the contents of '/home/ec2-user/actions-runner/_work/pytorch/pytorch' 2025-12-04T09:32:45.1535569Z ##[group]Initializing the repository 2025-12-04T09:32:45.1540193Z [command]/usr/bin/git init /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:32:45.1599824Z hint: Using 'master' as the name for the initial branch. This default branch name 2025-12-04T09:32:45.1600860Z hint: is subject to change. To configure the initial branch name to use in all 2025-12-04T09:32:45.1601908Z hint: of your new repositories, which will suppress this warning, call: 2025-12-04T09:32:45.1602489Z hint: 2025-12-04T09:32:45.1602817Z hint: git config --global init.defaultBranch 2025-12-04T09:32:45.1603226Z hint: 2025-12-04T09:32:45.1603617Z hint: Names commonly chosen instead of 'master' are 'main', 'trunk' and 2025-12-04T09:32:45.1604312Z hint: 'development'. The just-created branch can be renamed via this command: 2025-12-04T09:32:45.1604829Z hint: 2025-12-04T09:32:45.1605081Z hint: git branch -m 2025-12-04T09:32:45.1605373Z hint: 2025-12-04T09:32:45.1605803Z hint: Disable this message with "git config set advice.defaultBranchName false" 2025-12-04T09:32:45.1609820Z Initialized empty Git repository in /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/ 2025-12-04T09:32:45.1619349Z [command]/usr/bin/git remote add origin https://github.com/pytorch/pytorch 2025-12-04T09:32:45.1655958Z ##[endgroup] 2025-12-04T09:32:45.1656465Z ##[group]Disabling automatic garbage collection 2025-12-04T09:32:45.1659537Z [command]/usr/bin/git config --local gc.auto 0 2025-12-04T09:32:45.1687279Z ##[endgroup] 2025-12-04T09:32:45.1687733Z ##[group]Setting up auth 2025-12-04T09:32:45.1693933Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-12-04T09:32:45.1722402Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-12-04T09:32:45.2057507Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-12-04T09:32:45.2085037Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-12-04T09:32:45.2417640Z [command]/usr/bin/git config --local --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T09:32:45.2447611Z [command]/usr/bin/git submodule foreach --recursive git config --local --show-origin --name-only --get-regexp remote.origin.url 2025-12-04T09:32:45.2752378Z [command]/usr/bin/git config --local http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-12-04T09:32:45.2803978Z ##[endgroup] 2025-12-04T09:32:45.2804612Z ##[group]Fetching the repository 2025-12-04T09:32:45.2812805Z [command]/usr/bin/git -c protocol.version=2 fetch --prune --no-recurse-submodules origin +refs/heads/*:refs/remotes/origin/* +refs/tags/*:refs/tags/* 2025-12-04T09:33:40.2628617Z From https://github.com/pytorch/pytorch 2025-12-04T09:33:40.2629225Z * [new branch] 2.6.0.dev20241004+ -> origin/2.6.0.dev20241004+ 2025-12-04T09:33:40.2629948Z * [new branch] 2.9.1 -> origin/2.9.1 2025-12-04T09:33:40.2630652Z * [new branch] AaronWang04_addmmfusion_perftest -> origin/AaronWang04_addmmfusion_perftest 2025-12-04T09:33:40.2631401Z * [new branch] Flamefire-patch-1 -> origin/Flamefire-patch-1 2025-12-04T09:33:40.2632131Z * [new branch] HDCharles-2.6.0-release-notes -> origin/HDCharles-2.6.0-release-notes 2025-12-04T09:33:40.2632820Z * [new branch] HOPrintFunc -> origin/HOPrintFunc 2025-12-04T09:33:40.2634467Z * [new branch] IvanKobzarev/stack/1 -> origin/IvanKobzarev/stack/1 2025-12-04T09:33:40.2636596Z * [new branch] NicoshevSVE128 -> origin/NicoshevSVE128 2025-12-04T09:33:40.2637724Z * [new branch] PR-AOTInductorNoneBug -> origin/PR-AOTInductorNoneBug 2025-12-04T09:33:40.2639383Z * [new branch] PR-AOTInductorNoneBugFix -> origin/PR-AOTInductorNoneBugFix 2025-12-04T09:33:40.2640489Z * [new branch] PR-FixConfigsIssue -> origin/PR-FixConfigsIssue 2025-12-04T09:33:40.2641724Z * [new branch] PR-NoneBugFix-viable -> origin/PR-NoneBugFix-viable 2025-12-04T09:33:40.2643166Z * [new branch] PR-ResetToZero -> origin/PR-ResetToZero 2025-12-04T09:33:40.2644433Z * [new branch] Update-Flash-Packaging -> origin/Update-Flash-Packaging 2025-12-04T09:33:40.2645812Z * [new branch] VLA_exp -> origin/VLA_exp 2025-12-04T09:33:40.2647328Z * [new branch] activation_bench -> origin/activation_bench 2025-12-04T09:33:40.2648518Z * [new branch] addmm-heuristic -> origin/addmm-heuristic 2025-12-04T09:33:40.2650413Z * [new branch] adi/onednn_aarch64 -> origin/adi/onednn_aarch64 2025-12-04T09:33:40.2651661Z * [new branch] adi/test -> origin/adi/test 2025-12-04T09:33:40.2652935Z * [new branch] adi/test_bgemm -> origin/adi/test_bgemm 2025-12-04T09:33:40.2654319Z * [new branch] adi/test_m8g -> origin/adi/test_m8g 2025-12-04T09:33:40.2655529Z * [new branch] adi/test_onednn -> origin/adi/test_onednn 2025-12-04T09:33:40.2657008Z * [new branch] adi/test_onednn_v3.9 -> origin/adi/test_onednn_v3.9 2025-12-04T09:33:40.2658316Z * [new branch] adi/test_presve_change -> origin/adi/test_presve_change 2025-12-04T09:33:40.2659480Z * [new branch] adi/test_timm -> origin/adi/test_timm 2025-12-04T09:33:40.2661188Z * [new branch] adi/testpresve_change -> origin/adi/testpresve_change 2025-12-04T09:33:40.2663466Z * [new branch] aditew01/test/vec_bf16 -> origin/aditew01/test/vec_bf16 2025-12-04T09:33:40.2664724Z * [new branch] ah-globalfeedback-hook -> origin/ah-globalfeedback-hook 2025-12-04T09:33:40.2666599Z * [new branch] albanD-patch-1 -> origin/albanD-patch-1 2025-12-04T09:33:40.2667408Z * [new branch] also-surround-shimh -> origin/also-surround-shimh 2025-12-04T09:33:40.2669340Z * [new branch] angelayi/aot_compile -> origin/angelayi/aot_compile 2025-12-04T09:33:40.2670653Z * [new branch] angelayi/aoti_additional_files -> origin/angelayi/aoti_additional_files 2025-12-04T09:33:40.2671842Z * [new branch] angelayi/benchmark -> origin/angelayi/benchmark 2025-12-04T09:33:40.2673204Z * [new branch] angelayi/change_pytree_serialization -> origin/angelayi/change_pytree_serialization 2025-12-04T09:33:40.2674362Z * [new branch] angelayi/cpp_loader -> origin/angelayi/cpp_loader 2025-12-04T09:33:40.2675664Z * [new branch] angelayi/inductor_const -> origin/angelayi/inductor_const 2025-12-04T09:33:40.2676871Z * [new branch] angelayi/lstm -> origin/angelayi/lstm 2025-12-04T09:33:40.2678672Z * [new branch] angelayi/no_so_weight -> origin/angelayi/no_so_weight 2025-12-04T09:33:40.2680308Z * [new branch] angelayi/scan_layers -> origin/angelayi/scan_layers 2025-12-04T09:33:40.2681560Z * [new branch] angelayi/side_eff -> origin/angelayi/side_eff 2025-12-04T09:33:40.2682966Z * [new branch] angelayi/state_dict -> origin/angelayi/state_dict 2025-12-04T09:33:40.2684241Z * [new branch] angelayi/symint_input -> origin/angelayi/symint_input 2025-12-04T09:33:40.2685735Z * [new branch] angelayi/symm_mem -> origin/angelayi/symm_mem 2025-12-04T09:33:40.2686892Z * [new branch] angelayi/test_cpp -> origin/angelayi/test_cpp 2025-12-04T09:33:40.2688330Z * [new branch] angelayi/torch_size -> origin/angelayi/torch_size 2025-12-04T09:33:40.2689523Z * [new branch] annotate_assert -> origin/annotate_assert 2025-12-04T09:33:40.2690955Z * [new branch] annotate_fallback_kernel -> origin/annotate_fallback_kernel 2025-12-04T09:33:40.2692127Z * [new branch] annotation_deepcopy -> origin/annotation_deepcopy 2025-12-04T09:33:40.2693398Z * [new branch] annotation_dynamo -> origin/annotation_dynamo 2025-12-04T09:33:40.2694714Z * [new branch] aot_eager_stack_trace -> origin/aot_eager_stack_trace 2025-12-04T09:33:40.2695960Z * [new branch] aoti-cuda-alloc -> origin/aoti-cuda-alloc 2025-12-04T09:33:40.2697690Z * [new branch] aoti_const_device -> origin/aoti_const_device 2025-12-04T09:33:40.2698931Z * [new branch] aoti_fqn_name_interface -> origin/aoti_fqn_name_interface 2025-12-04T09:33:40.2700208Z * [new branch] aoti_package_weights_binary -> origin/aoti_package_weights_binary 2025-12-04T09:33:40.2701435Z * [new branch] aoti_target_windows -> origin/aoti_target_windows 2025-12-04T09:33:40.2703818Z * [new branch] arsh/feat/inductor_check_profiling -> origin/arsh/feat/inductor_check_profiling 2025-12-04T09:33:40.2704951Z * [new branch] async_tp -> origin/async_tp 2025-12-04T09:33:40.2706521Z * [new branch] atalman-inductor-perf-cu124 -> origin/atalman-inductor-perf-cu124 2025-12-04T09:33:40.2707776Z * [new branch] atalman-inductor-perf-cu124.1 -> origin/atalman-inductor-perf-cu124.1 2025-12-04T09:33:40.2709102Z * [new branch] atalman-patch-2 -> origin/atalman-patch-2 2025-12-04T09:33:40.2710573Z * [new branch] atalman-patch-3 -> origin/atalman-patch-3 2025-12-04T09:33:40.2711982Z * [new branch] atalman-patch-4 -> origin/atalman-patch-4 2025-12-04T09:33:40.2713255Z * [new branch] atalman-patch-5 -> origin/atalman-patch-5 2025-12-04T09:33:40.2714730Z * [new branch] atalman-patch-6 -> origin/atalman-patch-6 2025-12-04T09:33:40.2715958Z * [new branch] atalman-patch-7 -> origin/atalman-patch-7 2025-12-04T09:33:40.2717456Z * [new branch] atalman-patch-8 -> origin/atalman-patch-8 2025-12-04T09:33:40.2718671Z * [new branch] atalman_inductor_2.3.1 -> origin/atalman_inductor_2.3.1 2025-12-04T09:33:40.2719996Z * [new branch] atalman_inductor_2.4.0 -> origin/atalman_inductor_2.4.0 2025-12-04T09:33:40.2721350Z * [new branch] atalman_inductor_2.4.x -> origin/atalman_inductor_2.4.x 2025-12-04T09:33:40.2722885Z * [new branch] attention_benchmarking_clean -> origin/attention_benchmarking_clean 2025-12-04T09:33:40.2724582Z * [new branch] bahuang/dt_fix_scalar_add -> origin/bahuang/dt_fix_scalar_add 2025-12-04T09:33:40.2725779Z * [new branch] bahuang/fix_debug_mode -> origin/bahuang/fix_debug_mode 2025-12-04T09:33:40.2727027Z * [new branch] bahuang/fix_expand -> origin/bahuang/fix_expand 2025-12-04T09:33:40.2728282Z * [new branch] bahuang/test -> origin/bahuang/test 2025-12-04T09:33:40.2730268Z * [new branch] base/1.5 -> origin/base/1.5 2025-12-04T09:33:40.2731923Z * [new branch] batching_sdpa_efficient_attention -> origin/batching_sdpa_efficient_attention 2025-12-04T09:33:40.2733054Z * [new branch] bench_scaled_mm_ops -> origin/bench_scaled_mm_ops 2025-12-04T09:33:40.2734474Z * [new branch] benchmark-updates -> origin/benchmark-updates 2025-12-04T09:33:40.2735668Z * [new branch] benchmarking-script -> origin/benchmarking-script 2025-12-04T09:33:40.2737601Z * [new branch] bertmaher/pinbump26 -> origin/bertmaher/pinbump26 2025-12-04T09:33:40.2739425Z * [new branch] bertrand/cutlass -> origin/bertrand/cutlass 2025-12-04T09:33:40.2741176Z * [new branch] bf/bug-static-input -> origin/bf/bug-static-input 2025-12-04T09:33:40.2742319Z * [new branch] bf/cg-backend -> origin/bf/cg-backend 2025-12-04T09:33:40.2743561Z * [new branch] bf/cg-nccl-test -> origin/bf/cg-nccl-test 2025-12-04T09:33:40.2744824Z * [new branch] bf/cg-remove-check -> origin/bf/cg-remove-check 2025-12-04T09:33:40.2746134Z * [new branch] bf/clean-torchbench-hf -> origin/bf/clean-torchbench-hf 2025-12-04T09:33:40.2747333Z * [new branch] bf/combo-debug-log -> origin/bf/combo-debug-log 2025-12-04T09:33:40.2748565Z * [new branch] bf/cudagraph -> origin/bf/cudagraph 2025-12-04T09:33:40.2750522Z * [new branch] bf/cudagraph-disable-input-mutation -> origin/bf/cudagraph-disable-input-mutation 2025-12-04T09:33:40.2752100Z * [new branch] bf/cudagraph-enable-input-mutation-support-benchmark -> origin/bf/cudagraph-enable-input-mutation-support-benchmark 2025-12-04T09:33:40.2753171Z * [new branch] bf/cudagraph-partition -> origin/bf/cudagraph-partition 2025-12-04T09:33:40.2754324Z * [new branch] bf/donated-buffer-bench -> origin/bf/donated-buffer-bench 2025-12-04T09:33:40.2755658Z * [new branch] bf/dynamo-partition -> origin/bf/dynamo-partition 2025-12-04T09:33:40.2756917Z * [new branch] bf/lite -> origin/bf/lite 2025-12-04T09:33:40.2758330Z * [new branch] bf/pa-non-divisible -> origin/bf/pa-non-divisible 2025-12-04T09:33:40.2759663Z * [new branch] bf/partition-cache-free-symbols -> origin/bf/partition-cache-free-symbols 2025-12-04T09:33:40.2760997Z * [new branch] bf/partition-memory-plan -> origin/bf/partition-memory-plan 2025-12-04T09:33:40.2762304Z * [new branch] bf/partition-move-cpu -> origin/bf/partition-move-cpu 2025-12-04T09:33:40.2763682Z * [new branch] bf/partition-view-fallback -> origin/bf/partition-view-fallback 2025-12-04T09:33:40.2764920Z * [new branch] bf/remove-check-55b0c39d -> origin/bf/remove-check-55b0c39d 2025-12-04T09:33:40.2766149Z * [new branch] bf/timm-nov-26-2025 -> origin/bf/timm-nov-26-2025 2025-12-04T09:33:40.2767478Z * [new branch] bf/transformer-pin-4-57-3 -> origin/bf/transformer-pin-4-57-3 2025-12-04T09:33:40.2768777Z * [new branch] bisect_perf_hf_T5_3acc6eac492 -> origin/bisect_perf_hf_T5_3acc6eac492 2025-12-04T09:33:40.2770021Z * [new branch] bisect_perf_hf_T5_3fcf66f61fb -> origin/bisect_perf_hf_T5_3fcf66f61fb 2025-12-04T09:33:40.2771798Z * [new branch] bisect_perf_hf_T5_4009d154129 -> origin/bisect_perf_hf_T5_4009d154129 2025-12-04T09:33:40.2772940Z * [new branch] bisect_perf_hf_T5_40d0740e73d -> origin/bisect_perf_hf_T5_40d0740e73d 2025-12-04T09:33:40.2774195Z * [new branch] bisect_perf_hf_T5_5268754e -> origin/bisect_perf_hf_T5_5268754e 2025-12-04T09:33:40.2775434Z * [new branch] bisect_perf_hf_T5_7d89a8d385c -> origin/bisect_perf_hf_T5_7d89a8d385c 2025-12-04T09:33:40.2776691Z * [new branch] bisect_perf_hf_T5_b7a25c1ee7c -> origin/bisect_perf_hf_T5_b7a25c1ee7c 2025-12-04T09:33:40.2778048Z * [new branch] bisect_perf_hf_T5_c25b201583f -> origin/bisect_perf_hf_T5_c25b201583f 2025-12-04T09:33:40.2779329Z * [new branch] bisect_perf_hf_T5_c93e57efac0 -> origin/bisect_perf_hf_T5_c93e57efac0 2025-12-04T09:33:40.2780760Z * [new branch] bisect_perf_hf_T5_ca9813ea149 -> origin/bisect_perf_hf_T5_ca9813ea149 2025-12-04T09:33:40.2781845Z * [new branch] bisect_perf_hf_T5_d65f194a -> origin/bisect_perf_hf_T5_d65f194a 2025-12-04T09:33:40.2783065Z * [new branch] bisect_perf_hf_T5_da94ab0b -> origin/bisect_perf_hf_T5_da94ab0b 2025-12-04T09:33:40.2784353Z * [new branch] bisect_perf_hf_T5_da94ab0b_new -> origin/bisect_perf_hf_T5_da94ab0b_new 2025-12-04T09:33:40.2785602Z * [new branch] bisect_perf_hf_T5_db4e8a1d8a8 -> origin/bisect_perf_hf_T5_db4e8a1d8a8 2025-12-04T09:33:40.2786773Z * [new branch] bisect_perf_hf_T5_e0d97e936a2 -> origin/bisect_perf_hf_T5_e0d97e936a2 2025-12-04T09:33:40.2788080Z * [new branch] bisect_perf_hf_T5_f23621ec563 -> origin/bisect_perf_hf_T5_f23621ec563 2025-12-04T09:33:40.2790235Z * [new branch] brister/fx_device_type -> origin/brister/fx_device_type 2025-12-04T09:33:40.2791506Z * [new branch] brister/test_inductor_all_fx -> origin/brister/test_inductor_all_fx 2025-12-04T09:33:40.2792877Z * [new branch] brister/tiled_reduction_no_numel_check -> origin/brister/tiled_reduction_no_numel_check 2025-12-04T09:33:40.2794043Z * [new branch] bwd-backup -> origin/bwd-backup 2025-12-04T09:33:40.2795578Z * [new branch] c57382a49 -> origin/c57382a49 2025-12-04T09:33:40.2796923Z * [new branch] ca_0431d47eaa -> origin/ca_0431d47eaa 2025-12-04T09:33:40.2798134Z * [new branch] ca_fix_0431d47eaa -> origin/ca_fix_0431d47eaa 2025-12-04T09:33:40.2800170Z * [new branch] camyllh/test_setup_hooks_push -> origin/camyllh/test_setup_hooks_push 2025-12-04T09:33:40.2801372Z * [new branch] cccclai-patch-1 -> origin/cccclai-patch-1 2025-12-04T09:33:40.2802950Z * [new branch] cherry-pick-159969-by-pytorch_bot_bot_ -> origin/cherry-pick-159969-by-pytorch_bot_bot_ 2025-12-04T09:33:40.2804196Z * [new branch] cherry-pick-160586-by-pytorch_bot_bot_ -> origin/cherry-pick-160586-by-pytorch_bot_bot_ 2025-12-04T09:33:40.2805616Z * [new branch] cherry-pick-162208-by-pytorch_bot_bot_ -> origin/cherry-pick-162208-by-pytorch_bot_bot_ 2025-12-04T09:33:40.2806901Z * [new branch] cherry-pick-163169-by-pytorch_bot_bot_ -> origin/cherry-pick-163169-by-pytorch_bot_bot_ 2025-12-04T09:33:40.2808214Z * [new branch] cherry-pick-165086-by-pytorch_bot_bot_ -> origin/cherry-pick-165086-by-pytorch_bot_bot_ 2025-12-04T09:33:40.2809680Z * [new branch] cherry-pick-165514-by-pytorch_bot_bot_ -> origin/cherry-pick-165514-by-pytorch_bot_bot_ 2025-12-04T09:33:40.2810993Z * [new branch] cherry-pick-165601-by-pytorch_bot_bot_ -> origin/cherry-pick-165601-by-pytorch_bot_bot_ 2025-12-04T09:33:40.2812320Z * [new branch] cherry-pick-165667-by-pytorch_bot_bot_ -> origin/cherry-pick-165667-by-pytorch_bot_bot_ 2025-12-04T09:33:40.2813718Z * [new branch] cherry-pick-165815-by-pytorch_bot_bot_ -> origin/cherry-pick-165815-by-pytorch_bot_bot_ 2025-12-04T09:33:40.2815081Z * [new branch] cherry-pick-165922-by-pytorch_bot_bot_ -> origin/cherry-pick-165922-by-pytorch_bot_bot_ 2025-12-04T09:33:40.2816394Z * [new branch] cherry-pick-166148-by-pytorch_bot_bot_ -> origin/cherry-pick-166148-by-pytorch_bot_bot_ 2025-12-04T09:33:40.2817814Z * [new branch] cherry-pick-166181-by-pytorch_bot_bot_ -> origin/cherry-pick-166181-by-pytorch_bot_bot_ 2025-12-04T09:33:40.2819105Z * [new branch] cherry-pick-166404-by-pytorch_bot_bot_ -> origin/cherry-pick-166404-by-pytorch_bot_bot_ 2025-12-04T09:33:40.2820443Z * [new branch] cherry-pick-166427-by-pytorch_bot_bot_ -> origin/cherry-pick-166427-by-pytorch_bot_bot_ 2025-12-04T09:33:40.2822384Z * [new branch] cherry-pick-166480-by-pytorch_bot_bot_ -> origin/cherry-pick-166480-by-pytorch_bot_bot_ 2025-12-04T09:33:40.2823542Z * [new branch] cherry-pick-166570-by-pytorch_bot_bot_ -> origin/cherry-pick-166570-by-pytorch_bot_bot_ 2025-12-04T09:33:40.2824873Z * [new branch] cherry-pick-166993-by-pytorch_bot_bot_ -> origin/cherry-pick-166993-by-pytorch_bot_bot_ 2025-12-04T09:33:40.2826171Z * [new branch] cherry-pick-167111-by-pytorch_bot_bot_ -> origin/cherry-pick-167111-by-pytorch_bot_bot_ 2025-12-04T09:33:40.2827533Z * [new branch] cherry-pick-167478-by-pytorch_bot_bot_ -> origin/cherry-pick-167478-by-pytorch_bot_bot_ 2025-12-04T09:33:40.2828653Z * [new branch] cherry_pick_166036_166040 -> origin/cherry_pick_166036_166040 2025-12-04T09:33:40.2830093Z * [new branch] cherry_pick_166457 -> origin/cherry_pick_166457 2025-12-04T09:33:40.2831386Z * [new branch] cherrypick_166338 -> origin/cherrypick_166338 2025-12-04T09:33:40.2832804Z * [new branch] cherrypick_166458 -> origin/cherrypick_166458 2025-12-04T09:33:40.2833971Z * [new branch] cherrypick_166586 -> origin/cherrypick_166586 2025-12-04T09:33:40.2835325Z * [new branch] cherrypick_166956 -> origin/cherrypick_166956 2025-12-04T09:33:40.2836532Z * [new branch] ci_attn -> origin/ci_attn 2025-12-04T09:33:40.2837929Z * [new branch] codex-testing -> origin/codex-testing 2025-12-04T09:33:40.2840092Z * [new branch] codex/add-check_memory_overlap-helper-functions -> origin/codex/add-check_memory_overlap-helper-functions 2025-12-04T09:33:40.2841528Z * [new branch] codex/fix-issue-121219-in-pytorch -> origin/codex/fix-issue-121219-in-pytorch 2025-12-04T09:33:40.2842855Z * [new branch] codex/investigate-segfaults-in-get_tensor_storage_id -> origin/codex/investigate-segfaults-in-get_tensor_storage_id 2025-12-04T09:33:40.2844418Z * [new branch] codex/refactor-lintrunner-config-to-use-uv-run -> origin/codex/refactor-lintrunner-config-to-use-uv-run 2025-12-04T09:33:40.2845384Z * [new branch] compatiblpy39util -> origin/compatiblpy39util 2025-12-04T09:33:40.2846666Z * [new branch] cond_hop_device -> origin/cond_hop_device 2025-12-04T09:33:40.2848072Z * [new branch] context_test -> origin/context_test 2025-12-04T09:33:40.2850004Z * [new branch] copilot/code-style-cleanup-python-pip -> origin/copilot/code-style-cleanup-python-pip 2025-12-04T09:33:40.2851461Z * [new branch] cpio/fix_new_ami_tests -> origin/cpio/fix_new_ami_tests 2025-12-04T09:33:40.2852811Z * [new branch] cpp-docs-dependency-upgrade -> origin/cpp-docs-dependency-upgrade 2025-12-04T09:33:40.2854788Z * [new branch] crpa/typo-in-inductor_comm_lowering -> origin/crpa/typo-in-inductor_comm_lowering 2025-12-04T09:33:40.2856262Z * [new branch] csl/always_produce_xml -> origin/csl/always_produce_xml 2025-12-04T09:33:40.2857567Z * [new branch] csl/build_test_more_procs -> origin/csl/build_test_more_procs 2025-12-04T09:33:40.2858846Z * [new branch] csl/build_test_more_procs2 -> origin/csl/build_test_more_procs2 2025-12-04T09:33:40.2860059Z * [new branch] csl/clean_up -> origin/csl/clean_up 2025-12-04T09:33:40.2861381Z * [new branch] csl/fix_retry_segfault_exit -> origin/csl/fix_retry_segfault_exit 2025-12-04T09:33:40.2862484Z * [new branch] csl/katex -> origin/csl/katex 2025-12-04T09:33:40.2864089Z * [new branch] csl/larger_runner -> origin/csl/larger_runner 2025-12-04T09:33:40.2865644Z * [new branch] csl/lint_testing -> origin/csl/lint_testing 2025-12-04T09:33:40.2867233Z * [new branch] csl/lint_thing -> origin/csl/lint_thing 2025-12-04T09:33:40.2868636Z * [new branch] csl/lintrunner_stuff -> origin/csl/lintrunner_stuff 2025-12-04T09:33:40.2869890Z * [new branch] csl/manually_gen_json -> origin/csl/manually_gen_json 2025-12-04T09:33:40.2871184Z * [new branch] csl/mps_sharding -> origin/csl/mps_sharding 2025-12-04T09:33:40.2872608Z * [new branch] csl/multistage_docker -> origin/csl/multistage_docker 2025-12-04T09:33:40.2873801Z * [new branch] csl/print_timing -> origin/csl/print_timing 2025-12-04T09:33:40.2875021Z * [new branch] csl/remove_experiment -> origin/csl/remove_experiment 2025-12-04T09:33:40.2876333Z * [new branch] csl/remove_maybe_unused_var -> origin/csl/remove_maybe_unused_var 2025-12-04T09:33:40.2877746Z * [new branch] csl/remove_repo_specific_autolabel -> origin/csl/remove_repo_specific_autolabel 2025-12-04T09:33:40.2879012Z * [new branch] csl/remove_run_parallel -> origin/csl/remove_run_parallel 2025-12-04T09:33:40.2880209Z * [new branch] csl/remove_unused_vars -> origin/csl/remove_unused_vars 2025-12-04T09:33:40.2881437Z * [new branch] csl/revert_open -> origin/csl/revert_open 2025-12-04T09:33:40.2882727Z * [new branch] csl/skip_build -> origin/csl/skip_build 2025-12-04T09:33:40.2884050Z * [new branch] csl/smaller_avx_amx_runenrs -> origin/csl/smaller_avx_amx_runenrs 2025-12-04T09:33:40.2885299Z * [new branch] csl/td_job_level -> origin/csl/td_job_level 2025-12-04T09:33:40.2886655Z * [new branch] csl/test_cuda_build_large_runner -> origin/csl/test_cuda_build_large_runner 2025-12-04T09:33:40.2888061Z * [new branch] csl/test_owners_autograd_dispatch_nn -> origin/csl/test_owners_autograd_dispatch_nn 2025-12-04T09:33:40.2889317Z * [new branch] csl/test_owners_higher_confidence -> origin/csl/test_owners_higher_confidence 2025-12-04T09:33:40.2890583Z * [new branch] csl/upload_json_running -> origin/csl/upload_json_running 2025-12-04T09:33:40.2891828Z * [new branch] csl/win_sccache -> origin/csl/win_sccache 2025-12-04T09:33:40.2893079Z * [new branch] csl/xml_stuff -> origin/csl/xml_stuff 2025-12-04T09:33:40.2894469Z * [new branch] cublasrelax2 -> origin/cublasrelax2 2025-12-04T09:33:40.2895643Z * [new branch] cuda_mempool -> origin/cuda_mempool 2025-12-04T09:33:40.2897295Z * [new branch] custom_lowering_dict -> origin/custom_lowering_dict 2025-12-04T09:33:40.2899135Z * [new branch] d4l3k/debug_plane_frtrace -> origin/d4l3k/debug_plane_frtrace 2025-12-04T09:33:40.2900804Z * [new branch] daxia6/2.8o3 -> origin/daxia6/2.8o3 2025-12-04T09:33:40.2901993Z * [new branch] debug-guard -> origin/debug-guard 2025-12-04T09:33:40.2903481Z * [new branch] delete-quant-docs -> origin/delete-quant-docs 2025-12-04T09:33:40.2907749Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.0 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.0 2025-12-04T09:33:40.2909488Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.1 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.1 2025-12-04T09:33:40.2910700Z * [new branch] desertfire/test_cpp_wrapper -> origin/desertfire/test_cpp_wrapper 2025-12-04T09:33:40.2912074Z * [new branch] desertfire/triton-cpu-for-aarch64 -> origin/desertfire/triton-cpu-for-aarch64 2025-12-04T09:33:40.2914107Z * [new branch] dev/dhruva/flex_attn_opt -> origin/dev/dhruva/flex_attn_opt 2025-12-04T09:33:40.2916283Z * [new branch] dev/joona/MPSNDArrayAdd -> origin/dev/joona/MPSNDArrayAdd 2025-12-04T09:33:40.2917942Z * [new branch] dev/joona/Unranked -> origin/dev/joona/Unranked 2025-12-04T09:33:40.2919416Z * [new branch] dev/joona/cat -> origin/dev/joona/cat 2025-12-04T09:33:40.2920672Z * [new branch] dev/joona/embeddingbag -> origin/dev/joona/embeddingbag 2025-12-04T09:33:40.2922166Z * [new branch] dev/joona/fix_sdpa_memtest -> origin/dev/joona/fix_sdpa_memtest 2025-12-04T09:33:40.2923754Z * [new branch] dev/joona/getTensorsString -> origin/dev/joona/getTensorsString 2025-12-04T09:33:40.2925272Z * [new branch] dev/joona/mps_linear_macos14 -> origin/dev/joona/mps_linear_macos14 2025-12-04T09:33:40.2927054Z * [new branch] dev/joona/scalar_clamp -> origin/dev/joona/scalar_clamp 2025-12-04T09:33:40.2928853Z * [new branch] dev/joona/sdpa -> origin/dev/joona/sdpa 2025-12-04T09:33:40.2930870Z * [new branch] dev/joona/sdpa_api -> origin/dev/joona/sdpa_api 2025-12-04T09:33:40.2932436Z * [new branch] dev/joona/type_inf -> origin/dev/joona/type_inf 2025-12-04T09:33:40.2934013Z * [new branch] dev/joona/ulpAssertClose -> origin/dev/joona/ulpAssertClose 2025-12-04T09:33:40.2935394Z * [new branch] dev/joona/upsize3d -> origin/dev/joona/upsize3d 2025-12-04T09:33:40.2936632Z * [new branch] disp_counter -> origin/disp_counter 2025-12-04T09:33:40.2938309Z * [new branch] divyanshk-patch-1 -> origin/divyanshk-patch-1 2025-12-04T09:33:40.2939379Z * [new branch] docs -> origin/docs 2025-12-04T09:33:40.2940842Z * [new branch] documentation -> origin/documentation 2025-12-04T09:33:40.2942081Z * [new branch] eager_model_benchmarks -> origin/eager_model_benchmarks 2025-12-04T09:33:40.2943996Z * [new branch] embg/test_inductor_ci_control -> origin/embg/test_inductor_ci_control 2025-12-04T09:33:40.2945102Z * [new branch] embg/triton_l2_prefetch_128B -> origin/embg/triton_l2_prefetch_128B 2025-12-04T09:33:40.2946247Z * [new branch] embg/triton_l2_prefetch_256B -> origin/embg/triton_l2_prefetch_256B 2025-12-04T09:33:40.2947538Z * [new branch] eqy-patch-1 -> origin/eqy-patch-1 2025-12-04T09:33:40.2949003Z * [new branch] eqy-patch-2 -> origin/eqy-patch-2 2025-12-04T09:33:40.2950814Z * [new branch] eqy-patch-3 -> origin/eqy-patch-3 2025-12-04T09:33:40.2952147Z * [new branch] eqy-patch-4 -> origin/eqy-patch-4 2025-12-04T09:33:40.2953332Z * [new branch] eqy-patch-5 -> origin/eqy-patch-5 2025-12-04T09:33:40.2954676Z * [new branch] eqy-patch-6 -> origin/eqy-patch-6 2025-12-04T09:33:40.2956508Z * [new branch] exclamaforte/amd-ma -> origin/exclamaforte/amd-ma 2025-12-04T09:33:40.2957863Z * [new branch] exclamaforte/combo-kernels-perf-run -> origin/exclamaforte/combo-kernels-perf-run 2025-12-04T09:33:40.2959052Z * [new branch] exclamaforte/do_bench_refactor -> origin/exclamaforte/do_bench_refactor 2025-12-04T09:33:40.2960368Z * [new branch] exclamaforte/enable-mem-dep-fusion -> origin/exclamaforte/enable-mem-dep-fusion 2025-12-04T09:33:40.2961715Z * [new branch] exclamaforte/fix-exhaustive-autotuning -> origin/exclamaforte/fix-exhaustive-autotuning 2025-12-04T09:33:40.2963286Z * [new branch] exclamaforte/fix-trace-parsing-fx-svg -> origin/exclamaforte/fix-trace-parsing-fx-svg 2025-12-04T09:33:40.2964937Z * [new branch] exclamaforte/force-pointwise-cat-perf-run -> origin/exclamaforte/force-pointwise-cat-perf-run 2025-12-04T09:33:40.2966071Z * [new branch] exclamaforte/fusion-data -> origin/exclamaforte/fusion-data 2025-12-04T09:33:40.2967575Z * [new branch] exclamaforte/gemm-benchmark-run -> origin/exclamaforte/gemm-benchmark-run 2025-12-04T09:33:40.2968787Z * [new branch] exclamaforte/gemm-export-model -> origin/exclamaforte/gemm-export-model 2025-12-04T09:33:40.2969990Z * [new branch] exclamaforte/gemm-model -> origin/exclamaforte/gemm-model 2025-12-04T09:33:40.2971457Z * [new branch] exclamaforte/gemm-model-all-data-collection -> origin/exclamaforte/gemm-model-all-data-collection 2025-12-04T09:33:40.2972516Z * [new branch] exclamaforte/gemm-to-amd -> origin/exclamaforte/gemm-to-amd 2025-12-04T09:33:40.2973795Z * [new branch] exclamaforte/just-gemm-model -> origin/exclamaforte/just-gemm-model 2025-12-04T09:33:40.2975324Z * [new branch] exclamaforte/just-gemm-model-no-refactor -> origin/exclamaforte/just-gemm-model-no-refactor 2025-12-04T09:33:40.2976527Z * [new branch] exclamaforte/profile-diff-algo -> origin/exclamaforte/profile-diff-algo 2025-12-04T09:33:40.2977965Z * [new branch] exclamaforte/profiler-visualization -> origin/exclamaforte/profiler-visualization 2025-12-04T09:33:40.2979252Z * [new branch] exclamaforte/test_cpp_wrapper_mode -> origin/exclamaforte/test_cpp_wrapper_mode 2025-12-04T09:33:40.2980631Z * [new branch] exclamaforte/update-autotune-configs -> origin/exclamaforte/update-autotune-configs 2025-12-04T09:33:40.2982000Z * [new branch] exclamaforte/update-autotune-configs-2 -> origin/exclamaforte/update-autotune-configs-2 2025-12-04T09:33:40.2983124Z * [new branch] exec -> origin/exec 2025-12-04T09:33:40.2984699Z * [new branch] experimental-mosaic -> origin/experimental-mosaic 2025-12-04T09:33:40.2985977Z * [new branch] export-D61047529 -> origin/export-D61047529 2025-12-04T09:33:40.2987442Z * [new branch] export-D71412006 -> origin/export-D71412006 2025-12-04T09:33:40.2988830Z * [new branch] export-D73042989 -> origin/export-D73042989 2025-12-04T09:33:40.2990003Z * [new branch] export-D78957093 -> origin/export-D78957093 2025-12-04T09:33:40.2991273Z * [new branch] export-D78996107 -> origin/export-D78996107 2025-12-04T09:33:40.2992555Z * [new branch] export-D80823877 -> origin/export-D80823877 2025-12-04T09:33:40.2994043Z * [new branch] export-D80958642 -> origin/export-D80958642 2025-12-04T09:33:40.2995235Z * [new branch] export-D81054193 -> origin/export-D81054193 2025-12-04T09:33:40.2996800Z * [new branch] export-D81204584 -> origin/export-D81204584 2025-12-04T09:33:40.2997966Z * [new branch] export-D81429090 -> origin/export-D81429090 2025-12-04T09:33:40.2999436Z * [new branch] export-D82250826 -> origin/export-D82250826 2025-12-04T09:33:40.3000654Z * [new branch] export-D82253817 -> origin/export-D82253817 2025-12-04T09:33:40.3001881Z * [new branch] export-D83541846 -> origin/export-D83541846 2025-12-04T09:33:40.3003333Z * [new branch] export-D83627170 -> origin/export-D83627170 2025-12-04T09:33:40.3004549Z * [new branch] export-D83766701 -> origin/export-D83766701 2025-12-04T09:33:40.3005784Z * [new branch] export-D83768878 -> origin/export-D83768878 2025-12-04T09:33:40.3007045Z * [new branch] export-D83769447 -> origin/export-D83769447 2025-12-04T09:33:40.3008309Z * [new branch] export-D84089824 -> origin/export-D84089824 2025-12-04T09:33:40.3009545Z * [new branch] export-D84213020 -> origin/export-D84213020 2025-12-04T09:33:40.3011463Z * [new branch] export-D84373821 -> origin/export-D84373821 2025-12-04T09:33:40.3012995Z * [new branch] export-D84612194 -> origin/export-D84612194 2025-12-04T09:33:40.3014183Z * [new branch] export-D84890985 -> origin/export-D84890985 2025-12-04T09:33:40.3015361Z * [new branch] export-D85122326 -> origin/export-D85122326 2025-12-04T09:33:40.3016804Z * [new branch] export-D86256198 -> origin/export-D86256198 2025-12-04T09:33:40.3018239Z * [new branch] export-D86460608 -> origin/export-D86460608 2025-12-04T09:33:40.3019623Z * [new branch] export-D86474796 -> origin/export-D86474796 2025-12-04T09:33:40.3021067Z * [new branch] export-D86712396 -> origin/export-D86712396 2025-12-04T09:33:40.3022323Z * [new branch] export-D87022129 -> origin/export-D87022129 2025-12-04T09:33:40.3023747Z * [new branch] export-D87838959 -> origin/export-D87838959 2025-12-04T09:33:40.3025034Z * [new branch] export-D88319437 -> origin/export-D88319437 2025-12-04T09:33:40.3026748Z * [new branch] exported-model-train-idempotent -> origin/exported-model-train-idempotent 2025-12-04T09:33:40.3027939Z * [new branch] ezyang-titan-october -> origin/ezyang-titan-october 2025-12-04T09:33:40.3029187Z * [new branch] ezyang-titan-october2 -> origin/ezyang-titan-october2 2025-12-04T09:33:40.3030385Z * [new branch] ezyang-war -> origin/ezyang-war 2025-12-04T09:33:40.3032365Z * [new branch] ezyang/wip-aot-descriptors -> origin/ezyang/wip-aot-descriptors 2025-12-04T09:33:40.3033418Z * [new branch] fa_u8_brgemm -> origin/fa_u8_brgemm 2025-12-04T09:33:40.3035334Z * [new branch] fadeputr/sequence_fbgemm -> origin/fadeputr/sequence_fbgemm 2025-12-04T09:33:40.3036519Z * [new branch] fastmath_baseline -> origin/fastmath_baseline 2025-12-04T09:33:40.3038443Z * [new branch] fbcode/warm -> origin/fbcode/warm 2025-12-04T09:33:40.3040261Z * [new branch] fca -> origin/fca 2025-12-04T09:33:40.3041399Z * [new branch] fca2_ca5984c -> origin/fca2_ca5984c 2025-12-04T09:33:40.3042793Z * [new branch] fca5 -> origin/fca5 2025-12-04T09:33:40.3044573Z * [new branch] feature/justknobs-cpp -> origin/feature/justknobs-cpp 2025-12-04T09:33:40.3045842Z * [new branch] feature/numa-forkserver -> origin/feature/numa-forkserver 2025-12-04T09:33:40.3048215Z * [new branch] ffast_math_baseline -> origin/ffast_math_baseline 2025-12-04T09:33:40.3049387Z * [new branch] ffast_math_target -> origin/ffast_math_target 2025-12-04T09:33:40.3051296Z * [new branch] findhao/base_commit -> origin/findhao/base_commit 2025-12-04T09:33:40.3052554Z * [new branch] findhao/base_commit1 -> origin/findhao/base_commit1 2025-12-04T09:33:40.3054207Z * [new branch] findhao/multistream2 -> origin/findhao/multistream2 2025-12-04T09:33:40.3055100Z * [new branch] findhao/multistream5 -> origin/findhao/multistream5 2025-12-04T09:33:40.3056262Z * [new branch] findhao/multistream6 -> origin/findhao/multistream6 2025-12-04T09:33:40.3057610Z * [new branch] findhao/operatorbench3 -> origin/findhao/operatorbench3 2025-12-04T09:33:40.3058799Z * [new branch] findhao/operatorbench5 -> origin/findhao/operatorbench5 2025-12-04T09:33:40.3059998Z * [new branch] findhao/tritonparse -> origin/findhao/tritonparse 2025-12-04T09:33:40.3061452Z * [new branch] fix-ck-gemm-template-format -> origin/fix-ck-gemm-template-format 2025-12-04T09:33:40.3062787Z * [new branch] fix-config-ignore -> origin/fix-config-ignore 2025-12-04T09:33:40.3063946Z * [new branch] fix-dict-guard -> origin/fix-dict-guard 2025-12-04T09:33:40.3065376Z * [new branch] fix_addmm_issue -> origin/fix_addmm_issue 2025-12-04T09:33:40.3066628Z * [new branch] fix_amd_missing_cluster_dims -> origin/fix_amd_missing_cluster_dims 2025-12-04T09:33:40.3067809Z * [new branch] fix_bench_bwd_pass -> origin/fix_bench_bwd_pass 2025-12-04T09:33:40.3069223Z * [new branch] fix_mem_profiler_config -> origin/fix_mem_profiler_config 2025-12-04T09:33:40.3070389Z * [new branch] fix_nvrtc_discovery -> origin/fix_nvrtc_discovery 2025-12-04T09:33:40.3071603Z * [new branch] fix_op_runner -> origin/fix_op_runner 2025-12-04T09:33:40.3072979Z * [new branch] fix_ubn_159469 -> origin/fix_ubn_159469 2025-12-04T09:33:40.3074213Z * [new branch] fixes-triage -> origin/fixes-triage 2025-12-04T09:33:40.3075574Z * [new branch] fixflashinfer -> origin/fixflashinfer 2025-12-04T09:33:40.3076746Z * [new branch] flash_decoding_cpu -> origin/flash_decoding_cpu 2025-12-04T09:33:40.3077959Z * [new branch] flex-flash -> origin/flex-flash 2025-12-04T09:33:40.3079315Z * [new branch] flex_attention_functorch_grad -> origin/flex_attention_functorch_grad 2025-12-04T09:33:40.3081060Z * [new branch] flex_flash -> origin/flex_flash 2025-12-04T09:33:40.3083002Z * [new branch] fmassa/fix_memeff_sharding_rule -> origin/fmassa/fix_memeff_sharding_rule 2025-12-04T09:33:40.3084166Z * [new branch] fmassa/tests_comm_compute_scheduler -> origin/fmassa/tests_comm_compute_scheduler 2025-12-04T09:33:40.3085360Z * [new branch] forkserver_fix -> origin/forkserver_fix 2025-12-04T09:33:40.3086734Z * [new branch] fsdp2_trace_rules -> origin/fsdp2_trace_rules 2025-12-04T09:33:40.3087985Z * [new branch] fx_cpp -> origin/fx_cpp 2025-12-04T09:33:40.3089799Z * [new branch] fy/fix-win -> origin/fy/fix-win 2025-12-04T09:33:40.3091218Z * [new branch] galv-patch-1 -> origin/galv-patch-1 2025-12-04T09:33:40.3093345Z * [new branch] galv/cudagraphs-conditional-nodes-4 -> origin/galv/cudagraphs-conditional-nodes-4 2025-12-04T09:33:40.3095019Z * [new branch] georgehong/cmakelists-patch -> origin/georgehong/cmakelists-patch 2025-12-04T09:33:40.3098208Z * [new branch] gh/AlnisM/1/base -> origin/gh/AlnisM/1/base 2025-12-04T09:33:40.3120271Z * [new branch] gh/AlnisM/1/head -> origin/gh/AlnisM/1/head 2025-12-04T09:33:40.3120924Z * [new branch] gh/EikanWang/67/base -> origin/gh/EikanWang/67/base 2025-12-04T09:33:40.3121590Z * [new branch] gh/EikanWang/67/head -> origin/gh/EikanWang/67/head 2025-12-04T09:33:40.3122376Z * [new branch] gh/Gasoonjia/1/base -> origin/gh/Gasoonjia/1/base 2025-12-04T09:33:40.3123030Z * [new branch] gh/Gasoonjia/1/head -> origin/gh/Gasoonjia/1/head 2025-12-04T09:33:40.3123681Z * [new branch] gh/H-Huang/131/base -> origin/gh/H-Huang/131/base 2025-12-04T09:33:40.3124285Z * [new branch] gh/H-Huang/131/head -> origin/gh/H-Huang/131/head 2025-12-04T09:33:40.3124904Z * [new branch] gh/H-Huang/131/orig -> origin/gh/H-Huang/131/orig 2025-12-04T09:33:40.3125522Z * [new branch] gh/H-Huang/132/base -> origin/gh/H-Huang/132/base 2025-12-04T09:33:40.3126139Z * [new branch] gh/H-Huang/132/head -> origin/gh/H-Huang/132/head 2025-12-04T09:33:40.3126744Z * [new branch] gh/H-Huang/132/orig -> origin/gh/H-Huang/132/orig 2025-12-04T09:33:40.3127503Z * [new branch] gh/H-Huang/180/base -> origin/gh/H-Huang/180/base 2025-12-04T09:33:40.3128121Z * [new branch] gh/H-Huang/180/head -> origin/gh/H-Huang/180/head 2025-12-04T09:33:40.3128723Z * [new branch] gh/H-Huang/180/orig -> origin/gh/H-Huang/180/orig 2025-12-04T09:33:40.3129349Z * [new branch] gh/H-Huang/182/base -> origin/gh/H-Huang/182/base 2025-12-04T09:33:40.3129978Z * [new branch] gh/H-Huang/182/head -> origin/gh/H-Huang/182/head 2025-12-04T09:33:40.3130593Z * [new branch] gh/H-Huang/182/orig -> origin/gh/H-Huang/182/orig 2025-12-04T09:33:40.3131193Z * [new branch] gh/H-Huang/226/base -> origin/gh/H-Huang/226/base 2025-12-04T09:33:40.3131812Z * [new branch] gh/H-Huang/226/head -> origin/gh/H-Huang/226/head 2025-12-04T09:33:40.3132425Z * [new branch] gh/H-Huang/226/orig -> origin/gh/H-Huang/226/orig 2025-12-04T09:33:40.3133046Z * [new branch] gh/H-Huang/228/base -> origin/gh/H-Huang/228/base 2025-12-04T09:33:40.3133645Z * [new branch] gh/H-Huang/228/head -> origin/gh/H-Huang/228/head 2025-12-04T09:33:40.3134264Z * [new branch] gh/H-Huang/228/orig -> origin/gh/H-Huang/228/orig 2025-12-04T09:33:40.3135492Z * [new branch] gh/IvanKobzarev/150/base -> origin/gh/IvanKobzarev/150/base 2025-12-04T09:33:40.3136682Z * [new branch] gh/IvanKobzarev/150/head -> origin/gh/IvanKobzarev/150/head 2025-12-04T09:33:40.3138109Z * [new branch] gh/IvanKobzarev/150/orig -> origin/gh/IvanKobzarev/150/orig 2025-12-04T09:33:40.3140031Z * [new branch] gh/IvanKobzarev/157/base -> origin/gh/IvanKobzarev/157/base 2025-12-04T09:33:40.3141275Z * [new branch] gh/IvanKobzarev/157/head -> origin/gh/IvanKobzarev/157/head 2025-12-04T09:33:40.3142603Z * [new branch] gh/IvanKobzarev/157/orig -> origin/gh/IvanKobzarev/157/orig 2025-12-04T09:33:40.3144576Z * [new branch] gh/IvanKobzarev/159/base -> origin/gh/IvanKobzarev/159/base 2025-12-04T09:33:40.3145749Z * [new branch] gh/IvanKobzarev/159/head -> origin/gh/IvanKobzarev/159/head 2025-12-04T09:33:40.3147074Z * [new branch] gh/IvanKobzarev/159/orig -> origin/gh/IvanKobzarev/159/orig 2025-12-04T09:33:40.3148975Z * [new branch] gh/IvanKobzarev/162/base -> origin/gh/IvanKobzarev/162/base 2025-12-04T09:33:40.3150335Z * [new branch] gh/IvanKobzarev/162/head -> origin/gh/IvanKobzarev/162/head 2025-12-04T09:33:40.3151613Z * [new branch] gh/IvanKobzarev/162/orig -> origin/gh/IvanKobzarev/162/orig 2025-12-04T09:33:40.3153464Z * [new branch] gh/IvanKobzarev/163/base -> origin/gh/IvanKobzarev/163/base 2025-12-04T09:33:40.3154580Z * [new branch] gh/IvanKobzarev/163/head -> origin/gh/IvanKobzarev/163/head 2025-12-04T09:33:40.3155826Z * [new branch] gh/IvanKobzarev/163/orig -> origin/gh/IvanKobzarev/163/orig 2025-12-04T09:33:40.3157689Z * [new branch] gh/IvanKobzarev/166/base -> origin/gh/IvanKobzarev/166/base 2025-12-04T09:33:40.3158939Z * [new branch] gh/IvanKobzarev/166/head -> origin/gh/IvanKobzarev/166/head 2025-12-04T09:33:40.3160182Z * [new branch] gh/IvanKobzarev/166/orig -> origin/gh/IvanKobzarev/166/orig 2025-12-04T09:33:40.3162128Z * [new branch] gh/IvanKobzarev/167/base -> origin/gh/IvanKobzarev/167/base 2025-12-04T09:33:40.3163247Z * [new branch] gh/IvanKobzarev/167/head -> origin/gh/IvanKobzarev/167/head 2025-12-04T09:33:40.3164518Z * [new branch] gh/IvanKobzarev/167/orig -> origin/gh/IvanKobzarev/167/orig 2025-12-04T09:33:40.3166298Z * [new branch] gh/IvanKobzarev/168/base -> origin/gh/IvanKobzarev/168/base 2025-12-04T09:33:40.3167661Z * [new branch] gh/IvanKobzarev/168/head -> origin/gh/IvanKobzarev/168/head 2025-12-04T09:33:40.3168857Z * [new branch] gh/IvanKobzarev/168/orig -> origin/gh/IvanKobzarev/168/orig 2025-12-04T09:33:40.3170688Z * [new branch] gh/IvanKobzarev/169/base -> origin/gh/IvanKobzarev/169/base 2025-12-04T09:33:40.3171867Z * [new branch] gh/IvanKobzarev/169/head -> origin/gh/IvanKobzarev/169/head 2025-12-04T09:33:40.3173157Z * [new branch] gh/IvanKobzarev/169/orig -> origin/gh/IvanKobzarev/169/orig 2025-12-04T09:33:40.3174870Z * [new branch] gh/IvanKobzarev/170/base -> origin/gh/IvanKobzarev/170/base 2025-12-04T09:33:40.3176063Z * [new branch] gh/IvanKobzarev/170/head -> origin/gh/IvanKobzarev/170/head 2025-12-04T09:33:40.3177396Z * [new branch] gh/IvanKobzarev/170/orig -> origin/gh/IvanKobzarev/170/orig 2025-12-04T09:33:40.3179473Z * [new branch] gh/IvanKobzarev/171/base -> origin/gh/IvanKobzarev/171/base 2025-12-04T09:33:40.3180665Z * [new branch] gh/IvanKobzarev/171/head -> origin/gh/IvanKobzarev/171/head 2025-12-04T09:33:40.3181906Z * [new branch] gh/IvanKobzarev/171/orig -> origin/gh/IvanKobzarev/171/orig 2025-12-04T09:33:40.3183756Z * [new branch] gh/IvanKobzarev/172/base -> origin/gh/IvanKobzarev/172/base 2025-12-04T09:33:40.3185035Z * [new branch] gh/IvanKobzarev/172/head -> origin/gh/IvanKobzarev/172/head 2025-12-04T09:33:40.3186349Z * [new branch] gh/IvanKobzarev/172/orig -> origin/gh/IvanKobzarev/172/orig 2025-12-04T09:33:40.3188104Z * [new branch] gh/IvanKobzarev/173/base -> origin/gh/IvanKobzarev/173/base 2025-12-04T09:33:40.3189267Z * [new branch] gh/IvanKobzarev/173/head -> origin/gh/IvanKobzarev/173/head 2025-12-04T09:33:40.3190529Z * [new branch] gh/IvanKobzarev/173/orig -> origin/gh/IvanKobzarev/173/orig 2025-12-04T09:33:40.3192385Z * [new branch] gh/IvanKobzarev/174/base -> origin/gh/IvanKobzarev/174/base 2025-12-04T09:33:40.3193668Z * [new branch] gh/IvanKobzarev/174/head -> origin/gh/IvanKobzarev/174/head 2025-12-04T09:33:40.3194950Z * [new branch] gh/IvanKobzarev/174/orig -> origin/gh/IvanKobzarev/174/orig 2025-12-04T09:33:40.3197098Z * [new branch] gh/IvanKobzarev/175/base -> origin/gh/IvanKobzarev/175/base 2025-12-04T09:33:40.3198375Z * [new branch] gh/IvanKobzarev/175/head -> origin/gh/IvanKobzarev/175/head 2025-12-04T09:33:40.3199722Z * [new branch] gh/IvanKobzarev/175/orig -> origin/gh/IvanKobzarev/175/orig 2025-12-04T09:33:40.3201720Z * [new branch] gh/IvanKobzarev/176/base -> origin/gh/IvanKobzarev/176/base 2025-12-04T09:33:40.3202894Z * [new branch] gh/IvanKobzarev/176/head -> origin/gh/IvanKobzarev/176/head 2025-12-04T09:33:40.3204171Z * [new branch] gh/IvanKobzarev/176/orig -> origin/gh/IvanKobzarev/176/orig 2025-12-04T09:33:40.3206252Z * [new branch] gh/IvanKobzarev/177/base -> origin/gh/IvanKobzarev/177/base 2025-12-04T09:33:40.3207546Z * [new branch] gh/IvanKobzarev/177/head -> origin/gh/IvanKobzarev/177/head 2025-12-04T09:33:40.3208765Z * [new branch] gh/IvanKobzarev/177/orig -> origin/gh/IvanKobzarev/177/orig 2025-12-04T09:33:40.3210768Z * [new branch] gh/IvanKobzarev/178/base -> origin/gh/IvanKobzarev/178/base 2025-12-04T09:33:40.3212063Z * [new branch] gh/IvanKobzarev/178/head -> origin/gh/IvanKobzarev/178/head 2025-12-04T09:33:40.3213354Z * [new branch] gh/IvanKobzarev/178/orig -> origin/gh/IvanKobzarev/178/orig 2025-12-04T09:33:40.3215235Z * [new branch] gh/IvanKobzarev/179/base -> origin/gh/IvanKobzarev/179/base 2025-12-04T09:33:40.3216351Z * [new branch] gh/IvanKobzarev/179/head -> origin/gh/IvanKobzarev/179/head 2025-12-04T09:33:40.3218001Z * [new branch] gh/IvanKobzarev/179/orig -> origin/gh/IvanKobzarev/179/orig 2025-12-04T09:33:40.3219678Z * [new branch] gh/IvanKobzarev/180/base -> origin/gh/IvanKobzarev/180/base 2025-12-04T09:33:40.3220910Z * [new branch] gh/IvanKobzarev/180/head -> origin/gh/IvanKobzarev/180/head 2025-12-04T09:33:40.3222153Z * [new branch] gh/IvanKobzarev/180/orig -> origin/gh/IvanKobzarev/180/orig 2025-12-04T09:33:40.3224190Z * [new branch] gh/IvanKobzarev/181/base -> origin/gh/IvanKobzarev/181/base 2025-12-04T09:33:40.3225488Z * [new branch] gh/IvanKobzarev/181/head -> origin/gh/IvanKobzarev/181/head 2025-12-04T09:33:40.3226763Z * [new branch] gh/IvanKobzarev/181/orig -> origin/gh/IvanKobzarev/181/orig 2025-12-04T09:33:40.3228865Z * [new branch] gh/IvanKobzarev/182/base -> origin/gh/IvanKobzarev/182/base 2025-12-04T09:33:40.3230095Z * [new branch] gh/IvanKobzarev/182/head -> origin/gh/IvanKobzarev/182/head 2025-12-04T09:33:40.3231305Z * [new branch] gh/IvanKobzarev/182/orig -> origin/gh/IvanKobzarev/182/orig 2025-12-04T09:33:40.3233322Z * [new branch] gh/IvanKobzarev/183/base -> origin/gh/IvanKobzarev/183/base 2025-12-04T09:33:40.3234584Z * [new branch] gh/IvanKobzarev/183/head -> origin/gh/IvanKobzarev/183/head 2025-12-04T09:33:40.3236021Z * [new branch] gh/IvanKobzarev/183/orig -> origin/gh/IvanKobzarev/183/orig 2025-12-04T09:33:40.3237786Z * [new branch] gh/IvanKobzarev/184/base -> origin/gh/IvanKobzarev/184/base 2025-12-04T09:33:40.3239008Z * [new branch] gh/IvanKobzarev/184/head -> origin/gh/IvanKobzarev/184/head 2025-12-04T09:33:40.3240315Z * [new branch] gh/IvanKobzarev/184/orig -> origin/gh/IvanKobzarev/184/orig 2025-12-04T09:33:40.3242458Z * [new branch] gh/NikhilAPatel/1/base -> origin/gh/NikhilAPatel/1/base 2025-12-04T09:33:40.3243987Z * [new branch] gh/NikhilAPatel/1/head -> origin/gh/NikhilAPatel/1/head 2025-12-04T09:33:40.3245469Z * [new branch] gh/NikhilAPatel/2/base -> origin/gh/NikhilAPatel/2/base 2025-12-04T09:33:40.3246673Z * [new branch] gh/NikhilAPatel/2/head -> origin/gh/NikhilAPatel/2/head 2025-12-04T09:33:40.3248567Z * [new branch] gh/NikhilAPatel/4/base -> origin/gh/NikhilAPatel/4/base 2025-12-04T09:33:40.3249901Z * [new branch] gh/NikhilAPatel/4/head -> origin/gh/NikhilAPatel/4/head 2025-12-04T09:33:40.3251707Z * [new branch] gh/NikhilAPatel/5/base -> origin/gh/NikhilAPatel/5/base 2025-12-04T09:33:40.3252924Z * [new branch] gh/NikhilAPatel/5/head -> origin/gh/NikhilAPatel/5/head 2025-12-04T09:33:40.3254246Z * [new branch] gh/NikhilAPatel/5/orig -> origin/gh/NikhilAPatel/5/orig 2025-12-04T09:33:40.3256302Z * [new branch] gh/PaliC/17/base -> origin/gh/PaliC/17/base 2025-12-04T09:33:40.3257658Z * [new branch] gh/PaliC/17/head -> origin/gh/PaliC/17/head 2025-12-04T09:33:40.3259066Z * [new branch] gh/PaliC/17/orig -> origin/gh/PaliC/17/orig 2025-12-04T09:33:40.3260744Z * [new branch] gh/PaliC/18/base -> origin/gh/PaliC/18/base 2025-12-04T09:33:40.3261909Z * [new branch] gh/PaliC/18/head -> origin/gh/PaliC/18/head 2025-12-04T09:33:40.3263275Z * [new branch] gh/PaliC/18/orig -> origin/gh/PaliC/18/orig 2025-12-04T09:33:40.3265128Z * [new branch] gh/PaliC/20/base -> origin/gh/PaliC/20/base 2025-12-04T09:33:40.3266297Z * [new branch] gh/PaliC/20/head -> origin/gh/PaliC/20/head 2025-12-04T09:33:40.3267870Z * [new branch] gh/PaliC/20/orig -> origin/gh/PaliC/20/orig 2025-12-04T09:33:40.3269500Z * [new branch] gh/PaliC/21/base -> origin/gh/PaliC/21/base 2025-12-04T09:33:40.3270799Z * [new branch] gh/PaliC/21/head -> origin/gh/PaliC/21/head 2025-12-04T09:33:40.3271961Z * [new branch] gh/PaliC/21/orig -> origin/gh/PaliC/21/orig 2025-12-04T09:33:40.3273646Z * [new branch] gh/PaliC/23/base -> origin/gh/PaliC/23/base 2025-12-04T09:33:40.3274854Z * [new branch] gh/PaliC/23/head -> origin/gh/PaliC/23/head 2025-12-04T09:33:40.3276091Z * [new branch] gh/PaliC/23/orig -> origin/gh/PaliC/23/orig 2025-12-04T09:33:40.3277869Z * [new branch] gh/PaliC/24/base -> origin/gh/PaliC/24/base 2025-12-04T09:33:40.3279050Z * [new branch] gh/PaliC/24/head -> origin/gh/PaliC/24/head 2025-12-04T09:33:40.3280371Z * [new branch] gh/PaliC/24/orig -> origin/gh/PaliC/24/orig 2025-12-04T09:33:40.3282034Z * [new branch] gh/PaliC/25/head -> origin/gh/PaliC/25/head 2025-12-04T09:33:40.3283190Z * [new branch] gh/PaliC/25/next -> origin/gh/PaliC/25/next 2025-12-04T09:33:40.3284701Z * [new branch] gh/PaliC/25/orig -> origin/gh/PaliC/25/orig 2025-12-04T09:33:40.3286322Z * [new branch] gh/PaliC/26/head -> origin/gh/PaliC/26/head 2025-12-04T09:33:40.3287361Z * [new branch] gh/PaliC/26/next -> origin/gh/PaliC/26/next 2025-12-04T09:33:40.3288765Z * [new branch] gh/PaliC/26/orig -> origin/gh/PaliC/26/orig 2025-12-04T09:33:40.3290505Z * [new branch] gh/PaliC/27/next -> origin/gh/PaliC/27/next 2025-12-04T09:33:40.3292382Z * [new branch] gh/PaliC/28/head -> origin/gh/PaliC/28/head 2025-12-04T09:33:40.3293449Z * [new branch] gh/PaliC/28/next -> origin/gh/PaliC/28/next 2025-12-04T09:33:40.3294871Z * [new branch] gh/PaliC/28/orig -> origin/gh/PaliC/28/orig 2025-12-04T09:33:40.3296758Z * [new branch] gh/PaliC/29/head -> origin/gh/PaliC/29/head 2025-12-04T09:33:40.3297910Z * [new branch] gh/PaliC/29/next -> origin/gh/PaliC/29/next 2025-12-04T09:33:40.3299271Z * [new branch] gh/PaliC/29/orig -> origin/gh/PaliC/29/orig 2025-12-04T09:33:40.3301039Z * [new branch] gh/PaliC/30/head -> origin/gh/PaliC/30/head 2025-12-04T09:33:40.3302123Z * [new branch] gh/PaliC/30/next -> origin/gh/PaliC/30/next 2025-12-04T09:33:40.3303391Z * [new branch] gh/PaliC/30/orig -> origin/gh/PaliC/30/orig 2025-12-04T09:33:40.3305138Z * [new branch] gh/PaliC/31/head -> origin/gh/PaliC/31/head 2025-12-04T09:33:40.3306169Z * [new branch] gh/PaliC/31/next -> origin/gh/PaliC/31/next 2025-12-04T09:33:40.3307524Z * [new branch] gh/PaliC/31/orig -> origin/gh/PaliC/31/orig 2025-12-04T09:33:40.3309552Z * [new branch] gh/PaulZhang12/25/base -> origin/gh/PaulZhang12/25/base 2025-12-04T09:33:40.3310808Z * [new branch] gh/PaulZhang12/25/head -> origin/gh/PaulZhang12/25/head 2025-12-04T09:33:40.3312112Z * [new branch] gh/PaulZhang12/25/orig -> origin/gh/PaulZhang12/25/orig 2025-12-04T09:33:40.3313931Z * [new branch] gh/PaulZhang12/28/base -> origin/gh/PaulZhang12/28/base 2025-12-04T09:33:40.3315252Z * [new branch] gh/PaulZhang12/28/head -> origin/gh/PaulZhang12/28/head 2025-12-04T09:33:40.3316675Z * [new branch] gh/PaulZhang12/28/orig -> origin/gh/PaulZhang12/28/orig 2025-12-04T09:33:40.3318599Z * [new branch] gh/PaulZhang12/31/base -> origin/gh/PaulZhang12/31/base 2025-12-04T09:33:40.3319815Z * [new branch] gh/PaulZhang12/31/head -> origin/gh/PaulZhang12/31/head 2025-12-04T09:33:40.3322958Z * [new branch] gh/PaulZhang12/31/orig -> origin/gh/PaulZhang12/31/orig 2025-12-04T09:33:40.3323770Z * [new branch] gh/PaulZhang12/37/base -> origin/gh/PaulZhang12/37/base 2025-12-04T09:33:40.3324440Z * [new branch] gh/PaulZhang12/37/head -> origin/gh/PaulZhang12/37/head 2025-12-04T09:33:40.3325234Z * [new branch] gh/PaulZhang12/37/orig -> origin/gh/PaulZhang12/37/orig 2025-12-04T09:33:40.3327051Z * [new branch] gh/PaulZhang12/40/base -> origin/gh/PaulZhang12/40/base 2025-12-04T09:33:40.3328240Z * [new branch] gh/PaulZhang12/40/head -> origin/gh/PaulZhang12/40/head 2025-12-04T09:33:40.3329500Z * [new branch] gh/PaulZhang12/40/orig -> origin/gh/PaulZhang12/40/orig 2025-12-04T09:33:40.3331311Z * [new branch] gh/PaulZhang12/42/base -> origin/gh/PaulZhang12/42/base 2025-12-04T09:33:40.3332508Z * [new branch] gh/PaulZhang12/42/head -> origin/gh/PaulZhang12/42/head 2025-12-04T09:33:40.3334318Z * [new branch] gh/PaulZhang12/43/base -> origin/gh/PaulZhang12/43/base 2025-12-04T09:33:40.3335502Z * [new branch] gh/PaulZhang12/43/head -> origin/gh/PaulZhang12/43/head 2025-12-04T09:33:40.3336779Z * [new branch] gh/PaulZhang12/43/orig -> origin/gh/PaulZhang12/43/orig 2025-12-04T09:33:40.3338604Z * [new branch] gh/PaulZhang12/44/base -> origin/gh/PaulZhang12/44/base 2025-12-04T09:33:40.3339749Z * [new branch] gh/PaulZhang12/44/head -> origin/gh/PaulZhang12/44/head 2025-12-04T09:33:40.3341630Z * [new branch] gh/PaulZhang12/45/base -> origin/gh/PaulZhang12/45/base 2025-12-04T09:33:40.3342748Z * [new branch] gh/PaulZhang12/45/head -> origin/gh/PaulZhang12/45/head 2025-12-04T09:33:40.3343976Z * [new branch] gh/PaulZhang12/45/orig -> origin/gh/PaulZhang12/45/orig 2025-12-04T09:33:40.3345774Z * [new branch] gh/PaulZhang12/46/base -> origin/gh/PaulZhang12/46/base 2025-12-04T09:33:40.3347059Z * [new branch] gh/PaulZhang12/46/head -> origin/gh/PaulZhang12/46/head 2025-12-04T09:33:40.3348492Z * [new branch] gh/PaulZhang12/46/orig -> origin/gh/PaulZhang12/46/orig 2025-12-04T09:33:40.3350284Z * [new branch] gh/PaulZhang12/47/base -> origin/gh/PaulZhang12/47/base 2025-12-04T09:33:40.3351492Z * [new branch] gh/PaulZhang12/47/head -> origin/gh/PaulZhang12/47/head 2025-12-04T09:33:40.3352799Z * [new branch] gh/PaulZhang12/47/orig -> origin/gh/PaulZhang12/47/orig 2025-12-04T09:33:40.3354438Z * [new branch] gh/PaulZhang12/48/base -> origin/gh/PaulZhang12/48/base 2025-12-04T09:33:40.3355643Z * [new branch] gh/PaulZhang12/48/head -> origin/gh/PaulZhang12/48/head 2025-12-04T09:33:40.3356905Z * [new branch] gh/PaulZhang12/48/orig -> origin/gh/PaulZhang12/48/orig 2025-12-04T09:33:40.3358998Z * [new branch] gh/SamGinzburg/11/base -> origin/gh/SamGinzburg/11/base 2025-12-04T09:33:40.3360233Z * [new branch] gh/SamGinzburg/11/head -> origin/gh/SamGinzburg/11/head 2025-12-04T09:33:40.3362512Z * [new branch] gh/SherlockNoMad/1/base -> origin/gh/SherlockNoMad/1/base 2025-12-04T09:33:40.3363753Z * [new branch] gh/SherlockNoMad/1/head -> origin/gh/SherlockNoMad/1/head 2025-12-04T09:33:40.3365573Z * [new branch] gh/SherlockNoMad/10/base -> origin/gh/SherlockNoMad/10/base 2025-12-04T09:33:40.3366845Z * [new branch] gh/SherlockNoMad/10/head -> origin/gh/SherlockNoMad/10/head 2025-12-04T09:33:40.3368227Z * [new branch] gh/SherlockNoMad/10/orig -> origin/gh/SherlockNoMad/10/orig 2025-12-04T09:33:40.3369877Z * [new branch] gh/SherlockNoMad/11/base -> origin/gh/SherlockNoMad/11/base 2025-12-04T09:33:40.3371075Z * [new branch] gh/SherlockNoMad/11/head -> origin/gh/SherlockNoMad/11/head 2025-12-04T09:33:40.3372403Z * [new branch] gh/SherlockNoMad/11/orig -> origin/gh/SherlockNoMad/11/orig 2025-12-04T09:33:40.3373920Z * [new branch] gh/SherlockNoMad/12/base -> origin/gh/SherlockNoMad/12/base 2025-12-04T09:33:40.3375081Z * [new branch] gh/SherlockNoMad/12/head -> origin/gh/SherlockNoMad/12/head 2025-12-04T09:33:40.3376306Z * [new branch] gh/SherlockNoMad/12/orig -> origin/gh/SherlockNoMad/12/orig 2025-12-04T09:33:40.3378269Z * [new branch] gh/SherlockNoMad/15/base -> origin/gh/SherlockNoMad/15/base 2025-12-04T09:33:40.3379467Z * [new branch] gh/SherlockNoMad/15/head -> origin/gh/SherlockNoMad/15/head 2025-12-04T09:33:40.3380767Z * [new branch] gh/SherlockNoMad/15/orig -> origin/gh/SherlockNoMad/15/orig 2025-12-04T09:33:40.3382536Z * [new branch] gh/SherlockNoMad/17/base -> origin/gh/SherlockNoMad/17/base 2025-12-04T09:33:40.3383735Z * [new branch] gh/SherlockNoMad/17/head -> origin/gh/SherlockNoMad/17/head 2025-12-04T09:33:40.3384993Z * [new branch] gh/SherlockNoMad/17/orig -> origin/gh/SherlockNoMad/17/orig 2025-12-04T09:33:40.3386954Z * [new branch] gh/SherlockNoMad/18/base -> origin/gh/SherlockNoMad/18/base 2025-12-04T09:33:40.3388216Z * [new branch] gh/SherlockNoMad/18/head -> origin/gh/SherlockNoMad/18/head 2025-12-04T09:33:40.3389452Z * [new branch] gh/SherlockNoMad/18/orig -> origin/gh/SherlockNoMad/18/orig 2025-12-04T09:33:40.3391082Z * [new branch] gh/SherlockNoMad/19/base -> origin/gh/SherlockNoMad/19/base 2025-12-04T09:33:40.3392443Z * [new branch] gh/SherlockNoMad/19/head -> origin/gh/SherlockNoMad/19/head 2025-12-04T09:33:40.3393729Z * [new branch] gh/SherlockNoMad/19/orig -> origin/gh/SherlockNoMad/19/orig 2025-12-04T09:33:40.3395368Z * [new branch] gh/SherlockNoMad/2/base -> origin/gh/SherlockNoMad/2/base 2025-12-04T09:33:40.3396692Z * [new branch] gh/SherlockNoMad/2/head -> origin/gh/SherlockNoMad/2/head 2025-12-04T09:33:40.3398347Z * [new branch] gh/SherlockNoMad/20/base -> origin/gh/SherlockNoMad/20/base 2025-12-04T09:33:40.3399624Z * [new branch] gh/SherlockNoMad/20/head -> origin/gh/SherlockNoMad/20/head 2025-12-04T09:33:40.3400795Z * [new branch] gh/SherlockNoMad/20/orig -> origin/gh/SherlockNoMad/20/orig 2025-12-04T09:33:40.3402812Z * [new branch] gh/SherlockNoMad/21/base -> origin/gh/SherlockNoMad/21/base 2025-12-04T09:33:40.3404074Z * [new branch] gh/SherlockNoMad/21/head -> origin/gh/SherlockNoMad/21/head 2025-12-04T09:33:40.3405286Z * [new branch] gh/SherlockNoMad/21/orig -> origin/gh/SherlockNoMad/21/orig 2025-12-04T09:33:40.3406895Z * [new branch] gh/SherlockNoMad/3/base -> origin/gh/SherlockNoMad/3/base 2025-12-04T09:33:40.3408091Z * [new branch] gh/SherlockNoMad/3/head -> origin/gh/SherlockNoMad/3/head 2025-12-04T09:33:40.3409675Z * [new branch] gh/SherlockNoMad/4/base -> origin/gh/SherlockNoMad/4/base 2025-12-04T09:33:40.3410822Z * [new branch] gh/SherlockNoMad/4/head -> origin/gh/SherlockNoMad/4/head 2025-12-04T09:33:40.3412477Z * [new branch] gh/SherlockNoMad/5/base -> origin/gh/SherlockNoMad/5/base 2025-12-04T09:33:40.3413609Z * [new branch] gh/SherlockNoMad/5/head -> origin/gh/SherlockNoMad/5/head 2025-12-04T09:33:40.3416230Z * [new branch] gh/Sidharth123-cpu/24/base -> origin/gh/Sidharth123-cpu/24/base 2025-12-04T09:33:40.3417946Z * [new branch] gh/Sidharth123-cpu/25/base -> origin/gh/Sidharth123-cpu/25/base 2025-12-04T09:33:40.3419491Z * [new branch] gh/Sidharth123-cpu/26/base -> origin/gh/Sidharth123-cpu/26/base 2025-12-04T09:33:40.3421292Z * [new branch] gh/Sidharth123-cpu/27/base -> origin/gh/Sidharth123-cpu/27/base 2025-12-04T09:33:40.3423456Z * [new branch] gh/StrongerXi/1/base -> origin/gh/StrongerXi/1/base 2025-12-04T09:33:40.3424554Z * [new branch] gh/StrongerXi/1/head -> origin/gh/StrongerXi/1/head 2025-12-04T09:33:40.3426355Z * [new branch] gh/StrongerXi/71/base -> origin/gh/StrongerXi/71/base 2025-12-04T09:33:40.3427517Z * [new branch] gh/StrongerXi/71/head -> origin/gh/StrongerXi/71/head 2025-12-04T09:33:40.3429154Z * [new branch] gh/StrongerXi/72/base -> origin/gh/StrongerXi/72/base 2025-12-04T09:33:40.3430341Z * [new branch] gh/StrongerXi/72/head -> origin/gh/StrongerXi/72/head 2025-12-04T09:33:40.3432090Z * [new branch] gh/StrongerXi/73/base -> origin/gh/StrongerXi/73/base 2025-12-04T09:33:40.3433286Z * [new branch] gh/StrongerXi/73/head -> origin/gh/StrongerXi/73/head 2025-12-04T09:33:40.3434674Z * [new branch] gh/StrongerXi/73/orig -> origin/gh/StrongerXi/73/orig 2025-12-04T09:33:40.3436858Z * [new branch] gh/XilunWu/160/base -> origin/gh/XilunWu/160/base 2025-12-04T09:33:40.3438038Z * [new branch] gh/XilunWu/160/head -> origin/gh/XilunWu/160/head 2025-12-04T09:33:40.3439426Z * [new branch] gh/XilunWu/160/orig -> origin/gh/XilunWu/160/orig 2025-12-04T09:33:40.3441224Z * [new branch] gh/XilunWu/163/base -> origin/gh/XilunWu/163/base 2025-12-04T09:33:40.3442483Z * [new branch] gh/XilunWu/163/head -> origin/gh/XilunWu/163/head 2025-12-04T09:33:40.3443743Z * [new branch] gh/XilunWu/163/orig -> origin/gh/XilunWu/163/orig 2025-12-04T09:33:40.3445675Z * [new branch] gh/XilunWu/168/base -> origin/gh/XilunWu/168/base 2025-12-04T09:33:40.3446813Z * [new branch] gh/XilunWu/168/head -> origin/gh/XilunWu/168/head 2025-12-04T09:33:40.3448216Z * [new branch] gh/XilunWu/168/orig -> origin/gh/XilunWu/168/orig 2025-12-04T09:33:40.3449884Z * [new branch] gh/XilunWu/169/base -> origin/gh/XilunWu/169/base 2025-12-04T09:33:40.3451053Z * [new branch] gh/XilunWu/169/head -> origin/gh/XilunWu/169/head 2025-12-04T09:33:40.3452338Z * [new branch] gh/XilunWu/169/orig -> origin/gh/XilunWu/169/orig 2025-12-04T09:33:40.3453939Z * [new branch] gh/XilunWu/170/base -> origin/gh/XilunWu/170/base 2025-12-04T09:33:40.3455126Z * [new branch] gh/XilunWu/170/head -> origin/gh/XilunWu/170/head 2025-12-04T09:33:40.3456362Z * [new branch] gh/XilunWu/170/orig -> origin/gh/XilunWu/170/orig 2025-12-04T09:33:40.3458485Z * [new branch] gh/XilunWu/171/base -> origin/gh/XilunWu/171/base 2025-12-04T09:33:40.3459670Z * [new branch] gh/XilunWu/171/head -> origin/gh/XilunWu/171/head 2025-12-04T09:33:40.3460956Z * [new branch] gh/XilunWu/171/orig -> origin/gh/XilunWu/171/orig 2025-12-04T09:33:40.3462648Z * [new branch] gh/XilunWu/173/base -> origin/gh/XilunWu/173/base 2025-12-04T09:33:40.3463915Z * [new branch] gh/XilunWu/173/head -> origin/gh/XilunWu/173/head 2025-12-04T09:33:40.3465321Z * [new branch] gh/XilunWu/173/orig -> origin/gh/XilunWu/173/orig 2025-12-04T09:33:40.3466985Z * [new branch] gh/XilunWu/175/base -> origin/gh/XilunWu/175/base 2025-12-04T09:33:40.3468238Z * [new branch] gh/XilunWu/175/head -> origin/gh/XilunWu/175/head 2025-12-04T09:33:40.3469646Z * [new branch] gh/XilunWu/175/orig -> origin/gh/XilunWu/175/orig 2025-12-04T09:33:40.3471350Z * [new branch] gh/XilunWu/176/base -> origin/gh/XilunWu/176/base 2025-12-04T09:33:40.3472576Z * [new branch] gh/XilunWu/176/head -> origin/gh/XilunWu/176/head 2025-12-04T09:33:40.3474072Z * [new branch] gh/XilunWu/176/orig -> origin/gh/XilunWu/176/orig 2025-12-04T09:33:40.3476110Z * [new branch] gh/XuehaiPan/14/base -> origin/gh/XuehaiPan/14/base 2025-12-04T09:33:40.3477324Z * [new branch] gh/XuehaiPan/14/head -> origin/gh/XuehaiPan/14/head 2025-12-04T09:33:40.3478586Z * [new branch] gh/XuehaiPan/14/orig -> origin/gh/XuehaiPan/14/orig 2025-12-04T09:33:40.3480461Z * [new branch] gh/XuehaiPan/179/base -> origin/gh/XuehaiPan/179/base 2025-12-04T09:33:40.3481675Z * [new branch] gh/XuehaiPan/179/head -> origin/gh/XuehaiPan/179/head 2025-12-04T09:33:40.3483140Z * [new branch] gh/XuehaiPan/179/orig -> origin/gh/XuehaiPan/179/orig 2025-12-04T09:33:40.3484816Z * [new branch] gh/XuehaiPan/249/base -> origin/gh/XuehaiPan/249/base 2025-12-04T09:33:40.3486032Z * [new branch] gh/XuehaiPan/249/head -> origin/gh/XuehaiPan/249/head 2025-12-04T09:33:40.3487450Z * [new branch] gh/XuehaiPan/249/orig -> origin/gh/XuehaiPan/249/orig 2025-12-04T09:33:40.3489143Z * [new branch] gh/XuehaiPan/253/base -> origin/gh/XuehaiPan/253/base 2025-12-04T09:33:40.3490335Z * [new branch] gh/XuehaiPan/253/head -> origin/gh/XuehaiPan/253/head 2025-12-04T09:33:40.3491996Z * [new branch] gh/XuehaiPan/253/orig -> origin/gh/XuehaiPan/253/orig 2025-12-04T09:33:40.3493791Z * [new branch] gh/XuehaiPan/254/base -> origin/gh/XuehaiPan/254/base 2025-12-04T09:33:40.3494995Z * [new branch] gh/XuehaiPan/254/head -> origin/gh/XuehaiPan/254/head 2025-12-04T09:33:40.3496666Z * [new branch] gh/XuehaiPan/254/orig -> origin/gh/XuehaiPan/254/orig 2025-12-04T09:33:40.3498319Z * [new branch] gh/XuehaiPan/255/base -> origin/gh/XuehaiPan/255/base 2025-12-04T09:33:40.3499486Z * [new branch] gh/XuehaiPan/255/head -> origin/gh/XuehaiPan/255/head 2025-12-04T09:33:40.3500730Z * [new branch] gh/XuehaiPan/255/orig -> origin/gh/XuehaiPan/255/orig 2025-12-04T09:33:40.3502523Z * [new branch] gh/XuehaiPan/271/base -> origin/gh/XuehaiPan/271/base 2025-12-04T09:33:40.3503758Z * [new branch] gh/XuehaiPan/271/head -> origin/gh/XuehaiPan/271/head 2025-12-04T09:33:40.3504976Z * [new branch] gh/XuehaiPan/271/orig -> origin/gh/XuehaiPan/271/orig 2025-12-04T09:33:40.3506731Z * [new branch] gh/XuehaiPan/343/base -> origin/gh/XuehaiPan/343/base 2025-12-04T09:33:40.3507946Z * [new branch] gh/XuehaiPan/343/head -> origin/gh/XuehaiPan/343/head 2025-12-04T09:33:40.3509247Z * [new branch] gh/XuehaiPan/343/orig -> origin/gh/XuehaiPan/343/orig 2025-12-04T09:33:40.3511069Z * [new branch] gh/XuehaiPan/347/base -> origin/gh/XuehaiPan/347/base 2025-12-04T09:33:40.3512244Z * [new branch] gh/XuehaiPan/347/head -> origin/gh/XuehaiPan/347/head 2025-12-04T09:33:40.3513573Z * [new branch] gh/XuehaiPan/347/orig -> origin/gh/XuehaiPan/347/orig 2025-12-04T09:33:40.3515328Z * [new branch] gh/XuehaiPan/348/base -> origin/gh/XuehaiPan/348/base 2025-12-04T09:33:40.3516483Z * [new branch] gh/XuehaiPan/348/head -> origin/gh/XuehaiPan/348/head 2025-12-04T09:33:40.3517754Z * [new branch] gh/XuehaiPan/348/orig -> origin/gh/XuehaiPan/348/orig 2025-12-04T09:33:40.3519497Z * [new branch] gh/XuehaiPan/350/base -> origin/gh/XuehaiPan/350/base 2025-12-04T09:33:40.3520697Z * [new branch] gh/XuehaiPan/350/head -> origin/gh/XuehaiPan/350/head 2025-12-04T09:33:40.3521976Z * [new branch] gh/XuehaiPan/350/orig -> origin/gh/XuehaiPan/350/orig 2025-12-04T09:33:40.3523926Z * [new branch] gh/XuehaiPan/365/base -> origin/gh/XuehaiPan/365/base 2025-12-04T09:33:40.3525054Z * [new branch] gh/XuehaiPan/365/head -> origin/gh/XuehaiPan/365/head 2025-12-04T09:33:40.3526327Z * [new branch] gh/XuehaiPan/365/orig -> origin/gh/XuehaiPan/365/orig 2025-12-04T09:33:40.3528166Z * [new branch] gh/XuehaiPan/366/base -> origin/gh/XuehaiPan/366/base 2025-12-04T09:33:40.3529378Z * [new branch] gh/XuehaiPan/366/head -> origin/gh/XuehaiPan/366/head 2025-12-04T09:33:40.3531143Z * [new branch] gh/XuehaiPan/370/base -> origin/gh/XuehaiPan/370/base 2025-12-04T09:33:40.3532306Z * [new branch] gh/XuehaiPan/370/head -> origin/gh/XuehaiPan/370/head 2025-12-04T09:33:40.3533626Z * [new branch] gh/XuehaiPan/370/orig -> origin/gh/XuehaiPan/370/orig 2025-12-04T09:33:40.3535470Z * [new branch] gh/XuehaiPan/390/base -> origin/gh/XuehaiPan/390/base 2025-12-04T09:33:40.3536706Z * [new branch] gh/XuehaiPan/390/head -> origin/gh/XuehaiPan/390/head 2025-12-04T09:33:40.3538062Z * [new branch] gh/XuehaiPan/390/orig -> origin/gh/XuehaiPan/390/orig 2025-12-04T09:33:40.3539872Z * [new branch] gh/XuehaiPan/391/base -> origin/gh/XuehaiPan/391/base 2025-12-04T09:33:40.3541038Z * [new branch] gh/XuehaiPan/391/head -> origin/gh/XuehaiPan/391/head 2025-12-04T09:33:40.3542328Z * [new branch] gh/XuehaiPan/391/orig -> origin/gh/XuehaiPan/391/orig 2025-12-04T09:33:40.3544141Z * [new branch] gh/XuehaiPan/392/base -> origin/gh/XuehaiPan/392/base 2025-12-04T09:33:40.3545302Z * [new branch] gh/XuehaiPan/392/head -> origin/gh/XuehaiPan/392/head 2025-12-04T09:33:40.3546557Z * [new branch] gh/XuehaiPan/392/orig -> origin/gh/XuehaiPan/392/orig 2025-12-04T09:33:40.3548887Z * [new branch] gh/XuehaiPan/394/base -> origin/gh/XuehaiPan/394/base 2025-12-04T09:33:40.3550106Z * [new branch] gh/XuehaiPan/394/head -> origin/gh/XuehaiPan/394/head 2025-12-04T09:33:40.3551416Z * [new branch] gh/XuehaiPan/394/orig -> origin/gh/XuehaiPan/394/orig 2025-12-04T09:33:40.3553259Z * [new branch] gh/XuehaiPan/397/base -> origin/gh/XuehaiPan/397/base 2025-12-04T09:33:40.3554473Z * [new branch] gh/XuehaiPan/397/head -> origin/gh/XuehaiPan/397/head 2025-12-04T09:33:40.3555725Z * [new branch] gh/XuehaiPan/397/orig -> origin/gh/XuehaiPan/397/orig 2025-12-04T09:33:40.3557581Z * [new branch] gh/XuehaiPan/398/base -> origin/gh/XuehaiPan/398/base 2025-12-04T09:33:40.3558800Z * [new branch] gh/XuehaiPan/398/head -> origin/gh/XuehaiPan/398/head 2025-12-04T09:33:40.3560070Z * [new branch] gh/XuehaiPan/398/orig -> origin/gh/XuehaiPan/398/orig 2025-12-04T09:33:40.3561825Z * [new branch] gh/XuehaiPan/399/base -> origin/gh/XuehaiPan/399/base 2025-12-04T09:33:40.3562995Z * [new branch] gh/XuehaiPan/399/head -> origin/gh/XuehaiPan/399/head 2025-12-04T09:33:40.3564318Z * [new branch] gh/XuehaiPan/399/orig -> origin/gh/XuehaiPan/399/orig 2025-12-04T09:33:40.3566162Z * [new branch] gh/XuehaiPan/400/base -> origin/gh/XuehaiPan/400/base 2025-12-04T09:33:40.3567335Z * [new branch] gh/XuehaiPan/400/head -> origin/gh/XuehaiPan/400/head 2025-12-04T09:33:40.3568612Z * [new branch] gh/XuehaiPan/400/orig -> origin/gh/XuehaiPan/400/orig 2025-12-04T09:33:40.3570775Z * [new branch] gh/ZhiweiYan-96/39/base -> origin/gh/ZhiweiYan-96/39/base 2025-12-04T09:33:40.3571976Z * [new branch] gh/ZhiweiYan-96/39/head -> origin/gh/ZhiweiYan-96/39/head 2025-12-04T09:33:40.3573283Z * [new branch] gh/ZhiweiYan-96/39/orig -> origin/gh/ZhiweiYan-96/39/orig 2025-12-04T09:33:40.3575220Z * [new branch] gh/ZhiweiYan-96/44/base -> origin/gh/ZhiweiYan-96/44/base 2025-12-04T09:33:40.3576356Z * [new branch] gh/ZhiweiYan-96/44/head -> origin/gh/ZhiweiYan-96/44/head 2025-12-04T09:33:40.3578228Z * [new branch] gh/ZhiweiYan-96/45/base -> origin/gh/ZhiweiYan-96/45/base 2025-12-04T09:33:40.3579351Z * [new branch] gh/ZhiweiYan-96/45/head -> origin/gh/ZhiweiYan-96/45/head 2025-12-04T09:33:40.3581166Z * [new branch] gh/ZhiweiYan-96/49/base -> origin/gh/ZhiweiYan-96/49/base 2025-12-04T09:33:40.3582349Z * [new branch] gh/ZhiweiYan-96/49/head -> origin/gh/ZhiweiYan-96/49/head 2025-12-04T09:33:40.3584213Z * [new branch] gh/ZhiweiYan-96/62/base -> origin/gh/ZhiweiYan-96/62/base 2025-12-04T09:33:40.3585397Z * [new branch] gh/ZhiweiYan-96/62/head -> origin/gh/ZhiweiYan-96/62/head 2025-12-04T09:33:40.3587236Z * [new branch] gh/ZhiweiYan-96/66/base -> origin/gh/ZhiweiYan-96/66/base 2025-12-04T09:33:40.3588423Z * [new branch] gh/ZhiweiYan-96/66/head -> origin/gh/ZhiweiYan-96/66/head 2025-12-04T09:33:40.3590145Z * [new branch] gh/ZhiweiYan-96/67/base -> origin/gh/ZhiweiYan-96/67/base 2025-12-04T09:33:40.3591266Z * [new branch] gh/ZhiweiYan-96/67/head -> origin/gh/ZhiweiYan-96/67/head 2025-12-04T09:33:40.3593091Z * [new branch] gh/ZhiweiYan-96/68/base -> origin/gh/ZhiweiYan-96/68/base 2025-12-04T09:33:40.3594120Z * [new branch] gh/ZhiweiYan-96/68/head -> origin/gh/ZhiweiYan-96/68/head 2025-12-04T09:33:40.3595390Z * [new branch] gh/ZhiweiYan-96/68/orig -> origin/gh/ZhiweiYan-96/68/orig 2025-12-04T09:33:40.3597858Z * [new branch] gh/aakhundov/1/base -> origin/gh/aakhundov/1/base 2025-12-04T09:33:40.3599135Z * [new branch] gh/aakhundov/1/head -> origin/gh/aakhundov/1/head 2025-12-04T09:33:40.3600836Z * [new branch] gh/aakhundov/2/base -> origin/gh/aakhundov/2/base 2025-12-04T09:33:40.3602090Z * [new branch] gh/aakhundov/2/head -> origin/gh/aakhundov/2/head 2025-12-04T09:33:40.3603980Z * [new branch] gh/aditew01/openblas -> origin/gh/aditew01/openblas 2025-12-04T09:33:40.3605098Z * [new branch] gh/aditew01/sbgemm -> origin/gh/aditew01/sbgemm 2025-12-04T09:33:40.3606481Z * [new branch] gh/aditew01/vecbf16 -> origin/gh/aditew01/vecbf16 2025-12-04T09:33:40.3608557Z * [new branch] gh/albanD/4/base -> origin/gh/albanD/4/base 2025-12-04T09:33:40.3609749Z * [new branch] gh/albanD/4/head -> origin/gh/albanD/4/head 2025-12-04T09:33:40.3611143Z * [new branch] gh/albanD/4/orig -> origin/gh/albanD/4/orig 2025-12-04T09:33:40.3613254Z * [new branch] gh/alexbrauckmann/paddedtensor_faketensor_init -> origin/gh/alexbrauckmann/paddedtensor_faketensor_init 2025-12-04T09:33:40.3615140Z * [new branch] gh/alexsamardzic/12/base -> origin/gh/alexsamardzic/12/base 2025-12-04T09:33:40.3616295Z * [new branch] gh/alexsamardzic/12/head -> origin/gh/alexsamardzic/12/head 2025-12-04T09:33:40.3617680Z * [new branch] gh/alexsamardzic/12/orig -> origin/gh/alexsamardzic/12/orig 2025-12-04T09:33:40.3619521Z * [new branch] gh/alexsamardzic/14/base -> origin/gh/alexsamardzic/14/base 2025-12-04T09:33:40.3620814Z * [new branch] gh/alexsamardzic/14/head -> origin/gh/alexsamardzic/14/head 2025-12-04T09:33:40.3622165Z * [new branch] gh/alexsamardzic/14/orig -> origin/gh/alexsamardzic/14/orig 2025-12-04T09:33:40.3624046Z * [new branch] gh/alexsamardzic/15/base -> origin/gh/alexsamardzic/15/base 2025-12-04T09:33:40.3625287Z * [new branch] gh/alexsamardzic/15/head -> origin/gh/alexsamardzic/15/head 2025-12-04T09:33:40.3626703Z * [new branch] gh/alexsamardzic/15/orig -> origin/gh/alexsamardzic/15/orig 2025-12-04T09:33:40.3628811Z * [new branch] gh/amjames/18/base -> origin/gh/amjames/18/base 2025-12-04T09:33:40.3629985Z * [new branch] gh/amjames/18/head -> origin/gh/amjames/18/head 2025-12-04T09:33:40.3631232Z * [new branch] gh/amjames/18/orig -> origin/gh/amjames/18/orig 2025-12-04T09:33:40.3633627Z * [new branch] gh/andrewor14/35/base -> origin/gh/andrewor14/35/base 2025-12-04T09:33:40.3635105Z * [new branch] gh/andrewor14/35/head -> origin/gh/andrewor14/35/head 2025-12-04T09:33:40.3636945Z * [new branch] gh/andrewor14/35/orig -> origin/gh/andrewor14/35/orig 2025-12-04T09:33:40.3638308Z * [new branch] gh/andrewor14/50/base -> origin/gh/andrewor14/50/base 2025-12-04T09:33:40.3639797Z * [new branch] gh/andrewor14/50/head -> origin/gh/andrewor14/50/head 2025-12-04T09:33:40.3641010Z * [new branch] gh/andrewor14/50/orig -> origin/gh/andrewor14/50/orig 2025-12-04T09:33:40.3643284Z * [new branch] gh/andyanwang/30/base -> origin/gh/andyanwang/30/base 2025-12-04T09:33:40.3644851Z * [new branch] gh/andyanwang/30/orig -> origin/gh/andyanwang/30/orig 2025-12-04T09:33:40.3646521Z * [new branch] gh/andyanwang/31/base -> origin/gh/andyanwang/31/base 2025-12-04T09:33:40.3648122Z * [new branch] gh/andyanwang/31/orig -> origin/gh/andyanwang/31/orig 2025-12-04T09:33:40.3649917Z * [new branch] gh/andyanwang/39/base -> origin/gh/andyanwang/39/base 2025-12-04T09:33:40.3651231Z * [new branch] gh/andyanwang/39/head -> origin/gh/andyanwang/39/head 2025-12-04T09:33:40.3652529Z * [new branch] gh/andyanwang/39/orig -> origin/gh/andyanwang/39/orig 2025-12-04T09:33:40.3654582Z * [new branch] gh/andyanwang/42/base -> origin/gh/andyanwang/42/base 2025-12-04T09:33:40.3655700Z * [new branch] gh/andyanwang/42/head -> origin/gh/andyanwang/42/head 2025-12-04T09:33:40.3657227Z * [new branch] gh/andyanwang/42/orig -> origin/gh/andyanwang/42/orig 2025-12-04T09:33:40.3659131Z * [new branch] gh/andyanwang/45/base -> origin/gh/andyanwang/45/base 2025-12-04T09:33:40.3660385Z * [new branch] gh/andyanwang/45/head -> origin/gh/andyanwang/45/head 2025-12-04T09:33:40.3661784Z * [new branch] gh/andyanwang/45/orig -> origin/gh/andyanwang/45/orig 2025-12-04T09:33:40.3663921Z * [new branch] gh/angelayi/107/base -> origin/gh/angelayi/107/base 2025-12-04T09:33:40.3665068Z * [new branch] gh/angelayi/107/head -> origin/gh/angelayi/107/head 2025-12-04T09:33:40.3666943Z * [new branch] gh/angelayi/114/base -> origin/gh/angelayi/114/base 2025-12-04T09:33:40.3668225Z * [new branch] gh/angelayi/114/head -> origin/gh/angelayi/114/head 2025-12-04T09:33:40.3669488Z * [new branch] gh/angelayi/114/orig -> origin/gh/angelayi/114/orig 2025-12-04T09:33:40.3671290Z * [new branch] gh/angelayi/116/base -> origin/gh/angelayi/116/base 2025-12-04T09:33:40.3672608Z * [new branch] gh/angelayi/116/head -> origin/gh/angelayi/116/head 2025-12-04T09:33:40.3673905Z * [new branch] gh/angelayi/116/orig -> origin/gh/angelayi/116/orig 2025-12-04T09:33:40.3675828Z * [new branch] gh/angelayi/122/base -> origin/gh/angelayi/122/base 2025-12-04T09:33:40.3676924Z * [new branch] gh/angelayi/122/head -> origin/gh/angelayi/122/head 2025-12-04T09:33:40.3678358Z * [new branch] gh/angelayi/122/orig -> origin/gh/angelayi/122/orig 2025-12-04T09:33:40.3680188Z * [new branch] gh/angelayi/124/base -> origin/gh/angelayi/124/base 2025-12-04T09:33:40.3681391Z * [new branch] gh/angelayi/124/head -> origin/gh/angelayi/124/head 2025-12-04T09:33:40.3682620Z * [new branch] gh/angelayi/124/orig -> origin/gh/angelayi/124/orig 2025-12-04T09:33:40.3684569Z * [new branch] gh/angelayi/128/base -> origin/gh/angelayi/128/base 2025-12-04T09:33:40.3685746Z * [new branch] gh/angelayi/128/head -> origin/gh/angelayi/128/head 2025-12-04T09:33:40.3686983Z * [new branch] gh/angelayi/128/orig -> origin/gh/angelayi/128/orig 2025-12-04T09:33:40.3688823Z * [new branch] gh/angelayi/131/base -> origin/gh/angelayi/131/base 2025-12-04T09:33:40.3690278Z * [new branch] gh/angelayi/131/head -> origin/gh/angelayi/131/head 2025-12-04T09:33:40.3691483Z * [new branch] gh/angelayi/131/orig -> origin/gh/angelayi/131/orig 2025-12-04T09:33:40.3693595Z * [new branch] gh/angelayi/132/base -> origin/gh/angelayi/132/base 2025-12-04T09:33:40.3695033Z * [new branch] gh/angelayi/132/head -> origin/gh/angelayi/132/head 2025-12-04T09:33:40.3696727Z * [new branch] gh/angelayi/132/orig -> origin/gh/angelayi/132/orig 2025-12-04T09:33:40.3698485Z * [new branch] gh/angelayi/133/base -> origin/gh/angelayi/133/base 2025-12-04T09:33:40.3699658Z * [new branch] gh/angelayi/133/head -> origin/gh/angelayi/133/head 2025-12-04T09:33:40.3700920Z * [new branch] gh/angelayi/133/orig -> origin/gh/angelayi/133/orig 2025-12-04T09:33:40.3703104Z * [new branch] gh/angelayi/134/base -> origin/gh/angelayi/134/base 2025-12-04T09:33:40.3704632Z * [new branch] gh/angelayi/134/head -> origin/gh/angelayi/134/head 2025-12-04T09:33:40.3705881Z * [new branch] gh/angelayi/134/orig -> origin/gh/angelayi/134/orig 2025-12-04T09:33:40.3707998Z * [new branch] gh/angelayi/135/base -> origin/gh/angelayi/135/base 2025-12-04T09:33:40.3709452Z * [new branch] gh/angelayi/135/head -> origin/gh/angelayi/135/head 2025-12-04T09:33:40.3710671Z * [new branch] gh/angelayi/135/orig -> origin/gh/angelayi/135/orig 2025-12-04T09:33:40.3712447Z * [new branch] gh/angelayi/136/base -> origin/gh/angelayi/136/base 2025-12-04T09:33:40.3713663Z * [new branch] gh/angelayi/136/head -> origin/gh/angelayi/136/head 2025-12-04T09:33:40.3714904Z * [new branch] gh/angelayi/136/orig -> origin/gh/angelayi/136/orig 2025-12-04T09:33:40.3716803Z * [new branch] gh/angelayi/137/base -> origin/gh/angelayi/137/base 2025-12-04T09:33:40.3717938Z * [new branch] gh/angelayi/137/head -> origin/gh/angelayi/137/head 2025-12-04T09:33:40.3719548Z * [new branch] gh/angelayi/137/orig -> origin/gh/angelayi/137/orig 2025-12-04T09:33:40.3721262Z * [new branch] gh/angelayi/138/base -> origin/gh/angelayi/138/base 2025-12-04T09:33:40.3722698Z * [new branch] gh/angelayi/138/head -> origin/gh/angelayi/138/head 2025-12-04T09:33:40.3724313Z * [new branch] gh/angelayi/138/orig -> origin/gh/angelayi/138/orig 2025-12-04T09:33:40.3725734Z * [new branch] gh/angelayi/139/base -> origin/gh/angelayi/139/base 2025-12-04T09:33:40.3726926Z * [new branch] gh/angelayi/139/head -> origin/gh/angelayi/139/head 2025-12-04T09:33:40.3728212Z * [new branch] gh/angelayi/139/orig -> origin/gh/angelayi/139/orig 2025-12-04T09:33:40.3730090Z * [new branch] gh/angelayi/140/base -> origin/gh/angelayi/140/base 2025-12-04T09:33:40.3731356Z * [new branch] gh/angelayi/140/head -> origin/gh/angelayi/140/head 2025-12-04T09:33:40.3732616Z * [new branch] gh/angelayi/140/orig -> origin/gh/angelayi/140/orig 2025-12-04T09:33:40.3735029Z * [new branch] gh/angelayi/141/base -> origin/gh/angelayi/141/base 2025-12-04T09:33:40.3736165Z * [new branch] gh/angelayi/141/head -> origin/gh/angelayi/141/head 2025-12-04T09:33:40.3737645Z * [new branch] gh/angelayi/141/orig -> origin/gh/angelayi/141/orig 2025-12-04T09:33:40.3739469Z * [new branch] gh/angelayi/142/base -> origin/gh/angelayi/142/base 2025-12-04T09:33:40.3740651Z * [new branch] gh/angelayi/142/head -> origin/gh/angelayi/142/head 2025-12-04T09:33:40.3741933Z * [new branch] gh/angelayi/142/orig -> origin/gh/angelayi/142/orig 2025-12-04T09:33:40.3743758Z * [new branch] gh/angelayi/143/base -> origin/gh/angelayi/143/base 2025-12-04T09:33:40.3744934Z * [new branch] gh/angelayi/143/head -> origin/gh/angelayi/143/head 2025-12-04T09:33:40.3746175Z * [new branch] gh/angelayi/143/orig -> origin/gh/angelayi/143/orig 2025-12-04T09:33:40.3748025Z * [new branch] gh/angelayi/144/base -> origin/gh/angelayi/144/base 2025-12-04T09:33:40.3749435Z * [new branch] gh/angelayi/144/head -> origin/gh/angelayi/144/head 2025-12-04T09:33:40.3750612Z * [new branch] gh/angelayi/144/orig -> origin/gh/angelayi/144/orig 2025-12-04T09:33:40.3752980Z * [new branch] gh/anijain2305/753/base -> origin/gh/anijain2305/753/base 2025-12-04T09:33:40.3754105Z * [new branch] gh/anijain2305/753/head -> origin/gh/anijain2305/753/head 2025-12-04T09:33:40.3755388Z * [new branch] gh/anijain2305/753/orig -> origin/gh/anijain2305/753/orig 2025-12-04T09:33:40.3757363Z * [new branch] gh/anijain2305/810/base -> origin/gh/anijain2305/810/base 2025-12-04T09:33:40.3758538Z * [new branch] gh/anijain2305/810/head -> origin/gh/anijain2305/810/head 2025-12-04T09:33:40.3759817Z * [new branch] gh/anijain2305/810/orig -> origin/gh/anijain2305/810/orig 2025-12-04T09:33:40.3761610Z * [new branch] gh/anijain2305/854/base -> origin/gh/anijain2305/854/base 2025-12-04T09:33:40.3762874Z * [new branch] gh/anijain2305/854/head -> origin/gh/anijain2305/854/head 2025-12-04T09:33:40.3764158Z * [new branch] gh/anijain2305/854/orig -> origin/gh/anijain2305/854/orig 2025-12-04T09:33:40.3766174Z * [new branch] gh/anijain2305/864/base -> origin/gh/anijain2305/864/base 2025-12-04T09:33:40.3767352Z * [new branch] gh/anijain2305/864/head -> origin/gh/anijain2305/864/head 2025-12-04T09:33:40.3768664Z * [new branch] gh/anijain2305/864/orig -> origin/gh/anijain2305/864/orig 2025-12-04T09:33:40.3770534Z * [new branch] gh/anijain2305/870/base -> origin/gh/anijain2305/870/base 2025-12-04T09:33:40.3771637Z * [new branch] gh/anijain2305/870/head -> origin/gh/anijain2305/870/head 2025-12-04T09:33:40.3773113Z * [new branch] gh/anijain2305/870/orig -> origin/gh/anijain2305/870/orig 2025-12-04T09:33:40.3774901Z * [new branch] gh/anijain2305/873/base -> origin/gh/anijain2305/873/base 2025-12-04T09:33:40.3776015Z * [new branch] gh/anijain2305/873/head -> origin/gh/anijain2305/873/head 2025-12-04T09:33:40.3777332Z * [new branch] gh/anijain2305/873/orig -> origin/gh/anijain2305/873/orig 2025-12-04T09:33:40.3779317Z * [new branch] gh/anijain2305/894/base -> origin/gh/anijain2305/894/base 2025-12-04T09:33:40.3780488Z * [new branch] gh/anijain2305/894/head -> origin/gh/anijain2305/894/head 2025-12-04T09:33:40.3781776Z * [new branch] gh/anijain2305/894/orig -> origin/gh/anijain2305/894/orig 2025-12-04T09:33:40.3783628Z * [new branch] gh/anijain2305/895/base -> origin/gh/anijain2305/895/base 2025-12-04T09:33:40.3784812Z * [new branch] gh/anijain2305/895/head -> origin/gh/anijain2305/895/head 2025-12-04T09:33:40.3786166Z * [new branch] gh/anijain2305/895/orig -> origin/gh/anijain2305/895/orig 2025-12-04T09:33:40.3788008Z * [new branch] gh/anijain2305/910/base -> origin/gh/anijain2305/910/base 2025-12-04T09:33:40.3789150Z * [new branch] gh/anijain2305/910/head -> origin/gh/anijain2305/910/head 2025-12-04T09:33:40.3790619Z * [new branch] gh/anijain2305/910/orig -> origin/gh/anijain2305/910/orig 2025-12-04T09:33:40.3792383Z * [new branch] gh/anijain2305/919/base -> origin/gh/anijain2305/919/base 2025-12-04T09:33:40.3793622Z * [new branch] gh/anijain2305/919/head -> origin/gh/anijain2305/919/head 2025-12-04T09:33:40.3794935Z * [new branch] gh/anijain2305/919/orig -> origin/gh/anijain2305/919/orig 2025-12-04T09:33:40.3800629Z * [new branch] gh/anijain2305/922/base -> origin/gh/anijain2305/922/base 2025-12-04T09:33:40.3801984Z * [new branch] gh/anijain2305/922/head -> origin/gh/anijain2305/922/head 2025-12-04T09:33:40.3803297Z * [new branch] gh/anijain2305/922/orig -> origin/gh/anijain2305/922/orig 2025-12-04T09:33:40.3805229Z * [new branch] gh/anijain2305/932/base -> origin/gh/anijain2305/932/base 2025-12-04T09:33:40.3806561Z * [new branch] gh/anijain2305/932/head -> origin/gh/anijain2305/932/head 2025-12-04T09:33:40.3807872Z * [new branch] gh/anijain2305/932/orig -> origin/gh/anijain2305/932/orig 2025-12-04T09:33:40.3809689Z * [new branch] gh/anijain2305/940/base -> origin/gh/anijain2305/940/base 2025-12-04T09:33:40.3810862Z * [new branch] gh/anijain2305/940/head -> origin/gh/anijain2305/940/head 2025-12-04T09:33:40.3812157Z * [new branch] gh/anijain2305/940/orig -> origin/gh/anijain2305/940/orig 2025-12-04T09:33:40.3813962Z * [new branch] gh/anijain2305/941/base -> origin/gh/anijain2305/941/base 2025-12-04T09:33:40.3815239Z * [new branch] gh/anijain2305/941/head -> origin/gh/anijain2305/941/head 2025-12-04T09:33:40.3816536Z * [new branch] gh/anijain2305/941/orig -> origin/gh/anijain2305/941/orig 2025-12-04T09:33:40.3818568Z * [new branch] gh/anijain2305/942/base -> origin/gh/anijain2305/942/base 2025-12-04T09:33:40.3819782Z * [new branch] gh/anijain2305/942/head -> origin/gh/anijain2305/942/head 2025-12-04T09:33:40.3821237Z * [new branch] gh/anijain2305/942/orig -> origin/gh/anijain2305/942/orig 2025-12-04T09:33:40.3823016Z * [new branch] gh/anijain2305/943/base -> origin/gh/anijain2305/943/base 2025-12-04T09:33:40.3824221Z * [new branch] gh/anijain2305/943/head -> origin/gh/anijain2305/943/head 2025-12-04T09:33:40.3825520Z * [new branch] gh/anijain2305/943/orig -> origin/gh/anijain2305/943/orig 2025-12-04T09:33:40.3827810Z * [new branch] gh/anijain2305/944/base -> origin/gh/anijain2305/944/base 2025-12-04T09:33:40.3828986Z * [new branch] gh/anijain2305/944/head -> origin/gh/anijain2305/944/head 2025-12-04T09:33:40.3831135Z * [new branch] gh/anijain2305/944/orig -> origin/gh/anijain2305/944/orig 2025-12-04T09:33:40.3832895Z * [new branch] gh/anijain2305/945/base -> origin/gh/anijain2305/945/base 2025-12-04T09:33:40.3834155Z * [new branch] gh/anijain2305/945/head -> origin/gh/anijain2305/945/head 2025-12-04T09:33:40.3835423Z * [new branch] gh/anijain2305/945/orig -> origin/gh/anijain2305/945/orig 2025-12-04T09:33:40.3837457Z * [new branch] gh/anijain2305/946/base -> origin/gh/anijain2305/946/base 2025-12-04T09:33:40.3838656Z * [new branch] gh/anijain2305/946/head -> origin/gh/anijain2305/946/head 2025-12-04T09:33:40.3839943Z * [new branch] gh/anijain2305/946/orig -> origin/gh/anijain2305/946/orig 2025-12-04T09:33:40.3841901Z * [new branch] gh/anijain2305/947/base -> origin/gh/anijain2305/947/base 2025-12-04T09:33:40.3843233Z * [new branch] gh/anijain2305/947/head -> origin/gh/anijain2305/947/head 2025-12-04T09:33:40.3844303Z * [new branch] gh/anijain2305/947/orig -> origin/gh/anijain2305/947/orig 2025-12-04T09:33:40.3846294Z * [new branch] gh/anijain2305/948/base -> origin/gh/anijain2305/948/base 2025-12-04T09:33:40.3847458Z * [new branch] gh/anijain2305/948/head -> origin/gh/anijain2305/948/head 2025-12-04T09:33:40.3848689Z * [new branch] gh/anijain2305/948/orig -> origin/gh/anijain2305/948/orig 2025-12-04T09:33:40.3850534Z * [new branch] gh/anijain2305/949/base -> origin/gh/anijain2305/949/base 2025-12-04T09:33:40.3851731Z * [new branch] gh/anijain2305/949/head -> origin/gh/anijain2305/949/head 2025-12-04T09:33:40.3852997Z * [new branch] gh/anijain2305/949/orig -> origin/gh/anijain2305/949/orig 2025-12-04T09:33:40.3854971Z * [new branch] gh/anijain2305/950/base -> origin/gh/anijain2305/950/base 2025-12-04T09:33:40.3856158Z * [new branch] gh/anijain2305/950/head -> origin/gh/anijain2305/950/head 2025-12-04T09:33:40.3857541Z * [new branch] gh/anijain2305/950/orig -> origin/gh/anijain2305/950/orig 2025-12-04T09:33:40.3859434Z * [new branch] gh/anijain2305/951/base -> origin/gh/anijain2305/951/base 2025-12-04T09:33:40.3860616Z * [new branch] gh/anijain2305/951/head -> origin/gh/anijain2305/951/head 2025-12-04T09:33:40.3862115Z * [new branch] gh/anijain2305/951/orig -> origin/gh/anijain2305/951/orig 2025-12-04T09:33:40.3863853Z * [new branch] gh/anijain2305/952/base -> origin/gh/anijain2305/952/base 2025-12-04T09:33:40.3865036Z * [new branch] gh/anijain2305/952/head -> origin/gh/anijain2305/952/head 2025-12-04T09:33:40.3866286Z * [new branch] gh/anijain2305/952/orig -> origin/gh/anijain2305/952/orig 2025-12-04T09:33:40.3868116Z * [new branch] gh/anijain2305/953/base -> origin/gh/anijain2305/953/base 2025-12-04T09:33:40.3869292Z * [new branch] gh/anijain2305/953/head -> origin/gh/anijain2305/953/head 2025-12-04T09:33:40.3870596Z * [new branch] gh/anijain2305/953/orig -> origin/gh/anijain2305/953/orig 2025-12-04T09:33:40.3872515Z * [new branch] gh/anijain2305/954/base -> origin/gh/anijain2305/954/base 2025-12-04T09:33:40.3873801Z * [new branch] gh/anijain2305/954/head -> origin/gh/anijain2305/954/head 2025-12-04T09:33:40.3875115Z * [new branch] gh/anijain2305/954/orig -> origin/gh/anijain2305/954/orig 2025-12-04T09:33:40.3877044Z * [new branch] gh/anijain2305/955/base -> origin/gh/anijain2305/955/base 2025-12-04T09:33:40.3878315Z * [new branch] gh/anijain2305/955/head -> origin/gh/anijain2305/955/head 2025-12-04T09:33:40.3879571Z * [new branch] gh/anijain2305/955/orig -> origin/gh/anijain2305/955/orig 2025-12-04T09:33:40.3881574Z * [new branch] gh/anijain2305/956/base -> origin/gh/anijain2305/956/base 2025-12-04T09:33:40.3882795Z * [new branch] gh/anijain2305/956/head -> origin/gh/anijain2305/956/head 2025-12-04T09:33:40.3884097Z * [new branch] gh/anijain2305/956/orig -> origin/gh/anijain2305/956/orig 2025-12-04T09:33:40.3886079Z * [new branch] gh/anijain2305/957/base -> origin/gh/anijain2305/957/base 2025-12-04T09:33:40.3887236Z * [new branch] gh/anijain2305/957/head -> origin/gh/anijain2305/957/head 2025-12-04T09:33:40.3888524Z * [new branch] gh/anijain2305/957/orig -> origin/gh/anijain2305/957/orig 2025-12-04T09:33:40.3890743Z * [new branch] gh/anijain2305/958/base -> origin/gh/anijain2305/958/base 2025-12-04T09:33:40.3891909Z * [new branch] gh/anijain2305/958/head -> origin/gh/anijain2305/958/head 2025-12-04T09:33:40.3893072Z * [new branch] gh/anijain2305/958/orig -> origin/gh/anijain2305/958/orig 2025-12-04T09:33:40.3894909Z * [new branch] gh/anijain2305/959/base -> origin/gh/anijain2305/959/base 2025-12-04T09:33:40.3896309Z * [new branch] gh/anijain2305/959/head -> origin/gh/anijain2305/959/head 2025-12-04T09:33:40.3897636Z * [new branch] gh/anijain2305/959/orig -> origin/gh/anijain2305/959/orig 2025-12-04T09:33:40.3899560Z * [new branch] gh/anijain2305/960/base -> origin/gh/anijain2305/960/base 2025-12-04T09:33:40.3900783Z * [new branch] gh/anijain2305/960/head -> origin/gh/anijain2305/960/head 2025-12-04T09:33:40.3902053Z * [new branch] gh/anijain2305/960/orig -> origin/gh/anijain2305/960/orig 2025-12-04T09:33:40.3903959Z * [new branch] gh/anijain2305/961/base -> origin/gh/anijain2305/961/base 2025-12-04T09:33:40.3905138Z * [new branch] gh/anijain2305/961/head -> origin/gh/anijain2305/961/head 2025-12-04T09:33:40.3906473Z * [new branch] gh/anijain2305/961/orig -> origin/gh/anijain2305/961/orig 2025-12-04T09:33:40.3908504Z * [new branch] gh/anijain2305/962/base -> origin/gh/anijain2305/962/base 2025-12-04T09:33:40.3909684Z * [new branch] gh/anijain2305/962/head -> origin/gh/anijain2305/962/head 2025-12-04T09:33:40.3910930Z * [new branch] gh/anijain2305/962/orig -> origin/gh/anijain2305/962/orig 2025-12-04T09:33:40.3913307Z * [new branch] gh/anijain2305/963/base -> origin/gh/anijain2305/963/base 2025-12-04T09:33:40.3914917Z * [new branch] gh/anijain2305/963/head -> origin/gh/anijain2305/963/head 2025-12-04T09:33:40.3916226Z * [new branch] gh/anijain2305/963/orig -> origin/gh/anijain2305/963/orig 2025-12-04T09:33:40.3918128Z * [new branch] gh/anijain2305/964/base -> origin/gh/anijain2305/964/base 2025-12-04T09:33:40.3919257Z * [new branch] gh/anijain2305/964/head -> origin/gh/anijain2305/964/head 2025-12-04T09:33:40.3920507Z * [new branch] gh/anijain2305/964/orig -> origin/gh/anijain2305/964/orig 2025-12-04T09:33:40.3922363Z * [new branch] gh/anijain2305/965/base -> origin/gh/anijain2305/965/base 2025-12-04T09:33:40.3923607Z * [new branch] gh/anijain2305/965/head -> origin/gh/anijain2305/965/head 2025-12-04T09:33:40.3925318Z * [new branch] gh/anijain2305/965/orig -> origin/gh/anijain2305/965/orig 2025-12-04T09:33:40.3926997Z * [new branch] gh/anijain2305/966/base -> origin/gh/anijain2305/966/base 2025-12-04T09:33:40.3928803Z * [new branch] gh/anijain2305/966/head -> origin/gh/anijain2305/966/head 2025-12-04T09:33:40.3930029Z * [new branch] gh/anijain2305/966/orig -> origin/gh/anijain2305/966/orig 2025-12-04T09:33:40.3931945Z * [new branch] gh/anijain2305/967/base -> origin/gh/anijain2305/967/base 2025-12-04T09:33:40.3933099Z * [new branch] gh/anijain2305/967/head -> origin/gh/anijain2305/967/head 2025-12-04T09:33:40.3934629Z * [new branch] gh/anijain2305/967/orig -> origin/gh/anijain2305/967/orig 2025-12-04T09:33:40.3936388Z * [new branch] gh/anijain2305/968/base -> origin/gh/anijain2305/968/base 2025-12-04T09:33:40.3937666Z * [new branch] gh/anijain2305/968/head -> origin/gh/anijain2305/968/head 2025-12-04T09:33:40.3939075Z * [new branch] gh/anijain2305/968/orig -> origin/gh/anijain2305/968/orig 2025-12-04T09:33:40.3940807Z * [new branch] gh/anijain2305/969/base -> origin/gh/anijain2305/969/base 2025-12-04T09:33:40.3942053Z * [new branch] gh/anijain2305/969/head -> origin/gh/anijain2305/969/head 2025-12-04T09:33:40.3943588Z * [new branch] gh/anijain2305/969/orig -> origin/gh/anijain2305/969/orig 2025-12-04T09:33:40.3945390Z * [new branch] gh/anijain2305/970/base -> origin/gh/anijain2305/970/base 2025-12-04T09:33:40.3946698Z * [new branch] gh/anijain2305/970/head -> origin/gh/anijain2305/970/head 2025-12-04T09:33:40.3947969Z * [new branch] gh/anijain2305/970/orig -> origin/gh/anijain2305/970/orig 2025-12-04T09:33:40.3950214Z * [new branch] gh/anjali411/216/base -> origin/gh/anjali411/216/base 2025-12-04T09:33:40.3951437Z * [new branch] gh/anjali411/216/head -> origin/gh/anjali411/216/head 2025-12-04T09:33:40.3952698Z * [new branch] gh/anjali411/216/orig -> origin/gh/anjali411/216/orig 2025-12-04T09:33:40.3955010Z * [new branch] gh/anshul-si/1/base -> origin/gh/anshul-si/1/base 2025-12-04T09:33:40.3956725Z * [new branch] gh/anshul-si/1/head -> origin/gh/anshul-si/1/head 2025-12-04T09:33:40.3958700Z * [new branch] gh/anshul-si/2/base -> origin/gh/anshul-si/2/base 2025-12-04T09:33:40.3959397Z * [new branch] gh/anshul-si/2/head -> origin/gh/anshul-si/2/head 2025-12-04T09:33:40.3961177Z * [new branch] gh/anshul-si/3/base -> origin/gh/anshul-si/3/base 2025-12-04T09:33:40.3962321Z * [new branch] gh/anshul-si/3/head -> origin/gh/anshul-si/3/head 2025-12-04T09:33:40.3963948Z * [new branch] gh/anshul-si/4/base -> origin/gh/anshul-si/4/base 2025-12-04T09:33:40.3965005Z * [new branch] gh/anshul-si/4/head -> origin/gh/anshul-si/4/head 2025-12-04T09:33:40.3966624Z * [new branch] gh/anshul-si/5/base -> origin/gh/anshul-si/5/base 2025-12-04T09:33:40.3967819Z * [new branch] gh/anshul-si/5/head -> origin/gh/anshul-si/5/head 2025-12-04T09:33:40.3969793Z * [new branch] gh/anshul-si/53/base -> origin/gh/anshul-si/53/base 2025-12-04T09:33:40.3971010Z * [new branch] gh/anshul-si/53/head -> origin/gh/anshul-si/53/head 2025-12-04T09:33:40.3972888Z * [new branch] gh/anshul-si/58/base -> origin/gh/anshul-si/58/base 2025-12-04T09:33:40.3974059Z * [new branch] gh/anshul-si/58/head -> origin/gh/anshul-si/58/head 2025-12-04T09:33:40.3975789Z * [new branch] gh/anshul-si/66/base -> origin/gh/anshul-si/66/base 2025-12-04T09:33:40.3977071Z * [new branch] gh/anshul-si/66/head -> origin/gh/anshul-si/66/head 2025-12-04T09:33:40.3978526Z * [new branch] gh/anshul-si/66/orig -> origin/gh/anshul-si/66/orig 2025-12-04T09:33:40.3980094Z * [new branch] gh/anshul-si/67/base -> origin/gh/anshul-si/67/base 2025-12-04T09:33:40.3981265Z * [new branch] gh/anshul-si/67/head -> origin/gh/anshul-si/67/head 2025-12-04T09:33:40.3982571Z * [new branch] gh/anshul-si/67/orig -> origin/gh/anshul-si/67/orig 2025-12-04T09:33:40.3984532Z * [new branch] gh/anshul-si/68/base -> origin/gh/anshul-si/68/base 2025-12-04T09:33:40.3985637Z * [new branch] gh/anshul-si/68/head -> origin/gh/anshul-si/68/head 2025-12-04T09:33:40.3986865Z * [new branch] gh/anshul-si/68/orig -> origin/gh/anshul-si/68/orig 2025-12-04T09:33:40.3988931Z * [new branch] gh/anshul-si/69/base -> origin/gh/anshul-si/69/base 2025-12-04T09:33:40.3990047Z * [new branch] gh/anshul-si/69/head -> origin/gh/anshul-si/69/head 2025-12-04T09:33:40.3991320Z * [new branch] gh/anshul-si/69/orig -> origin/gh/anshul-si/69/orig 2025-12-04T09:33:40.3993216Z * [new branch] gh/anshul-si/70/base -> origin/gh/anshul-si/70/base 2025-12-04T09:33:40.3994493Z * [new branch] gh/anshul-si/70/head -> origin/gh/anshul-si/70/head 2025-12-04T09:33:40.3995857Z * [new branch] gh/anshul-si/70/orig -> origin/gh/anshul-si/70/orig 2025-12-04T09:33:40.3997800Z * [new branch] gh/anshul-si/71/base -> origin/gh/anshul-si/71/base 2025-12-04T09:33:40.3999005Z * [new branch] gh/anshul-si/71/head -> origin/gh/anshul-si/71/head 2025-12-04T09:33:40.4000303Z * [new branch] gh/anshul-si/71/orig -> origin/gh/anshul-si/71/orig 2025-12-04T09:33:40.4002180Z * [new branch] gh/anshul-si/72/base -> origin/gh/anshul-si/72/base 2025-12-04T09:33:40.4003414Z * [new branch] gh/anshul-si/72/head -> origin/gh/anshul-si/72/head 2025-12-04T09:33:40.4004698Z * [new branch] gh/anshul-si/72/orig -> origin/gh/anshul-si/72/orig 2025-12-04T09:33:40.4006509Z * [new branch] gh/anshul-si/73/base -> origin/gh/anshul-si/73/base 2025-12-04T09:33:40.4007733Z * [new branch] gh/anshul-si/73/head -> origin/gh/anshul-si/73/head 2025-12-04T09:33:40.4009008Z * [new branch] gh/anshul-si/73/orig -> origin/gh/anshul-si/73/orig 2025-12-04T09:33:40.4011377Z * [new branch] gh/aorenste/132/base -> origin/gh/aorenste/132/base 2025-12-04T09:33:40.4012585Z * [new branch] gh/aorenste/132/head -> origin/gh/aorenste/132/head 2025-12-04T09:33:40.4014639Z * [new branch] gh/aorenste/134/base -> origin/gh/aorenste/134/base 2025-12-04T09:33:40.4016066Z * [new branch] gh/aorenste/134/head -> origin/gh/aorenste/134/head 2025-12-04T09:33:40.4017395Z * [new branch] gh/aorenste/134/orig -> origin/gh/aorenste/134/orig 2025-12-04T09:33:40.4019300Z * [new branch] gh/aorenste/139/base -> origin/gh/aorenste/139/base 2025-12-04T09:33:40.4020508Z * [new branch] gh/aorenste/139/head -> origin/gh/aorenste/139/head 2025-12-04T09:33:40.4021788Z * [new branch] gh/aorenste/139/orig -> origin/gh/aorenste/139/orig 2025-12-04T09:33:40.4023568Z * [new branch] gh/aorenste/141/base -> origin/gh/aorenste/141/base 2025-12-04T09:33:40.4024754Z * [new branch] gh/aorenste/141/head -> origin/gh/aorenste/141/head 2025-12-04T09:33:40.4027147Z * [new branch] gh/aorenste/145/base -> origin/gh/aorenste/145/base 2025-12-04T09:33:40.4028472Z * [new branch] gh/aorenste/145/head -> origin/gh/aorenste/145/head 2025-12-04T09:33:40.4029964Z * [new branch] gh/aorenste/145/orig -> origin/gh/aorenste/145/orig 2025-12-04T09:33:40.4031824Z * [new branch] gh/aorenste/146/base -> origin/gh/aorenste/146/base 2025-12-04T09:33:40.4033128Z * [new branch] gh/aorenste/146/head -> origin/gh/aorenste/146/head 2025-12-04T09:33:40.4034421Z * [new branch] gh/aorenste/146/orig -> origin/gh/aorenste/146/orig 2025-12-04T09:33:40.4036336Z * [new branch] gh/aorenste/147/base -> origin/gh/aorenste/147/base 2025-12-04T09:33:40.4037639Z * [new branch] gh/aorenste/147/head -> origin/gh/aorenste/147/head 2025-12-04T09:33:40.4038893Z * [new branch] gh/aorenste/147/orig -> origin/gh/aorenste/147/orig 2025-12-04T09:33:40.4040699Z * [new branch] gh/aorenste/148/base -> origin/gh/aorenste/148/base 2025-12-04T09:33:40.4041959Z * [new branch] gh/aorenste/148/head -> origin/gh/aorenste/148/head 2025-12-04T09:33:40.4043384Z * [new branch] gh/aorenste/148/orig -> origin/gh/aorenste/148/orig 2025-12-04T09:33:40.4045094Z * [new branch] gh/aorenste/149/base -> origin/gh/aorenste/149/base 2025-12-04T09:33:40.4046372Z * [new branch] gh/aorenste/149/head -> origin/gh/aorenste/149/head 2025-12-04T09:33:40.4047670Z * [new branch] gh/aorenste/149/orig -> origin/gh/aorenste/149/orig 2025-12-04T09:33:40.4049605Z * [new branch] gh/aorenste/150/base -> origin/gh/aorenste/150/base 2025-12-04T09:33:40.4050719Z * [new branch] gh/aorenste/150/head -> origin/gh/aorenste/150/head 2025-12-04T09:33:40.4052150Z * [new branch] gh/aorenste/150/orig -> origin/gh/aorenste/150/orig 2025-12-04T09:33:40.4053710Z * [new branch] gh/aorenste/151/base -> origin/gh/aorenste/151/base 2025-12-04T09:33:40.4054907Z * [new branch] gh/aorenste/151/head -> origin/gh/aorenste/151/head 2025-12-04T09:33:40.4056192Z * [new branch] gh/aorenste/151/orig -> origin/gh/aorenste/151/orig 2025-12-04T09:33:40.4058141Z * [new branch] gh/aorenste/152/base -> origin/gh/aorenste/152/base 2025-12-04T09:33:40.4059312Z * [new branch] gh/aorenste/152/head -> origin/gh/aorenste/152/head 2025-12-04T09:33:40.4060580Z * [new branch] gh/aorenste/152/orig -> origin/gh/aorenste/152/orig 2025-12-04T09:33:40.4062213Z * [new branch] gh/aorenste/153/base -> origin/gh/aorenste/153/base 2025-12-04T09:33:40.4063440Z * [new branch] gh/aorenste/153/head -> origin/gh/aorenste/153/head 2025-12-04T09:33:40.4064731Z * [new branch] gh/aorenste/153/orig -> origin/gh/aorenste/153/orig 2025-12-04T09:33:40.4066637Z * [new branch] gh/aorenste/154/base -> origin/gh/aorenste/154/base 2025-12-04T09:33:40.4067665Z * [new branch] gh/aorenste/154/head -> origin/gh/aorenste/154/head 2025-12-04T09:33:40.4068825Z * [new branch] gh/aorenste/154/orig -> origin/gh/aorenste/154/orig 2025-12-04T09:33:40.4070426Z * [new branch] gh/aorenste/155/base -> origin/gh/aorenste/155/base 2025-12-04T09:33:40.4071709Z * [new branch] gh/aorenste/155/head -> origin/gh/aorenste/155/head 2025-12-04T09:33:40.4073131Z * [new branch] gh/aorenste/155/orig -> origin/gh/aorenste/155/orig 2025-12-04T09:33:40.4074383Z * [new branch] gh/aorenste/156/base -> origin/gh/aorenste/156/base 2025-12-04T09:33:40.4075670Z * [new branch] gh/aorenste/156/head -> origin/gh/aorenste/156/head 2025-12-04T09:33:40.4076853Z * [new branch] gh/aorenste/156/orig -> origin/gh/aorenste/156/orig 2025-12-04T09:33:40.4078927Z * [new branch] gh/aorenste/157/base -> origin/gh/aorenste/157/base 2025-12-04T09:33:40.4080363Z * [new branch] gh/aorenste/157/head -> origin/gh/aorenste/157/head 2025-12-04T09:33:40.4081522Z * [new branch] gh/aorenste/157/orig -> origin/gh/aorenste/157/orig 2025-12-04T09:33:40.4083193Z * [new branch] gh/aorenste/158/base -> origin/gh/aorenste/158/base 2025-12-04T09:33:40.4084412Z * [new branch] gh/aorenste/158/head -> origin/gh/aorenste/158/head 2025-12-04T09:33:40.4085561Z * [new branch] gh/aorenste/158/orig -> origin/gh/aorenste/158/orig 2025-12-04T09:33:40.4087224Z * [new branch] gh/aorenste/159/base -> origin/gh/aorenste/159/base 2025-12-04T09:33:40.4088516Z * [new branch] gh/aorenste/159/head -> origin/gh/aorenste/159/head 2025-12-04T09:33:40.4089674Z * [new branch] gh/aorenste/159/orig -> origin/gh/aorenste/159/orig 2025-12-04T09:33:40.4091811Z * [new branch] gh/avikchaudhuri/1/base -> origin/gh/avikchaudhuri/1/base 2025-12-04T09:33:40.4093098Z * [new branch] gh/avikchaudhuri/1/head -> origin/gh/avikchaudhuri/1/head 2025-12-04T09:33:40.4094712Z * [new branch] gh/avikchaudhuri/2/base -> origin/gh/avikchaudhuri/2/base 2025-12-04T09:33:40.4096637Z * [new branch] gh/avikchaudhuri/2/head -> origin/gh/avikchaudhuri/2/head 2025-12-04T09:33:40.4100470Z * [new branch] gh/avikchaudhuri/2/orig -> origin/gh/avikchaudhuri/2/orig 2025-12-04T09:33:40.4103043Z * [new branch] gh/bdhirsh/666/base -> origin/gh/bdhirsh/666/base 2025-12-04T09:33:40.4104122Z * [new branch] gh/bdhirsh/666/head -> origin/gh/bdhirsh/666/head 2025-12-04T09:33:40.4105562Z * [new branch] gh/bdhirsh/666/orig -> origin/gh/bdhirsh/666/orig 2025-12-04T09:33:40.4107318Z * [new branch] gh/bdhirsh/668/base -> origin/gh/bdhirsh/668/base 2025-12-04T09:33:40.4108526Z * [new branch] gh/bdhirsh/668/head -> origin/gh/bdhirsh/668/head 2025-12-04T09:33:40.4109764Z * [new branch] gh/bdhirsh/668/orig -> origin/gh/bdhirsh/668/orig 2025-12-04T09:33:40.4112238Z * [new branch] gh/bdhirsh/669/base -> origin/gh/bdhirsh/669/base 2025-12-04T09:33:40.4113451Z * [new branch] gh/bdhirsh/669/head -> origin/gh/bdhirsh/669/head 2025-12-04T09:33:40.4114703Z * [new branch] gh/bdhirsh/669/orig -> origin/gh/bdhirsh/669/orig 2025-12-04T09:33:40.4116822Z * [new branch] gh/bdhirsh/670/base -> origin/gh/bdhirsh/670/base 2025-12-04T09:33:40.4118183Z * [new branch] gh/bdhirsh/670/head -> origin/gh/bdhirsh/670/head 2025-12-04T09:33:40.4119425Z * [new branch] gh/bdhirsh/670/orig -> origin/gh/bdhirsh/670/orig 2025-12-04T09:33:40.4121277Z * [new branch] gh/bdhirsh/672/base -> origin/gh/bdhirsh/672/base 2025-12-04T09:33:40.4122407Z * [new branch] gh/bdhirsh/672/head -> origin/gh/bdhirsh/672/head 2025-12-04T09:33:40.4123671Z * [new branch] gh/bdhirsh/672/orig -> origin/gh/bdhirsh/672/orig 2025-12-04T09:33:40.4125908Z * [new branch] gh/bdhirsh/675/base -> origin/gh/bdhirsh/675/base 2025-12-04T09:33:40.4127399Z * [new branch] gh/bdhirsh/675/head -> origin/gh/bdhirsh/675/head 2025-12-04T09:33:40.4128573Z * [new branch] gh/bdhirsh/675/orig -> origin/gh/bdhirsh/675/orig 2025-12-04T09:33:40.4130395Z * [new branch] gh/bdhirsh/676/base -> origin/gh/bdhirsh/676/base 2025-12-04T09:33:40.4131644Z * [new branch] gh/bdhirsh/676/head -> origin/gh/bdhirsh/676/head 2025-12-04T09:33:40.4132913Z * [new branch] gh/bdhirsh/676/orig -> origin/gh/bdhirsh/676/orig 2025-12-04T09:33:40.4134757Z * [new branch] gh/bdhirsh/677/base -> origin/gh/bdhirsh/677/base 2025-12-04T09:33:40.4136500Z * [new branch] gh/bdhirsh/677/head -> origin/gh/bdhirsh/677/head 2025-12-04T09:33:40.4137967Z * [new branch] gh/bdhirsh/677/orig -> origin/gh/bdhirsh/677/orig 2025-12-04T09:33:40.4139775Z * [new branch] gh/bdhirsh/678/base -> origin/gh/bdhirsh/678/base 2025-12-04T09:33:40.4141146Z * [new branch] gh/bdhirsh/678/head -> origin/gh/bdhirsh/678/head 2025-12-04T09:33:40.4142494Z * [new branch] gh/bdhirsh/678/orig -> origin/gh/bdhirsh/678/orig 2025-12-04T09:33:40.4144429Z * [new branch] gh/bdhirsh/679/base -> origin/gh/bdhirsh/679/base 2025-12-04T09:33:40.4145761Z * [new branch] gh/bdhirsh/679/head -> origin/gh/bdhirsh/679/head 2025-12-04T09:33:40.4147055Z * [new branch] gh/bdhirsh/679/orig -> origin/gh/bdhirsh/679/orig 2025-12-04T09:33:40.4148820Z * [new branch] gh/bdhirsh/680/base -> origin/gh/bdhirsh/680/base 2025-12-04T09:33:40.4150164Z * [new branch] gh/bdhirsh/680/head -> origin/gh/bdhirsh/680/head 2025-12-04T09:33:40.4151435Z * [new branch] gh/bdhirsh/680/orig -> origin/gh/bdhirsh/680/orig 2025-12-04T09:33:40.4153106Z * [new branch] gh/bdhirsh/681/base -> origin/gh/bdhirsh/681/base 2025-12-04T09:33:40.4154536Z * [new branch] gh/bdhirsh/681/head -> origin/gh/bdhirsh/681/head 2025-12-04T09:33:40.4155844Z * [new branch] gh/bdhirsh/681/orig -> origin/gh/bdhirsh/681/orig 2025-12-04T09:33:40.4157926Z * [new branch] gh/benjaminglass1/101/base -> origin/gh/benjaminglass1/101/base 2025-12-04T09:33:40.4159164Z * [new branch] gh/benjaminglass1/101/head -> origin/gh/benjaminglass1/101/head 2025-12-04T09:33:40.4160480Z * [new branch] gh/benjaminglass1/101/orig -> origin/gh/benjaminglass1/101/orig 2025-12-04T09:33:40.4162245Z * [new branch] gh/benjaminglass1/102/base -> origin/gh/benjaminglass1/102/base 2025-12-04T09:33:40.4163536Z * [new branch] gh/benjaminglass1/102/head -> origin/gh/benjaminglass1/102/head 2025-12-04T09:33:40.4164836Z * [new branch] gh/benjaminglass1/102/orig -> origin/gh/benjaminglass1/102/orig 2025-12-04T09:33:40.4166506Z * [new branch] gh/benjaminglass1/106/base -> origin/gh/benjaminglass1/106/base 2025-12-04T09:33:40.4167800Z * [new branch] gh/benjaminglass1/106/head -> origin/gh/benjaminglass1/106/head 2025-12-04T09:33:40.4169086Z * [new branch] gh/benjaminglass1/106/orig -> origin/gh/benjaminglass1/106/orig 2025-12-04T09:33:40.4170925Z * [new branch] gh/benjaminglass1/107/base -> origin/gh/benjaminglass1/107/base 2025-12-04T09:33:40.4172150Z * [new branch] gh/benjaminglass1/107/head -> origin/gh/benjaminglass1/107/head 2025-12-04T09:33:40.4173457Z * [new branch] gh/benjaminglass1/107/orig -> origin/gh/benjaminglass1/107/orig 2025-12-04T09:33:40.4175186Z * [new branch] gh/benjaminglass1/108/base -> origin/gh/benjaminglass1/108/base 2025-12-04T09:33:40.4176443Z * [new branch] gh/benjaminglass1/108/head -> origin/gh/benjaminglass1/108/head 2025-12-04T09:33:40.4177824Z * [new branch] gh/benjaminglass1/108/orig -> origin/gh/benjaminglass1/108/orig 2025-12-04T09:33:40.4179512Z * [new branch] gh/benjaminglass1/109/base -> origin/gh/benjaminglass1/109/base 2025-12-04T09:33:40.4180767Z * [new branch] gh/benjaminglass1/109/head -> origin/gh/benjaminglass1/109/head 2025-12-04T09:33:40.4182080Z * [new branch] gh/benjaminglass1/109/orig -> origin/gh/benjaminglass1/109/orig 2025-12-04T09:33:40.4183750Z * [new branch] gh/benjaminglass1/97/base -> origin/gh/benjaminglass1/97/base 2025-12-04T09:33:40.4185020Z * [new branch] gh/benjaminglass1/97/head -> origin/gh/benjaminglass1/97/head 2025-12-04T09:33:40.4186608Z * [new branch] gh/benjaminglass1/97/orig -> origin/gh/benjaminglass1/97/orig 2025-12-04T09:33:40.4188650Z * [new branch] gh/bobrenjc93/570/base -> origin/gh/bobrenjc93/570/base 2025-12-04T09:33:40.4189984Z * [new branch] gh/bobrenjc93/570/head -> origin/gh/bobrenjc93/570/head 2025-12-04T09:33:40.4191305Z * [new branch] gh/bobrenjc93/570/orig -> origin/gh/bobrenjc93/570/orig 2025-12-04T09:33:40.4192920Z * [new branch] gh/bobrenjc93/604/base -> origin/gh/bobrenjc93/604/base 2025-12-04T09:33:40.4194211Z * [new branch] gh/bobrenjc93/604/head -> origin/gh/bobrenjc93/604/head 2025-12-04T09:33:40.4195491Z * [new branch] gh/bobrenjc93/604/orig -> origin/gh/bobrenjc93/604/orig 2025-12-04T09:33:40.4197342Z * [new branch] gh/bobrenjc93/638/base -> origin/gh/bobrenjc93/638/base 2025-12-04T09:33:40.4198611Z * [new branch] gh/bobrenjc93/638/head -> origin/gh/bobrenjc93/638/head 2025-12-04T09:33:40.4199821Z * [new branch] gh/bobrenjc93/638/orig -> origin/gh/bobrenjc93/638/orig 2025-12-04T09:33:40.4201498Z * [new branch] gh/bobrenjc93/653/base -> origin/gh/bobrenjc93/653/base 2025-12-04T09:33:40.4202762Z * [new branch] gh/bobrenjc93/653/head -> origin/gh/bobrenjc93/653/head 2025-12-04T09:33:40.4204143Z * [new branch] gh/bobrenjc93/653/orig -> origin/gh/bobrenjc93/653/orig 2025-12-04T09:33:40.4206046Z * [new branch] gh/bobrenjc93/654/base -> origin/gh/bobrenjc93/654/base 2025-12-04T09:33:40.4207226Z * [new branch] gh/bobrenjc93/654/head -> origin/gh/bobrenjc93/654/head 2025-12-04T09:33:40.4208445Z * [new branch] gh/bobrenjc93/654/orig -> origin/gh/bobrenjc93/654/orig 2025-12-04T09:33:40.4210080Z * [new branch] gh/bobrenjc93/657/base -> origin/gh/bobrenjc93/657/base 2025-12-04T09:33:40.4211373Z * [new branch] gh/bobrenjc93/657/head -> origin/gh/bobrenjc93/657/head 2025-12-04T09:33:40.4212626Z * [new branch] gh/bobrenjc93/657/orig -> origin/gh/bobrenjc93/657/orig 2025-12-04T09:33:40.4214797Z * [new branch] gh/bobrenjc93/672/base -> origin/gh/bobrenjc93/672/base 2025-12-04T09:33:40.4215984Z * [new branch] gh/bobrenjc93/672/head -> origin/gh/bobrenjc93/672/head 2025-12-04T09:33:40.4217344Z * [new branch] gh/bobrenjc93/672/orig -> origin/gh/bobrenjc93/672/orig 2025-12-04T09:33:40.4219137Z * [new branch] gh/bobrenjc93/679/base -> origin/gh/bobrenjc93/679/base 2025-12-04T09:33:40.4220663Z * [new branch] gh/bobrenjc93/679/head -> origin/gh/bobrenjc93/679/head 2025-12-04T09:33:40.4222096Z * [new branch] gh/bobrenjc93/679/orig -> origin/gh/bobrenjc93/679/orig 2025-12-04T09:33:40.4223880Z * [new branch] gh/bobrenjc93/680/base -> origin/gh/bobrenjc93/680/base 2025-12-04T09:33:40.4225118Z * [new branch] gh/bobrenjc93/680/head -> origin/gh/bobrenjc93/680/head 2025-12-04T09:33:40.4226425Z * [new branch] gh/bobrenjc93/680/orig -> origin/gh/bobrenjc93/680/orig 2025-12-04T09:33:40.4227921Z * [new branch] gh/bobrenjc93/681/base -> origin/gh/bobrenjc93/681/base 2025-12-04T09:33:40.4229210Z * [new branch] gh/bobrenjc93/681/head -> origin/gh/bobrenjc93/681/head 2025-12-04T09:33:40.4230487Z * [new branch] gh/bobrenjc93/681/orig -> origin/gh/bobrenjc93/681/orig 2025-12-04T09:33:40.4231973Z * [new branch] gh/bobrenjc93/682/base -> origin/gh/bobrenjc93/682/base 2025-12-04T09:33:40.4233302Z * [new branch] gh/bobrenjc93/682/head -> origin/gh/bobrenjc93/682/head 2025-12-04T09:33:40.4234557Z * [new branch] gh/bobrenjc93/682/orig -> origin/gh/bobrenjc93/682/orig 2025-12-04T09:33:40.4236272Z * [new branch] gh/bobrenjc93/683/base -> origin/gh/bobrenjc93/683/base 2025-12-04T09:33:40.4237506Z * [new branch] gh/bobrenjc93/683/head -> origin/gh/bobrenjc93/683/head 2025-12-04T09:33:40.4238929Z * [new branch] gh/bobrenjc93/683/orig -> origin/gh/bobrenjc93/683/orig 2025-12-04T09:33:40.4240679Z * [new branch] gh/bobrenjc93/684/base -> origin/gh/bobrenjc93/684/base 2025-12-04T09:33:40.4242153Z * [new branch] gh/bobrenjc93/684/head -> origin/gh/bobrenjc93/684/head 2025-12-04T09:33:40.4243622Z * [new branch] gh/bobrenjc93/684/orig -> origin/gh/bobrenjc93/684/orig 2025-12-04T09:33:40.4245161Z * [new branch] gh/bobrenjc93/685/base -> origin/gh/bobrenjc93/685/base 2025-12-04T09:33:40.4246747Z * [new branch] gh/bobrenjc93/685/head -> origin/gh/bobrenjc93/685/head 2025-12-04T09:33:40.4248363Z * [new branch] gh/bobrenjc93/685/orig -> origin/gh/bobrenjc93/685/orig 2025-12-04T09:33:40.4250287Z * [new branch] gh/bobrenjc93/686/base -> origin/gh/bobrenjc93/686/base 2025-12-04T09:33:40.4253869Z * [new branch] gh/bobrenjc93/686/head -> origin/gh/bobrenjc93/686/head 2025-12-04T09:33:40.4254135Z * [new branch] gh/bobrenjc93/686/orig -> origin/gh/bobrenjc93/686/orig 2025-12-04T09:33:40.4254951Z * [new branch] gh/bobrenjc93/687/base -> origin/gh/bobrenjc93/687/base 2025-12-04T09:33:40.4255883Z * [new branch] gh/bobrenjc93/687/head -> origin/gh/bobrenjc93/687/head 2025-12-04T09:33:40.4257381Z * [new branch] gh/bobrenjc93/687/orig -> origin/gh/bobrenjc93/687/orig 2025-12-04T09:33:40.4259721Z * [new branch] gh/bobrenjc93/688/base -> origin/gh/bobrenjc93/688/base 2025-12-04T09:33:40.4261004Z * [new branch] gh/bobrenjc93/688/head -> origin/gh/bobrenjc93/688/head 2025-12-04T09:33:40.4262309Z * [new branch] gh/bobrenjc93/688/orig -> origin/gh/bobrenjc93/688/orig 2025-12-04T09:33:40.4263959Z * [new branch] gh/bobrenjc93/689/base -> origin/gh/bobrenjc93/689/base 2025-12-04T09:33:40.4265342Z * [new branch] gh/bobrenjc93/689/head -> origin/gh/bobrenjc93/689/head 2025-12-04T09:33:40.4266650Z * [new branch] gh/bobrenjc93/689/orig -> origin/gh/bobrenjc93/689/orig 2025-12-04T09:33:40.4268271Z * [new branch] gh/bobrenjc93/690/base -> origin/gh/bobrenjc93/690/base 2025-12-04T09:33:40.4269510Z * [new branch] gh/bobrenjc93/690/head -> origin/gh/bobrenjc93/690/head 2025-12-04T09:33:40.4270771Z * [new branch] gh/bobrenjc93/690/orig -> origin/gh/bobrenjc93/690/orig 2025-12-04T09:33:40.4273288Z * [new branch] gh/bobrenjc93/691/base -> origin/gh/bobrenjc93/691/base 2025-12-04T09:33:40.4274923Z * [new branch] gh/bobrenjc93/691/head -> origin/gh/bobrenjc93/691/head 2025-12-04T09:33:40.4276756Z * [new branch] gh/bobrenjc93/691/orig -> origin/gh/bobrenjc93/691/orig 2025-12-04T09:33:40.4279241Z * [new branch] gh/bobrenjc93/692/base -> origin/gh/bobrenjc93/692/base 2025-12-04T09:33:40.4280579Z * [new branch] gh/bobrenjc93/692/head -> origin/gh/bobrenjc93/692/head 2025-12-04T09:33:40.4281869Z * [new branch] gh/bobrenjc93/692/orig -> origin/gh/bobrenjc93/692/orig 2025-12-04T09:33:40.4283475Z * [new branch] gh/bobrenjc93/693/base -> origin/gh/bobrenjc93/693/base 2025-12-04T09:33:40.4284793Z * [new branch] gh/bobrenjc93/693/head -> origin/gh/bobrenjc93/693/head 2025-12-04T09:33:40.4286158Z * [new branch] gh/bobrenjc93/693/orig -> origin/gh/bobrenjc93/693/orig 2025-12-04T09:33:40.4287901Z * [new branch] gh/bobrenjc93/694/base -> origin/gh/bobrenjc93/694/base 2025-12-04T09:33:40.4289181Z * [new branch] gh/bobrenjc93/694/head -> origin/gh/bobrenjc93/694/head 2025-12-04T09:33:40.4290538Z * [new branch] gh/bobrenjc93/694/orig -> origin/gh/bobrenjc93/694/orig 2025-12-04T09:33:40.4292105Z * [new branch] gh/bobrenjc93/695/base -> origin/gh/bobrenjc93/695/base 2025-12-04T09:33:40.4293462Z * [new branch] gh/bobrenjc93/695/head -> origin/gh/bobrenjc93/695/head 2025-12-04T09:33:40.4294769Z * [new branch] gh/bobrenjc93/695/orig -> origin/gh/bobrenjc93/695/orig 2025-12-04T09:33:40.4297129Z * [new branch] gh/c00w/23/base -> origin/gh/c00w/23/base 2025-12-04T09:33:40.4298548Z * [new branch] gh/c00w/23/head -> origin/gh/c00w/23/head 2025-12-04T09:33:40.4300318Z * [new branch] gh/c00w/53/base -> origin/gh/c00w/53/base 2025-12-04T09:33:40.4301532Z * [new branch] gh/c00w/53/head -> origin/gh/c00w/53/head 2025-12-04T09:33:40.4302762Z * [new branch] gh/c00w/53/orig -> origin/gh/c00w/53/orig 2025-12-04T09:33:40.4304267Z * [new branch] gh/c00w/54/base -> origin/gh/c00w/54/base 2025-12-04T09:33:40.4305608Z * [new branch] gh/c00w/54/head -> origin/gh/c00w/54/head 2025-12-04T09:33:40.4306900Z * [new branch] gh/c00w/54/orig -> origin/gh/c00w/54/orig 2025-12-04T09:33:40.4308605Z * [new branch] gh/c00w/56/base -> origin/gh/c00w/56/base 2025-12-04T09:33:40.4309941Z * [new branch] gh/c00w/56/head -> origin/gh/c00w/56/head 2025-12-04T09:33:40.4311254Z * [new branch] gh/c00w/56/orig -> origin/gh/c00w/56/orig 2025-12-04T09:33:40.4312853Z * [new branch] gh/c00w/57/base -> origin/gh/c00w/57/base 2025-12-04T09:33:40.4314103Z * [new branch] gh/c00w/57/head -> origin/gh/c00w/57/head 2025-12-04T09:33:40.4315415Z * [new branch] gh/c00w/57/orig -> origin/gh/c00w/57/orig 2025-12-04T09:33:40.4316947Z * [new branch] gh/c00w/58/base -> origin/gh/c00w/58/base 2025-12-04T09:33:40.4318264Z * [new branch] gh/c00w/58/head -> origin/gh/c00w/58/head 2025-12-04T09:33:40.4319544Z * [new branch] gh/c00w/58/orig -> origin/gh/c00w/58/orig 2025-12-04T09:33:40.4321611Z * [new branch] gh/clee2000/1/base -> origin/gh/clee2000/1/base 2025-12-04T09:33:40.4322959Z * [new branch] gh/clee2000/1/head -> origin/gh/clee2000/1/head 2025-12-04T09:33:40.4324208Z * [new branch] gh/clee2000/1/orig -> origin/gh/clee2000/1/orig 2025-12-04T09:33:40.4326977Z * [new branch] gh/coconutruben/1/base -> origin/gh/coconutruben/1/base 2025-12-04T09:33:40.4328436Z * [new branch] gh/coconutruben/1/head -> origin/gh/coconutruben/1/head 2025-12-04T09:33:40.4330501Z * [new branch] gh/coconutruben/55/base -> origin/gh/coconutruben/55/base 2025-12-04T09:33:40.4331792Z * [new branch] gh/coconutruben/55/head -> origin/gh/coconutruben/55/head 2025-12-04T09:33:40.4333110Z * [new branch] gh/coconutruben/55/orig -> origin/gh/coconutruben/55/orig 2025-12-04T09:33:40.4334963Z * [new branch] gh/coconutruben/57/base -> origin/gh/coconutruben/57/base 2025-12-04T09:33:40.4336425Z * [new branch] gh/coconutruben/57/head -> origin/gh/coconutruben/57/head 2025-12-04T09:33:40.4337960Z * [new branch] gh/coconutruben/57/orig -> origin/gh/coconutruben/57/orig 2025-12-04T09:33:40.4339680Z * [new branch] gh/coconutruben/70/base -> origin/gh/coconutruben/70/base 2025-12-04T09:33:40.4341012Z * [new branch] gh/coconutruben/70/head -> origin/gh/coconutruben/70/head 2025-12-04T09:33:40.4342369Z * [new branch] gh/coconutruben/70/orig -> origin/gh/coconutruben/70/orig 2025-12-04T09:33:40.4343932Z * [new branch] gh/coconutruben/71/base -> origin/gh/coconutruben/71/base 2025-12-04T09:33:40.4345318Z * [new branch] gh/coconutruben/71/head -> origin/gh/coconutruben/71/head 2025-12-04T09:33:40.4346607Z * [new branch] gh/coconutruben/71/orig -> origin/gh/coconutruben/71/orig 2025-12-04T09:33:40.4348175Z * [new branch] gh/coconutruben/72/base -> origin/gh/coconutruben/72/base 2025-12-04T09:33:40.4349679Z * [new branch] gh/coconutruben/72/head -> origin/gh/coconutruben/72/head 2025-12-04T09:33:40.4350780Z * [new branch] gh/coconutruben/72/orig -> origin/gh/coconutruben/72/orig 2025-12-04T09:33:40.4352287Z * [new branch] gh/coconutruben/73/base -> origin/gh/coconutruben/73/base 2025-12-04T09:33:40.4353566Z * [new branch] gh/coconutruben/73/head -> origin/gh/coconutruben/73/head 2025-12-04T09:33:40.4354868Z * [new branch] gh/coconutruben/73/orig -> origin/gh/coconutruben/73/orig 2025-12-04T09:33:40.4356724Z * [new branch] gh/coconutruben/74/base -> origin/gh/coconutruben/74/base 2025-12-04T09:33:40.4358166Z * [new branch] gh/coconutruben/74/head -> origin/gh/coconutruben/74/head 2025-12-04T09:33:40.4359465Z * [new branch] gh/coconutruben/74/orig -> origin/gh/coconutruben/74/orig 2025-12-04T09:33:40.4361413Z * [new branch] gh/coconutruben/79/base -> origin/gh/coconutruben/79/base 2025-12-04T09:33:40.4363013Z * [new branch] gh/coconutruben/79/head -> origin/gh/coconutruben/79/head 2025-12-04T09:33:40.4364243Z * [new branch] gh/coconutruben/79/orig -> origin/gh/coconutruben/79/orig 2025-12-04T09:33:40.4366028Z * [new branch] gh/coconutruben/80/base -> origin/gh/coconutruben/80/base 2025-12-04T09:33:40.4367332Z * [new branch] gh/coconutruben/80/head -> origin/gh/coconutruben/80/head 2025-12-04T09:33:40.4368655Z * [new branch] gh/coconutruben/80/orig -> origin/gh/coconutruben/80/orig 2025-12-04T09:33:40.4370425Z * [new branch] gh/coconutruben/82/base -> origin/gh/coconutruben/82/base 2025-12-04T09:33:40.4371638Z * [new branch] gh/coconutruben/82/head -> origin/gh/coconutruben/82/head 2025-12-04T09:33:40.4372842Z * [new branch] gh/coconutruben/82/orig -> origin/gh/coconutruben/82/orig 2025-12-04T09:33:40.4374704Z * [new branch] gh/coconutruben/83/base -> origin/gh/coconutruben/83/base 2025-12-04T09:33:40.4375947Z * [new branch] gh/coconutruben/83/head -> origin/gh/coconutruben/83/head 2025-12-04T09:33:40.4377329Z * [new branch] gh/coconutruben/83/orig -> origin/gh/coconutruben/83/orig 2025-12-04T09:33:40.4379311Z * [new branch] gh/coconutruben/84/base -> origin/gh/coconutruben/84/base 2025-12-04T09:33:40.4381050Z * [new branch] gh/coconutruben/84/head -> origin/gh/coconutruben/84/head 2025-12-04T09:33:40.4381884Z * [new branch] gh/coconutruben/84/orig -> origin/gh/coconutruben/84/orig 2025-12-04T09:33:40.4383686Z * [new branch] gh/coconutruben/85/base -> origin/gh/coconutruben/85/base 2025-12-04T09:33:40.4385029Z * [new branch] gh/coconutruben/85/head -> origin/gh/coconutruben/85/head 2025-12-04T09:33:40.4386353Z * [new branch] gh/coconutruben/85/orig -> origin/gh/coconutruben/85/orig 2025-12-04T09:33:40.4388067Z * [new branch] gh/coconutruben/86/base -> origin/gh/coconutruben/86/base 2025-12-04T09:33:40.4389365Z * [new branch] gh/coconutruben/86/head -> origin/gh/coconutruben/86/head 2025-12-04T09:33:40.4390647Z * [new branch] gh/coconutruben/86/orig -> origin/gh/coconutruben/86/orig 2025-12-04T09:33:40.4392777Z * [new branch] gh/colinchan15/1/base -> origin/gh/colinchan15/1/base 2025-12-04T09:33:40.4394062Z * [new branch] gh/colinchan15/1/head -> origin/gh/colinchan15/1/head 2025-12-04T09:33:40.4395624Z * [new branch] gh/colinchan15/2/base -> origin/gh/colinchan15/2/base 2025-12-04T09:33:40.4397191Z * [new branch] gh/colinchan15/2/head -> origin/gh/colinchan15/2/head 2025-12-04T09:33:40.4398700Z * [new branch] gh/colinchan15/3/base -> origin/gh/colinchan15/3/base 2025-12-04T09:33:40.4399956Z * [new branch] gh/colinchan15/3/head -> origin/gh/colinchan15/3/head 2025-12-04T09:33:40.4401976Z * [new branch] gh/colinchan15/6/base -> origin/gh/colinchan15/6/base 2025-12-04T09:33:40.4403241Z * [new branch] gh/colinchan15/6/head -> origin/gh/colinchan15/6/head 2025-12-04T09:33:40.4405232Z * [new branch] gh/d4l3k/1/base -> origin/gh/d4l3k/1/base 2025-12-04T09:33:40.4406489Z * [new branch] gh/d4l3k/1/head -> origin/gh/d4l3k/1/head 2025-12-04T09:33:40.4408246Z * [new branch] gh/d4l3k/2/base -> origin/gh/d4l3k/2/base 2025-12-04T09:33:40.4409490Z * [new branch] gh/d4l3k/2/head -> origin/gh/d4l3k/2/head 2025-12-04T09:33:40.4410712Z * [new branch] gh/d4l3k/2/orig -> origin/gh/d4l3k/2/orig 2025-12-04T09:33:40.4412487Z * [new branch] gh/d4l3k/3/base -> origin/gh/d4l3k/3/base 2025-12-04T09:33:40.4413701Z * [new branch] gh/d4l3k/3/head -> origin/gh/d4l3k/3/head 2025-12-04T09:33:40.4415092Z * [new branch] gh/d4l3k/3/orig -> origin/gh/d4l3k/3/orig 2025-12-04T09:33:40.4417102Z * [new branch] gh/d4l3k/4/base -> origin/gh/d4l3k/4/base 2025-12-04T09:33:40.4418520Z * [new branch] gh/d4l3k/4/head -> origin/gh/d4l3k/4/head 2025-12-04T09:33:40.4419756Z * [new branch] gh/d4l3k/4/orig -> origin/gh/d4l3k/4/orig 2025-12-04T09:33:40.4421411Z * [new branch] gh/d4l3k/5/base -> origin/gh/d4l3k/5/base 2025-12-04T09:33:40.4422649Z * [new branch] gh/d4l3k/5/orig -> origin/gh/d4l3k/5/orig 2025-12-04T09:33:40.4424869Z * [new branch] gh/davidberard98/392/base -> origin/gh/davidberard98/392/base 2025-12-04T09:33:40.4426105Z * [new branch] gh/davidberard98/392/head -> origin/gh/davidberard98/392/head 2025-12-04T09:33:40.4427409Z * [new branch] gh/davidberard98/392/orig -> origin/gh/davidberard98/392/orig 2025-12-04T09:33:40.4429345Z * [new branch] gh/davidberard98/399/base -> origin/gh/davidberard98/399/base 2025-12-04T09:33:40.4430704Z * [new branch] gh/davidberard98/399/head -> origin/gh/davidberard98/399/head 2025-12-04T09:33:40.4431991Z * [new branch] gh/davidberard98/399/orig -> origin/gh/davidberard98/399/orig 2025-12-04T09:33:40.4434020Z * [new branch] gh/desertfire/605/base -> origin/gh/desertfire/605/base 2025-12-04T09:33:40.4435255Z * [new branch] gh/desertfire/605/head -> origin/gh/desertfire/605/head 2025-12-04T09:33:40.4436537Z * [new branch] gh/desertfire/605/orig -> origin/gh/desertfire/605/orig 2025-12-04T09:33:40.4438207Z * [new branch] gh/desertfire/606/base -> origin/gh/desertfire/606/base 2025-12-04T09:33:40.4439461Z * [new branch] gh/desertfire/606/head -> origin/gh/desertfire/606/head 2025-12-04T09:33:40.4441422Z * [new branch] gh/desertfire/606/orig -> origin/gh/desertfire/606/orig 2025-12-04T09:33:40.4443125Z * [new branch] gh/desertfire/607/base -> origin/gh/desertfire/607/base 2025-12-04T09:33:40.4444344Z * [new branch] gh/desertfire/607/head -> origin/gh/desertfire/607/head 2025-12-04T09:33:40.4445617Z * [new branch] gh/desertfire/607/orig -> origin/gh/desertfire/607/orig 2025-12-04T09:33:40.4447420Z * [new branch] gh/desertfire/608/base -> origin/gh/desertfire/608/base 2025-12-04T09:33:40.4448671Z * [new branch] gh/desertfire/608/head -> origin/gh/desertfire/608/head 2025-12-04T09:33:40.4450001Z * [new branch] gh/desertfire/608/orig -> origin/gh/desertfire/608/orig 2025-12-04T09:33:40.4451645Z * [new branch] gh/desertfire/609/base -> origin/gh/desertfire/609/base 2025-12-04T09:33:40.4452954Z * [new branch] gh/desertfire/609/head -> origin/gh/desertfire/609/head 2025-12-04T09:33:40.4454292Z * [new branch] gh/desertfire/609/orig -> origin/gh/desertfire/609/orig 2025-12-04T09:33:40.4456172Z * [new branch] gh/desertfire/610/base -> origin/gh/desertfire/610/base 2025-12-04T09:33:40.4457558Z * [new branch] gh/desertfire/610/head -> origin/gh/desertfire/610/head 2025-12-04T09:33:40.4458991Z * [new branch] gh/desertfire/610/orig -> origin/gh/desertfire/610/orig 2025-12-04T09:33:40.4460613Z * [new branch] gh/desertfire/611/base -> origin/gh/desertfire/611/base 2025-12-04T09:33:40.4461948Z * [new branch] gh/desertfire/611/head -> origin/gh/desertfire/611/head 2025-12-04T09:33:40.4463275Z * [new branch] gh/desertfire/611/orig -> origin/gh/desertfire/611/orig 2025-12-04T09:33:40.4465095Z * [new branch] gh/desertfire/612/base -> origin/gh/desertfire/612/base 2025-12-04T09:33:40.4466505Z * [new branch] gh/desertfire/612/head -> origin/gh/desertfire/612/head 2025-12-04T09:33:40.4467653Z * [new branch] gh/desertfire/612/orig -> origin/gh/desertfire/612/orig 2025-12-04T09:33:40.4469365Z * [new branch] gh/desertfire/613/base -> origin/gh/desertfire/613/base 2025-12-04T09:33:40.4470803Z * [new branch] gh/desertfire/613/head -> origin/gh/desertfire/613/head 2025-12-04T09:33:40.4472118Z * [new branch] gh/desertfire/613/orig -> origin/gh/desertfire/613/orig 2025-12-04T09:33:40.4473967Z * [new branch] gh/desertfire/614/base -> origin/gh/desertfire/614/base 2025-12-04T09:33:40.4475373Z * [new branch] gh/desertfire/614/head -> origin/gh/desertfire/614/head 2025-12-04T09:33:40.4476679Z * [new branch] gh/desertfire/614/orig -> origin/gh/desertfire/614/orig 2025-12-04T09:33:40.4478442Z * [new branch] gh/desertfire/615/base -> origin/gh/desertfire/615/base 2025-12-04T09:33:40.4480012Z * [new branch] gh/desertfire/615/head -> origin/gh/desertfire/615/head 2025-12-04T09:33:40.4481254Z * [new branch] gh/desertfire/615/orig -> origin/gh/desertfire/615/orig 2025-12-04T09:33:40.4482885Z * [new branch] gh/desertfire/616/base -> origin/gh/desertfire/616/base 2025-12-04T09:33:40.4484322Z * [new branch] gh/desertfire/616/head -> origin/gh/desertfire/616/head 2025-12-04T09:33:40.4485510Z * [new branch] gh/desertfire/616/orig -> origin/gh/desertfire/616/orig 2025-12-04T09:33:40.4487032Z * [new branch] gh/desertfire/617/base -> origin/gh/desertfire/617/base 2025-12-04T09:33:40.4508526Z * [new branch] gh/desertfire/617/head -> origin/gh/desertfire/617/head 2025-12-04T09:33:40.4509017Z * [new branch] gh/desertfire/617/orig -> origin/gh/desertfire/617/orig 2025-12-04T09:33:40.4509270Z * [new branch] gh/dharakk/1/base -> origin/gh/dharakk/1/base 2025-12-04T09:33:40.4509545Z * [new branch] gh/dharakk/1/head -> origin/gh/dharakk/1/head 2025-12-04T09:33:40.4509794Z * [new branch] gh/drisspg/170/base -> origin/gh/drisspg/170/base 2025-12-04T09:33:40.4510092Z * [new branch] gh/drisspg/170/head -> origin/gh/drisspg/170/head 2025-12-04T09:33:40.4510359Z * [new branch] gh/drisspg/170/orig -> origin/gh/drisspg/170/orig 2025-12-04T09:33:40.4510602Z * [new branch] gh/drisspg/182/base -> origin/gh/drisspg/182/base 2025-12-04T09:33:40.4510846Z * [new branch] gh/drisspg/182/head -> origin/gh/drisspg/182/head 2025-12-04T09:33:40.4511101Z * [new branch] gh/drisspg/183/base -> origin/gh/drisspg/183/base 2025-12-04T09:33:40.4511394Z * [new branch] gh/drisspg/183/head -> origin/gh/drisspg/183/head 2025-12-04T09:33:40.4511653Z * [new branch] gh/drisspg/184/base -> origin/gh/drisspg/184/base 2025-12-04T09:33:40.4511895Z * [new branch] gh/drisspg/184/head -> origin/gh/drisspg/184/head 2025-12-04T09:33:40.4512141Z * [new branch] gh/drisspg/185/base -> origin/gh/drisspg/185/base 2025-12-04T09:33:40.4512395Z * [new branch] gh/drisspg/185/head -> origin/gh/drisspg/185/head 2025-12-04T09:33:40.4513694Z * [new branch] gh/drisspg/194/base -> origin/gh/drisspg/194/base 2025-12-04T09:33:40.4514931Z * [new branch] gh/drisspg/194/head -> origin/gh/drisspg/194/head 2025-12-04T09:33:40.4516303Z * [new branch] gh/drisspg/194/orig -> origin/gh/drisspg/194/orig 2025-12-04T09:33:40.4517971Z * [new branch] gh/drisspg/200/base -> origin/gh/drisspg/200/base 2025-12-04T09:33:40.4519275Z * [new branch] gh/drisspg/200/head -> origin/gh/drisspg/200/head 2025-12-04T09:33:40.4520643Z * [new branch] gh/drisspg/200/orig -> origin/gh/drisspg/200/orig 2025-12-04T09:33:40.4522706Z * [new branch] gh/drisspg/218/base -> origin/gh/drisspg/218/base 2025-12-04T09:33:40.4523963Z * [new branch] gh/drisspg/218/head -> origin/gh/drisspg/218/head 2025-12-04T09:33:40.4525202Z * [new branch] gh/drisspg/218/orig -> origin/gh/drisspg/218/orig 2025-12-04T09:33:40.4526871Z * [new branch] gh/drisspg/219/base -> origin/gh/drisspg/219/base 2025-12-04T09:33:40.4528141Z * [new branch] gh/drisspg/219/head -> origin/gh/drisspg/219/head 2025-12-04T09:33:40.4529441Z * [new branch] gh/drisspg/219/orig -> origin/gh/drisspg/219/orig 2025-12-04T09:33:40.4531067Z * [new branch] gh/drisspg/220/base -> origin/gh/drisspg/220/base 2025-12-04T09:33:40.4532341Z * [new branch] gh/drisspg/220/head -> origin/gh/drisspg/220/head 2025-12-04T09:33:40.4533711Z * [new branch] gh/drisspg/220/orig -> origin/gh/drisspg/220/orig 2025-12-04T09:33:40.4535864Z * [new branch] gh/drisspg/221/base -> origin/gh/drisspg/221/base 2025-12-04T09:33:40.4537235Z * [new branch] gh/drisspg/221/head -> origin/gh/drisspg/221/head 2025-12-04T09:33:40.4538507Z * [new branch] gh/drisspg/221/orig -> origin/gh/drisspg/221/orig 2025-12-04T09:33:40.4540208Z * [new branch] gh/drisspg/222/base -> origin/gh/drisspg/222/base 2025-12-04T09:33:40.4541461Z * [new branch] gh/drisspg/222/head -> origin/gh/drisspg/222/head 2025-12-04T09:33:40.4542731Z * [new branch] gh/drisspg/222/orig -> origin/gh/drisspg/222/orig 2025-12-04T09:33:40.4544387Z * [new branch] gh/drisspg/223/base -> origin/gh/drisspg/223/base 2025-12-04T09:33:40.4545622Z * [new branch] gh/drisspg/223/head -> origin/gh/drisspg/223/head 2025-12-04T09:33:40.4546895Z * [new branch] gh/drisspg/223/orig -> origin/gh/drisspg/223/orig 2025-12-04T09:33:40.4548599Z * [new branch] gh/drisspg/224/base -> origin/gh/drisspg/224/base 2025-12-04T09:33:40.4549801Z * [new branch] gh/drisspg/224/head -> origin/gh/drisspg/224/head 2025-12-04T09:33:40.4551160Z * [new branch] gh/drisspg/224/orig -> origin/gh/drisspg/224/orig 2025-12-04T09:33:40.4552879Z * [new branch] gh/drisspg/225/base -> origin/gh/drisspg/225/base 2025-12-04T09:33:40.4554105Z * [new branch] gh/drisspg/225/head -> origin/gh/drisspg/225/head 2025-12-04T09:33:40.4555395Z * [new branch] gh/drisspg/225/orig -> origin/gh/drisspg/225/orig 2025-12-04T09:33:40.4557017Z * [new branch] gh/drisspg/226/base -> origin/gh/drisspg/226/base 2025-12-04T09:33:40.4558271Z * [new branch] gh/drisspg/226/head -> origin/gh/drisspg/226/head 2025-12-04T09:33:40.4559548Z * [new branch] gh/drisspg/226/orig -> origin/gh/drisspg/226/orig 2025-12-04T09:33:40.4561649Z * [new branch] gh/drisspg/227/base -> origin/gh/drisspg/227/base 2025-12-04T09:33:40.4562913Z * [new branch] gh/drisspg/227/head -> origin/gh/drisspg/227/head 2025-12-04T09:33:40.4564235Z * [new branch] gh/drisspg/227/orig -> origin/gh/drisspg/227/orig 2025-12-04T09:33:40.4565955Z * [new branch] gh/drisspg/228/base -> origin/gh/drisspg/228/base 2025-12-04T09:33:40.4567214Z * [new branch] gh/drisspg/228/head -> origin/gh/drisspg/228/head 2025-12-04T09:33:40.4568577Z * [new branch] gh/drisspg/228/orig -> origin/gh/drisspg/228/orig 2025-12-04T09:33:40.4570310Z * [new branch] gh/drisspg/229/base -> origin/gh/drisspg/229/base 2025-12-04T09:33:40.4571521Z * [new branch] gh/drisspg/229/head -> origin/gh/drisspg/229/head 2025-12-04T09:33:40.4572885Z * [new branch] gh/drisspg/229/orig -> origin/gh/drisspg/229/orig 2025-12-04T09:33:40.4574546Z * [new branch] gh/drisspg/230/base -> origin/gh/drisspg/230/base 2025-12-04T09:33:40.4575842Z * [new branch] gh/drisspg/230/head -> origin/gh/drisspg/230/head 2025-12-04T09:33:40.4577132Z * [new branch] gh/drisspg/230/orig -> origin/gh/drisspg/230/orig 2025-12-04T09:33:40.4579335Z * [new branch] gh/dsjohns2/1/base -> origin/gh/dsjohns2/1/base 2025-12-04T09:33:40.4580700Z * [new branch] gh/dsjohns2/1/head -> origin/gh/dsjohns2/1/head 2025-12-04T09:33:40.4582703Z * [new branch] gh/dzmitry-huba/1/base -> origin/gh/dzmitry-huba/1/base 2025-12-04T09:33:40.4584129Z * [new branch] gh/dzmitry-huba/1/head -> origin/gh/dzmitry-huba/1/head 2025-12-04T09:33:40.4586030Z * [new branch] gh/dzmitry-huba/12/base -> origin/gh/dzmitry-huba/12/base 2025-12-04T09:33:40.4587416Z * [new branch] gh/dzmitry-huba/12/head -> origin/gh/dzmitry-huba/12/head 2025-12-04T09:33:40.4588716Z * [new branch] gh/dzmitry-huba/12/orig -> origin/gh/dzmitry-huba/12/orig 2025-12-04T09:33:40.4590556Z * [new branch] gh/dzmitry-huba/13/base -> origin/gh/dzmitry-huba/13/base 2025-12-04T09:33:40.4591907Z * [new branch] gh/dzmitry-huba/13/head -> origin/gh/dzmitry-huba/13/head 2025-12-04T09:33:40.4593147Z * [new branch] gh/dzmitry-huba/13/orig -> origin/gh/dzmitry-huba/13/orig 2025-12-04T09:33:40.4594804Z * [new branch] gh/dzmitry-huba/14/base -> origin/gh/dzmitry-huba/14/base 2025-12-04T09:33:40.4596261Z * [new branch] gh/dzmitry-huba/14/head -> origin/gh/dzmitry-huba/14/head 2025-12-04T09:33:40.4597559Z * [new branch] gh/dzmitry-huba/14/orig -> origin/gh/dzmitry-huba/14/orig 2025-12-04T09:33:40.4599328Z * [new branch] gh/dzmitry-huba/15/base -> origin/gh/dzmitry-huba/15/base 2025-12-04T09:33:40.4600899Z * [new branch] gh/dzmitry-huba/15/head -> origin/gh/dzmitry-huba/15/head 2025-12-04T09:33:40.4602751Z * [new branch] gh/dzmitry-huba/15/orig -> origin/gh/dzmitry-huba/15/orig 2025-12-04T09:33:40.4604459Z * [new branch] gh/dzmitry-huba/16/base -> origin/gh/dzmitry-huba/16/base 2025-12-04T09:33:40.4606036Z * [new branch] gh/dzmitry-huba/16/head -> origin/gh/dzmitry-huba/16/head 2025-12-04T09:33:40.4607372Z * [new branch] gh/dzmitry-huba/16/orig -> origin/gh/dzmitry-huba/16/orig 2025-12-04T09:33:40.4609295Z * [new branch] gh/dzmitry-huba/17/base -> origin/gh/dzmitry-huba/17/base 2025-12-04T09:33:40.4610523Z * [new branch] gh/dzmitry-huba/17/head -> origin/gh/dzmitry-huba/17/head 2025-12-04T09:33:40.4612016Z * [new branch] gh/dzmitry-huba/17/orig -> origin/gh/dzmitry-huba/17/orig 2025-12-04T09:33:40.4613591Z * [new branch] gh/dzmitry-huba/2/base -> origin/gh/dzmitry-huba/2/base 2025-12-04T09:33:40.4614680Z * [new branch] gh/dzmitry-huba/2/head -> origin/gh/dzmitry-huba/2/head 2025-12-04T09:33:40.4616388Z * [new branch] gh/dzmitry-huba/3/base -> origin/gh/dzmitry-huba/3/base 2025-12-04T09:33:40.4617560Z * [new branch] gh/dzmitry-huba/3/head -> origin/gh/dzmitry-huba/3/head 2025-12-04T09:33:40.4620108Z * [new branch] gh/eellison/808/base -> origin/gh/eellison/808/base 2025-12-04T09:33:40.4621673Z * [new branch] gh/eellison/808/head -> origin/gh/eellison/808/head 2025-12-04T09:33:40.4622949Z * [new branch] gh/eellison/808/orig -> origin/gh/eellison/808/orig 2025-12-04T09:33:40.4625149Z * [new branch] gh/eellison/822/base -> origin/gh/eellison/822/base 2025-12-04T09:33:40.4627171Z * [new branch] gh/eellison/822/head -> origin/gh/eellison/822/head 2025-12-04T09:33:40.4628181Z * [new branch] gh/eellison/822/orig -> origin/gh/eellison/822/orig 2025-12-04T09:33:40.4630053Z * [new branch] gh/eellison/823/base -> origin/gh/eellison/823/base 2025-12-04T09:33:40.4631427Z * [new branch] gh/eellison/823/head -> origin/gh/eellison/823/head 2025-12-04T09:33:40.4632671Z * [new branch] gh/eellison/823/orig -> origin/gh/eellison/823/orig 2025-12-04T09:33:40.4634491Z * [new branch] gh/eellison/862/base -> origin/gh/eellison/862/base 2025-12-04T09:33:40.4635680Z * [new branch] gh/eellison/862/head -> origin/gh/eellison/862/head 2025-12-04T09:33:40.4636941Z * [new branch] gh/eellison/862/orig -> origin/gh/eellison/862/orig 2025-12-04T09:33:40.4638921Z * [new branch] gh/eellison/863/base -> origin/gh/eellison/863/base 2025-12-04T09:33:40.4640149Z * [new branch] gh/eellison/863/head -> origin/gh/eellison/863/head 2025-12-04T09:33:40.4641521Z * [new branch] gh/eellison/863/orig -> origin/gh/eellison/863/orig 2025-12-04T09:33:40.4643250Z * [new branch] gh/eellison/864/base -> origin/gh/eellison/864/base 2025-12-04T09:33:40.4644389Z * [new branch] gh/eellison/864/head -> origin/gh/eellison/864/head 2025-12-04T09:33:40.4645967Z * [new branch] gh/eellison/864/orig -> origin/gh/eellison/864/orig 2025-12-04T09:33:40.4648478Z * [new branch] gh/eellison/865/base -> origin/gh/eellison/865/base 2025-12-04T09:33:40.4649563Z * [new branch] gh/eellison/865/head -> origin/gh/eellison/865/head 2025-12-04T09:33:40.4650787Z * [new branch] gh/eellison/865/orig -> origin/gh/eellison/865/orig 2025-12-04T09:33:40.4652610Z * [new branch] gh/eellison/866/base -> origin/gh/eellison/866/base 2025-12-04T09:33:40.4653964Z * [new branch] gh/eellison/866/head -> origin/gh/eellison/866/head 2025-12-04T09:33:40.4655058Z * [new branch] gh/eellison/866/orig -> origin/gh/eellison/866/orig 2025-12-04T09:33:40.4657112Z * [new branch] gh/eellison/867/base -> origin/gh/eellison/867/base 2025-12-04T09:33:40.4658383Z * [new branch] gh/eellison/867/head -> origin/gh/eellison/867/head 2025-12-04T09:33:40.4659742Z * [new branch] gh/eellison/867/orig -> origin/gh/eellison/867/orig 2025-12-04T09:33:40.4661653Z * [new branch] gh/eellison/868/base -> origin/gh/eellison/868/base 2025-12-04T09:33:40.4663262Z * [new branch] gh/eellison/868/head -> origin/gh/eellison/868/head 2025-12-04T09:33:40.4664520Z * [new branch] gh/eellison/868/orig -> origin/gh/eellison/868/orig 2025-12-04T09:33:40.4666228Z * [new branch] gh/eellison/869/base -> origin/gh/eellison/869/base 2025-12-04T09:33:40.4667480Z * [new branch] gh/eellison/869/head -> origin/gh/eellison/869/head 2025-12-04T09:33:40.4668735Z * [new branch] gh/eellison/869/orig -> origin/gh/eellison/869/orig 2025-12-04T09:33:40.4670422Z * [new branch] gh/eellison/870/base -> origin/gh/eellison/870/base 2025-12-04T09:33:40.4671606Z * [new branch] gh/eellison/870/head -> origin/gh/eellison/870/head 2025-12-04T09:33:40.4672853Z * [new branch] gh/eellison/870/orig -> origin/gh/eellison/870/orig 2025-12-04T09:33:40.4674698Z * [new branch] gh/eellison/871/base -> origin/gh/eellison/871/base 2025-12-04T09:33:40.4675906Z * [new branch] gh/eellison/871/head -> origin/gh/eellison/871/head 2025-12-04T09:33:40.4677221Z * [new branch] gh/eellison/871/orig -> origin/gh/eellison/871/orig 2025-12-04T09:33:40.4679023Z * [new branch] gh/eellison/872/base -> origin/gh/eellison/872/base 2025-12-04T09:33:40.4680238Z * [new branch] gh/eellison/872/head -> origin/gh/eellison/872/head 2025-12-04T09:33:40.4681508Z * [new branch] gh/eellison/872/orig -> origin/gh/eellison/872/orig 2025-12-04T09:33:40.4683395Z * [new branch] gh/eellison/873/base -> origin/gh/eellison/873/base 2025-12-04T09:33:40.4684616Z * [new branch] gh/eellison/873/head -> origin/gh/eellison/873/head 2025-12-04T09:33:40.4685884Z * [new branch] gh/eellison/873/orig -> origin/gh/eellison/873/orig 2025-12-04T09:33:40.4687576Z * [new branch] gh/eellison/874/base -> origin/gh/eellison/874/base 2025-12-04T09:33:40.4688861Z * [new branch] gh/eellison/874/head -> origin/gh/eellison/874/head 2025-12-04T09:33:40.4690128Z * [new branch] gh/eellison/874/orig -> origin/gh/eellison/874/orig 2025-12-04T09:33:40.4692446Z * [new branch] gh/eellison/875/base -> origin/gh/eellison/875/base 2025-12-04T09:33:40.4693886Z * [new branch] gh/eellison/875/head -> origin/gh/eellison/875/head 2025-12-04T09:33:40.4695583Z * [new branch] gh/eellison/875/orig -> origin/gh/eellison/875/orig 2025-12-04T09:33:40.4697201Z * [new branch] gh/eellison/876/base -> origin/gh/eellison/876/base 2025-12-04T09:33:40.4698525Z * [new branch] gh/eellison/876/head -> origin/gh/eellison/876/head 2025-12-04T09:33:40.4699759Z * [new branch] gh/eellison/876/orig -> origin/gh/eellison/876/orig 2025-12-04T09:33:40.4701548Z * [new branch] gh/eellison/877/base -> origin/gh/eellison/877/base 2025-12-04T09:33:40.4702779Z * [new branch] gh/eellison/877/head -> origin/gh/eellison/877/head 2025-12-04T09:33:40.4704026Z * [new branch] gh/eellison/877/orig -> origin/gh/eellison/877/orig 2025-12-04T09:33:40.4705889Z * [new branch] gh/eellison/878/base -> origin/gh/eellison/878/base 2025-12-04T09:33:40.4707126Z * [new branch] gh/eellison/878/head -> origin/gh/eellison/878/head 2025-12-04T09:33:40.4708391Z * [new branch] gh/eellison/878/orig -> origin/gh/eellison/878/orig 2025-12-04T09:33:40.4710260Z * [new branch] gh/eellison/879/base -> origin/gh/eellison/879/base 2025-12-04T09:33:40.4711527Z * [new branch] gh/eellison/879/head -> origin/gh/eellison/879/head 2025-12-04T09:33:40.4712805Z * [new branch] gh/eellison/879/orig -> origin/gh/eellison/879/orig 2025-12-04T09:33:40.4714374Z * [new branch] gh/eellison/880/base -> origin/gh/eellison/880/base 2025-12-04T09:33:40.4715698Z * [new branch] gh/eellison/880/head -> origin/gh/eellison/880/head 2025-12-04T09:33:40.4717031Z * [new branch] gh/eellison/880/orig -> origin/gh/eellison/880/orig 2025-12-04T09:33:40.4718778Z * [new branch] gh/eellison/881/base -> origin/gh/eellison/881/base 2025-12-04T09:33:40.4720010Z * [new branch] gh/eellison/881/head -> origin/gh/eellison/881/head 2025-12-04T09:33:40.4721283Z * [new branch] gh/eellison/881/orig -> origin/gh/eellison/881/orig 2025-12-04T09:33:40.4722975Z * [new branch] gh/eellison/882/base -> origin/gh/eellison/882/base 2025-12-04T09:33:40.4724242Z * [new branch] gh/eellison/882/head -> origin/gh/eellison/882/head 2025-12-04T09:33:40.4725683Z * [new branch] gh/eellison/882/orig -> origin/gh/eellison/882/orig 2025-12-04T09:33:40.4727427Z * [new branch] gh/eellison/883/base -> origin/gh/eellison/883/base 2025-12-04T09:33:40.4728718Z * [new branch] gh/eellison/883/head -> origin/gh/eellison/883/head 2025-12-04T09:33:40.4730036Z * [new branch] gh/eellison/883/orig -> origin/gh/eellison/883/orig 2025-12-04T09:33:40.4731533Z * [new branch] gh/eellison/884/base -> origin/gh/eellison/884/base 2025-12-04T09:33:40.4732854Z * [new branch] gh/eellison/884/head -> origin/gh/eellison/884/head 2025-12-04T09:33:40.4734069Z * [new branch] gh/eellison/884/orig -> origin/gh/eellison/884/orig 2025-12-04T09:33:40.4736071Z * [new branch] gh/etaf/147/base -> origin/gh/etaf/147/base 2025-12-04T09:33:40.4737492Z * [new branch] gh/etaf/147/head -> origin/gh/etaf/147/head 2025-12-04T09:33:40.4739469Z * [new branch] gh/etaf/154/base -> origin/gh/etaf/154/base 2025-12-04T09:33:40.4740796Z * [new branch] gh/etaf/154/head -> origin/gh/etaf/154/head 2025-12-04T09:33:40.4742042Z * [new branch] gh/etaf/154/orig -> origin/gh/etaf/154/orig 2025-12-04T09:33:40.4743759Z * [new branch] gh/etaf/156/base -> origin/gh/etaf/156/base 2025-12-04T09:33:40.4745025Z * [new branch] gh/etaf/156/head -> origin/gh/etaf/156/head 2025-12-04T09:33:40.4746308Z * [new branch] gh/etaf/156/orig -> origin/gh/etaf/156/orig 2025-12-04T09:33:40.4748199Z * [new branch] gh/etaf/157/base -> origin/gh/etaf/157/base 2025-12-04T09:33:40.4749505Z * [new branch] gh/etaf/157/head -> origin/gh/etaf/157/head 2025-12-04T09:33:40.4750754Z * [new branch] gh/etaf/157/orig -> origin/gh/etaf/157/orig 2025-12-04T09:33:40.4752569Z * [new branch] gh/etaf/158/base -> origin/gh/etaf/158/base 2025-12-04T09:33:40.4753888Z * [new branch] gh/etaf/158/head -> origin/gh/etaf/158/head 2025-12-04T09:33:40.4755170Z * [new branch] gh/etaf/158/orig -> origin/gh/etaf/158/orig 2025-12-04T09:33:40.4756973Z * [new branch] gh/etaf/159/base -> origin/gh/etaf/159/base 2025-12-04T09:33:40.4758288Z * [new branch] gh/etaf/159/head -> origin/gh/etaf/159/head 2025-12-04T09:33:40.4759527Z * [new branch] gh/etaf/159/orig -> origin/gh/etaf/159/orig 2025-12-04T09:33:40.4761473Z * [new branch] gh/etaf/160/base -> origin/gh/etaf/160/base 2025-12-04T09:33:40.4762817Z * [new branch] gh/etaf/160/head -> origin/gh/etaf/160/head 2025-12-04T09:33:40.4764133Z * [new branch] gh/etaf/160/orig -> origin/gh/etaf/160/orig 2025-12-04T09:33:40.4765833Z * [new branch] gh/etaf/161/base -> origin/gh/etaf/161/base 2025-12-04T09:33:40.4767256Z * [new branch] gh/etaf/161/head -> origin/gh/etaf/161/head 2025-12-04T09:33:40.4768496Z * [new branch] gh/etaf/161/orig -> origin/gh/etaf/161/orig 2025-12-04T09:33:40.4770199Z * [new branch] gh/etaf/166/base -> origin/gh/etaf/166/base 2025-12-04T09:33:40.4771633Z * [new branch] gh/etaf/166/head -> origin/gh/etaf/166/head 2025-12-04T09:33:40.4772976Z * [new branch] gh/etaf/166/orig -> origin/gh/etaf/166/orig 2025-12-04T09:33:40.4774573Z * [new branch] gh/etaf/167/base -> origin/gh/etaf/167/base 2025-12-04T09:33:40.4775842Z * [new branch] gh/etaf/167/head -> origin/gh/etaf/167/head 2025-12-04T09:33:40.4777169Z * [new branch] gh/etaf/167/orig -> origin/gh/etaf/167/orig 2025-12-04T09:33:40.4779134Z * [new branch] gh/etaf/168/base -> origin/gh/etaf/168/base 2025-12-04T09:33:40.4780502Z * [new branch] gh/etaf/168/head -> origin/gh/etaf/168/head 2025-12-04T09:33:40.4781744Z * [new branch] gh/etaf/168/orig -> origin/gh/etaf/168/orig 2025-12-04T09:33:40.4783627Z * [new branch] gh/etaf/172/base -> origin/gh/etaf/172/base 2025-12-04T09:33:40.4784891Z * [new branch] gh/etaf/172/head -> origin/gh/etaf/172/head 2025-12-04T09:33:40.4786155Z * [new branch] gh/etaf/172/orig -> origin/gh/etaf/172/orig 2025-12-04T09:33:40.4788082Z * [new branch] gh/etaf/173/base -> origin/gh/etaf/173/base 2025-12-04T09:33:40.4789460Z * [new branch] gh/etaf/173/head -> origin/gh/etaf/173/head 2025-12-04T09:33:40.4790703Z * [new branch] gh/etaf/173/orig -> origin/gh/etaf/173/orig 2025-12-04T09:33:40.4792939Z * [new branch] gh/etaf/174/base -> origin/gh/etaf/174/base 2025-12-04T09:33:40.4794232Z * [new branch] gh/etaf/174/head -> origin/gh/etaf/174/head 2025-12-04T09:33:40.4796183Z * [new branch] gh/etaf/175/base -> origin/gh/etaf/175/base 2025-12-04T09:33:40.4797977Z * [new branch] gh/etaf/175/head -> origin/gh/etaf/175/head 2025-12-04T09:33:40.4798851Z * [new branch] gh/etaf/175/orig -> origin/gh/etaf/175/orig 2025-12-04T09:33:40.4800804Z * [new branch] gh/etaf/176/base -> origin/gh/etaf/176/base 2025-12-04T09:33:40.4802159Z * [new branch] gh/etaf/176/head -> origin/gh/etaf/176/head 2025-12-04T09:33:40.4803441Z * [new branch] gh/etaf/176/orig -> origin/gh/etaf/176/orig 2025-12-04T09:33:40.4805679Z * [new branch] gh/etaf/177/base -> origin/gh/etaf/177/base 2025-12-04T09:33:40.4807215Z * [new branch] gh/etaf/177/head -> origin/gh/etaf/177/head 2025-12-04T09:33:40.4808571Z * [new branch] gh/etaf/177/orig -> origin/gh/etaf/177/orig 2025-12-04T09:33:40.4810503Z * [new branch] gh/etaf/178/base -> origin/gh/etaf/178/base 2025-12-04T09:33:40.4811900Z * [new branch] gh/etaf/178/head -> origin/gh/etaf/178/head 2025-12-04T09:33:40.4813166Z * [new branch] gh/etaf/178/orig -> origin/gh/etaf/178/orig 2025-12-04T09:33:40.4815009Z * [new branch] gh/etaf/179/base -> origin/gh/etaf/179/base 2025-12-04T09:33:40.4816288Z * [new branch] gh/etaf/179/head -> origin/gh/etaf/179/head 2025-12-04T09:33:40.4817661Z * [new branch] gh/etaf/179/orig -> origin/gh/etaf/179/orig 2025-12-04T09:33:40.4819243Z * [new branch] gh/etaf/180/base -> origin/gh/etaf/180/base 2025-12-04T09:33:40.4820544Z * [new branch] gh/etaf/180/head -> origin/gh/etaf/180/head 2025-12-04T09:33:40.4821769Z * [new branch] gh/etaf/180/orig -> origin/gh/etaf/180/orig 2025-12-04T09:33:40.4823854Z * [new branch] gh/exclamaforte/1/base -> origin/gh/exclamaforte/1/base 2025-12-04T09:33:40.4825135Z * [new branch] gh/exclamaforte/1/head -> origin/gh/exclamaforte/1/head 2025-12-04T09:33:40.4826838Z * [new branch] gh/exclamaforte/2/base -> origin/gh/exclamaforte/2/base 2025-12-04T09:33:40.4827900Z * [new branch] gh/exclamaforte/2/head -> origin/gh/exclamaforte/2/head 2025-12-04T09:33:40.4829544Z * [new branch] gh/exclamaforte/3/base -> origin/gh/exclamaforte/3/base 2025-12-04T09:33:40.4831002Z * [new branch] gh/exclamaforte/3/head -> origin/gh/exclamaforte/3/head 2025-12-04T09:33:40.4832668Z * [new branch] gh/exclamaforte/4/base -> origin/gh/exclamaforte/4/base 2025-12-04T09:33:40.4833932Z * [new branch] gh/exclamaforte/4/head -> origin/gh/exclamaforte/4/head 2025-12-04T09:33:40.4835995Z * [new branch] gh/ezyang/2374/base -> origin/gh/ezyang/2374/base 2025-12-04T09:33:40.4837308Z * [new branch] gh/ezyang/2374/head -> origin/gh/ezyang/2374/head 2025-12-04T09:33:40.4839188Z * [new branch] gh/ezyang/2374/orig -> origin/gh/ezyang/2374/orig 2025-12-04T09:33:40.4840794Z * [new branch] gh/ezyang/2973/base -> origin/gh/ezyang/2973/base 2025-12-04T09:33:40.4842068Z * [new branch] gh/ezyang/2973/head -> origin/gh/ezyang/2973/head 2025-12-04T09:33:40.4843376Z * [new branch] gh/ezyang/2973/orig -> origin/gh/ezyang/2973/orig 2025-12-04T09:33:40.4845049Z * [new branch] gh/ezyang/2974/base -> origin/gh/ezyang/2974/base 2025-12-04T09:33:40.4846348Z * [new branch] gh/ezyang/2974/head -> origin/gh/ezyang/2974/head 2025-12-04T09:33:40.4847797Z * [new branch] gh/ezyang/2974/orig -> origin/gh/ezyang/2974/orig 2025-12-04T09:33:40.4849452Z * [new branch] gh/ezyang/3131/base -> origin/gh/ezyang/3131/base 2025-12-04T09:33:40.4850740Z * [new branch] gh/ezyang/3131/head -> origin/gh/ezyang/3131/head 2025-12-04T09:33:40.4851987Z * [new branch] gh/ezyang/3131/orig -> origin/gh/ezyang/3131/orig 2025-12-04T09:33:40.4853644Z * [new branch] gh/ezyang/3139/base -> origin/gh/ezyang/3139/base 2025-12-04T09:33:40.4854887Z * [new branch] gh/ezyang/3139/head -> origin/gh/ezyang/3139/head 2025-12-04T09:33:40.4856190Z * [new branch] gh/ezyang/3139/orig -> origin/gh/ezyang/3139/orig 2025-12-04T09:33:40.4857912Z * [new branch] gh/ezyang/3140/base -> origin/gh/ezyang/3140/base 2025-12-04T09:33:40.4859153Z * [new branch] gh/ezyang/3140/head -> origin/gh/ezyang/3140/head 2025-12-04T09:33:40.4860452Z * [new branch] gh/ezyang/3140/orig -> origin/gh/ezyang/3140/orig 2025-12-04T09:33:40.4862094Z * [new branch] gh/ezyang/3143/base -> origin/gh/ezyang/3143/base 2025-12-04T09:33:40.4863360Z * [new branch] gh/ezyang/3143/head -> origin/gh/ezyang/3143/head 2025-12-04T09:33:40.4864707Z * [new branch] gh/ezyang/3143/orig -> origin/gh/ezyang/3143/orig 2025-12-04T09:33:40.4866449Z * [new branch] gh/ezyang/3144/base -> origin/gh/ezyang/3144/base 2025-12-04T09:33:40.4867744Z * [new branch] gh/ezyang/3144/head -> origin/gh/ezyang/3144/head 2025-12-04T09:33:40.4869451Z * [new branch] gh/ezyang/3144/orig -> origin/gh/ezyang/3144/orig 2025-12-04T09:33:40.4871158Z * [new branch] gh/ezyang/3167/base -> origin/gh/ezyang/3167/base 2025-12-04T09:33:40.4872402Z * [new branch] gh/ezyang/3167/head -> origin/gh/ezyang/3167/head 2025-12-04T09:33:40.4873746Z * [new branch] gh/ezyang/3167/orig -> origin/gh/ezyang/3167/orig 2025-12-04T09:33:40.4875844Z * [new branch] gh/ezyang/3173/base -> origin/gh/ezyang/3173/base 2025-12-04T09:33:40.4877122Z * [new branch] gh/ezyang/3173/head -> origin/gh/ezyang/3173/head 2025-12-04T09:33:40.4878494Z * [new branch] gh/ezyang/3173/orig -> origin/gh/ezyang/3173/orig 2025-12-04T09:33:40.4880160Z * [new branch] gh/ezyang/3175/base -> origin/gh/ezyang/3175/base 2025-12-04T09:33:40.4881400Z * [new branch] gh/ezyang/3175/head -> origin/gh/ezyang/3175/head 2025-12-04T09:33:40.4882768Z * [new branch] gh/ezyang/3175/orig -> origin/gh/ezyang/3175/orig 2025-12-04T09:33:40.4884431Z * [new branch] gh/ezyang/3182/base -> origin/gh/ezyang/3182/base 2025-12-04T09:33:40.4885707Z * [new branch] gh/ezyang/3182/head -> origin/gh/ezyang/3182/head 2025-12-04T09:33:40.4887000Z * [new branch] gh/ezyang/3182/orig -> origin/gh/ezyang/3182/orig 2025-12-04T09:33:40.4888673Z * [new branch] gh/ezyang/3185/base -> origin/gh/ezyang/3185/base 2025-12-04T09:33:40.4889960Z * [new branch] gh/ezyang/3185/head -> origin/gh/ezyang/3185/head 2025-12-04T09:33:40.4891219Z * [new branch] gh/ezyang/3185/orig -> origin/gh/ezyang/3185/orig 2025-12-04T09:33:40.4892832Z * [new branch] gh/ezyang/3189/base -> origin/gh/ezyang/3189/base 2025-12-04T09:33:40.4894103Z * [new branch] gh/ezyang/3189/head -> origin/gh/ezyang/3189/head 2025-12-04T09:33:40.4895374Z * [new branch] gh/ezyang/3189/orig -> origin/gh/ezyang/3189/orig 2025-12-04T09:33:40.4899553Z * [new branch] gh/ezyang/3191/base -> origin/gh/ezyang/3191/base 2025-12-04T09:33:40.4900865Z * [new branch] gh/ezyang/3191/head -> origin/gh/ezyang/3191/head 2025-12-04T09:33:40.4902246Z * [new branch] gh/ezyang/3191/orig -> origin/gh/ezyang/3191/orig 2025-12-04T09:33:40.4904382Z * [new branch] gh/ezyang/3192/base -> origin/gh/ezyang/3192/base 2025-12-04T09:33:40.4905666Z * [new branch] gh/ezyang/3192/head -> origin/gh/ezyang/3192/head 2025-12-04T09:33:40.4906996Z * [new branch] gh/ezyang/3192/orig -> origin/gh/ezyang/3192/orig 2025-12-04T09:33:40.4908754Z * [new branch] gh/ezyang/3193/base -> origin/gh/ezyang/3193/base 2025-12-04T09:33:40.4910061Z * [new branch] gh/ezyang/3193/head -> origin/gh/ezyang/3193/head 2025-12-04T09:33:40.4911370Z * [new branch] gh/ezyang/3193/orig -> origin/gh/ezyang/3193/orig 2025-12-04T09:33:40.4913099Z * [new branch] gh/ezyang/3194/base -> origin/gh/ezyang/3194/base 2025-12-04T09:33:40.4914411Z * [new branch] gh/ezyang/3194/head -> origin/gh/ezyang/3194/head 2025-12-04T09:33:40.4915641Z * [new branch] gh/ezyang/3194/orig -> origin/gh/ezyang/3194/orig 2025-12-04T09:33:40.4917303Z * [new branch] gh/ezyang/3195/base -> origin/gh/ezyang/3195/base 2025-12-04T09:33:40.4918596Z * [new branch] gh/ezyang/3195/head -> origin/gh/ezyang/3195/head 2025-12-04T09:33:40.4919943Z * [new branch] gh/ezyang/3195/orig -> origin/gh/ezyang/3195/orig 2025-12-04T09:33:40.4921662Z * [new branch] gh/ezyang/3196/base -> origin/gh/ezyang/3196/base 2025-12-04T09:33:40.4922919Z * [new branch] gh/ezyang/3196/head -> origin/gh/ezyang/3196/head 2025-12-04T09:33:40.4924265Z * [new branch] gh/ezyang/3196/orig -> origin/gh/ezyang/3196/orig 2025-12-04T09:33:40.4925910Z * [new branch] gh/ezyang/3197/base -> origin/gh/ezyang/3197/base 2025-12-04T09:33:40.4927169Z * [new branch] gh/ezyang/3197/head -> origin/gh/ezyang/3197/head 2025-12-04T09:33:40.4928444Z * [new branch] gh/ezyang/3197/orig -> origin/gh/ezyang/3197/orig 2025-12-04T09:33:40.4930114Z * [new branch] gh/ezyang/3198/base -> origin/gh/ezyang/3198/base 2025-12-04T09:33:40.4931477Z * [new branch] gh/ezyang/3198/head -> origin/gh/ezyang/3198/head 2025-12-04T09:33:40.4932766Z * [new branch] gh/ezyang/3198/orig -> origin/gh/ezyang/3198/orig 2025-12-04T09:33:40.4934512Z * [new branch] gh/ezyang/3199/base -> origin/gh/ezyang/3199/base 2025-12-04T09:33:40.4935681Z * [new branch] gh/ezyang/3199/head -> origin/gh/ezyang/3199/head 2025-12-04T09:33:40.4937163Z * [new branch] gh/ezyang/3199/orig -> origin/gh/ezyang/3199/orig 2025-12-04T09:33:40.4939053Z * [new branch] gh/ezyang/3200/base -> origin/gh/ezyang/3200/base 2025-12-04T09:33:40.4940281Z * [new branch] gh/ezyang/3200/head -> origin/gh/ezyang/3200/head 2025-12-04T09:33:40.4941589Z * [new branch] gh/ezyang/3200/orig -> origin/gh/ezyang/3200/orig 2025-12-04T09:33:40.4943263Z * [new branch] gh/ezyang/3201/base -> origin/gh/ezyang/3201/base 2025-12-04T09:33:40.4944644Z * [new branch] gh/ezyang/3201/head -> origin/gh/ezyang/3201/head 2025-12-04T09:33:40.4945752Z * [new branch] gh/ezyang/3201/orig -> origin/gh/ezyang/3201/orig 2025-12-04T09:33:40.4947668Z * [new branch] gh/ezyang/3202/base -> origin/gh/ezyang/3202/base 2025-12-04T09:33:40.4948711Z * [new branch] gh/ezyang/3202/head -> origin/gh/ezyang/3202/head 2025-12-04T09:33:40.4949923Z * [new branch] gh/ezyang/3202/orig -> origin/gh/ezyang/3202/orig 2025-12-04T09:33:40.4951636Z * [new branch] gh/ezyang/3203/base -> origin/gh/ezyang/3203/base 2025-12-04T09:33:40.4952915Z * [new branch] gh/ezyang/3203/head -> origin/gh/ezyang/3203/head 2025-12-04T09:33:40.4954403Z * [new branch] gh/ezyang/3203/orig -> origin/gh/ezyang/3203/orig 2025-12-04T09:33:40.4956120Z * [new branch] gh/ezyang/3204/base -> origin/gh/ezyang/3204/base 2025-12-04T09:33:40.4957385Z * [new branch] gh/ezyang/3204/head -> origin/gh/ezyang/3204/head 2025-12-04T09:33:40.4958671Z * [new branch] gh/ezyang/3204/orig -> origin/gh/ezyang/3204/orig 2025-12-04T09:33:40.4960372Z * [new branch] gh/ezyang/3205/base -> origin/gh/ezyang/3205/base 2025-12-04T09:33:40.4961623Z * [new branch] gh/ezyang/3205/head -> origin/gh/ezyang/3205/head 2025-12-04T09:33:40.4962868Z * [new branch] gh/ezyang/3205/orig -> origin/gh/ezyang/3205/orig 2025-12-04T09:33:40.4964499Z * [new branch] gh/ezyang/3206/base -> origin/gh/ezyang/3206/base 2025-12-04T09:33:40.4965782Z * [new branch] gh/ezyang/3206/head -> origin/gh/ezyang/3206/head 2025-12-04T09:33:40.4967109Z * [new branch] gh/ezyang/3206/orig -> origin/gh/ezyang/3206/orig 2025-12-04T09:33:40.4968828Z * [new branch] gh/ezyang/3207/base -> origin/gh/ezyang/3207/base 2025-12-04T09:33:40.4970047Z * [new branch] gh/ezyang/3207/head -> origin/gh/ezyang/3207/head 2025-12-04T09:33:40.4971429Z * [new branch] gh/ezyang/3207/orig -> origin/gh/ezyang/3207/orig 2025-12-04T09:33:40.4973115Z * [new branch] gh/ezyang/3208/base -> origin/gh/ezyang/3208/base 2025-12-04T09:33:40.4974429Z * [new branch] gh/ezyang/3208/head -> origin/gh/ezyang/3208/head 2025-12-04T09:33:40.4975690Z * [new branch] gh/ezyang/3208/orig -> origin/gh/ezyang/3208/orig 2025-12-04T09:33:40.4977521Z * [new branch] gh/ezyang/3209/base -> origin/gh/ezyang/3209/base 2025-12-04T09:33:40.4978839Z * [new branch] gh/ezyang/3209/head -> origin/gh/ezyang/3209/head 2025-12-04T09:33:40.4980056Z * [new branch] gh/ezyang/3209/orig -> origin/gh/ezyang/3209/orig 2025-12-04T09:33:40.4982023Z * [new branch] gh/fadara01/3/base -> origin/gh/fadara01/3/base 2025-12-04T09:33:40.4983334Z * [new branch] gh/fadara01/3/head -> origin/gh/fadara01/3/head 2025-12-04T09:33:40.4984647Z * [new branch] gh/fadara01/3/orig -> origin/gh/fadara01/3/orig 2025-12-04T09:33:40.4986369Z * [new branch] gh/fadara01/5/base -> origin/gh/fadara01/5/base 2025-12-04T09:33:40.4987724Z * [new branch] gh/fadara01/5/head -> origin/gh/fadara01/5/head 2025-12-04T09:33:40.4988983Z * [new branch] gh/fadara01/5/orig -> origin/gh/fadara01/5/orig 2025-12-04T09:33:40.4990647Z * [new branch] gh/fadara01/6/base -> origin/gh/fadara01/6/base 2025-12-04T09:33:40.4991928Z * [new branch] gh/fadara01/6/head -> origin/gh/fadara01/6/head 2025-12-04T09:33:40.4993182Z * [new branch] gh/fadara01/6/orig -> origin/gh/fadara01/6/orig 2025-12-04T09:33:40.4995004Z * [new branch] gh/fadara01/7/base -> origin/gh/fadara01/7/base 2025-12-04T09:33:40.4996294Z * [new branch] gh/fadara01/7/head -> origin/gh/fadara01/7/head 2025-12-04T09:33:40.4997748Z * [new branch] gh/fadara01/7/orig -> origin/gh/fadara01/7/orig 2025-12-04T09:33:40.4999401Z * [new branch] gh/fadara01/8/base -> origin/gh/fadara01/8/base 2025-12-04T09:33:40.5000647Z * [new branch] gh/fadara01/8/head -> origin/gh/fadara01/8/head 2025-12-04T09:33:40.5001894Z * [new branch] gh/fadara01/8/orig -> origin/gh/fadara01/8/orig 2025-12-04T09:33:40.5003609Z * [new branch] gh/fadara01/9/base -> origin/gh/fadara01/9/base 2025-12-04T09:33:40.5004996Z * [new branch] gh/fadara01/9/head -> origin/gh/fadara01/9/head 2025-12-04T09:33:40.5006305Z * [new branch] gh/fadara01/9/orig -> origin/gh/fadara01/9/orig 2025-12-04T09:33:40.5008310Z * [new branch] gh/fduwjj/182/base -> origin/gh/fduwjj/182/base 2025-12-04T09:33:40.5009597Z * [new branch] gh/fduwjj/182/head -> origin/gh/fduwjj/182/head 2025-12-04T09:33:40.5010846Z * [new branch] gh/fduwjj/182/orig -> origin/gh/fduwjj/182/orig 2025-12-04T09:33:40.5012546Z * [new branch] gh/fduwjj/211/base -> origin/gh/fduwjj/211/base 2025-12-04T09:33:40.5013885Z * [new branch] gh/fduwjj/211/head -> origin/gh/fduwjj/211/head 2025-12-04T09:33:40.5015146Z * [new branch] gh/fduwjj/211/orig -> origin/gh/fduwjj/211/orig 2025-12-04T09:33:40.5016822Z * [new branch] gh/fduwjj/212/base -> origin/gh/fduwjj/212/base 2025-12-04T09:33:40.5018723Z * [new branch] gh/fduwjj/212/head -> origin/gh/fduwjj/212/head 2025-12-04T09:33:40.5019783Z * [new branch] gh/fduwjj/212/orig -> origin/gh/fduwjj/212/orig 2025-12-04T09:33:40.5022087Z * [new branch] gh/fduwjj/213/base -> origin/gh/fduwjj/213/base 2025-12-04T09:33:40.5023355Z * [new branch] gh/fduwjj/213/head -> origin/gh/fduwjj/213/head 2025-12-04T09:33:40.5024668Z * [new branch] gh/fduwjj/213/orig -> origin/gh/fduwjj/213/orig 2025-12-04T09:33:40.5026437Z * [new branch] gh/fduwjj/226/base -> origin/gh/fduwjj/226/base 2025-12-04T09:33:40.5027649Z * [new branch] gh/fduwjj/226/head -> origin/gh/fduwjj/226/head 2025-12-04T09:33:40.5029002Z * [new branch] gh/fduwjj/226/orig -> origin/gh/fduwjj/226/orig 2025-12-04T09:33:40.5030856Z * [new branch] gh/fduwjj/229/base -> origin/gh/fduwjj/229/base 2025-12-04T09:33:40.5032059Z * [new branch] gh/fduwjj/229/head -> origin/gh/fduwjj/229/head 2025-12-04T09:33:40.5033358Z * [new branch] gh/fduwjj/229/orig -> origin/gh/fduwjj/229/orig 2025-12-04T09:33:40.5035113Z * [new branch] gh/fduwjj/233/base -> origin/gh/fduwjj/233/base 2025-12-04T09:33:40.5036360Z * [new branch] gh/fduwjj/233/head -> origin/gh/fduwjj/233/head 2025-12-04T09:33:40.5037646Z * [new branch] gh/fduwjj/233/orig -> origin/gh/fduwjj/233/orig 2025-12-04T09:33:40.5039501Z * [new branch] gh/fduwjj/234/base -> origin/gh/fduwjj/234/base 2025-12-04T09:33:40.5040761Z * [new branch] gh/fduwjj/234/head -> origin/gh/fduwjj/234/head 2025-12-04T09:33:40.5041989Z * [new branch] gh/fduwjj/234/orig -> origin/gh/fduwjj/234/orig 2025-12-04T09:33:40.5043647Z * [new branch] gh/fduwjj/235/base -> origin/gh/fduwjj/235/base 2025-12-04T09:33:40.5044947Z * [new branch] gh/fduwjj/235/head -> origin/gh/fduwjj/235/head 2025-12-04T09:33:40.5046200Z * [new branch] gh/fduwjj/235/orig -> origin/gh/fduwjj/235/orig 2025-12-04T09:33:40.5047933Z * [new branch] gh/fduwjj/236/base -> origin/gh/fduwjj/236/base 2025-12-04T09:33:40.5049039Z * [new branch] gh/fduwjj/236/head -> origin/gh/fduwjj/236/head 2025-12-04T09:33:40.5050314Z * [new branch] gh/fduwjj/236/orig -> origin/gh/fduwjj/236/orig 2025-12-04T09:33:40.5051842Z * [new branch] gh/fduwjj/237/base -> origin/gh/fduwjj/237/base 2025-12-04T09:33:40.5053086Z * [new branch] gh/fduwjj/237/head -> origin/gh/fduwjj/237/head 2025-12-04T09:33:40.5054357Z * [new branch] gh/fduwjj/237/orig -> origin/gh/fduwjj/237/orig 2025-12-04T09:33:40.5056135Z * [new branch] gh/fduwjj/238/base -> origin/gh/fduwjj/238/base 2025-12-04T09:33:40.5057576Z * [new branch] gh/fduwjj/238/head -> origin/gh/fduwjj/238/head 2025-12-04T09:33:40.5058875Z * [new branch] gh/fduwjj/238/orig -> origin/gh/fduwjj/238/orig 2025-12-04T09:33:40.5060618Z * [new branch] gh/fduwjj/239/base -> origin/gh/fduwjj/239/base 2025-12-04T09:33:40.5061995Z * [new branch] gh/fduwjj/239/head -> origin/gh/fduwjj/239/head 2025-12-04T09:33:40.5063223Z * [new branch] gh/fduwjj/239/orig -> origin/gh/fduwjj/239/orig 2025-12-04T09:33:40.5065232Z * [new branch] gh/fegin/332/base -> origin/gh/fegin/332/base 2025-12-04T09:33:40.5066536Z * [new branch] gh/fegin/332/head -> origin/gh/fegin/332/head 2025-12-04T09:33:40.5067833Z * [new branch] gh/fegin/332/orig -> origin/gh/fegin/332/orig 2025-12-04T09:33:40.5069541Z * [new branch] gh/fegin/333/base -> origin/gh/fegin/333/base 2025-12-04T09:33:40.5070790Z * [new branch] gh/fegin/333/head -> origin/gh/fegin/333/head 2025-12-04T09:33:40.5072112Z * [new branch] gh/fegin/333/orig -> origin/gh/fegin/333/orig 2025-12-04T09:33:40.5073924Z * [new branch] gh/fegin/334/base -> origin/gh/fegin/334/base 2025-12-04T09:33:40.5075165Z * [new branch] gh/fegin/334/head -> origin/gh/fegin/334/head 2025-12-04T09:33:40.5076576Z * [new branch] gh/fegin/334/orig -> origin/gh/fegin/334/orig 2025-12-04T09:33:40.5078242Z * [new branch] gh/fegin/335/base -> origin/gh/fegin/335/base 2025-12-04T09:33:40.5079516Z * [new branch] gh/fegin/335/head -> origin/gh/fegin/335/head 2025-12-04T09:33:40.5080758Z * [new branch] gh/fegin/335/orig -> origin/gh/fegin/335/orig 2025-12-04T09:33:40.5082760Z * [new branch] gh/fffrog/160/base -> origin/gh/fffrog/160/base 2025-12-04T09:33:40.5084031Z * [new branch] gh/fffrog/160/head -> origin/gh/fffrog/160/head 2025-12-04T09:33:40.5085704Z * [new branch] gh/fffrog/177/base -> origin/gh/fffrog/177/base 2025-12-04T09:33:40.5086937Z * [new branch] gh/fffrog/177/head -> origin/gh/fffrog/177/head 2025-12-04T09:33:40.5088245Z * [new branch] gh/fffrog/177/orig -> origin/gh/fffrog/177/orig 2025-12-04T09:33:40.5089964Z * [new branch] gh/fffrog/178/base -> origin/gh/fffrog/178/base 2025-12-04T09:33:40.5091231Z * [new branch] gh/fffrog/178/head -> origin/gh/fffrog/178/head 2025-12-04T09:33:40.5092520Z * [new branch] gh/fffrog/178/orig -> origin/gh/fffrog/178/orig 2025-12-04T09:33:40.5094201Z * [new branch] gh/fffrog/181/base -> origin/gh/fffrog/181/base 2025-12-04T09:33:40.5095417Z * [new branch] gh/fffrog/181/head -> origin/gh/fffrog/181/head 2025-12-04T09:33:40.5096966Z * [new branch] gh/fffrog/181/orig -> origin/gh/fffrog/181/orig 2025-12-04T09:33:40.5098781Z * [new branch] gh/fffrog/183/base -> origin/gh/fffrog/183/base 2025-12-04T09:33:40.5100201Z * [new branch] gh/fffrog/183/head -> origin/gh/fffrog/183/head 2025-12-04T09:33:40.5101485Z * [new branch] gh/fffrog/183/orig -> origin/gh/fffrog/183/orig 2025-12-04T09:33:40.5103567Z * [new branch] gh/fxdawnn/10/base -> origin/gh/fxdawnn/10/base 2025-12-04T09:33:40.5104771Z * [new branch] gh/fxdawnn/10/head -> origin/gh/fxdawnn/10/head 2025-12-04T09:33:40.5106385Z * [new branch] gh/fxdawnn/10/orig -> origin/gh/fxdawnn/10/orig 2025-12-04T09:33:40.5107973Z * [new branch] gh/fxdawnn/11/base -> origin/gh/fxdawnn/11/base 2025-12-04T09:33:40.5109180Z * [new branch] gh/fxdawnn/11/head -> origin/gh/fxdawnn/11/head 2025-12-04T09:33:40.5110433Z * [new branch] gh/fxdawnn/11/orig -> origin/gh/fxdawnn/11/orig 2025-12-04T09:33:40.5112742Z * [new branch] gh/fxdawnn/12/base -> origin/gh/fxdawnn/12/base 2025-12-04T09:33:40.5113989Z * [new branch] gh/fxdawnn/12/head -> origin/gh/fxdawnn/12/head 2025-12-04T09:33:40.5115098Z * [new branch] gh/fxdawnn/12/orig -> origin/gh/fxdawnn/12/orig 2025-12-04T09:33:40.5117670Z * [new branch] gh/fxdawnn/13/base -> origin/gh/fxdawnn/13/base 2025-12-04T09:33:40.5118754Z * [new branch] gh/fxdawnn/13/head -> origin/gh/fxdawnn/13/head 2025-12-04T09:33:40.5120431Z * [new branch] gh/fxdawnn/13/orig -> origin/gh/fxdawnn/13/orig 2025-12-04T09:33:40.5122322Z * [new branch] gh/fxdawnn/14/base -> origin/gh/fxdawnn/14/base 2025-12-04T09:33:40.5123479Z * [new branch] gh/fxdawnn/14/head -> origin/gh/fxdawnn/14/head 2025-12-04T09:33:40.5124937Z * [new branch] gh/fxdawnn/14/orig -> origin/gh/fxdawnn/14/orig 2025-12-04T09:33:40.5126751Z * [new branch] gh/fxdawnn/15/base -> origin/gh/fxdawnn/15/base 2025-12-04T09:33:40.5128011Z * [new branch] gh/fxdawnn/15/head -> origin/gh/fxdawnn/15/head 2025-12-04T09:33:40.5129267Z * [new branch] gh/fxdawnn/15/orig -> origin/gh/fxdawnn/15/orig 2025-12-04T09:33:40.5131164Z * [new branch] gh/fxdawnn/6/base -> origin/gh/fxdawnn/6/base 2025-12-04T09:33:40.5132324Z * [new branch] gh/fxdawnn/6/head -> origin/gh/fxdawnn/6/head 2025-12-04T09:33:40.5133633Z * [new branch] gh/fxdawnn/6/orig -> origin/gh/fxdawnn/6/orig 2025-12-04T09:33:40.5135499Z * [new branch] gh/fxdawnn/7/base -> origin/gh/fxdawnn/7/base 2025-12-04T09:33:40.5136817Z * [new branch] gh/fxdawnn/7/head -> origin/gh/fxdawnn/7/head 2025-12-04T09:33:40.5138160Z * [new branch] gh/fxdawnn/7/orig -> origin/gh/fxdawnn/7/orig 2025-12-04T09:33:40.5140062Z * [new branch] gh/fxdawnn/9/base -> origin/gh/fxdawnn/9/base 2025-12-04T09:33:40.5141276Z * [new branch] gh/fxdawnn/9/head -> origin/gh/fxdawnn/9/head 2025-12-04T09:33:40.5142713Z * [new branch] gh/fxdawnn/9/orig -> origin/gh/fxdawnn/9/orig 2025-12-04T09:33:40.5144760Z * [new branch] gh/galv/1/base -> origin/gh/galv/1/base 2025-12-04T09:33:40.5146032Z * [new branch] gh/galv/1/head -> origin/gh/galv/1/head 2025-12-04T09:33:40.5147597Z * [new branch] gh/galv/1/orig -> origin/gh/galv/1/orig 2025-12-04T09:33:40.5149407Z * [new branch] gh/galv/2/base -> origin/gh/galv/2/base 2025-12-04T09:33:40.5150589Z * [new branch] gh/galv/2/head -> origin/gh/galv/2/head 2025-12-04T09:33:40.5152188Z * [new branch] gh/galv/2/orig -> origin/gh/galv/2/orig 2025-12-04T09:33:40.5154084Z * [new branch] gh/galv/3/base -> origin/gh/galv/3/base 2025-12-04T09:33:40.5155158Z * [new branch] gh/galv/3/head -> origin/gh/galv/3/head 2025-12-04T09:33:40.5156806Z * [new branch] gh/galv/3/orig -> origin/gh/galv/3/orig 2025-12-04T09:33:40.5159203Z * [new branch] gh/guangyey/134/base -> origin/gh/guangyey/134/base 2025-12-04T09:33:40.5160449Z * [new branch] gh/guangyey/134/head -> origin/gh/guangyey/134/head 2025-12-04T09:33:40.5161726Z * [new branch] gh/guangyey/134/orig -> origin/gh/guangyey/134/orig 2025-12-04T09:33:40.5163618Z * [new branch] gh/guangyey/163/base -> origin/gh/guangyey/163/base 2025-12-04T09:33:40.5164775Z * [new branch] gh/guangyey/163/head -> origin/gh/guangyey/163/head 2025-12-04T09:33:40.5166073Z * [new branch] gh/guangyey/163/orig -> origin/gh/guangyey/163/orig 2025-12-04T09:33:40.5167952Z * [new branch] gh/guangyey/168/base -> origin/gh/guangyey/168/base 2025-12-04T09:33:40.5169244Z * [new branch] gh/guangyey/168/head -> origin/gh/guangyey/168/head 2025-12-04T09:33:40.5170548Z * [new branch] gh/guangyey/168/orig -> origin/gh/guangyey/168/orig 2025-12-04T09:33:40.5172413Z * [new branch] gh/guangyey/169/base -> origin/gh/guangyey/169/base 2025-12-04T09:33:40.5173647Z * [new branch] gh/guangyey/169/head -> origin/gh/guangyey/169/head 2025-12-04T09:33:40.5174886Z * [new branch] gh/guangyey/169/orig -> origin/gh/guangyey/169/orig 2025-12-04T09:33:40.5176827Z * [new branch] gh/guangyey/170/base -> origin/gh/guangyey/170/base 2025-12-04T09:33:40.5178221Z * [new branch] gh/guangyey/170/head -> origin/gh/guangyey/170/head 2025-12-04T09:33:40.5179490Z * [new branch] gh/guangyey/170/orig -> origin/gh/guangyey/170/orig 2025-12-04T09:33:40.5181378Z * [new branch] gh/guangyey/171/base -> origin/gh/guangyey/171/base 2025-12-04T09:33:40.5182564Z * [new branch] gh/guangyey/171/head -> origin/gh/guangyey/171/head 2025-12-04T09:33:40.5183839Z * [new branch] gh/guangyey/171/orig -> origin/gh/guangyey/171/orig 2025-12-04T09:33:40.5185765Z * [new branch] gh/guangyey/178/base -> origin/gh/guangyey/178/base 2025-12-04T09:33:40.5187445Z * [new branch] gh/guangyey/178/head -> origin/gh/guangyey/178/head 2025-12-04T09:33:40.5188113Z * [new branch] gh/guangyey/178/orig -> origin/gh/guangyey/178/orig 2025-12-04T09:33:40.5189871Z * [new branch] gh/guangyey/182/base -> origin/gh/guangyey/182/base 2025-12-04T09:33:40.5191145Z * [new branch] gh/guangyey/182/head -> origin/gh/guangyey/182/head 2025-12-04T09:33:40.5192477Z * [new branch] gh/guangyey/182/orig -> origin/gh/guangyey/182/orig 2025-12-04T09:33:40.5194136Z * [new branch] gh/guangyey/183/base -> origin/gh/guangyey/183/base 2025-12-04T09:33:40.5195392Z * [new branch] gh/guangyey/183/head -> origin/gh/guangyey/183/head 2025-12-04T09:33:40.5196886Z * [new branch] gh/guangyey/183/orig -> origin/gh/guangyey/183/orig 2025-12-04T09:33:40.5199108Z * [new branch] gh/guangyey/185/base -> origin/gh/guangyey/185/base 2025-12-04T09:33:40.5200385Z * [new branch] gh/guangyey/185/head -> origin/gh/guangyey/185/head 2025-12-04T09:33:40.5201639Z * [new branch] gh/guangyey/185/orig -> origin/gh/guangyey/185/orig 2025-12-04T09:33:40.5203414Z * [new branch] gh/guangyey/186/base -> origin/gh/guangyey/186/base 2025-12-04T09:33:40.5204690Z * [new branch] gh/guangyey/186/head -> origin/gh/guangyey/186/head 2025-12-04T09:33:40.5205989Z * [new branch] gh/guangyey/186/orig -> origin/gh/guangyey/186/orig 2025-12-04T09:33:40.5207598Z * [new branch] gh/guangyey/187/base -> origin/gh/guangyey/187/base 2025-12-04T09:33:40.5208882Z * [new branch] gh/guangyey/187/head -> origin/gh/guangyey/187/head 2025-12-04T09:33:40.5210158Z * [new branch] gh/guangyey/187/orig -> origin/gh/guangyey/187/orig 2025-12-04T09:33:40.5211940Z * [new branch] gh/guangyey/188/base -> origin/gh/guangyey/188/base 2025-12-04T09:33:40.5213280Z * [new branch] gh/guangyey/188/head -> origin/gh/guangyey/188/head 2025-12-04T09:33:40.5214545Z * [new branch] gh/guangyey/188/orig -> origin/gh/guangyey/188/orig 2025-12-04T09:33:40.5216241Z * [new branch] gh/guangyey/190/base -> origin/gh/guangyey/190/base 2025-12-04T09:33:40.5217614Z * [new branch] gh/guangyey/190/head -> origin/gh/guangyey/190/head 2025-12-04T09:33:40.5218968Z * [new branch] gh/guangyey/190/orig -> origin/gh/guangyey/190/orig 2025-12-04T09:33:40.5220703Z * [new branch] gh/guangyey/208/base -> origin/gh/guangyey/208/base 2025-12-04T09:33:40.5221960Z * [new branch] gh/guangyey/208/head -> origin/gh/guangyey/208/head 2025-12-04T09:33:40.5223296Z * [new branch] gh/guangyey/208/orig -> origin/gh/guangyey/208/orig 2025-12-04T09:33:40.5224915Z * [new branch] gh/guangyey/228/base -> origin/gh/guangyey/228/base 2025-12-04T09:33:40.5226207Z * [new branch] gh/guangyey/228/head -> origin/gh/guangyey/228/head 2025-12-04T09:33:40.5227484Z * [new branch] gh/guangyey/228/orig -> origin/gh/guangyey/228/orig 2025-12-04T09:33:40.5229751Z * [new branch] gh/guangyey/230/base -> origin/gh/guangyey/230/base 2025-12-04T09:33:40.5231008Z * [new branch] gh/guangyey/230/head -> origin/gh/guangyey/230/head 2025-12-04T09:33:40.5232297Z * [new branch] gh/guangyey/230/orig -> origin/gh/guangyey/230/orig 2025-12-04T09:33:40.5234040Z * [new branch] gh/guangyey/231/base -> origin/gh/guangyey/231/base 2025-12-04T09:33:40.5235285Z * [new branch] gh/guangyey/231/head -> origin/gh/guangyey/231/head 2025-12-04T09:33:40.5236585Z * [new branch] gh/guangyey/231/orig -> origin/gh/guangyey/231/orig 2025-12-04T09:33:40.5238358Z * [new branch] gh/guangyey/232/base -> origin/gh/guangyey/232/base 2025-12-04T09:33:40.5239596Z * [new branch] gh/guangyey/232/head -> origin/gh/guangyey/232/head 2025-12-04T09:33:40.5240865Z * [new branch] gh/guangyey/232/orig -> origin/gh/guangyey/232/orig 2025-12-04T09:33:40.5242558Z * [new branch] gh/guangyey/233/base -> origin/gh/guangyey/233/base 2025-12-04T09:33:40.5243835Z * [new branch] gh/guangyey/233/head -> origin/gh/guangyey/233/head 2025-12-04T09:33:40.5245080Z * [new branch] gh/guangyey/233/orig -> origin/gh/guangyey/233/orig 2025-12-04T09:33:40.5246905Z * [new branch] gh/guangyey/234/base -> origin/gh/guangyey/234/base 2025-12-04T09:33:40.5248220Z * [new branch] gh/guangyey/234/head -> origin/gh/guangyey/234/head 2025-12-04T09:33:40.5249464Z * [new branch] gh/guangyey/234/orig -> origin/gh/guangyey/234/orig 2025-12-04T09:33:40.5251149Z * [new branch] gh/guangyey/235/base -> origin/gh/guangyey/235/base 2025-12-04T09:33:40.5252412Z * [new branch] gh/guangyey/235/head -> origin/gh/guangyey/235/head 2025-12-04T09:33:40.5253656Z * [new branch] gh/guangyey/235/orig -> origin/gh/guangyey/235/orig 2025-12-04T09:33:40.5255372Z * [new branch] gh/guangyey/236/base -> origin/gh/guangyey/236/base 2025-12-04T09:33:40.5256714Z * [new branch] gh/guangyey/236/head -> origin/gh/guangyey/236/head 2025-12-04T09:33:40.5258223Z * [new branch] gh/guangyey/236/orig -> origin/gh/guangyey/236/orig 2025-12-04T09:33:40.5259856Z * [new branch] gh/guangyey/237/base -> origin/gh/guangyey/237/base 2025-12-04T09:33:40.5261127Z * [new branch] gh/guangyey/237/head -> origin/gh/guangyey/237/head 2025-12-04T09:33:40.5262381Z * [new branch] gh/guangyey/237/orig -> origin/gh/guangyey/237/orig 2025-12-04T09:33:40.5264197Z * [new branch] gh/guangyey/238/base -> origin/gh/guangyey/238/base 2025-12-04T09:33:40.5265468Z * [new branch] gh/guangyey/238/head -> origin/gh/guangyey/238/head 2025-12-04T09:33:40.5267144Z * [new branch] gh/guangyey/239/base -> origin/gh/guangyey/239/base 2025-12-04T09:33:40.5268515Z * [new branch] gh/guangyey/239/head -> origin/gh/guangyey/239/head 2025-12-04T09:33:40.5269761Z * [new branch] gh/guangyey/239/orig -> origin/gh/guangyey/239/orig 2025-12-04T09:33:40.5271495Z * [new branch] gh/guangyey/240/base -> origin/gh/guangyey/240/base 2025-12-04T09:33:40.5272707Z * [new branch] gh/guangyey/240/head -> origin/gh/guangyey/240/head 2025-12-04T09:33:40.5274010Z * [new branch] gh/guangyey/240/orig -> origin/gh/guangyey/240/orig 2025-12-04T09:33:40.5276268Z * [new branch] gh/guangyey/241/base -> origin/gh/guangyey/241/base 2025-12-04T09:33:40.5277512Z * [new branch] gh/guangyey/241/head -> origin/gh/guangyey/241/head 2025-12-04T09:33:40.5278798Z * [new branch] gh/guangyey/241/orig -> origin/gh/guangyey/241/orig 2025-12-04T09:33:40.5280501Z * [new branch] gh/guangyey/242/base -> origin/gh/guangyey/242/base 2025-12-04T09:33:40.5281889Z * [new branch] gh/guangyey/242/head -> origin/gh/guangyey/242/head 2025-12-04T09:33:40.5283185Z * [new branch] gh/guangyey/242/orig -> origin/gh/guangyey/242/orig 2025-12-04T09:33:40.5284878Z * [new branch] gh/guangyey/243/base -> origin/gh/guangyey/243/base 2025-12-04T09:33:40.5286181Z * [new branch] gh/guangyey/243/head -> origin/gh/guangyey/243/head 2025-12-04T09:33:40.5287424Z * [new branch] gh/guangyey/243/orig -> origin/gh/guangyey/243/orig 2025-12-04T09:33:40.5289269Z * [new branch] gh/guangyey/244/base -> origin/gh/guangyey/244/base 2025-12-04T09:33:40.5290560Z * [new branch] gh/guangyey/244/head -> origin/gh/guangyey/244/head 2025-12-04T09:33:40.5291852Z * [new branch] gh/guangyey/244/orig -> origin/gh/guangyey/244/orig 2025-12-04T09:33:40.5293609Z * [new branch] gh/guangyey/245/base -> origin/gh/guangyey/245/base 2025-12-04T09:33:40.5294835Z * [new branch] gh/guangyey/245/head -> origin/gh/guangyey/245/head 2025-12-04T09:33:40.5296246Z * [new branch] gh/guangyey/245/orig -> origin/gh/guangyey/245/orig 2025-12-04T09:33:40.5300981Z * [new branch] gh/guangyey/246/base -> origin/gh/guangyey/246/base 2025-12-04T09:33:40.5302381Z * [new branch] gh/guangyey/246/head -> origin/gh/guangyey/246/head 2025-12-04T09:33:40.5303667Z * [new branch] gh/guangyey/246/orig -> origin/gh/guangyey/246/orig 2025-12-04T09:33:40.5306044Z * [new branch] gh/guangyey/247/base -> origin/gh/guangyey/247/base 2025-12-04T09:33:40.5307443Z * [new branch] gh/guangyey/247/head -> origin/gh/guangyey/247/head 2025-12-04T09:33:40.5308793Z * [new branch] gh/guangyey/247/orig -> origin/gh/guangyey/247/orig 2025-12-04T09:33:40.5310541Z * [new branch] gh/guangyey/248/base -> origin/gh/guangyey/248/base 2025-12-04T09:33:40.5311882Z * [new branch] gh/guangyey/248/head -> origin/gh/guangyey/248/head 2025-12-04T09:33:40.5313078Z * [new branch] gh/guangyey/248/orig -> origin/gh/guangyey/248/orig 2025-12-04T09:33:40.5314783Z * [new branch] gh/guangyey/249/base -> origin/gh/guangyey/249/base 2025-12-04T09:33:40.5316158Z * [new branch] gh/guangyey/249/head -> origin/gh/guangyey/249/head 2025-12-04T09:33:40.5317434Z * [new branch] gh/guangyey/249/orig -> origin/gh/guangyey/249/orig 2025-12-04T09:33:40.5319127Z * [new branch] gh/guangyey/250/base -> origin/gh/guangyey/250/base 2025-12-04T09:33:40.5320539Z * [new branch] gh/guangyey/250/head -> origin/gh/guangyey/250/head 2025-12-04T09:33:40.5321865Z * [new branch] gh/guangyey/250/orig -> origin/gh/guangyey/250/orig 2025-12-04T09:33:40.5323521Z * [new branch] gh/guangyey/251/base -> origin/gh/guangyey/251/base 2025-12-04T09:33:40.5324855Z * [new branch] gh/guangyey/251/head -> origin/gh/guangyey/251/head 2025-12-04T09:33:40.5326636Z * [new branch] gh/guangyey/251/orig -> origin/gh/guangyey/251/orig 2025-12-04T09:33:40.5328400Z * [new branch] gh/guangyey/252/base -> origin/gh/guangyey/252/base 2025-12-04T09:33:40.5329662Z * [new branch] gh/guangyey/252/head -> origin/gh/guangyey/252/head 2025-12-04T09:33:40.5331268Z * [new branch] gh/guangyey/252/orig -> origin/gh/guangyey/252/orig 2025-12-04T09:33:40.5333006Z * [new branch] gh/guangyey/253/base -> origin/gh/guangyey/253/base 2025-12-04T09:33:40.5334289Z * [new branch] gh/guangyey/253/head -> origin/gh/guangyey/253/head 2025-12-04T09:33:40.5335581Z * [new branch] gh/guangyey/253/orig -> origin/gh/guangyey/253/orig 2025-12-04T09:33:40.5337349Z * [new branch] gh/guangyey/254/base -> origin/gh/guangyey/254/base 2025-12-04T09:33:40.5338832Z * [new branch] gh/guangyey/254/head -> origin/gh/guangyey/254/head 2025-12-04T09:33:40.5340065Z * [new branch] gh/guangyey/254/orig -> origin/gh/guangyey/254/orig 2025-12-04T09:33:40.5341786Z * [new branch] gh/guangyey/255/base -> origin/gh/guangyey/255/base 2025-12-04T09:33:40.5343030Z * [new branch] gh/guangyey/255/head -> origin/gh/guangyey/255/head 2025-12-04T09:33:40.5344339Z * [new branch] gh/guangyey/255/orig -> origin/gh/guangyey/255/orig 2025-12-04T09:33:40.5346558Z * [new branch] gh/guilhermeleobas/107/base -> origin/gh/guilhermeleobas/107/base 2025-12-04T09:33:40.5347822Z * [new branch] gh/guilhermeleobas/107/head -> origin/gh/guilhermeleobas/107/head 2025-12-04T09:33:40.5349340Z * [new branch] gh/guilhermeleobas/107/orig -> origin/gh/guilhermeleobas/107/orig 2025-12-04T09:33:40.5350955Z * [new branch] gh/guilhermeleobas/108/base -> origin/gh/guilhermeleobas/108/base 2025-12-04T09:33:40.5352183Z * [new branch] gh/guilhermeleobas/108/head -> origin/gh/guilhermeleobas/108/head 2025-12-04T09:33:40.5353489Z * [new branch] gh/guilhermeleobas/108/orig -> origin/gh/guilhermeleobas/108/orig 2025-12-04T09:33:40.5355476Z * [new branch] gh/guilhermeleobas/150/base -> origin/gh/guilhermeleobas/150/base 2025-12-04T09:33:40.5358069Z * [new branch] gh/guilhermeleobas/150/head -> origin/gh/guilhermeleobas/150/head 2025-12-04T09:33:40.5359327Z * [new branch] gh/guilhermeleobas/150/orig -> origin/gh/guilhermeleobas/150/orig 2025-12-04T09:33:40.5361013Z * [new branch] gh/guilhermeleobas/168/base -> origin/gh/guilhermeleobas/168/base 2025-12-04T09:33:40.5362305Z * [new branch] gh/guilhermeleobas/168/head -> origin/gh/guilhermeleobas/168/head 2025-12-04T09:33:40.5363577Z * [new branch] gh/guilhermeleobas/168/orig -> origin/gh/guilhermeleobas/168/orig 2025-12-04T09:33:40.5366276Z * [new branch] gh/guilhermeleobas/169/base -> origin/gh/guilhermeleobas/169/base 2025-12-04T09:33:40.5366604Z * [new branch] gh/guilhermeleobas/169/head -> origin/gh/guilhermeleobas/169/head 2025-12-04T09:33:40.5367758Z * [new branch] gh/guilhermeleobas/169/orig -> origin/gh/guilhermeleobas/169/orig 2025-12-04T09:33:40.5369956Z * [new branch] gh/guilhermeleobas/170/base -> origin/gh/guilhermeleobas/170/base 2025-12-04T09:33:40.5371443Z * [new branch] gh/guilhermeleobas/170/head -> origin/gh/guilhermeleobas/170/head 2025-12-04T09:33:40.5372515Z * [new branch] gh/guilhermeleobas/170/orig -> origin/gh/guilhermeleobas/170/orig 2025-12-04T09:33:40.5374348Z * [new branch] gh/guilhermeleobas/171/base -> origin/gh/guilhermeleobas/171/base 2025-12-04T09:33:40.5375589Z * [new branch] gh/guilhermeleobas/171/head -> origin/gh/guilhermeleobas/171/head 2025-12-04T09:33:40.5376918Z * [new branch] gh/guilhermeleobas/171/orig -> origin/gh/guilhermeleobas/171/orig 2025-12-04T09:33:40.5378638Z * [new branch] gh/guilhermeleobas/173/base -> origin/gh/guilhermeleobas/173/base 2025-12-04T09:33:40.5380391Z * [new branch] gh/guilhermeleobas/173/head -> origin/gh/guilhermeleobas/173/head 2025-12-04T09:33:40.5381629Z * [new branch] gh/guilhermeleobas/173/orig -> origin/gh/guilhermeleobas/173/orig 2025-12-04T09:33:40.5383375Z * [new branch] gh/guilhermeleobas/193/base -> origin/gh/guilhermeleobas/193/base 2025-12-04T09:33:40.5384662Z * [new branch] gh/guilhermeleobas/193/head -> origin/gh/guilhermeleobas/193/head 2025-12-04T09:33:40.5386068Z * [new branch] gh/guilhermeleobas/193/orig -> origin/gh/guilhermeleobas/193/orig 2025-12-04T09:33:40.5387758Z * [new branch] gh/guilhermeleobas/204/base -> origin/gh/guilhermeleobas/204/base 2025-12-04T09:33:40.5389042Z * [new branch] gh/guilhermeleobas/204/head -> origin/gh/guilhermeleobas/204/head 2025-12-04T09:33:40.5390292Z * [new branch] gh/guilhermeleobas/204/orig -> origin/gh/guilhermeleobas/204/orig 2025-12-04T09:33:40.5392090Z * [new branch] gh/guilhermeleobas/211/base -> origin/gh/guilhermeleobas/211/base 2025-12-04T09:33:40.5393345Z * [new branch] gh/guilhermeleobas/211/head -> origin/gh/guilhermeleobas/211/head 2025-12-04T09:33:40.5394662Z * [new branch] gh/guilhermeleobas/211/orig -> origin/gh/guilhermeleobas/211/orig 2025-12-04T09:33:40.5396410Z * [new branch] gh/guilhermeleobas/226/base -> origin/gh/guilhermeleobas/226/base 2025-12-04T09:33:40.5397765Z * [new branch] gh/guilhermeleobas/226/head -> origin/gh/guilhermeleobas/226/head 2025-12-04T09:33:40.5399003Z * [new branch] gh/guilhermeleobas/226/orig -> origin/gh/guilhermeleobas/226/orig 2025-12-04T09:33:40.5400693Z * [new branch] gh/guilhermeleobas/236/base -> origin/gh/guilhermeleobas/236/base 2025-12-04T09:33:40.5402014Z * [new branch] gh/guilhermeleobas/236/head -> origin/gh/guilhermeleobas/236/head 2025-12-04T09:33:40.5403286Z * [new branch] gh/guilhermeleobas/236/orig -> origin/gh/guilhermeleobas/236/orig 2025-12-04T09:33:40.5405006Z * [new branch] gh/guilhermeleobas/247/base -> origin/gh/guilhermeleobas/247/base 2025-12-04T09:33:40.5406257Z * [new branch] gh/guilhermeleobas/247/head -> origin/gh/guilhermeleobas/247/head 2025-12-04T09:33:40.5407558Z * [new branch] gh/guilhermeleobas/247/orig -> origin/gh/guilhermeleobas/247/orig 2025-12-04T09:33:40.5409345Z * [new branch] gh/guilhermeleobas/248/base -> origin/gh/guilhermeleobas/248/base 2025-12-04T09:33:40.5412464Z * [new branch] gh/guilhermeleobas/248/head -> origin/gh/guilhermeleobas/248/head 2025-12-04T09:33:40.5412764Z * [new branch] gh/guilhermeleobas/248/orig -> origin/gh/guilhermeleobas/248/orig 2025-12-04T09:33:40.5414469Z * [new branch] gh/guilhermeleobas/250/base -> origin/gh/guilhermeleobas/250/base 2025-12-04T09:33:40.5415210Z * [new branch] gh/guilhermeleobas/250/head -> origin/gh/guilhermeleobas/250/head 2025-12-04T09:33:40.5416670Z * [new branch] gh/guilhermeleobas/250/orig -> origin/gh/guilhermeleobas/250/orig 2025-12-04T09:33:40.5418879Z * [new branch] gh/guilhermeleobas/253/base -> origin/gh/guilhermeleobas/253/base 2025-12-04T09:33:40.5420169Z * [new branch] gh/guilhermeleobas/253/head -> origin/gh/guilhermeleobas/253/head 2025-12-04T09:33:40.5421475Z * [new branch] gh/guilhermeleobas/253/orig -> origin/gh/guilhermeleobas/253/orig 2025-12-04T09:33:40.5423245Z * [new branch] gh/guilhermeleobas/254/base -> origin/gh/guilhermeleobas/254/base 2025-12-04T09:33:40.5424510Z * [new branch] gh/guilhermeleobas/254/head -> origin/gh/guilhermeleobas/254/head 2025-12-04T09:33:40.5425740Z * [new branch] gh/guilhermeleobas/254/orig -> origin/gh/guilhermeleobas/254/orig 2025-12-04T09:33:40.5427546Z * [new branch] gh/guilhermeleobas/255/base -> origin/gh/guilhermeleobas/255/base 2025-12-04T09:33:40.5428832Z * [new branch] gh/guilhermeleobas/255/head -> origin/gh/guilhermeleobas/255/head 2025-12-04T09:33:40.5430156Z * [new branch] gh/guilhermeleobas/255/orig -> origin/gh/guilhermeleobas/255/orig 2025-12-04T09:33:40.5431855Z * [new branch] gh/guilhermeleobas/256/base -> origin/gh/guilhermeleobas/256/base 2025-12-04T09:33:40.5433337Z * [new branch] gh/guilhermeleobas/256/head -> origin/gh/guilhermeleobas/256/head 2025-12-04T09:33:40.5434494Z * [new branch] gh/guilhermeleobas/256/orig -> origin/gh/guilhermeleobas/256/orig 2025-12-04T09:33:40.5436200Z * [new branch] gh/guilhermeleobas/257/base -> origin/gh/guilhermeleobas/257/base 2025-12-04T09:33:40.5437474Z * [new branch] gh/guilhermeleobas/257/head -> origin/gh/guilhermeleobas/257/head 2025-12-04T09:33:40.5438886Z * [new branch] gh/guilhermeleobas/257/orig -> origin/gh/guilhermeleobas/257/orig 2025-12-04T09:33:40.5440857Z * [new branch] gh/guilhermeleobas/258/base -> origin/gh/guilhermeleobas/258/base 2025-12-04T09:33:40.5441698Z * [new branch] gh/guilhermeleobas/258/head -> origin/gh/guilhermeleobas/258/head 2025-12-04T09:33:40.5443142Z * [new branch] gh/guilhermeleobas/258/orig -> origin/gh/guilhermeleobas/258/orig 2025-12-04T09:33:40.5444970Z * [new branch] gh/guilhermeleobas/259/base -> origin/gh/guilhermeleobas/259/base 2025-12-04T09:33:40.5446216Z * [new branch] gh/guilhermeleobas/259/head -> origin/gh/guilhermeleobas/259/head 2025-12-04T09:33:40.5447609Z * [new branch] gh/guilhermeleobas/259/orig -> origin/gh/guilhermeleobas/259/orig 2025-12-04T09:33:40.5449351Z * [new branch] gh/guilhermeleobas/260/base -> origin/gh/guilhermeleobas/260/base 2025-12-04T09:33:40.5450587Z * [new branch] gh/guilhermeleobas/260/head -> origin/gh/guilhermeleobas/260/head 2025-12-04T09:33:40.5451854Z * [new branch] gh/guilhermeleobas/260/orig -> origin/gh/guilhermeleobas/260/orig 2025-12-04T09:33:40.5453565Z * [new branch] gh/guilhermeleobas/261/base -> origin/gh/guilhermeleobas/261/base 2025-12-04T09:33:40.5454876Z * [new branch] gh/guilhermeleobas/261/head -> origin/gh/guilhermeleobas/261/head 2025-12-04T09:33:40.5456130Z * [new branch] gh/guilhermeleobas/261/orig -> origin/gh/guilhermeleobas/261/orig 2025-12-04T09:33:40.5457995Z * [new branch] gh/guilhermeleobas/262/base -> origin/gh/guilhermeleobas/262/base 2025-12-04T09:33:40.5459415Z * [new branch] gh/guilhermeleobas/262/head -> origin/gh/guilhermeleobas/262/head 2025-12-04T09:33:40.5460603Z * [new branch] gh/guilhermeleobas/262/orig -> origin/gh/guilhermeleobas/262/orig 2025-12-04T09:33:40.5462522Z * [new branch] gh/guilhermeleobas/263/base -> origin/gh/guilhermeleobas/263/base 2025-12-04T09:33:40.5463697Z * [new branch] gh/guilhermeleobas/263/head -> origin/gh/guilhermeleobas/263/head 2025-12-04T09:33:40.5464980Z * [new branch] gh/guilhermeleobas/263/orig -> origin/gh/guilhermeleobas/263/orig 2025-12-04T09:33:40.5466758Z * [new branch] gh/guilhermeleobas/264/base -> origin/gh/guilhermeleobas/264/base 2025-12-04T09:33:40.5468055Z * [new branch] gh/guilhermeleobas/264/head -> origin/gh/guilhermeleobas/264/head 2025-12-04T09:33:40.5469330Z * [new branch] gh/guilhermeleobas/264/orig -> origin/gh/guilhermeleobas/264/orig 2025-12-04T09:33:40.5471062Z * [new branch] gh/guilhermeleobas/265/base -> origin/gh/guilhermeleobas/265/base 2025-12-04T09:33:40.5472311Z * [new branch] gh/guilhermeleobas/265/head -> origin/gh/guilhermeleobas/265/head 2025-12-04T09:33:40.5473605Z * [new branch] gh/guilhermeleobas/265/orig -> origin/gh/guilhermeleobas/265/orig 2025-12-04T09:33:40.5475338Z * [new branch] gh/guilhermeleobas/266/base -> origin/gh/guilhermeleobas/266/base 2025-12-04T09:33:40.5476573Z * [new branch] gh/guilhermeleobas/266/head -> origin/gh/guilhermeleobas/266/head 2025-12-04T09:33:40.5477832Z * [new branch] gh/guilhermeleobas/266/orig -> origin/gh/guilhermeleobas/266/orig 2025-12-04T09:33:40.5479756Z * [new branch] gh/guilhermeleobas/267/base -> origin/gh/guilhermeleobas/267/base 2025-12-04T09:33:40.5480991Z * [new branch] gh/guilhermeleobas/267/head -> origin/gh/guilhermeleobas/267/head 2025-12-04T09:33:40.5482290Z * [new branch] gh/guilhermeleobas/267/orig -> origin/gh/guilhermeleobas/267/orig 2025-12-04T09:33:40.5484322Z * [new branch] gh/hameerabbasi/1/base -> origin/gh/hameerabbasi/1/base 2025-12-04T09:33:40.5485613Z * [new branch] gh/hameerabbasi/1/head -> origin/gh/hameerabbasi/1/head 2025-12-04T09:33:40.5487207Z * [new branch] gh/hameerabbasi/2/base -> origin/gh/hameerabbasi/2/base 2025-12-04T09:33:40.5488527Z * [new branch] gh/hameerabbasi/2/head -> origin/gh/hameerabbasi/2/head 2025-12-04T09:33:40.5489846Z * [new branch] gh/hameerabbasi/2/orig -> origin/gh/hameerabbasi/2/orig 2025-12-04T09:33:40.5491418Z * [new branch] gh/hameerabbasi/3/base -> origin/gh/hameerabbasi/3/base 2025-12-04T09:33:40.5492796Z * [new branch] gh/hameerabbasi/3/head -> origin/gh/hameerabbasi/3/head 2025-12-04T09:33:40.5494254Z * [new branch] gh/hameerabbasi/3/orig -> origin/gh/hameerabbasi/3/orig 2025-12-04T09:33:40.5495951Z * [new branch] gh/hameerabbasi/4/base -> origin/gh/hameerabbasi/4/base 2025-12-04T09:33:40.5498102Z * [new branch] gh/hameerabbasi/4/head -> origin/gh/hameerabbasi/4/head 2025-12-04T09:33:40.5499147Z * [new branch] gh/hameerabbasi/4/orig -> origin/gh/hameerabbasi/4/orig 2025-12-04T09:33:40.5501095Z * [new branch] gh/huydhn/1/next -> origin/gh/huydhn/1/next 2025-12-04T09:33:40.5503006Z * [new branch] gh/huydhn/2/next -> origin/gh/huydhn/2/next 2025-12-04T09:33:40.5504198Z * [new branch] gh/huydhn/3/next -> origin/gh/huydhn/3/next 2025-12-04T09:33:40.5506336Z * [new branch] gh/huydhn/4/next -> origin/gh/huydhn/4/next 2025-12-04T09:33:40.5507943Z * [new branch] gh/huydhn/5/next -> origin/gh/huydhn/5/next 2025-12-04T09:33:40.5509614Z * [new branch] gh/huydhn/6/next -> origin/gh/huydhn/6/next 2025-12-04T09:33:40.5511653Z * [new branch] gh/int3/97/base -> origin/gh/int3/97/base 2025-12-04T09:33:40.5513036Z * [new branch] gh/int3/97/head -> origin/gh/int3/97/head 2025-12-04T09:33:40.5515190Z * [new branch] gh/isuruf/101/base -> origin/gh/isuruf/101/base 2025-12-04T09:33:40.5516386Z * [new branch] gh/isuruf/101/head -> origin/gh/isuruf/101/head 2025-12-04T09:33:40.5518064Z * [new branch] gh/isuruf/146/base -> origin/gh/isuruf/146/base 2025-12-04T09:33:40.5519287Z * [new branch] gh/isuruf/146/head -> origin/gh/isuruf/146/head 2025-12-04T09:33:40.5520570Z * [new branch] gh/isuruf/146/orig -> origin/gh/isuruf/146/orig 2025-12-04T09:33:40.5522207Z * [new branch] gh/isuruf/158/base -> origin/gh/isuruf/158/base 2025-12-04T09:33:40.5523467Z * [new branch] gh/isuruf/158/head -> origin/gh/isuruf/158/head 2025-12-04T09:33:40.5525104Z * [new branch] gh/isuruf/159/base -> origin/gh/isuruf/159/base 2025-12-04T09:33:40.5526349Z * [new branch] gh/isuruf/159/head -> origin/gh/isuruf/159/head 2025-12-04T09:33:40.5528066Z * [new branch] gh/isuruf/160/base -> origin/gh/isuruf/160/base 2025-12-04T09:33:40.5529289Z * [new branch] gh/isuruf/160/head -> origin/gh/isuruf/160/head 2025-12-04T09:33:40.5530589Z * [new branch] gh/isuruf/160/orig -> origin/gh/isuruf/160/orig 2025-12-04T09:33:40.5532288Z * [new branch] gh/isuruf/81/base -> origin/gh/isuruf/81/base 2025-12-04T09:33:40.5533480Z * [new branch] gh/isuruf/81/head -> origin/gh/isuruf/81/head 2025-12-04T09:33:40.5534734Z * [new branch] gh/isuruf/81/orig -> origin/gh/isuruf/81/orig 2025-12-04T09:33:40.5537308Z * [new branch] gh/jamesjwu/176/base -> origin/gh/jamesjwu/176/base 2025-12-04T09:33:40.5538649Z * [new branch] gh/jamesjwu/176/head -> origin/gh/jamesjwu/176/head 2025-12-04T09:33:40.5539875Z * [new branch] gh/jamesjwu/176/orig -> origin/gh/jamesjwu/176/orig 2025-12-04T09:33:40.5541726Z * [new branch] gh/jamesjwu/187/base -> origin/gh/jamesjwu/187/base 2025-12-04T09:33:40.5542963Z * [new branch] gh/jamesjwu/187/head -> origin/gh/jamesjwu/187/head 2025-12-04T09:33:40.5544220Z * [new branch] gh/jamesjwu/187/orig -> origin/gh/jamesjwu/187/orig 2025-12-04T09:33:40.5545915Z * [new branch] gh/jamesjwu/196/base -> origin/gh/jamesjwu/196/base 2025-12-04T09:33:40.5547185Z * [new branch] gh/jamesjwu/196/head -> origin/gh/jamesjwu/196/head 2025-12-04T09:33:40.5548480Z * [new branch] gh/jamesjwu/196/orig -> origin/gh/jamesjwu/196/orig 2025-12-04T09:33:40.5550592Z * [new branch] gh/jamesjwu/198/base -> origin/gh/jamesjwu/198/base 2025-12-04T09:33:40.5551880Z * [new branch] gh/jamesjwu/198/head -> origin/gh/jamesjwu/198/head 2025-12-04T09:33:40.5553139Z * [new branch] gh/jamesjwu/198/orig -> origin/gh/jamesjwu/198/orig 2025-12-04T09:33:40.5554862Z * [new branch] gh/jamesjwu/207/base -> origin/gh/jamesjwu/207/base 2025-12-04T09:33:40.5556306Z * [new branch] gh/jamesjwu/207/head -> origin/gh/jamesjwu/207/head 2025-12-04T09:33:40.5557564Z * [new branch] gh/jamesjwu/207/orig -> origin/gh/jamesjwu/207/orig 2025-12-04T09:33:40.5559565Z * [new branch] gh/jamesjwu/208/base -> origin/gh/jamesjwu/208/base 2025-12-04T09:33:40.5561263Z * [new branch] gh/jamesjwu/208/head -> origin/gh/jamesjwu/208/head 2025-12-04T09:33:40.5562526Z * [new branch] gh/jamesjwu/208/orig -> origin/gh/jamesjwu/208/orig 2025-12-04T09:33:40.5564276Z * [new branch] gh/jamesjwu/52/base -> origin/gh/jamesjwu/52/base 2025-12-04T09:33:40.5565589Z * [new branch] gh/jamesjwu/52/head -> origin/gh/jamesjwu/52/head 2025-12-04T09:33:40.5567226Z * [new branch] gh/jamesjwu/53/base -> origin/gh/jamesjwu/53/base 2025-12-04T09:33:40.5568343Z * [new branch] gh/jamesjwu/53/head -> origin/gh/jamesjwu/53/head 2025-12-04T09:33:40.5569856Z * [new branch] gh/jamesjwu/54/base -> origin/gh/jamesjwu/54/base 2025-12-04T09:33:40.5571073Z * [new branch] gh/jamesjwu/54/head -> origin/gh/jamesjwu/54/head 2025-12-04T09:33:40.5572600Z * [new branch] gh/jamesjwu/55/base -> origin/gh/jamesjwu/55/base 2025-12-04T09:33:40.5573783Z * [new branch] gh/jamesjwu/55/head -> origin/gh/jamesjwu/55/head 2025-12-04T09:33:40.5575412Z * [new branch] gh/jamesjwu/56/base -> origin/gh/jamesjwu/56/base 2025-12-04T09:33:40.5576650Z * [new branch] gh/jamesjwu/56/head -> origin/gh/jamesjwu/56/head 2025-12-04T09:33:40.5578263Z * [new branch] gh/jamesjwu/57/base -> origin/gh/jamesjwu/57/base 2025-12-04T09:33:40.5579528Z * [new branch] gh/jamesjwu/57/head -> origin/gh/jamesjwu/57/head 2025-12-04T09:33:40.5581056Z * [new branch] gh/jamesjwu/58/base -> origin/gh/jamesjwu/58/base 2025-12-04T09:33:40.5582271Z * [new branch] gh/jamesjwu/58/head -> origin/gh/jamesjwu/58/head 2025-12-04T09:33:40.5583815Z * [new branch] gh/jamesjwu/59/base -> origin/gh/jamesjwu/59/base 2025-12-04T09:33:40.5585056Z * [new branch] gh/jamesjwu/59/head -> origin/gh/jamesjwu/59/head 2025-12-04T09:33:40.5586636Z * [new branch] gh/jamesjwu/60/base -> origin/gh/jamesjwu/60/base 2025-12-04T09:33:40.5587829Z * [new branch] gh/jamesjwu/60/head -> origin/gh/jamesjwu/60/head 2025-12-04T09:33:40.5589785Z * [new branch] gh/jamesjwu/61/base -> origin/gh/jamesjwu/61/base 2025-12-04T09:33:40.5591099Z * [new branch] gh/jamesjwu/61/head -> origin/gh/jamesjwu/61/head 2025-12-04T09:33:40.5592611Z * [new branch] gh/jamesjwu/62/base -> origin/gh/jamesjwu/62/base 2025-12-04T09:33:40.5593788Z * [new branch] gh/jamesjwu/62/head -> origin/gh/jamesjwu/62/head 2025-12-04T09:33:40.5595251Z * [new branch] gh/jamesjwu/63/base -> origin/gh/jamesjwu/63/base 2025-12-04T09:33:40.5596911Z * [new branch] gh/jamesjwu/63/head -> origin/gh/jamesjwu/63/head 2025-12-04T09:33:40.5599041Z * [new branch] gh/jamesjwu/64/base -> origin/gh/jamesjwu/64/base 2025-12-04T09:33:40.5600315Z * [new branch] gh/jamesjwu/64/head -> origin/gh/jamesjwu/64/head 2025-12-04T09:33:40.5601893Z * [new branch] gh/jamesjwu/65/base -> origin/gh/jamesjwu/65/base 2025-12-04T09:33:40.5603053Z * [new branch] gh/jamesjwu/65/head -> origin/gh/jamesjwu/65/head 2025-12-04T09:33:40.5605165Z * [new branch] gh/janeyx99/165/base -> origin/gh/janeyx99/165/base 2025-12-04T09:33:40.5606607Z * [new branch] gh/janeyx99/165/head -> origin/gh/janeyx99/165/head 2025-12-04T09:33:40.5607887Z * [new branch] gh/janeyx99/165/orig -> origin/gh/janeyx99/165/orig 2025-12-04T09:33:40.5609389Z * [new branch] gh/janeyx99/201/base -> origin/gh/janeyx99/201/base 2025-12-04T09:33:40.5610671Z * [new branch] gh/janeyx99/201/head -> origin/gh/janeyx99/201/head 2025-12-04T09:33:40.5611891Z * [new branch] gh/janeyx99/201/orig -> origin/gh/janeyx99/201/orig 2025-12-04T09:33:40.5613889Z * [new branch] gh/janeyx99/225/base -> origin/gh/janeyx99/225/base 2025-12-04T09:33:40.5615177Z * [new branch] gh/janeyx99/225/head -> origin/gh/janeyx99/225/head 2025-12-04T09:33:40.5616439Z * [new branch] gh/janeyx99/225/orig -> origin/gh/janeyx99/225/orig 2025-12-04T09:33:40.5618883Z * [new branch] gh/janeyx99/299/base -> origin/gh/janeyx99/299/base 2025-12-04T09:33:40.5620249Z * [new branch] gh/janeyx99/299/head -> origin/gh/janeyx99/299/head 2025-12-04T09:33:40.5621476Z * [new branch] gh/janeyx99/299/orig -> origin/gh/janeyx99/299/orig 2025-12-04T09:33:40.5623563Z * [new branch] gh/janeyx99/302/base -> origin/gh/janeyx99/302/base 2025-12-04T09:33:40.5624963Z * [new branch] gh/janeyx99/302/head -> origin/gh/janeyx99/302/head 2025-12-04T09:33:40.5626525Z * [new branch] gh/janeyx99/303/base -> origin/gh/janeyx99/303/base 2025-12-04T09:33:40.5627654Z * [new branch] gh/janeyx99/303/head -> origin/gh/janeyx99/303/head 2025-12-04T09:33:40.5629389Z * [new branch] gh/janeyx99/305/base -> origin/gh/janeyx99/305/base 2025-12-04T09:33:40.5630690Z * [new branch] gh/janeyx99/305/head -> origin/gh/janeyx99/305/head 2025-12-04T09:33:40.5632201Z * [new branch] gh/janeyx99/306/base -> origin/gh/janeyx99/306/base 2025-12-04T09:33:40.5633403Z * [new branch] gh/janeyx99/306/head -> origin/gh/janeyx99/306/head 2025-12-04T09:33:40.5635058Z * [new branch] gh/janeyx99/314/base -> origin/gh/janeyx99/314/base 2025-12-04T09:33:40.5636422Z * [new branch] gh/janeyx99/314/head -> origin/gh/janeyx99/314/head 2025-12-04T09:33:40.5637738Z * [new branch] gh/janeyx99/314/orig -> origin/gh/janeyx99/314/orig 2025-12-04T09:33:40.5639467Z * [new branch] gh/janeyx99/315/base -> origin/gh/janeyx99/315/base 2025-12-04T09:33:40.5640830Z * [new branch] gh/janeyx99/315/head -> origin/gh/janeyx99/315/head 2025-12-04T09:33:40.5642159Z * [new branch] gh/janeyx99/315/orig -> origin/gh/janeyx99/315/orig 2025-12-04T09:33:40.5643840Z * [new branch] gh/janeyx99/316/base -> origin/gh/janeyx99/316/base 2025-12-04T09:33:40.5645130Z * [new branch] gh/janeyx99/316/head -> origin/gh/janeyx99/316/head 2025-12-04T09:33:40.5646414Z * [new branch] gh/janeyx99/316/orig -> origin/gh/janeyx99/316/orig 2025-12-04T09:33:40.5648301Z * [new branch] gh/janeyx99/317/base -> origin/gh/janeyx99/317/base 2025-12-04T09:33:40.5649534Z * [new branch] gh/janeyx99/317/head -> origin/gh/janeyx99/317/head 2025-12-04T09:33:40.5650749Z * [new branch] gh/janeyx99/317/orig -> origin/gh/janeyx99/317/orig 2025-12-04T09:33:40.5652486Z * [new branch] gh/janeyx99/325/base -> origin/gh/janeyx99/325/base 2025-12-04T09:33:40.5653751Z * [new branch] gh/janeyx99/325/head -> origin/gh/janeyx99/325/head 2025-12-04T09:33:40.5655035Z * [new branch] gh/janeyx99/325/orig -> origin/gh/janeyx99/325/orig 2025-12-04T09:33:40.5657242Z * [new branch] gh/janeyx99/327/base -> origin/gh/janeyx99/327/base 2025-12-04T09:33:40.5658689Z * [new branch] gh/janeyx99/327/head -> origin/gh/janeyx99/327/head 2025-12-04T09:33:40.5660001Z * [new branch] gh/janeyx99/327/orig -> origin/gh/janeyx99/327/orig 2025-12-04T09:33:40.5661719Z * [new branch] gh/janeyx99/328/base -> origin/gh/janeyx99/328/base 2025-12-04T09:33:40.5663040Z * [new branch] gh/janeyx99/328/head -> origin/gh/janeyx99/328/head 2025-12-04T09:33:40.5664361Z * [new branch] gh/janeyx99/328/orig -> origin/gh/janeyx99/328/orig 2025-12-04T09:33:40.5665901Z * [new branch] gh/janeyx99/329/base -> origin/gh/janeyx99/329/base 2025-12-04T09:33:40.5667217Z * [new branch] gh/janeyx99/329/head -> origin/gh/janeyx99/329/head 2025-12-04T09:33:40.5668475Z * [new branch] gh/janeyx99/329/orig -> origin/gh/janeyx99/329/orig 2025-12-04T09:33:40.5670700Z * [new branch] gh/janeyx99/330/base -> origin/gh/janeyx99/330/base 2025-12-04T09:33:40.5672029Z * [new branch] gh/janeyx99/330/head -> origin/gh/janeyx99/330/head 2025-12-04T09:33:40.5673247Z * [new branch] gh/janeyx99/330/orig -> origin/gh/janeyx99/330/orig 2025-12-04T09:33:40.5675148Z * [new branch] gh/janeyx99/331/base -> origin/gh/janeyx99/331/base 2025-12-04T09:33:40.5676440Z * [new branch] gh/janeyx99/331/head -> origin/gh/janeyx99/331/head 2025-12-04T09:33:40.5677689Z * [new branch] gh/janeyx99/331/orig -> origin/gh/janeyx99/331/orig 2025-12-04T09:33:40.5679432Z * [new branch] gh/janeyx99/332/base -> origin/gh/janeyx99/332/base 2025-12-04T09:33:40.5680700Z * [new branch] gh/janeyx99/332/head -> origin/gh/janeyx99/332/head 2025-12-04T09:33:40.5681955Z * [new branch] gh/janeyx99/332/orig -> origin/gh/janeyx99/332/orig 2025-12-04T09:33:40.5683531Z * [new branch] gh/janeyx99/333/base -> origin/gh/janeyx99/333/base 2025-12-04T09:33:40.5684763Z * [new branch] gh/janeyx99/333/head -> origin/gh/janeyx99/333/head 2025-12-04T09:33:40.5685995Z * [new branch] gh/janeyx99/333/orig -> origin/gh/janeyx99/333/orig 2025-12-04T09:33:40.5687851Z * [new branch] gh/janeyx99/88/base -> origin/gh/janeyx99/88/base 2025-12-04T09:33:40.5689304Z * [new branch] gh/janeyx99/88/head -> origin/gh/janeyx99/88/head 2025-12-04T09:33:40.5690301Z * [new branch] gh/janeyx99/88/orig -> origin/gh/janeyx99/88/orig 2025-12-04T09:33:40.5692422Z * [new branch] gh/jansel/360/base -> origin/gh/jansel/360/base 2025-12-04T09:33:40.5694151Z * [new branch] gh/jansel/360/head -> origin/gh/jansel/360/head 2025-12-04T09:33:40.5695833Z * [new branch] gh/jansel/451/base -> origin/gh/jansel/451/base 2025-12-04T09:33:40.5700066Z * [new branch] gh/jansel/451/head -> origin/gh/jansel/451/head 2025-12-04T09:33:40.5701407Z * [new branch] gh/jansel/451/orig -> origin/gh/jansel/451/orig 2025-12-04T09:33:40.5703052Z * [new branch] gh/jansel/462/base -> origin/gh/jansel/462/base 2025-12-04T09:33:40.5704326Z * [new branch] gh/jansel/462/head -> origin/gh/jansel/462/head 2025-12-04T09:33:40.5705562Z * [new branch] gh/jansel/462/orig -> origin/gh/jansel/462/orig 2025-12-04T09:33:40.5707252Z * [new branch] gh/jansel/533/base -> origin/gh/jansel/533/base 2025-12-04T09:33:40.5708494Z * [new branch] gh/jansel/533/head -> origin/gh/jansel/533/head 2025-12-04T09:33:40.5709771Z * [new branch] gh/jansel/533/orig -> origin/gh/jansel/533/orig 2025-12-04T09:33:40.5711440Z * [new branch] gh/jansel/552/base -> origin/gh/jansel/552/base 2025-12-04T09:33:40.5712742Z * [new branch] gh/jansel/552/head -> origin/gh/jansel/552/head 2025-12-04T09:33:40.5714002Z * [new branch] gh/jansel/552/orig -> origin/gh/jansel/552/orig 2025-12-04T09:33:40.5715649Z * [new branch] gh/jansel/553/base -> origin/gh/jansel/553/base 2025-12-04T09:33:40.5716969Z * [new branch] gh/jansel/553/head -> origin/gh/jansel/553/head 2025-12-04T09:33:40.5718168Z * [new branch] gh/jansel/553/orig -> origin/gh/jansel/553/orig 2025-12-04T09:33:40.5719804Z * [new branch] gh/jansel/554/base -> origin/gh/jansel/554/base 2025-12-04T09:33:40.5721076Z * [new branch] gh/jansel/554/head -> origin/gh/jansel/554/head 2025-12-04T09:33:40.5722345Z * [new branch] gh/jansel/554/orig -> origin/gh/jansel/554/orig 2025-12-04T09:33:40.5724047Z * [new branch] gh/jansel/555/base -> origin/gh/jansel/555/base 2025-12-04T09:33:40.5725397Z * [new branch] gh/jansel/555/head -> origin/gh/jansel/555/head 2025-12-04T09:33:40.5726633Z * [new branch] gh/jansel/555/orig -> origin/gh/jansel/555/orig 2025-12-04T09:33:40.5728360Z * [new branch] gh/jansel/556/base -> origin/gh/jansel/556/base 2025-12-04T09:33:40.5729715Z * [new branch] gh/jansel/556/head -> origin/gh/jansel/556/head 2025-12-04T09:33:40.5730993Z * [new branch] gh/jansel/556/orig -> origin/gh/jansel/556/orig 2025-12-04T09:33:40.5732832Z * [new branch] gh/jansel/557/base -> origin/gh/jansel/557/base 2025-12-04T09:33:40.5733920Z * [new branch] gh/jansel/557/head -> origin/gh/jansel/557/head 2025-12-04T09:33:40.5735161Z * [new branch] gh/jansel/557/orig -> origin/gh/jansel/557/orig 2025-12-04T09:33:40.5736815Z * [new branch] gh/jansel/558/base -> origin/gh/jansel/558/base 2025-12-04T09:33:40.5738213Z * [new branch] gh/jansel/558/head -> origin/gh/jansel/558/head 2025-12-04T09:33:40.5739468Z * [new branch] gh/jansel/558/orig -> origin/gh/jansel/558/orig 2025-12-04T09:33:40.5741125Z * [new branch] gh/jansel/559/base -> origin/gh/jansel/559/base 2025-12-04T09:33:40.5742371Z * [new branch] gh/jansel/559/head -> origin/gh/jansel/559/head 2025-12-04T09:33:40.5743605Z * [new branch] gh/jansel/559/orig -> origin/gh/jansel/559/orig 2025-12-04T09:33:40.5745733Z * [new branch] gh/jansel/560/base -> origin/gh/jansel/560/base 2025-12-04T09:33:40.5747087Z * [new branch] gh/jansel/560/head -> origin/gh/jansel/560/head 2025-12-04T09:33:40.5748376Z * [new branch] gh/jansel/560/orig -> origin/gh/jansel/560/orig 2025-12-04T09:33:40.5750017Z * [new branch] gh/jansel/561/base -> origin/gh/jansel/561/base 2025-12-04T09:33:40.5751284Z * [new branch] gh/jansel/561/head -> origin/gh/jansel/561/head 2025-12-04T09:33:40.5752519Z * [new branch] gh/jansel/561/orig -> origin/gh/jansel/561/orig 2025-12-04T09:33:40.5754236Z * [new branch] gh/jansel/562/base -> origin/gh/jansel/562/base 2025-12-04T09:33:40.5755507Z * [new branch] gh/jansel/562/head -> origin/gh/jansel/562/head 2025-12-04T09:33:40.5756718Z * [new branch] gh/jansel/562/orig -> origin/gh/jansel/562/orig 2025-12-04T09:33:40.5758670Z * [new branch] gh/jansel/563/base -> origin/gh/jansel/563/base 2025-12-04T09:33:40.5759525Z * [new branch] gh/jansel/563/head -> origin/gh/jansel/563/head 2025-12-04T09:33:40.5760932Z * [new branch] gh/jansel/563/orig -> origin/gh/jansel/563/orig 2025-12-04T09:33:40.5762998Z * [new branch] gh/jansel/564/base -> origin/gh/jansel/564/base 2025-12-04T09:33:40.5764383Z * [new branch] gh/jansel/564/head -> origin/gh/jansel/564/head 2025-12-04T09:33:40.5765654Z * [new branch] gh/jansel/564/orig -> origin/gh/jansel/564/orig 2025-12-04T09:33:40.5767365Z * [new branch] gh/jansel/565/base -> origin/gh/jansel/565/base 2025-12-04T09:33:40.5768660Z * [new branch] gh/jansel/565/head -> origin/gh/jansel/565/head 2025-12-04T09:33:40.5769937Z * [new branch] gh/jansel/565/orig -> origin/gh/jansel/565/orig 2025-12-04T09:33:40.5771814Z * [new branch] gh/jansel/566/base -> origin/gh/jansel/566/base 2025-12-04T09:33:40.5773063Z * [new branch] gh/jansel/566/head -> origin/gh/jansel/566/head 2025-12-04T09:33:40.5774791Z * [new branch] gh/jansel/566/orig -> origin/gh/jansel/566/orig 2025-12-04T09:33:40.5776561Z * [new branch] gh/jansel/567/base -> origin/gh/jansel/567/base 2025-12-04T09:33:40.5778009Z * [new branch] gh/jansel/567/head -> origin/gh/jansel/567/head 2025-12-04T09:33:40.5779208Z * [new branch] gh/jansel/567/orig -> origin/gh/jansel/567/orig 2025-12-04T09:33:40.5780949Z * [new branch] gh/jansel/568/base -> origin/gh/jansel/568/base 2025-12-04T09:33:40.5782354Z * [new branch] gh/jansel/568/head -> origin/gh/jansel/568/head 2025-12-04T09:33:40.5783596Z * [new branch] gh/jansel/568/orig -> origin/gh/jansel/568/orig 2025-12-04T09:33:40.5785295Z * [new branch] gh/jansel/569/base -> origin/gh/jansel/569/base 2025-12-04T09:33:40.5786550Z * [new branch] gh/jansel/569/head -> origin/gh/jansel/569/head 2025-12-04T09:33:40.5787936Z * [new branch] gh/jansel/569/orig -> origin/gh/jansel/569/orig 2025-12-04T09:33:40.5789649Z * [new branch] gh/jansel/570/base -> origin/gh/jansel/570/base 2025-12-04T09:33:40.5790888Z * [new branch] gh/jansel/570/head -> origin/gh/jansel/570/head 2025-12-04T09:33:40.5792143Z * [new branch] gh/jansel/570/orig -> origin/gh/jansel/570/orig 2025-12-04T09:33:40.5794311Z * [new branch] gh/jansel/571/base -> origin/gh/jansel/571/base 2025-12-04T09:33:40.5795608Z * [new branch] gh/jansel/571/head -> origin/gh/jansel/571/head 2025-12-04T09:33:40.5797060Z * [new branch] gh/jansel/571/orig -> origin/gh/jansel/571/orig 2025-12-04T09:33:40.5798673Z * [new branch] gh/jansel/572/base -> origin/gh/jansel/572/base 2025-12-04T09:33:40.5800044Z * [new branch] gh/jansel/572/head -> origin/gh/jansel/572/head 2025-12-04T09:33:40.5801284Z * [new branch] gh/jansel/572/orig -> origin/gh/jansel/572/orig 2025-12-04T09:33:40.5803058Z * [new branch] gh/jansel/573/base -> origin/gh/jansel/573/base 2025-12-04T09:33:40.5804337Z * [new branch] gh/jansel/573/head -> origin/gh/jansel/573/head 2025-12-04T09:33:40.5805622Z * [new branch] gh/jansel/573/orig -> origin/gh/jansel/573/orig 2025-12-04T09:33:40.5807979Z * [new branch] gh/jansel/574/base -> origin/gh/jansel/574/base 2025-12-04T09:33:40.5809212Z * [new branch] gh/jansel/574/head -> origin/gh/jansel/574/head 2025-12-04T09:33:40.5810522Z * [new branch] gh/jansel/574/orig -> origin/gh/jansel/574/orig 2025-12-04T09:33:40.5812235Z * [new branch] gh/jansel/575/base -> origin/gh/jansel/575/base 2025-12-04T09:33:40.5813535Z * [new branch] gh/jansel/575/head -> origin/gh/jansel/575/head 2025-12-04T09:33:40.5814799Z * [new branch] gh/jansel/575/orig -> origin/gh/jansel/575/orig 2025-12-04T09:33:40.5816601Z * [new branch] gh/jansel/576/base -> origin/gh/jansel/576/base 2025-12-04T09:33:40.5818174Z * [new branch] gh/jansel/576/head -> origin/gh/jansel/576/head 2025-12-04T09:33:40.5819421Z * [new branch] gh/jansel/576/orig -> origin/gh/jansel/576/orig 2025-12-04T09:33:40.5821567Z * [new branch] gh/jbschlosser/247/base -> origin/gh/jbschlosser/247/base 2025-12-04T09:33:40.5822844Z * [new branch] gh/jbschlosser/247/head -> origin/gh/jbschlosser/247/head 2025-12-04T09:33:40.5824139Z * [new branch] gh/jbschlosser/247/orig -> origin/gh/jbschlosser/247/orig 2025-12-04T09:33:40.5825925Z * [new branch] gh/jbschlosser/250/base -> origin/gh/jbschlosser/250/base 2025-12-04T09:33:40.5827082Z * [new branch] gh/jbschlosser/250/head -> origin/gh/jbschlosser/250/head 2025-12-04T09:33:40.5828347Z * [new branch] gh/jbschlosser/250/orig -> origin/gh/jbschlosser/250/orig 2025-12-04T09:33:40.5830637Z * [new branch] gh/jerryzh168/1/base -> origin/gh/jerryzh168/1/base 2025-12-04T09:33:40.5831783Z * [new branch] gh/jerryzh168/1/head -> origin/gh/jerryzh168/1/head 2025-12-04T09:33:40.5833419Z * [new branch] gh/jerryzh168/1/orig -> origin/gh/jerryzh168/1/orig 2025-12-04T09:33:40.5835499Z * [new branch] gh/jiayisunx/59/base -> origin/gh/jiayisunx/59/base 2025-12-04T09:33:40.5836871Z * [new branch] gh/jiayisunx/59/head -> origin/gh/jiayisunx/59/head 2025-12-04T09:33:40.5838182Z * [new branch] gh/jiayisunx/59/orig -> origin/gh/jiayisunx/59/orig 2025-12-04T09:33:40.5839768Z * [new branch] gh/jiayisunx/61/base -> origin/gh/jiayisunx/61/base 2025-12-04T09:33:40.5840996Z * [new branch] gh/jiayisunx/61/head -> origin/gh/jiayisunx/61/head 2025-12-04T09:33:40.5842298Z * [new branch] gh/jiayisunx/61/orig -> origin/gh/jiayisunx/61/orig 2025-12-04T09:33:40.5844026Z * [new branch] gh/jiayisunx/68/base -> origin/gh/jiayisunx/68/base 2025-12-04T09:33:40.5845233Z * [new branch] gh/jiayisunx/68/head -> origin/gh/jiayisunx/68/head 2025-12-04T09:33:40.5846538Z * [new branch] gh/jiayisunx/68/orig -> origin/gh/jiayisunx/68/orig 2025-12-04T09:33:40.5848171Z * [new branch] gh/jiayisunx/77/base -> origin/gh/jiayisunx/77/base 2025-12-04T09:33:40.5849470Z * [new branch] gh/jiayisunx/77/head -> origin/gh/jiayisunx/77/head 2025-12-04T09:33:40.5850794Z * [new branch] gh/jiayisunx/77/orig -> origin/gh/jiayisunx/77/orig 2025-12-04T09:33:40.5852559Z * [new branch] gh/jiayisunx/78/base -> origin/gh/jiayisunx/78/base 2025-12-04T09:33:40.5853782Z * [new branch] gh/jiayisunx/78/head -> origin/gh/jiayisunx/78/head 2025-12-04T09:33:40.5855060Z * [new branch] gh/jiayisunx/78/orig -> origin/gh/jiayisunx/78/orig 2025-12-04T09:33:40.5856702Z * [new branch] gh/jiayisunx/79/base -> origin/gh/jiayisunx/79/base 2025-12-04T09:33:40.5858190Z * [new branch] gh/jiayisunx/79/head -> origin/gh/jiayisunx/79/head 2025-12-04T09:33:40.5859423Z * [new branch] gh/jiayisunx/79/orig -> origin/gh/jiayisunx/79/orig 2025-12-04T09:33:40.5861111Z * [new branch] gh/jiayisunx/82/base -> origin/gh/jiayisunx/82/base 2025-12-04T09:33:40.5862809Z * [new branch] gh/jiayisunx/82/head -> origin/gh/jiayisunx/82/head 2025-12-04T09:33:40.5864071Z * [new branch] gh/jiayisunx/82/orig -> origin/gh/jiayisunx/82/orig 2025-12-04T09:33:40.5865768Z * [new branch] gh/jiayisunx/83/base -> origin/gh/jiayisunx/83/base 2025-12-04T09:33:40.5867072Z * [new branch] gh/jiayisunx/83/head -> origin/gh/jiayisunx/83/head 2025-12-04T09:33:40.5868402Z * [new branch] gh/jiayisunx/83/orig -> origin/gh/jiayisunx/83/orig 2025-12-04T09:33:40.5870013Z * [new branch] gh/jiayisunx/84/base -> origin/gh/jiayisunx/84/base 2025-12-04T09:33:40.5871253Z * [new branch] gh/jiayisunx/84/head -> origin/gh/jiayisunx/84/head 2025-12-04T09:33:40.5872516Z * [new branch] gh/jiayisunx/84/orig -> origin/gh/jiayisunx/84/orig 2025-12-04T09:33:40.5874626Z * [new branch] gh/jiayisunx/85/base -> origin/gh/jiayisunx/85/base 2025-12-04T09:33:40.5875898Z * [new branch] gh/jiayisunx/85/head -> origin/gh/jiayisunx/85/head 2025-12-04T09:33:40.5877156Z * [new branch] gh/jiayisunx/85/orig -> origin/gh/jiayisunx/85/orig 2025-12-04T09:33:40.5878806Z * [new branch] gh/jiayisunx/86/base -> origin/gh/jiayisunx/86/base 2025-12-04T09:33:40.5880089Z * [new branch] gh/jiayisunx/86/head -> origin/gh/jiayisunx/86/head 2025-12-04T09:33:40.5881485Z * [new branch] gh/jiayisunx/86/orig -> origin/gh/jiayisunx/86/orig 2025-12-04T09:33:40.5883070Z * [new branch] gh/jiayisunx/87/base -> origin/gh/jiayisunx/87/base 2025-12-04T09:33:40.5884325Z * [new branch] gh/jiayisunx/87/head -> origin/gh/jiayisunx/87/head 2025-12-04T09:33:40.5885742Z * [new branch] gh/jiayisunx/87/orig -> origin/gh/jiayisunx/87/orig 2025-12-04T09:33:40.5887886Z * [new branch] gh/jiayisunx/88/base -> origin/gh/jiayisunx/88/base 2025-12-04T09:33:40.5889146Z * [new branch] gh/jiayisunx/88/head -> origin/gh/jiayisunx/88/head 2025-12-04T09:33:40.5890423Z * [new branch] gh/jiayisunx/88/orig -> origin/gh/jiayisunx/88/orig 2025-12-04T09:33:40.5892047Z * [new branch] gh/jiayisunx/89/base -> origin/gh/jiayisunx/89/base 2025-12-04T09:33:40.5893277Z * [new branch] gh/jiayisunx/89/head -> origin/gh/jiayisunx/89/head 2025-12-04T09:33:40.5894547Z * [new branch] gh/jiayisunx/89/orig -> origin/gh/jiayisunx/89/orig 2025-12-04T09:33:40.5896368Z * [new branch] gh/jiayisunx/90/base -> origin/gh/jiayisunx/90/base 2025-12-04T09:33:40.5897760Z * [new branch] gh/jiayisunx/90/head -> origin/gh/jiayisunx/90/head 2025-12-04T09:33:40.5899001Z * [new branch] gh/jiayisunx/90/orig -> origin/gh/jiayisunx/90/orig 2025-12-04T09:33:40.5900875Z * [new branch] gh/jjwu@meta.com/1/base -> origin/gh/jjwu@meta.com/1/base 2025-12-04T09:33:40.5902198Z * [new branch] gh/jjwu@meta.com/1/head -> origin/gh/jjwu@meta.com/1/head 2025-12-04T09:33:40.5904130Z * [new branch] gh/jturney/1/base -> origin/gh/jturney/1/base 2025-12-04T09:33:40.5905417Z * [new branch] gh/jturney/1/head -> origin/gh/jturney/1/head 2025-12-04T09:33:40.5906665Z * [new branch] gh/jturney/1/orig -> origin/gh/jturney/1/orig 2025-12-04T09:33:40.5908337Z * [new branch] gh/jturney/2/base -> origin/gh/jturney/2/base 2025-12-04T09:33:40.5910071Z * [new branch] gh/jturney/2/head -> origin/gh/jturney/2/head 2025-12-04T09:33:40.5911449Z * [new branch] gh/jturney/2/orig -> origin/gh/jturney/2/orig 2025-12-04T09:33:40.5913709Z * [new branch] gh/karthickai/10/base -> origin/gh/karthickai/10/base 2025-12-04T09:33:40.5915032Z * [new branch] gh/karthickai/10/head -> origin/gh/karthickai/10/head 2025-12-04T09:33:40.5916328Z * [new branch] gh/karthickai/10/orig -> origin/gh/karthickai/10/orig 2025-12-04T09:33:40.5918586Z * [new branch] gh/karthickai/11/base -> origin/gh/karthickai/11/base 2025-12-04T09:33:40.5919975Z * [new branch] gh/karthickai/11/head -> origin/gh/karthickai/11/head 2025-12-04T09:33:40.5921235Z * [new branch] gh/karthickai/11/orig -> origin/gh/karthickai/11/orig 2025-12-04T09:33:40.5923411Z * [new branch] gh/karthickai/12/base -> origin/gh/karthickai/12/base 2025-12-04T09:33:40.5924753Z * [new branch] gh/karthickai/12/head -> origin/gh/karthickai/12/head 2025-12-04T09:33:40.5926097Z * [new branch] gh/karthickai/12/orig -> origin/gh/karthickai/12/orig 2025-12-04T09:33:40.5927792Z * [new branch] gh/karthickai/13/base -> origin/gh/karthickai/13/base 2025-12-04T09:33:40.5929110Z * [new branch] gh/karthickai/13/head -> origin/gh/karthickai/13/head 2025-12-04T09:33:40.5930406Z * [new branch] gh/karthickai/13/orig -> origin/gh/karthickai/13/orig 2025-12-04T09:33:40.5932351Z * [new branch] gh/karthickai/14/base -> origin/gh/karthickai/14/base 2025-12-04T09:33:40.5933765Z * [new branch] gh/karthickai/14/head -> origin/gh/karthickai/14/head 2025-12-04T09:33:40.5935131Z * [new branch] gh/karthickai/14/orig -> origin/gh/karthickai/14/orig 2025-12-04T09:33:40.5937216Z * [new branch] gh/karthickai/15/base -> origin/gh/karthickai/15/base 2025-12-04T09:33:40.5938601Z * [new branch] gh/karthickai/15/head -> origin/gh/karthickai/15/head 2025-12-04T09:33:40.5939851Z * [new branch] gh/karthickai/15/orig -> origin/gh/karthickai/15/orig 2025-12-04T09:33:40.5941486Z * [new branch] gh/karthickai/16/base -> origin/gh/karthickai/16/base 2025-12-04T09:33:40.5942794Z * [new branch] gh/karthickai/16/head -> origin/gh/karthickai/16/head 2025-12-04T09:33:40.5944021Z * [new branch] gh/karthickai/16/orig -> origin/gh/karthickai/16/orig 2025-12-04T09:33:40.5945661Z * [new branch] gh/karthickai/17/base -> origin/gh/karthickai/17/base 2025-12-04T09:33:40.5946862Z * [new branch] gh/karthickai/17/head -> origin/gh/karthickai/17/head 2025-12-04T09:33:40.5948104Z * [new branch] gh/karthickai/17/orig -> origin/gh/karthickai/17/orig 2025-12-04T09:33:40.5949875Z * [new branch] gh/karthickai/18/base -> origin/gh/karthickai/18/base 2025-12-04T09:33:40.5951587Z * [new branch] gh/karthickai/18/head -> origin/gh/karthickai/18/head 2025-12-04T09:33:40.5953052Z * [new branch] gh/karthickai/18/orig -> origin/gh/karthickai/18/orig 2025-12-04T09:33:40.5954927Z * [new branch] gh/karthickai/19/base -> origin/gh/karthickai/19/base 2025-12-04T09:33:40.5956231Z * [new branch] gh/karthickai/19/head -> origin/gh/karthickai/19/head 2025-12-04T09:33:40.5957564Z * [new branch] gh/karthickai/19/orig -> origin/gh/karthickai/19/orig 2025-12-04T09:33:40.5960121Z * [new branch] gh/karthickai/20/base -> origin/gh/karthickai/20/base 2025-12-04T09:33:40.5961912Z * [new branch] gh/karthickai/20/head -> origin/gh/karthickai/20/head 2025-12-04T09:33:40.5963204Z * [new branch] gh/karthickai/20/orig -> origin/gh/karthickai/20/orig 2025-12-04T09:33:40.5965045Z * [new branch] gh/karthickai/21/base -> origin/gh/karthickai/21/base 2025-12-04T09:33:40.5966549Z * [new branch] gh/karthickai/21/head -> origin/gh/karthickai/21/head 2025-12-04T09:33:40.5967851Z * [new branch] gh/karthickai/21/orig -> origin/gh/karthickai/21/orig 2025-12-04T09:33:40.5969748Z * [new branch] gh/karthickai/22/base -> origin/gh/karthickai/22/base 2025-12-04T09:33:40.5970959Z * [new branch] gh/karthickai/22/head -> origin/gh/karthickai/22/head 2025-12-04T09:33:40.5972178Z * [new branch] gh/karthickai/22/orig -> origin/gh/karthickai/22/orig 2025-12-04T09:33:40.5974187Z * [new branch] gh/karthickai/23/base -> origin/gh/karthickai/23/base 2025-12-04T09:33:40.5975612Z * [new branch] gh/karthickai/23/head -> origin/gh/karthickai/23/head 2025-12-04T09:33:40.5976909Z * [new branch] gh/karthickai/23/orig -> origin/gh/karthickai/23/orig 2025-12-04T09:33:40.5978758Z * [new branch] gh/karthickai/24/base -> origin/gh/karthickai/24/base 2025-12-04T09:33:40.5980096Z * [new branch] gh/karthickai/24/head -> origin/gh/karthickai/24/head 2025-12-04T09:33:40.5981335Z * [new branch] gh/karthickai/24/orig -> origin/gh/karthickai/24/orig 2025-12-04T09:33:40.5983588Z * [new branch] gh/karthickai/25/base -> origin/gh/karthickai/25/base 2025-12-04T09:33:40.5984986Z * [new branch] gh/karthickai/25/head -> origin/gh/karthickai/25/head 2025-12-04T09:33:40.5986283Z * [new branch] gh/karthickai/25/orig -> origin/gh/karthickai/25/orig 2025-12-04T09:33:40.5988333Z * [new branch] gh/karthickai/26/base -> origin/gh/karthickai/26/base 2025-12-04T09:33:40.5989942Z * [new branch] gh/karthickai/26/head -> origin/gh/karthickai/26/head 2025-12-04T09:33:40.5991078Z * [new branch] gh/karthickai/26/orig -> origin/gh/karthickai/26/orig 2025-12-04T09:33:40.5994564Z * [new branch] gh/karthickai/6/base -> origin/gh/karthickai/6/base 2025-12-04T09:33:40.5996504Z * [new branch] gh/karthickai/6/head -> origin/gh/karthickai/6/head 2025-12-04T09:33:40.5997992Z * [new branch] gh/karthickai/6/orig -> origin/gh/karthickai/6/orig 2025-12-04T09:33:40.5999995Z * [new branch] gh/krocki/1/base -> origin/gh/krocki/1/base 2025-12-04T09:33:40.6001333Z * [new branch] gh/krocki/1/head -> origin/gh/krocki/1/head 2025-12-04T09:33:40.6002569Z * [new branch] gh/krocki/1/orig -> origin/gh/krocki/1/orig 2025-12-04T09:33:40.6004264Z * [new branch] gh/krocki/2/base -> origin/gh/krocki/2/base 2025-12-04T09:33:40.6005574Z * [new branch] gh/krocki/2/head -> origin/gh/krocki/2/head 2025-12-04T09:33:40.6006836Z * [new branch] gh/krocki/2/orig -> origin/gh/krocki/2/orig 2025-12-04T09:33:40.6008811Z * [new branch] gh/kurtamohler/60/base -> origin/gh/kurtamohler/60/base 2025-12-04T09:33:40.6010097Z * [new branch] gh/kurtamohler/60/head -> origin/gh/kurtamohler/60/head 2025-12-04T09:33:40.6011408Z * [new branch] gh/kurtamohler/60/orig -> origin/gh/kurtamohler/60/orig 2025-12-04T09:33:40.6013111Z * [new branch] gh/kurtamohler/61/base -> origin/gh/kurtamohler/61/base 2025-12-04T09:33:40.6014350Z * [new branch] gh/kurtamohler/61/head -> origin/gh/kurtamohler/61/head 2025-12-04T09:33:40.6015660Z * [new branch] gh/kurtamohler/61/orig -> origin/gh/kurtamohler/61/orig 2025-12-04T09:33:40.6017395Z * [new branch] gh/kurtamohler/62/base -> origin/gh/kurtamohler/62/base 2025-12-04T09:33:40.6018705Z * [new branch] gh/kurtamohler/62/head -> origin/gh/kurtamohler/62/head 2025-12-04T09:33:40.6019949Z * [new branch] gh/kurtamohler/62/orig -> origin/gh/kurtamohler/62/orig 2025-12-04T09:33:40.6021953Z * [new branch] gh/kurtamohler/63/base -> origin/gh/kurtamohler/63/base 2025-12-04T09:33:40.6023249Z * [new branch] gh/kurtamohler/63/head -> origin/gh/kurtamohler/63/head 2025-12-04T09:33:40.6024494Z * [new branch] gh/kurtamohler/63/orig -> origin/gh/kurtamohler/63/orig 2025-12-04T09:33:40.6026233Z * [new branch] gh/kurtamohler/64/base -> origin/gh/kurtamohler/64/base 2025-12-04T09:33:40.6027495Z * [new branch] gh/kurtamohler/64/head -> origin/gh/kurtamohler/64/head 2025-12-04T09:33:40.6028912Z * [new branch] gh/kurtamohler/64/orig -> origin/gh/kurtamohler/64/orig 2025-12-04T09:33:40.6030663Z * [new branch] gh/kurtamohler/65/base -> origin/gh/kurtamohler/65/base 2025-12-04T09:33:40.6031910Z * [new branch] gh/kurtamohler/65/head -> origin/gh/kurtamohler/65/head 2025-12-04T09:33:40.6033172Z * [new branch] gh/kurtamohler/65/orig -> origin/gh/kurtamohler/65/orig 2025-12-04T09:33:40.6034832Z * [new branch] gh/kurtamohler/66/base -> origin/gh/kurtamohler/66/base 2025-12-04T09:33:40.6036170Z * [new branch] gh/kurtamohler/66/head -> origin/gh/kurtamohler/66/head 2025-12-04T09:33:40.6037426Z * [new branch] gh/kurtamohler/66/orig -> origin/gh/kurtamohler/66/orig 2025-12-04T09:33:40.6039048Z * [new branch] gh/kurtamohler/67/base -> origin/gh/kurtamohler/67/base 2025-12-04T09:33:40.6040319Z * [new branch] gh/kurtamohler/67/head -> origin/gh/kurtamohler/67/head 2025-12-04T09:33:40.6041645Z * [new branch] gh/kurtamohler/67/orig -> origin/gh/kurtamohler/67/orig 2025-12-04T09:33:40.6043682Z * [new branch] gh/kwen2501/130/base -> origin/gh/kwen2501/130/base 2025-12-04T09:33:40.6045161Z * [new branch] gh/kwen2501/130/head -> origin/gh/kwen2501/130/head 2025-12-04T09:33:40.6046549Z * [new branch] gh/kwen2501/130/orig -> origin/gh/kwen2501/130/orig 2025-12-04T09:33:40.6048230Z * [new branch] gh/kwen2501/170/base -> origin/gh/kwen2501/170/base 2025-12-04T09:33:40.6049503Z * [new branch] gh/kwen2501/170/head -> origin/gh/kwen2501/170/head 2025-12-04T09:33:40.6051291Z * [new branch] gh/kwen2501/187/base -> origin/gh/kwen2501/187/base 2025-12-04T09:33:40.6052623Z * [new branch] gh/kwen2501/187/head -> origin/gh/kwen2501/187/head 2025-12-04T09:33:40.6053921Z * [new branch] gh/kwen2501/187/orig -> origin/gh/kwen2501/187/orig 2025-12-04T09:33:40.6055544Z * [new branch] gh/kwen2501/188/base -> origin/gh/kwen2501/188/base 2025-12-04T09:33:40.6056856Z * [new branch] gh/kwen2501/188/head -> origin/gh/kwen2501/188/head 2025-12-04T09:33:40.6058271Z * [new branch] gh/kwen2501/188/orig -> origin/gh/kwen2501/188/orig 2025-12-04T09:33:40.6059934Z * [new branch] gh/kwen2501/211/base -> origin/gh/kwen2501/211/base 2025-12-04T09:33:40.6061163Z * [new branch] gh/kwen2501/211/head -> origin/gh/kwen2501/211/head 2025-12-04T09:33:40.6062952Z * [new branch] gh/kwen2501/224/base -> origin/gh/kwen2501/224/base 2025-12-04T09:33:40.6064308Z * [new branch] gh/kwen2501/224/head -> origin/gh/kwen2501/224/head 2025-12-04T09:33:40.6065502Z * [new branch] gh/kwen2501/224/orig -> origin/gh/kwen2501/224/orig 2025-12-04T09:33:40.6067224Z * [new branch] gh/kwen2501/228/base -> origin/gh/kwen2501/228/base 2025-12-04T09:33:40.6068541Z * [new branch] gh/kwen2501/228/head -> origin/gh/kwen2501/228/head 2025-12-04T09:33:40.6069817Z * [new branch] gh/kwen2501/228/orig -> origin/gh/kwen2501/228/orig 2025-12-04T09:33:40.6071656Z * [new branch] gh/kwen2501/234/base -> origin/gh/kwen2501/234/base 2025-12-04T09:33:40.6072878Z * [new branch] gh/kwen2501/234/head -> origin/gh/kwen2501/234/head 2025-12-04T09:33:40.6074145Z * [new branch] gh/kwen2501/234/orig -> origin/gh/kwen2501/234/orig 2025-12-04T09:33:40.6075859Z * [new branch] gh/kwen2501/235/base -> origin/gh/kwen2501/235/base 2025-12-04T09:33:40.6077109Z * [new branch] gh/kwen2501/235/head -> origin/gh/kwen2501/235/head 2025-12-04T09:33:40.6078367Z * [new branch] gh/kwen2501/235/orig -> origin/gh/kwen2501/235/orig 2025-12-04T09:33:40.6080098Z * [new branch] gh/kwen2501/236/base -> origin/gh/kwen2501/236/base 2025-12-04T09:33:40.6081366Z * [new branch] gh/kwen2501/236/head -> origin/gh/kwen2501/236/head 2025-12-04T09:33:40.6082639Z * [new branch] gh/kwen2501/236/orig -> origin/gh/kwen2501/236/orig 2025-12-04T09:33:40.6084297Z * [new branch] gh/kwen2501/237/base -> origin/gh/kwen2501/237/base 2025-12-04T09:33:40.6085526Z * [new branch] gh/kwen2501/237/head -> origin/gh/kwen2501/237/head 2025-12-04T09:33:40.6086794Z * [new branch] gh/kwen2501/237/orig -> origin/gh/kwen2501/237/orig 2025-12-04T09:33:40.6088445Z * [new branch] gh/kwen2501/238/base -> origin/gh/kwen2501/238/base 2025-12-04T09:33:40.6089676Z * [new branch] gh/kwen2501/238/head -> origin/gh/kwen2501/238/head 2025-12-04T09:33:40.6090974Z * [new branch] gh/kwen2501/238/orig -> origin/gh/kwen2501/238/orig 2025-12-04T09:33:40.6092698Z * [new branch] gh/kwen2501/240/base -> origin/gh/kwen2501/240/base 2025-12-04T09:33:40.6093891Z * [new branch] gh/kwen2501/240/head -> origin/gh/kwen2501/240/head 2025-12-04T09:33:40.6095140Z * [new branch] gh/kwen2501/240/orig -> origin/gh/kwen2501/240/orig 2025-12-04T09:33:40.6100073Z * [new branch] gh/kwen2501/241/base -> origin/gh/kwen2501/241/base 2025-12-04T09:33:40.6101423Z * [new branch] gh/kwen2501/241/head -> origin/gh/kwen2501/241/head 2025-12-04T09:33:40.6102975Z * [new branch] gh/kwen2501/241/orig -> origin/gh/kwen2501/241/orig 2025-12-04T09:33:40.6104402Z * [new branch] gh/kwen2501/247/base -> origin/gh/kwen2501/247/base 2025-12-04T09:33:40.6105672Z * [new branch] gh/kwen2501/247/head -> origin/gh/kwen2501/247/head 2025-12-04T09:33:40.6106903Z * [new branch] gh/kwen2501/247/orig -> origin/gh/kwen2501/247/orig 2025-12-04T09:33:40.6108611Z * [new branch] gh/kwen2501/252/base -> origin/gh/kwen2501/252/base 2025-12-04T09:33:40.6109903Z * [new branch] gh/kwen2501/252/head -> origin/gh/kwen2501/252/head 2025-12-04T09:33:40.6111229Z * [new branch] gh/kwen2501/252/orig -> origin/gh/kwen2501/252/orig 2025-12-04T09:33:40.6113411Z * [new branch] gh/kwen2501/259/base -> origin/gh/kwen2501/259/base 2025-12-04T09:33:40.6114802Z * [new branch] gh/kwen2501/259/head -> origin/gh/kwen2501/259/head 2025-12-04T09:33:40.6116122Z * [new branch] gh/kwen2501/259/orig -> origin/gh/kwen2501/259/orig 2025-12-04T09:33:40.6118459Z * [new branch] gh/kwen2501/260/base -> origin/gh/kwen2501/260/base 2025-12-04T09:33:40.6119816Z * [new branch] gh/kwen2501/260/head -> origin/gh/kwen2501/260/head 2025-12-04T09:33:40.6121164Z * [new branch] gh/kwen2501/260/orig -> origin/gh/kwen2501/260/orig 2025-12-04T09:33:40.6122916Z * [new branch] gh/kwen2501/268/base -> origin/gh/kwen2501/268/base 2025-12-04T09:33:40.6124169Z * [new branch] gh/kwen2501/268/head -> origin/gh/kwen2501/268/head 2025-12-04T09:33:40.6125470Z * [new branch] gh/kwen2501/268/orig -> origin/gh/kwen2501/268/orig 2025-12-04T09:33:40.6127241Z * [new branch] gh/kwen2501/269/base -> origin/gh/kwen2501/269/base 2025-12-04T09:33:40.6128573Z * [new branch] gh/kwen2501/269/head -> origin/gh/kwen2501/269/head 2025-12-04T09:33:40.6129797Z * [new branch] gh/kwen2501/269/orig -> origin/gh/kwen2501/269/orig 2025-12-04T09:33:40.6131629Z * [new branch] gh/kwen2501/270/base -> origin/gh/kwen2501/270/base 2025-12-04T09:33:40.6133011Z * [new branch] gh/kwen2501/270/head -> origin/gh/kwen2501/270/head 2025-12-04T09:33:40.6134265Z * [new branch] gh/kwen2501/270/orig -> origin/gh/kwen2501/270/orig 2025-12-04T09:33:40.6136144Z * [new branch] gh/kwen2501/271/base -> origin/gh/kwen2501/271/base 2025-12-04T09:33:40.6137577Z * [new branch] gh/kwen2501/271/head -> origin/gh/kwen2501/271/head 2025-12-04T09:33:40.6138939Z * [new branch] gh/kwen2501/271/orig -> origin/gh/kwen2501/271/orig 2025-12-04T09:33:40.6140721Z * [new branch] gh/kwen2501/274/base -> origin/gh/kwen2501/274/base 2025-12-04T09:33:40.6142122Z * [new branch] gh/kwen2501/274/head -> origin/gh/kwen2501/274/head 2025-12-04T09:33:40.6143439Z * [new branch] gh/kwen2501/274/orig -> origin/gh/kwen2501/274/orig 2025-12-04T09:33:40.6145251Z * [new branch] gh/kwen2501/275/base -> origin/gh/kwen2501/275/base 2025-12-04T09:33:40.6146679Z * [new branch] gh/kwen2501/275/head -> origin/gh/kwen2501/275/head 2025-12-04T09:33:40.6148069Z * [new branch] gh/kwen2501/275/orig -> origin/gh/kwen2501/275/orig 2025-12-04T09:33:40.6150288Z * [new branch] gh/kwen2501/276/base -> origin/gh/kwen2501/276/base 2025-12-04T09:33:40.6151530Z * [new branch] gh/kwen2501/276/head -> origin/gh/kwen2501/276/head 2025-12-04T09:33:40.6152864Z * [new branch] gh/kwen2501/276/orig -> origin/gh/kwen2501/276/orig 2025-12-04T09:33:40.6154701Z * [new branch] gh/kwen2501/277/base -> origin/gh/kwen2501/277/base 2025-12-04T09:33:40.6155965Z * [new branch] gh/kwen2501/277/head -> origin/gh/kwen2501/277/head 2025-12-04T09:33:40.6157220Z * [new branch] gh/kwen2501/277/orig -> origin/gh/kwen2501/277/orig 2025-12-04T09:33:40.6158948Z * [new branch] gh/kwen2501/278/base -> origin/gh/kwen2501/278/base 2025-12-04T09:33:40.6160214Z * [new branch] gh/kwen2501/278/head -> origin/gh/kwen2501/278/head 2025-12-04T09:33:40.6161456Z * [new branch] gh/kwen2501/278/orig -> origin/gh/kwen2501/278/orig 2025-12-04T09:33:40.6163309Z * [new branch] gh/kwen2501/279/base -> origin/gh/kwen2501/279/base 2025-12-04T09:33:40.6164690Z * [new branch] gh/kwen2501/279/head -> origin/gh/kwen2501/279/head 2025-12-04T09:33:40.6166030Z * [new branch] gh/kwen2501/279/orig -> origin/gh/kwen2501/279/orig 2025-12-04T09:33:40.6167772Z * [new branch] gh/kwen2501/280/base -> origin/gh/kwen2501/280/base 2025-12-04T09:33:40.6169083Z * [new branch] gh/kwen2501/280/head -> origin/gh/kwen2501/280/head 2025-12-04T09:33:40.6170427Z * [new branch] gh/kwen2501/280/orig -> origin/gh/kwen2501/280/orig 2025-12-04T09:33:40.6172284Z * [new branch] gh/kwen2501/281/base -> origin/gh/kwen2501/281/base 2025-12-04T09:33:40.6173533Z * [new branch] gh/kwen2501/281/head -> origin/gh/kwen2501/281/head 2025-12-04T09:33:40.6174804Z * [new branch] gh/kwen2501/281/orig -> origin/gh/kwen2501/281/orig 2025-12-04T09:33:40.6176560Z * [new branch] gh/kwen2501/282/base -> origin/gh/kwen2501/282/base 2025-12-04T09:33:40.6178018Z * [new branch] gh/kwen2501/282/head -> origin/gh/kwen2501/282/head 2025-12-04T09:33:40.6179279Z * [new branch] gh/kwen2501/282/orig -> origin/gh/kwen2501/282/orig 2025-12-04T09:33:40.6181094Z * [new branch] gh/kwen2501/283/base -> origin/gh/kwen2501/283/base 2025-12-04T09:33:40.6182562Z * [new branch] gh/kwen2501/283/head -> origin/gh/kwen2501/283/head 2025-12-04T09:33:40.6183887Z * [new branch] gh/kwen2501/283/orig -> origin/gh/kwen2501/283/orig 2025-12-04T09:33:40.6186153Z * [new branch] gh/kwen2501/284/base -> origin/gh/kwen2501/284/base 2025-12-04T09:33:40.6187472Z * [new branch] gh/kwen2501/284/head -> origin/gh/kwen2501/284/head 2025-12-04T09:33:40.6188792Z * [new branch] gh/kwen2501/284/orig -> origin/gh/kwen2501/284/orig 2025-12-04T09:33:40.6190595Z * [new branch] gh/kwen2501/285/base -> origin/gh/kwen2501/285/base 2025-12-04T09:33:40.6191854Z * [new branch] gh/kwen2501/285/head -> origin/gh/kwen2501/285/head 2025-12-04T09:33:40.6193133Z * [new branch] gh/kwen2501/285/orig -> origin/gh/kwen2501/285/orig 2025-12-04T09:33:40.6194942Z * [new branch] gh/kwen2501/286/base -> origin/gh/kwen2501/286/base 2025-12-04T09:33:40.6196422Z * [new branch] gh/kwen2501/286/head -> origin/gh/kwen2501/286/head 2025-12-04T09:33:40.6197848Z * [new branch] gh/kwen2501/286/orig -> origin/gh/kwen2501/286/orig 2025-12-04T09:33:40.6199454Z * [new branch] gh/kwen2501/287/base -> origin/gh/kwen2501/287/base 2025-12-04T09:33:40.6200844Z * [new branch] gh/kwen2501/287/head -> origin/gh/kwen2501/287/head 2025-12-04T09:33:40.6201954Z * [new branch] gh/kwen2501/287/orig -> origin/gh/kwen2501/287/orig 2025-12-04T09:33:40.6203719Z * [new branch] gh/kwen2501/288/base -> origin/gh/kwen2501/288/base 2025-12-04T09:33:40.6205033Z * [new branch] gh/kwen2501/288/head -> origin/gh/kwen2501/288/head 2025-12-04T09:33:40.6206358Z * [new branch] gh/kwen2501/288/orig -> origin/gh/kwen2501/288/orig 2025-12-04T09:33:40.6208433Z * [new branch] gh/laithsakka/251/base -> origin/gh/laithsakka/251/base 2025-12-04T09:33:40.6210220Z * [new branch] gh/laithsakka/251/head -> origin/gh/laithsakka/251/head 2025-12-04T09:33:40.6211511Z * [new branch] gh/laithsakka/251/orig -> origin/gh/laithsakka/251/orig 2025-12-04T09:33:40.6213192Z * [new branch] gh/laithsakka/276/base -> origin/gh/laithsakka/276/base 2025-12-04T09:33:40.6214459Z * [new branch] gh/laithsakka/276/head -> origin/gh/laithsakka/276/head 2025-12-04T09:33:40.6215775Z * [new branch] gh/laithsakka/276/orig -> origin/gh/laithsakka/276/orig 2025-12-04T09:33:40.6217611Z * [new branch] gh/laithsakka/28/base -> origin/gh/laithsakka/28/base 2025-12-04T09:33:40.6219128Z * [new branch] gh/laithsakka/29/base -> origin/gh/laithsakka/29/base 2025-12-04T09:33:40.6220599Z * [new branch] gh/laithsakka/30/base -> origin/gh/laithsakka/30/base 2025-12-04T09:33:40.6221860Z * [new branch] gh/laithsakka/30/head -> origin/gh/laithsakka/30/head 2025-12-04T09:33:40.6223569Z * [new branch] gh/laithsakka/31/base -> origin/gh/laithsakka/31/base 2025-12-04T09:33:40.6224674Z * [new branch] gh/laithsakka/31/head -> origin/gh/laithsakka/31/head 2025-12-04T09:33:40.6227042Z * [new branch] gh/laithsakka/313/base -> origin/gh/laithsakka/313/base 2025-12-04T09:33:40.6228231Z * [new branch] gh/laithsakka/313/head -> origin/gh/laithsakka/313/head 2025-12-04T09:33:40.6229562Z * [new branch] gh/laithsakka/313/orig -> origin/gh/laithsakka/313/orig 2025-12-04T09:33:40.6231498Z * [new branch] gh/laithsakka/316/base -> origin/gh/laithsakka/316/base 2025-12-04T09:33:40.6232607Z * [new branch] gh/laithsakka/316/head -> origin/gh/laithsakka/316/head 2025-12-04T09:33:40.6234160Z * [new branch] gh/laithsakka/316/orig -> origin/gh/laithsakka/316/orig 2025-12-04T09:33:40.6235850Z * [new branch] gh/laithsakka/317/base -> origin/gh/laithsakka/317/base 2025-12-04T09:33:40.6236955Z * [new branch] gh/laithsakka/317/head -> origin/gh/laithsakka/317/head 2025-12-04T09:33:40.6238115Z * [new branch] gh/laithsakka/317/orig -> origin/gh/laithsakka/317/orig 2025-12-04T09:33:40.6240145Z * [new branch] gh/laithsakka/319/base -> origin/gh/laithsakka/319/base 2025-12-04T09:33:40.6241377Z * [new branch] gh/laithsakka/319/head -> origin/gh/laithsakka/319/head 2025-12-04T09:33:40.6242639Z * [new branch] gh/laithsakka/319/orig -> origin/gh/laithsakka/319/orig 2025-12-04T09:33:40.6244314Z * [new branch] gh/laithsakka/32/base -> origin/gh/laithsakka/32/base 2025-12-04T09:33:40.6245389Z * [new branch] gh/laithsakka/32/head -> origin/gh/laithsakka/32/head 2025-12-04T09:33:40.6247319Z * [new branch] gh/laithsakka/320/base -> origin/gh/laithsakka/320/base 2025-12-04T09:33:40.6248362Z * [new branch] gh/laithsakka/320/head -> origin/gh/laithsakka/320/head 2025-12-04T09:33:40.6249646Z * [new branch] gh/laithsakka/320/orig -> origin/gh/laithsakka/320/orig 2025-12-04T09:33:40.6251730Z * [new branch] gh/laithsakka/321/base -> origin/gh/laithsakka/321/base 2025-12-04T09:33:40.6253026Z * [new branch] gh/laithsakka/321/head -> origin/gh/laithsakka/321/head 2025-12-04T09:33:40.6254189Z * [new branch] gh/laithsakka/321/orig -> origin/gh/laithsakka/321/orig 2025-12-04T09:33:40.6256243Z * [new branch] gh/laithsakka/322/base -> origin/gh/laithsakka/322/base 2025-12-04T09:33:40.6257773Z * [new branch] gh/laithsakka/322/head -> origin/gh/laithsakka/322/head 2025-12-04T09:33:40.6259050Z * [new branch] gh/laithsakka/322/orig -> origin/gh/laithsakka/322/orig 2025-12-04T09:33:40.6260856Z * [new branch] gh/laithsakka/323/base -> origin/gh/laithsakka/323/base 2025-12-04T09:33:40.6262050Z * [new branch] gh/laithsakka/323/head -> origin/gh/laithsakka/323/head 2025-12-04T09:33:40.6263367Z * [new branch] gh/laithsakka/323/orig -> origin/gh/laithsakka/323/orig 2025-12-04T09:33:40.6265338Z * [new branch] gh/laithsakka/324/base -> origin/gh/laithsakka/324/base 2025-12-04T09:33:40.6266428Z * [new branch] gh/laithsakka/324/head -> origin/gh/laithsakka/324/head 2025-12-04T09:33:40.6267633Z * [new branch] gh/laithsakka/324/orig -> origin/gh/laithsakka/324/orig 2025-12-04T09:33:40.6269531Z * [new branch] gh/laithsakka/325/base -> origin/gh/laithsakka/325/base 2025-12-04T09:33:40.6270727Z * [new branch] gh/laithsakka/325/head -> origin/gh/laithsakka/325/head 2025-12-04T09:33:40.6271949Z * [new branch] gh/laithsakka/325/orig -> origin/gh/laithsakka/325/orig 2025-12-04T09:33:40.6274076Z * [new branch] gh/laithsakka/326/base -> origin/gh/laithsakka/326/base 2025-12-04T09:33:40.6275332Z * [new branch] gh/laithsakka/326/head -> origin/gh/laithsakka/326/head 2025-12-04T09:33:40.6276604Z * [new branch] gh/laithsakka/326/orig -> origin/gh/laithsakka/326/orig 2025-12-04T09:33:40.6278535Z * [new branch] gh/laithsakka/327/base -> origin/gh/laithsakka/327/base 2025-12-04T09:33:40.6279772Z * [new branch] gh/laithsakka/327/head -> origin/gh/laithsakka/327/head 2025-12-04T09:33:40.6281258Z * [new branch] gh/laithsakka/327/orig -> origin/gh/laithsakka/327/orig 2025-12-04T09:33:40.6282998Z * [new branch] gh/laithsakka/328/base -> origin/gh/laithsakka/328/base 2025-12-04T09:33:40.6284736Z * [new branch] gh/laithsakka/328/head -> origin/gh/laithsakka/328/head 2025-12-04T09:33:40.6285909Z * [new branch] gh/laithsakka/328/orig -> origin/gh/laithsakka/328/orig 2025-12-04T09:33:40.6287977Z * [new branch] gh/liangel/4/base -> origin/gh/liangel/4/base 2025-12-04T09:33:40.6289124Z * [new branch] gh/liangel/4/head -> origin/gh/liangel/4/head 2025-12-04T09:33:40.6290395Z * [new branch] gh/liangel/4/orig -> origin/gh/liangel/4/orig 2025-12-04T09:33:40.6294705Z * [new branch] gh/lucaskabela/1/base -> origin/gh/lucaskabela/1/base 2025-12-04T09:33:40.6295898Z * [new branch] gh/lucaskabela/1/head -> origin/gh/lucaskabela/1/head 2025-12-04T09:33:40.6298564Z * [new branch] gh/lw/4/base -> origin/gh/lw/4/base 2025-12-04T09:33:40.6299601Z * [new branch] gh/lw/4/head -> origin/gh/lw/4/head 2025-12-04T09:33:40.6300985Z * [new branch] gh/lw/4/orig -> origin/gh/lw/4/orig 2025-12-04T09:33:40.6302638Z * [new branch] gh/lw/5/base -> origin/gh/lw/5/base 2025-12-04T09:33:40.6303824Z * [new branch] gh/lw/5/head -> origin/gh/lw/5/head 2025-12-04T09:33:40.6305165Z * [new branch] gh/lw/5/orig -> origin/gh/lw/5/orig 2025-12-04T09:33:40.6306852Z * [new branch] gh/lw/6/base -> origin/gh/lw/6/base 2025-12-04T09:33:40.6308192Z * [new branch] gh/lw/6/head -> origin/gh/lw/6/head 2025-12-04T09:33:40.6309257Z * [new branch] gh/lw/6/orig -> origin/gh/lw/6/orig 2025-12-04T09:33:40.6311469Z * [new branch] gh/malfet/14/base -> origin/gh/malfet/14/base 2025-12-04T09:33:40.6313150Z * [new branch] gh/malfet/417/base -> origin/gh/malfet/417/base 2025-12-04T09:33:40.6314303Z * [new branch] gh/malfet/417/head -> origin/gh/malfet/417/head 2025-12-04T09:33:40.6315568Z * [new branch] gh/malfet/417/orig -> origin/gh/malfet/417/orig 2025-12-04T09:33:40.6317365Z * [new branch] gh/malfet/506/base -> origin/gh/malfet/506/base 2025-12-04T09:33:40.6318521Z * [new branch] gh/malfet/506/head -> origin/gh/malfet/506/head 2025-12-04T09:33:40.6319794Z * [new branch] gh/malfet/506/orig -> origin/gh/malfet/506/orig 2025-12-04T09:33:40.6321584Z * [new branch] gh/malfet/517/base -> origin/gh/malfet/517/base 2025-12-04T09:33:40.6322789Z * [new branch] gh/malfet/517/head -> origin/gh/malfet/517/head 2025-12-04T09:33:40.6324557Z * [new branch] gh/malfet/528/base -> origin/gh/malfet/528/base 2025-12-04T09:33:40.6325793Z * [new branch] gh/malfet/528/head -> origin/gh/malfet/528/head 2025-12-04T09:33:40.6327042Z * [new branch] gh/malfet/528/orig -> origin/gh/malfet/528/orig 2025-12-04T09:33:40.6328897Z * [new branch] gh/malfet/537/base -> origin/gh/malfet/537/base 2025-12-04T09:33:40.6330060Z * [new branch] gh/malfet/537/head -> origin/gh/malfet/537/head 2025-12-04T09:33:40.6331576Z * [new branch] gh/malfet/537/orig -> origin/gh/malfet/537/orig 2025-12-04T09:33:40.6333201Z * [new branch] gh/malfet/546/base -> origin/gh/malfet/546/base 2025-12-04T09:33:40.6334467Z * [new branch] gh/malfet/546/head -> origin/gh/malfet/546/head 2025-12-04T09:33:40.6335675Z * [new branch] gh/malfet/546/orig -> origin/gh/malfet/546/orig 2025-12-04T09:33:40.6337425Z * [new branch] gh/malfet/565/base -> origin/gh/malfet/565/base 2025-12-04T09:33:40.6338539Z * [new branch] gh/malfet/565/head -> origin/gh/malfet/565/head 2025-12-04T09:33:40.6339800Z * [new branch] gh/malfet/565/orig -> origin/gh/malfet/565/orig 2025-12-04T09:33:40.6341644Z * [new branch] gh/malfet/575/base -> origin/gh/malfet/575/base 2025-12-04T09:33:40.6342937Z * [new branch] gh/malfet/575/head -> origin/gh/malfet/575/head 2025-12-04T09:33:40.6344169Z * [new branch] gh/malfet/575/orig -> origin/gh/malfet/575/orig 2025-12-04T09:33:40.6345959Z * [new branch] gh/malfet/580/base -> origin/gh/malfet/580/base 2025-12-04T09:33:40.6347129Z * [new branch] gh/malfet/580/head -> origin/gh/malfet/580/head 2025-12-04T09:33:40.6348391Z * [new branch] gh/malfet/580/orig -> origin/gh/malfet/580/orig 2025-12-04T09:33:40.6350153Z * [new branch] gh/malfet/581/base -> origin/gh/malfet/581/base 2025-12-04T09:33:40.6351292Z * [new branch] gh/malfet/581/head -> origin/gh/malfet/581/head 2025-12-04T09:33:40.6352552Z * [new branch] gh/malfet/581/orig -> origin/gh/malfet/581/orig 2025-12-04T09:33:40.6354294Z * [new branch] gh/malfet/583/base -> origin/gh/malfet/583/base 2025-12-04T09:33:40.6355390Z * [new branch] gh/malfet/583/head -> origin/gh/malfet/583/head 2025-12-04T09:33:40.6356647Z * [new branch] gh/malfet/583/orig -> origin/gh/malfet/583/orig 2025-12-04T09:33:40.6358389Z * [new branch] gh/malfet/586/base -> origin/gh/malfet/586/base 2025-12-04T09:33:40.6359919Z * [new branch] gh/malfet/586/head -> origin/gh/malfet/586/head 2025-12-04T09:33:40.6360981Z * [new branch] gh/malfet/586/orig -> origin/gh/malfet/586/orig 2025-12-04T09:33:40.6362738Z * [new branch] gh/malfet/587/base -> origin/gh/malfet/587/base 2025-12-04T09:33:40.6363866Z * [new branch] gh/malfet/587/head -> origin/gh/malfet/587/head 2025-12-04T09:33:40.6365300Z * [new branch] gh/malfet/587/orig -> origin/gh/malfet/587/orig 2025-12-04T09:33:40.6367073Z * [new branch] gh/malfet/588/base -> origin/gh/malfet/588/base 2025-12-04T09:33:40.6368284Z * [new branch] gh/malfet/588/head -> origin/gh/malfet/588/head 2025-12-04T09:33:40.6369773Z * [new branch] gh/malfet/588/orig -> origin/gh/malfet/588/orig 2025-12-04T09:33:40.6371520Z * [new branch] gh/malfet/589/base -> origin/gh/malfet/589/base 2025-12-04T09:33:40.6372677Z * [new branch] gh/malfet/589/head -> origin/gh/malfet/589/head 2025-12-04T09:33:40.6373946Z * [new branch] gh/malfet/589/orig -> origin/gh/malfet/589/orig 2025-12-04T09:33:40.6375629Z * [new branch] gh/malfet/590/base -> origin/gh/malfet/590/base 2025-12-04T09:33:40.6376986Z * [new branch] gh/malfet/590/head -> origin/gh/malfet/590/head 2025-12-04T09:33:40.6378806Z * [new branch] gh/malfet/590/orig -> origin/gh/malfet/590/orig 2025-12-04T09:33:40.6380971Z * [new branch] gh/malfet/591/base -> origin/gh/malfet/591/base 2025-12-04T09:33:40.6382147Z * [new branch] gh/malfet/591/head -> origin/gh/malfet/591/head 2025-12-04T09:33:40.6383522Z * [new branch] gh/malfet/591/orig -> origin/gh/malfet/591/orig 2025-12-04T09:33:40.6385309Z * [new branch] gh/malfet/592/base -> origin/gh/malfet/592/base 2025-12-04T09:33:40.6386482Z * [new branch] gh/malfet/592/head -> origin/gh/malfet/592/head 2025-12-04T09:33:40.6387729Z * [new branch] gh/malfet/592/orig -> origin/gh/malfet/592/orig 2025-12-04T09:33:40.6389565Z * [new branch] gh/malfet/593/base -> origin/gh/malfet/593/base 2025-12-04T09:33:40.6390666Z * [new branch] gh/malfet/593/head -> origin/gh/malfet/593/head 2025-12-04T09:33:40.6392517Z * [new branch] gh/malfet/593/orig -> origin/gh/malfet/593/orig 2025-12-04T09:33:40.6394322Z * [new branch] gh/malfet/594/base -> origin/gh/malfet/594/base 2025-12-04T09:33:40.6395568Z * [new branch] gh/malfet/594/head -> origin/gh/malfet/594/head 2025-12-04T09:33:40.6397207Z * [new branch] gh/malfet/594/orig -> origin/gh/malfet/594/orig 2025-12-04T09:33:40.6398852Z * [new branch] gh/malfet/595/base -> origin/gh/malfet/595/base 2025-12-04T09:33:40.6399951Z * [new branch] gh/malfet/595/head -> origin/gh/malfet/595/head 2025-12-04T09:33:40.6401210Z * [new branch] gh/malfet/595/orig -> origin/gh/malfet/595/orig 2025-12-04T09:33:40.6403004Z * [new branch] gh/malfet/596/base -> origin/gh/malfet/596/base 2025-12-04T09:33:40.6404180Z * [new branch] gh/malfet/596/head -> origin/gh/malfet/596/head 2025-12-04T09:33:40.6405463Z * [new branch] gh/malfet/596/orig -> origin/gh/malfet/596/orig 2025-12-04T09:33:40.6407267Z * [new branch] gh/malfet/597/base -> origin/gh/malfet/597/base 2025-12-04T09:33:40.6408473Z * [new branch] gh/malfet/597/head -> origin/gh/malfet/597/head 2025-12-04T09:33:40.6409825Z * [new branch] gh/malfet/597/orig -> origin/gh/malfet/597/orig 2025-12-04T09:33:40.6411648Z * [new branch] gh/malfet/598/base -> origin/gh/malfet/598/base 2025-12-04T09:33:40.6413013Z * [new branch] gh/malfet/598/head -> origin/gh/malfet/598/head 2025-12-04T09:33:40.6414186Z * [new branch] gh/malfet/598/orig -> origin/gh/malfet/598/orig 2025-12-04T09:33:40.6416004Z * [new branch] gh/malfet/599/base -> origin/gh/malfet/599/base 2025-12-04T09:33:40.6417224Z * [new branch] gh/malfet/599/head -> origin/gh/malfet/599/head 2025-12-04T09:33:40.6418662Z * [new branch] gh/malfet/599/orig -> origin/gh/malfet/599/orig 2025-12-04T09:33:40.6420335Z * [new branch] gh/malfet/600/base -> origin/gh/malfet/600/base 2025-12-04T09:33:40.6421435Z * [new branch] gh/malfet/600/head -> origin/gh/malfet/600/head 2025-12-04T09:33:40.6422715Z * [new branch] gh/malfet/600/orig -> origin/gh/malfet/600/orig 2025-12-04T09:33:40.6424631Z * [new branch] gh/malfet/601/base -> origin/gh/malfet/601/base 2025-12-04T09:33:40.6425838Z * [new branch] gh/malfet/601/head -> origin/gh/malfet/601/head 2025-12-04T09:33:40.6427083Z * [new branch] gh/malfet/601/orig -> origin/gh/malfet/601/orig 2025-12-04T09:33:40.6429053Z * [new branch] gh/malfet/602/base -> origin/gh/malfet/602/base 2025-12-04T09:33:40.6430233Z * [new branch] gh/malfet/602/head -> origin/gh/malfet/602/head 2025-12-04T09:33:40.6431455Z * [new branch] gh/malfet/602/orig -> origin/gh/malfet/602/orig 2025-12-04T09:33:40.6433183Z * [new branch] gh/malfet/603/base -> origin/gh/malfet/603/base 2025-12-04T09:33:40.6434301Z * [new branch] gh/malfet/603/head -> origin/gh/malfet/603/head 2025-12-04T09:33:40.6435554Z * [new branch] gh/malfet/603/orig -> origin/gh/malfet/603/orig 2025-12-04T09:33:40.6437367Z * [new branch] gh/malfet/604/base -> origin/gh/malfet/604/base 2025-12-04T09:33:40.6438503Z * [new branch] gh/malfet/604/head -> origin/gh/malfet/604/head 2025-12-04T09:33:40.6439749Z * [new branch] gh/malfet/604/orig -> origin/gh/malfet/604/orig 2025-12-04T09:33:40.6441576Z * [new branch] gh/malfet/605/base -> origin/gh/malfet/605/base 2025-12-04T09:33:40.6442737Z * [new branch] gh/malfet/605/head -> origin/gh/malfet/605/head 2025-12-04T09:33:40.6444174Z * [new branch] gh/malfet/605/orig -> origin/gh/malfet/605/orig 2025-12-04T09:33:40.6445973Z * [new branch] gh/malfet/606/base -> origin/gh/malfet/606/base 2025-12-04T09:33:40.6447898Z * [new branch] gh/malfet/606/head -> origin/gh/malfet/606/head 2025-12-04T09:33:40.6449132Z * [new branch] gh/malfet/606/orig -> origin/gh/malfet/606/orig 2025-12-04T09:33:40.6450928Z * [new branch] gh/malfet/607/base -> origin/gh/malfet/607/base 2025-12-04T09:33:40.6452067Z * [new branch] gh/malfet/607/head -> origin/gh/malfet/607/head 2025-12-04T09:33:40.6453565Z * [new branch] gh/malfet/607/orig -> origin/gh/malfet/607/orig 2025-12-04T09:33:40.6455292Z * [new branch] gh/malfet/608/base -> origin/gh/malfet/608/base 2025-12-04T09:33:40.6456431Z * [new branch] gh/malfet/608/head -> origin/gh/malfet/608/head 2025-12-04T09:33:40.6457930Z * [new branch] gh/malfet/608/orig -> origin/gh/malfet/608/orig 2025-12-04T09:33:40.6459677Z * [new branch] gh/malfet/609/base -> origin/gh/malfet/609/base 2025-12-04T09:33:40.6460865Z * [new branch] gh/malfet/609/head -> origin/gh/malfet/609/head 2025-12-04T09:33:40.6462249Z * [new branch] gh/malfet/609/orig -> origin/gh/malfet/609/orig 2025-12-04T09:33:40.6464097Z * [new branch] gh/malfet/610/base -> origin/gh/malfet/610/base 2025-12-04T09:33:40.6465378Z * [new branch] gh/malfet/610/head -> origin/gh/malfet/610/head 2025-12-04T09:33:40.6466641Z * [new branch] gh/malfet/610/orig -> origin/gh/malfet/610/orig 2025-12-04T09:33:40.6468446Z * [new branch] gh/malfet/611/base -> origin/gh/malfet/611/base 2025-12-04T09:33:40.6469558Z * [new branch] gh/malfet/611/head -> origin/gh/malfet/611/head 2025-12-04T09:33:40.6470820Z * [new branch] gh/malfet/611/orig -> origin/gh/malfet/611/orig 2025-12-04T09:33:40.6472513Z * [new branch] gh/malfet/612/base -> origin/gh/malfet/612/base 2025-12-04T09:33:40.6473659Z * [new branch] gh/malfet/612/head -> origin/gh/malfet/612/head 2025-12-04T09:33:40.6475147Z * [new branch] gh/malfet/612/orig -> origin/gh/malfet/612/orig 2025-12-04T09:33:40.6477367Z * [new branch] gh/malfet/64/base -> origin/gh/malfet/64/base 2025-12-04T09:33:40.6479118Z * [new branch] gh/malfet/64/head -> origin/gh/malfet/64/head 2025-12-04T09:33:40.6481186Z * [new branch] gh/manuelcandales/11/base -> origin/gh/manuelcandales/11/base 2025-12-04T09:33:40.6482467Z * [new branch] gh/manuelcandales/11/head -> origin/gh/manuelcandales/11/head 2025-12-04T09:33:40.6483714Z * [new branch] gh/manuelcandales/11/orig -> origin/gh/manuelcandales/11/orig 2025-12-04T09:33:40.6486098Z * [new branch] gh/markkm/1/base -> origin/gh/markkm/1/base 2025-12-04T09:33:40.6488633Z * [new branch] gh/masnesral/1/base -> origin/gh/masnesral/1/base 2025-12-04T09:33:40.6489783Z * [new branch] gh/masnesral/1/head -> origin/gh/masnesral/1/head 2025-12-04T09:33:40.6491048Z * [new branch] gh/masnesral/1/orig -> origin/gh/masnesral/1/orig 2025-12-04T09:33:40.6493228Z * [new branch] gh/mhorowitz/0/base -> origin/gh/mhorowitz/0/base 2025-12-04T09:33:40.6494438Z * [new branch] gh/mhorowitz/0/head -> origin/gh/mhorowitz/0/head 2025-12-04T09:33:40.6496142Z * [new branch] gh/mhorowitz/1/base -> origin/gh/mhorowitz/1/base 2025-12-04T09:33:40.6497754Z * [new branch] gh/mhorowitz/1/head -> origin/gh/mhorowitz/1/head 2025-12-04T09:33:40.6499269Z * [new branch] gh/mhorowitz/2/base -> origin/gh/mhorowitz/2/base 2025-12-04T09:33:40.6500485Z * [new branch] gh/mhorowitz/2/head -> origin/gh/mhorowitz/2/head 2025-12-04T09:33:40.6502064Z * [new branch] gh/mhorowitz/3/base -> origin/gh/mhorowitz/3/base 2025-12-04T09:33:40.6503162Z * [new branch] gh/mhorowitz/3/head -> origin/gh/mhorowitz/3/head 2025-12-04T09:33:40.6504792Z * [new branch] gh/mhorowitz/4/base -> origin/gh/mhorowitz/4/base 2025-12-04T09:33:40.6505890Z * [new branch] gh/mhorowitz/4/head -> origin/gh/mhorowitz/4/head 2025-12-04T09:33:40.6507456Z * [new branch] gh/mhorowitz/5/base -> origin/gh/mhorowitz/5/base 2025-12-04T09:33:40.6508549Z * [new branch] gh/mhorowitz/5/head -> origin/gh/mhorowitz/5/head 2025-12-04T09:33:40.6527942Z * [new branch] gh/mhorowitz/6/base -> origin/gh/mhorowitz/6/base 2025-12-04T09:33:40.6528769Z * [new branch] gh/mhorowitz/6/head -> origin/gh/mhorowitz/6/head 2025-12-04T09:33:40.6529504Z * [new branch] gh/mikaylagawarecki/234/base -> origin/gh/mikaylagawarecki/234/base 2025-12-04T09:33:40.6530295Z * [new branch] gh/mikaylagawarecki/234/head -> origin/gh/mikaylagawarecki/234/head 2025-12-04T09:33:40.6531079Z * [new branch] gh/mikaylagawarecki/235/base -> origin/gh/mikaylagawarecki/235/base 2025-12-04T09:33:40.6532019Z * [new branch] gh/mikaylagawarecki/235/head -> origin/gh/mikaylagawarecki/235/head 2025-12-04T09:33:40.6532804Z * [new branch] gh/mikaylagawarecki/236/base -> origin/gh/mikaylagawarecki/236/base 2025-12-04T09:33:40.6533579Z * [new branch] gh/mikaylagawarecki/236/head -> origin/gh/mikaylagawarecki/236/head 2025-12-04T09:33:40.6534350Z * [new branch] gh/mikaylagawarecki/237/base -> origin/gh/mikaylagawarecki/237/base 2025-12-04T09:33:40.6535125Z * [new branch] gh/mikaylagawarecki/237/head -> origin/gh/mikaylagawarecki/237/head 2025-12-04T09:33:40.6535900Z * [new branch] gh/mikaylagawarecki/238/base -> origin/gh/mikaylagawarecki/238/base 2025-12-04T09:33:40.6536674Z * [new branch] gh/mikaylagawarecki/238/head -> origin/gh/mikaylagawarecki/238/head 2025-12-04T09:33:40.6537520Z * [new branch] gh/mikaylagawarecki/336/base -> origin/gh/mikaylagawarecki/336/base 2025-12-04T09:33:40.6538315Z * [new branch] gh/mikaylagawarecki/336/head -> origin/gh/mikaylagawarecki/336/head 2025-12-04T09:33:40.6539086Z * [new branch] gh/mikaylagawarecki/336/orig -> origin/gh/mikaylagawarecki/336/orig 2025-12-04T09:33:40.6539859Z * [new branch] gh/mikaylagawarecki/341/base -> origin/gh/mikaylagawarecki/341/base 2025-12-04T09:33:40.6540713Z * [new branch] gh/mikaylagawarecki/341/head -> origin/gh/mikaylagawarecki/341/head 2025-12-04T09:33:40.6541488Z * [new branch] gh/mikaylagawarecki/341/orig -> origin/gh/mikaylagawarecki/341/orig 2025-12-04T09:33:40.6542269Z * [new branch] gh/mikaylagawarecki/342/base -> origin/gh/mikaylagawarecki/342/base 2025-12-04T09:33:40.6543040Z * [new branch] gh/mikaylagawarecki/342/head -> origin/gh/mikaylagawarecki/342/head 2025-12-04T09:33:40.6543806Z * [new branch] gh/mikaylagawarecki/342/orig -> origin/gh/mikaylagawarecki/342/orig 2025-12-04T09:33:40.6544584Z * [new branch] gh/mikaylagawarecki/345/base -> origin/gh/mikaylagawarecki/345/base 2025-12-04T09:33:40.6545356Z * [new branch] gh/mikaylagawarecki/345/head -> origin/gh/mikaylagawarecki/345/head 2025-12-04T09:33:40.6546130Z * [new branch] gh/mikaylagawarecki/345/orig -> origin/gh/mikaylagawarecki/345/orig 2025-12-04T09:33:40.6546895Z * [new branch] gh/mikaylagawarecki/346/base -> origin/gh/mikaylagawarecki/346/base 2025-12-04T09:33:40.6547664Z * [new branch] gh/mikaylagawarecki/346/head -> origin/gh/mikaylagawarecki/346/head 2025-12-04T09:33:40.6548497Z * [new branch] gh/mikaylagawarecki/346/orig -> origin/gh/mikaylagawarecki/346/orig 2025-12-04T09:33:40.6550413Z * [new branch] gh/mikaylagawarecki/347/base -> origin/gh/mikaylagawarecki/347/base 2025-12-04T09:33:40.6551477Z * [new branch] gh/mikaylagawarecki/347/head -> origin/gh/mikaylagawarecki/347/head 2025-12-04T09:33:40.6552704Z * [new branch] gh/mikaylagawarecki/347/orig -> origin/gh/mikaylagawarecki/347/orig 2025-12-04T09:33:40.6554567Z * [new branch] gh/mikaylagawarecki/350/base -> origin/gh/mikaylagawarecki/350/base 2025-12-04T09:33:40.6555745Z * [new branch] gh/mikaylagawarecki/350/head -> origin/gh/mikaylagawarecki/350/head 2025-12-04T09:33:40.6557067Z * [new branch] gh/mikaylagawarecki/350/orig -> origin/gh/mikaylagawarecki/350/orig 2025-12-04T09:33:40.6559304Z * [new branch] gh/mikaylagawarecki/351/base -> origin/gh/mikaylagawarecki/351/base 2025-12-04T09:33:40.6560595Z * [new branch] gh/mikaylagawarecki/351/head -> origin/gh/mikaylagawarecki/351/head 2025-12-04T09:33:40.6561911Z * [new branch] gh/mikaylagawarecki/351/orig -> origin/gh/mikaylagawarecki/351/orig 2025-12-04T09:33:40.6564036Z * [new branch] gh/mikaylagawarecki/352/base -> origin/gh/mikaylagawarecki/352/base 2025-12-04T09:33:40.6565381Z * [new branch] gh/mikaylagawarecki/352/head -> origin/gh/mikaylagawarecki/352/head 2025-12-04T09:33:40.6566744Z * [new branch] gh/mikaylagawarecki/352/orig -> origin/gh/mikaylagawarecki/352/orig 2025-12-04T09:33:40.6568674Z * [new branch] gh/mikaylagawarecki/353/base -> origin/gh/mikaylagawarecki/353/base 2025-12-04T09:33:40.6570218Z * [new branch] gh/mikaylagawarecki/353/head -> origin/gh/mikaylagawarecki/353/head 2025-12-04T09:33:40.6571452Z * [new branch] gh/mikaylagawarecki/353/orig -> origin/gh/mikaylagawarecki/353/orig 2025-12-04T09:33:40.6573142Z * [new branch] gh/mikaylagawarecki/354/base -> origin/gh/mikaylagawarecki/354/base 2025-12-04T09:33:40.6574407Z * [new branch] gh/mikaylagawarecki/354/head -> origin/gh/mikaylagawarecki/354/head 2025-12-04T09:33:40.6575693Z * [new branch] gh/mikaylagawarecki/354/orig -> origin/gh/mikaylagawarecki/354/orig 2025-12-04T09:33:40.6578041Z * [new branch] gh/mikaylagawarecki/356/base -> origin/gh/mikaylagawarecki/356/base 2025-12-04T09:33:40.6579281Z * [new branch] gh/mikaylagawarecki/356/head -> origin/gh/mikaylagawarecki/356/head 2025-12-04T09:33:40.6580560Z * [new branch] gh/mikaylagawarecki/356/orig -> origin/gh/mikaylagawarecki/356/orig 2025-12-04T09:33:40.6582359Z * [new branch] gh/mikaylagawarecki/357/base -> origin/gh/mikaylagawarecki/357/base 2025-12-04T09:33:40.6583587Z * [new branch] gh/mikaylagawarecki/357/head -> origin/gh/mikaylagawarecki/357/head 2025-12-04T09:33:40.6584807Z * [new branch] gh/mikaylagawarecki/357/orig -> origin/gh/mikaylagawarecki/357/orig 2025-12-04T09:33:40.6586717Z * [new branch] gh/mikaylagawarecki/359/base -> origin/gh/mikaylagawarecki/359/base 2025-12-04T09:33:40.6587996Z * [new branch] gh/mikaylagawarecki/359/head -> origin/gh/mikaylagawarecki/359/head 2025-12-04T09:33:40.6589277Z * [new branch] gh/mikaylagawarecki/359/orig -> origin/gh/mikaylagawarecki/359/orig 2025-12-04T09:33:40.6591076Z * [new branch] gh/mikaylagawarecki/360/base -> origin/gh/mikaylagawarecki/360/base 2025-12-04T09:33:40.6592310Z * [new branch] gh/mikaylagawarecki/360/head -> origin/gh/mikaylagawarecki/360/head 2025-12-04T09:33:40.6593562Z * [new branch] gh/mikaylagawarecki/360/orig -> origin/gh/mikaylagawarecki/360/orig 2025-12-04T09:33:40.6595448Z * [new branch] gh/mikaylagawarecki/361/base -> origin/gh/mikaylagawarecki/361/base 2025-12-04T09:33:40.6599734Z * [new branch] gh/mikaylagawarecki/361/head -> origin/gh/mikaylagawarecki/361/head 2025-12-04T09:33:40.6601112Z * [new branch] gh/mikaylagawarecki/361/orig -> origin/gh/mikaylagawarecki/361/orig 2025-12-04T09:33:40.6603091Z * [new branch] gh/mikaylagawarecki/362/base -> origin/gh/mikaylagawarecki/362/base 2025-12-04T09:33:40.6604539Z * [new branch] gh/mikaylagawarecki/362/head -> origin/gh/mikaylagawarecki/362/head 2025-12-04T09:33:40.6605844Z * [new branch] gh/mikaylagawarecki/362/orig -> origin/gh/mikaylagawarecki/362/orig 2025-12-04T09:33:40.6608082Z * [new branch] gh/mikaylagawarecki/363/base -> origin/gh/mikaylagawarecki/363/base 2025-12-04T09:33:40.6609443Z * [new branch] gh/mikaylagawarecki/363/head -> origin/gh/mikaylagawarecki/363/head 2025-12-04T09:33:40.6610721Z * [new branch] gh/mikaylagawarecki/363/orig -> origin/gh/mikaylagawarecki/363/orig 2025-12-04T09:33:40.6612979Z * [new branch] gh/mikaylagawarecki/364/base -> origin/gh/mikaylagawarecki/364/base 2025-12-04T09:33:40.6614223Z * [new branch] gh/mikaylagawarecki/364/head -> origin/gh/mikaylagawarecki/364/head 2025-12-04T09:33:40.6615551Z * [new branch] gh/mikaylagawarecki/364/orig -> origin/gh/mikaylagawarecki/364/orig 2025-12-04T09:33:40.6617692Z * [new branch] gh/mikaylagawarecki/365/base -> origin/gh/mikaylagawarecki/365/base 2025-12-04T09:33:40.6619023Z * [new branch] gh/mikaylagawarecki/365/head -> origin/gh/mikaylagawarecki/365/head 2025-12-04T09:33:40.6620330Z * [new branch] gh/mikaylagawarecki/365/orig -> origin/gh/mikaylagawarecki/365/orig 2025-12-04T09:33:40.6622345Z * [new branch] gh/mikaylagawarecki/366/base -> origin/gh/mikaylagawarecki/366/base 2025-12-04T09:33:40.6623986Z * [new branch] gh/mikaylagawarecki/366/head -> origin/gh/mikaylagawarecki/366/head 2025-12-04T09:33:40.6625238Z * [new branch] gh/mikaylagawarecki/366/orig -> origin/gh/mikaylagawarecki/366/orig 2025-12-04T09:33:40.6627113Z * [new branch] gh/mikaylagawarecki/367/base -> origin/gh/mikaylagawarecki/367/base 2025-12-04T09:33:40.6628348Z * [new branch] gh/mikaylagawarecki/367/head -> origin/gh/mikaylagawarecki/367/head 2025-12-04T09:33:40.6629611Z * [new branch] gh/mikaylagawarecki/367/orig -> origin/gh/mikaylagawarecki/367/orig 2025-12-04T09:33:40.6631556Z * [new branch] gh/mikaylagawarecki/368/base -> origin/gh/mikaylagawarecki/368/base 2025-12-04T09:33:40.6632785Z * [new branch] gh/mikaylagawarecki/368/head -> origin/gh/mikaylagawarecki/368/head 2025-12-04T09:33:40.6634164Z * [new branch] gh/mikaylagawarecki/368/orig -> origin/gh/mikaylagawarecki/368/orig 2025-12-04T09:33:40.6635977Z * [new branch] gh/mikaylagawarecki/369/base -> origin/gh/mikaylagawarecki/369/base 2025-12-04T09:33:40.6637190Z * [new branch] gh/mikaylagawarecki/369/head -> origin/gh/mikaylagawarecki/369/head 2025-12-04T09:33:40.6638454Z * [new branch] gh/mikaylagawarecki/369/orig -> origin/gh/mikaylagawarecki/369/orig 2025-12-04T09:33:40.6640502Z * [new branch] gh/mikaylagawarecki/370/base -> origin/gh/mikaylagawarecki/370/base 2025-12-04T09:33:40.6641739Z * [new branch] gh/mikaylagawarecki/370/head -> origin/gh/mikaylagawarecki/370/head 2025-12-04T09:33:40.6643005Z * [new branch] gh/mikaylagawarecki/370/orig -> origin/gh/mikaylagawarecki/370/orig 2025-12-04T09:33:40.6644862Z * [new branch] gh/mikaylagawarecki/371/base -> origin/gh/mikaylagawarecki/371/base 2025-12-04T09:33:40.6646035Z * [new branch] gh/mikaylagawarecki/371/head -> origin/gh/mikaylagawarecki/371/head 2025-12-04T09:33:40.6647309Z * [new branch] gh/mikaylagawarecki/371/orig -> origin/gh/mikaylagawarecki/371/orig 2025-12-04T09:33:40.6649161Z * [new branch] gh/mikaylagawarecki/372/base -> origin/gh/mikaylagawarecki/372/base 2025-12-04T09:33:40.6650314Z * [new branch] gh/mikaylagawarecki/372/head -> origin/gh/mikaylagawarecki/372/head 2025-12-04T09:33:40.6651568Z * [new branch] gh/mikaylagawarecki/372/orig -> origin/gh/mikaylagawarecki/372/orig 2025-12-04T09:33:40.6653329Z * [new branch] gh/mikaylagawarecki/373/base -> origin/gh/mikaylagawarecki/373/base 2025-12-04T09:33:40.6654598Z * [new branch] gh/mikaylagawarecki/373/head -> origin/gh/mikaylagawarecki/373/head 2025-12-04T09:33:40.6655811Z * [new branch] gh/mikaylagawarecki/373/orig -> origin/gh/mikaylagawarecki/373/orig 2025-12-04T09:33:40.6657950Z * [new branch] gh/mikaylagawarecki/374/base -> origin/gh/mikaylagawarecki/374/base 2025-12-04T09:33:40.6659151Z * [new branch] gh/mikaylagawarecki/374/head -> origin/gh/mikaylagawarecki/374/head 2025-12-04T09:33:40.6660429Z * [new branch] gh/mikaylagawarecki/374/orig -> origin/gh/mikaylagawarecki/374/orig 2025-12-04T09:33:40.6662190Z * [new branch] gh/mikaylagawarecki/375/base -> origin/gh/mikaylagawarecki/375/base 2025-12-04T09:33:40.6663464Z * [new branch] gh/mikaylagawarecki/375/head -> origin/gh/mikaylagawarecki/375/head 2025-12-04T09:33:40.6664783Z * [new branch] gh/mikaylagawarecki/375/orig -> origin/gh/mikaylagawarecki/375/orig 2025-12-04T09:33:40.6666621Z * [new branch] gh/mikaylagawarecki/376/base -> origin/gh/mikaylagawarecki/376/base 2025-12-04T09:33:40.6667946Z * [new branch] gh/mikaylagawarecki/376/head -> origin/gh/mikaylagawarecki/376/head 2025-12-04T09:33:40.6669153Z * [new branch] gh/mikaylagawarecki/376/orig -> origin/gh/mikaylagawarecki/376/orig 2025-12-04T09:33:40.6671010Z * [new branch] gh/mikaylagawarecki/377/base -> origin/gh/mikaylagawarecki/377/base 2025-12-04T09:33:40.6672320Z * [new branch] gh/mikaylagawarecki/377/head -> origin/gh/mikaylagawarecki/377/head 2025-12-04T09:33:40.6673604Z * [new branch] gh/mikaylagawarecki/377/orig -> origin/gh/mikaylagawarecki/377/orig 2025-12-04T09:33:40.6675513Z * [new branch] gh/mikaylagawarecki/378/base -> origin/gh/mikaylagawarecki/378/base 2025-12-04T09:33:40.6676811Z * [new branch] gh/mikaylagawarecki/378/head -> origin/gh/mikaylagawarecki/378/head 2025-12-04T09:33:40.6678058Z * [new branch] gh/mikaylagawarecki/378/orig -> origin/gh/mikaylagawarecki/378/orig 2025-12-04T09:33:40.6679884Z * [new branch] gh/mikaylagawarecki/379/base -> origin/gh/mikaylagawarecki/379/base 2025-12-04T09:33:40.6681083Z * [new branch] gh/mikaylagawarecki/379/head -> origin/gh/mikaylagawarecki/379/head 2025-12-04T09:33:40.6682379Z * [new branch] gh/mikaylagawarecki/379/orig -> origin/gh/mikaylagawarecki/379/orig 2025-12-04T09:33:40.6684010Z * [new branch] gh/mikaylagawarecki/380/base -> origin/gh/mikaylagawarecki/380/base 2025-12-04T09:33:40.6685188Z * [new branch] gh/mikaylagawarecki/380/head -> origin/gh/mikaylagawarecki/380/head 2025-12-04T09:33:40.6686425Z * [new branch] gh/mikaylagawarecki/380/orig -> origin/gh/mikaylagawarecki/380/orig 2025-12-04T09:33:40.6688532Z * [new branch] gh/mikaylagawarecki/381/base -> origin/gh/mikaylagawarecki/381/base 2025-12-04T09:33:40.6689694Z * [new branch] gh/mikaylagawarecki/381/head -> origin/gh/mikaylagawarecki/381/head 2025-12-04T09:33:40.6690944Z * [new branch] gh/mikaylagawarecki/381/orig -> origin/gh/mikaylagawarecki/381/orig 2025-12-04T09:33:40.6692710Z * [new branch] gh/mikaylagawarecki/382/base -> origin/gh/mikaylagawarecki/382/base 2025-12-04T09:33:40.6693945Z * [new branch] gh/mikaylagawarecki/382/head -> origin/gh/mikaylagawarecki/382/head 2025-12-04T09:33:40.6695156Z * [new branch] gh/mikaylagawarecki/382/orig -> origin/gh/mikaylagawarecki/382/orig 2025-12-04T09:33:40.6697736Z * [new branch] gh/mikaylagawarecki/383/base -> origin/gh/mikaylagawarecki/383/base 2025-12-04T09:33:40.6699109Z * [new branch] gh/mikaylagawarecki/383/head -> origin/gh/mikaylagawarecki/383/head 2025-12-04T09:33:40.6700390Z * [new branch] gh/mikaylagawarecki/383/orig -> origin/gh/mikaylagawarecki/383/orig 2025-12-04T09:33:40.6702184Z * [new branch] gh/mikaylagawarecki/384/base -> origin/gh/mikaylagawarecki/384/base 2025-12-04T09:33:40.6703394Z * [new branch] gh/mikaylagawarecki/384/head -> origin/gh/mikaylagawarecki/384/head 2025-12-04T09:33:40.6704663Z * [new branch] gh/mikaylagawarecki/384/orig -> origin/gh/mikaylagawarecki/384/orig 2025-12-04T09:33:40.6706489Z * [new branch] gh/mikaylagawarecki/385/base -> origin/gh/mikaylagawarecki/385/base 2025-12-04T09:33:40.6707752Z * [new branch] gh/mikaylagawarecki/385/head -> origin/gh/mikaylagawarecki/385/head 2025-12-04T09:33:40.6709001Z * [new branch] gh/mikaylagawarecki/385/orig -> origin/gh/mikaylagawarecki/385/orig 2025-12-04T09:33:40.6711111Z * [new branch] gh/mikaylagawarecki/386/base -> origin/gh/mikaylagawarecki/386/base 2025-12-04T09:33:40.6712569Z * [new branch] gh/mikaylagawarecki/386/head -> origin/gh/mikaylagawarecki/386/head 2025-12-04T09:33:40.6713908Z * [new branch] gh/mikaylagawarecki/386/orig -> origin/gh/mikaylagawarecki/386/orig 2025-12-04T09:33:40.6715956Z * [new branch] gh/mikaylagawarecki/387/base -> origin/gh/mikaylagawarecki/387/base 2025-12-04T09:33:40.6717006Z * [new branch] gh/mikaylagawarecki/387/head -> origin/gh/mikaylagawarecki/387/head 2025-12-04T09:33:40.6718269Z * [new branch] gh/mikaylagawarecki/387/orig -> origin/gh/mikaylagawarecki/387/orig 2025-12-04T09:33:40.6719953Z * [new branch] gh/mikaylagawarecki/388/base -> origin/gh/mikaylagawarecki/388/base 2025-12-04T09:33:40.6721122Z * [new branch] gh/mikaylagawarecki/388/head -> origin/gh/mikaylagawarecki/388/head 2025-12-04T09:33:40.6722451Z * [new branch] gh/mikaylagawarecki/388/orig -> origin/gh/mikaylagawarecki/388/orig 2025-12-04T09:33:40.6724353Z * [new branch] gh/mikaylagawarecki/389/base -> origin/gh/mikaylagawarecki/389/base 2025-12-04T09:33:40.6725532Z * [new branch] gh/mikaylagawarecki/389/head -> origin/gh/mikaylagawarecki/389/head 2025-12-04T09:33:40.6726792Z * [new branch] gh/mikaylagawarecki/389/orig -> origin/gh/mikaylagawarecki/389/orig 2025-12-04T09:33:40.6728754Z * [new branch] gh/mikaylagawarecki/390/base -> origin/gh/mikaylagawarecki/390/base 2025-12-04T09:33:40.6729896Z * [new branch] gh/mikaylagawarecki/390/head -> origin/gh/mikaylagawarecki/390/head 2025-12-04T09:33:40.6731111Z * [new branch] gh/mikaylagawarecki/390/orig -> origin/gh/mikaylagawarecki/390/orig 2025-12-04T09:33:40.6733118Z * [new branch] gh/mikaylagawarecki/391/base -> origin/gh/mikaylagawarecki/391/base 2025-12-04T09:33:40.6734364Z * [new branch] gh/mikaylagawarecki/391/head -> origin/gh/mikaylagawarecki/391/head 2025-12-04T09:33:40.6736197Z * [new branch] gh/mikaylagawarecki/391/orig -> origin/gh/mikaylagawarecki/391/orig 2025-12-04T09:33:40.6738231Z * [new branch] gh/mikaylagawarecki/392/base -> origin/gh/mikaylagawarecki/392/base 2025-12-04T09:33:40.6739436Z * [new branch] gh/mikaylagawarecki/392/head -> origin/gh/mikaylagawarecki/392/head 2025-12-04T09:33:40.6740806Z * [new branch] gh/mikaylagawarecki/392/orig -> origin/gh/mikaylagawarecki/392/orig 2025-12-04T09:33:40.6742846Z * [new branch] gh/mlazos/41/base -> origin/gh/mlazos/41/base 2025-12-04T09:33:40.6744009Z * [new branch] gh/mlazos/41/head -> origin/gh/mlazos/41/head 2025-12-04T09:33:40.6745452Z * [new branch] gh/mlazos/41/orig -> origin/gh/mlazos/41/orig 2025-12-04T09:33:40.6747208Z * [new branch] gh/mlazos/42/base -> origin/gh/mlazos/42/base 2025-12-04T09:33:40.6748277Z * [new branch] gh/mlazos/42/head -> origin/gh/mlazos/42/head 2025-12-04T09:33:40.6749515Z * [new branch] gh/mlazos/42/orig -> origin/gh/mlazos/42/orig 2025-12-04T09:33:40.6751618Z * [new branch] gh/mlazos/43/base -> origin/gh/mlazos/43/base 2025-12-04T09:33:40.6753013Z * [new branch] gh/mlazos/43/head -> origin/gh/mlazos/43/head 2025-12-04T09:33:40.6754175Z * [new branch] gh/mlazos/43/orig -> origin/gh/mlazos/43/orig 2025-12-04T09:33:40.6755775Z * [new branch] gh/mlazos/44/base -> origin/gh/mlazos/44/base 2025-12-04T09:33:40.6756966Z * [new branch] gh/mlazos/44/head -> origin/gh/mlazos/44/head 2025-12-04T09:33:40.6758226Z * [new branch] gh/mlazos/44/orig -> origin/gh/mlazos/44/orig 2025-12-04T09:33:40.6759949Z * [new branch] gh/mlazos/47/base -> origin/gh/mlazos/47/base 2025-12-04T09:33:40.6761097Z * [new branch] gh/mlazos/47/head -> origin/gh/mlazos/47/head 2025-12-04T09:33:40.6762554Z * [new branch] gh/mlazos/47/orig -> origin/gh/mlazos/47/orig 2025-12-04T09:33:40.6764166Z * [new branch] gh/mlazos/48/base -> origin/gh/mlazos/48/base 2025-12-04T09:33:40.6765397Z * [new branch] gh/mlazos/48/head -> origin/gh/mlazos/48/head 2025-12-04T09:33:40.6766598Z * [new branch] gh/mlazos/48/orig -> origin/gh/mlazos/48/orig 2025-12-04T09:33:40.6768328Z * [new branch] gh/mlazos/49/base -> origin/gh/mlazos/49/base 2025-12-04T09:33:40.6769851Z * [new branch] gh/mlazos/49/head -> origin/gh/mlazos/49/head 2025-12-04T09:33:40.6770824Z * [new branch] gh/mlazos/49/orig -> origin/gh/mlazos/49/orig 2025-12-04T09:33:40.6772624Z * [new branch] gh/mlazos/50/base -> origin/gh/mlazos/50/base 2025-12-04T09:33:40.6773685Z * [new branch] gh/mlazos/50/head -> origin/gh/mlazos/50/head 2025-12-04T09:33:40.6774942Z * [new branch] gh/mlazos/50/orig -> origin/gh/mlazos/50/orig 2025-12-04T09:33:40.6776962Z * [new branch] gh/mlazos/51/base -> origin/gh/mlazos/51/base 2025-12-04T09:33:40.6777844Z * [new branch] gh/mlazos/51/head -> origin/gh/mlazos/51/head 2025-12-04T09:33:40.6779264Z * [new branch] gh/mlazos/51/orig -> origin/gh/mlazos/51/orig 2025-12-04T09:33:40.6780965Z * [new branch] gh/mlazos/52/base -> origin/gh/mlazos/52/base 2025-12-04T09:33:40.6782670Z * [new branch] gh/mlazos/52/head -> origin/gh/mlazos/52/head 2025-12-04T09:33:40.6783931Z * [new branch] gh/mlazos/52/orig -> origin/gh/mlazos/52/orig 2025-12-04T09:33:40.6785728Z * [new branch] gh/mlazos/53/base -> origin/gh/mlazos/53/base 2025-12-04T09:33:40.6786931Z * [new branch] gh/mlazos/53/head -> origin/gh/mlazos/53/head 2025-12-04T09:33:40.6788177Z * [new branch] gh/mlazos/53/orig -> origin/gh/mlazos/53/orig 2025-12-04T09:33:40.6789962Z * [new branch] gh/mlazos/54/base -> origin/gh/mlazos/54/base 2025-12-04T09:33:40.6791134Z * [new branch] gh/mlazos/54/head -> origin/gh/mlazos/54/head 2025-12-04T09:33:40.6792358Z * [new branch] gh/mlazos/54/orig -> origin/gh/mlazos/54/orig 2025-12-04T09:33:40.6794121Z * [new branch] gh/mlazos/55/base -> origin/gh/mlazos/55/base 2025-12-04T09:33:40.6795310Z * [new branch] gh/mlazos/55/head -> origin/gh/mlazos/55/head 2025-12-04T09:33:40.6796932Z * [new branch] gh/mlazos/55/orig -> origin/gh/mlazos/55/orig 2025-12-04T09:33:40.6798583Z * [new branch] gh/mlazos/56/base -> origin/gh/mlazos/56/base 2025-12-04T09:33:40.6799781Z * [new branch] gh/mlazos/56/head -> origin/gh/mlazos/56/head 2025-12-04T09:33:40.6801048Z * [new branch] gh/mlazos/56/orig -> origin/gh/mlazos/56/orig 2025-12-04T09:33:40.6802825Z * [new branch] gh/mlazos/57/base -> origin/gh/mlazos/57/base 2025-12-04T09:33:40.6804490Z * [new branch] gh/mlazos/57/head -> origin/gh/mlazos/57/head 2025-12-04T09:33:40.6805636Z * [new branch] gh/mlazos/57/orig -> origin/gh/mlazos/57/orig 2025-12-04T09:33:40.6807416Z * [new branch] gh/mlazos/58/base -> origin/gh/mlazos/58/base 2025-12-04T09:33:40.6808614Z * [new branch] gh/mlazos/58/head -> origin/gh/mlazos/58/head 2025-12-04T09:33:40.6809988Z * [new branch] gh/mlazos/58/orig -> origin/gh/mlazos/58/orig 2025-12-04T09:33:40.6811658Z * [new branch] gh/mlazos/59/base -> origin/gh/mlazos/59/base 2025-12-04T09:33:40.6812865Z * [new branch] gh/mlazos/59/head -> origin/gh/mlazos/59/head 2025-12-04T09:33:40.6814239Z * [new branch] gh/mlazos/59/orig -> origin/gh/mlazos/59/orig 2025-12-04T09:33:40.6817833Z * [new branch] gh/mlazos/60/base -> origin/gh/mlazos/60/base 2025-12-04T09:33:40.6818594Z * [new branch] gh/mlazos/60/head -> origin/gh/mlazos/60/head 2025-12-04T09:33:40.6819219Z * [new branch] gh/mlazos/60/orig -> origin/gh/mlazos/60/orig 2025-12-04T09:33:40.6821134Z * [new branch] gh/mlazos/61/base -> origin/gh/mlazos/61/base 2025-12-04T09:33:40.6822343Z * [new branch] gh/mlazos/61/head -> origin/gh/mlazos/61/head 2025-12-04T09:33:40.6823594Z * [new branch] gh/mlazos/61/orig -> origin/gh/mlazos/61/orig 2025-12-04T09:33:40.6825450Z * [new branch] gh/mlazos/62/base -> origin/gh/mlazos/62/base 2025-12-04T09:33:40.6826683Z * [new branch] gh/mlazos/62/head -> origin/gh/mlazos/62/head 2025-12-04T09:33:40.6827954Z * [new branch] gh/mlazos/62/orig -> origin/gh/mlazos/62/orig 2025-12-04T09:33:40.6829796Z * [new branch] gh/mlazos/63/base -> origin/gh/mlazos/63/base 2025-12-04T09:33:40.6831066Z * [new branch] gh/mlazos/63/head -> origin/gh/mlazos/63/head 2025-12-04T09:33:40.6832509Z * [new branch] gh/mlazos/63/orig -> origin/gh/mlazos/63/orig 2025-12-04T09:33:40.6834193Z * [new branch] gh/mlazos/64/base -> origin/gh/mlazos/64/base 2025-12-04T09:33:40.6835388Z * [new branch] gh/mlazos/64/head -> origin/gh/mlazos/64/head 2025-12-04T09:33:40.6836669Z * [new branch] gh/mlazos/64/orig -> origin/gh/mlazos/64/orig 2025-12-04T09:33:40.6838415Z * [new branch] gh/mlazos/65/base -> origin/gh/mlazos/65/base 2025-12-04T09:33:40.6839607Z * [new branch] gh/mlazos/65/head -> origin/gh/mlazos/65/head 2025-12-04T09:33:40.6840845Z * [new branch] gh/mlazos/65/orig -> origin/gh/mlazos/65/orig 2025-12-04T09:33:40.6842644Z * [new branch] gh/mlazos/66/base -> origin/gh/mlazos/66/base 2025-12-04T09:33:40.6843867Z * [new branch] gh/mlazos/66/head -> origin/gh/mlazos/66/head 2025-12-04T09:33:40.6845694Z * [new branch] gh/mlazos/66/orig -> origin/gh/mlazos/66/orig 2025-12-04T09:33:40.6847850Z * [new branch] gh/mlazos/67/base -> origin/gh/mlazos/67/base 2025-12-04T09:33:40.6849632Z * [new branch] gh/mlazos/67/head -> origin/gh/mlazos/67/head 2025-12-04T09:33:40.6850915Z * [new branch] gh/mlazos/67/orig -> origin/gh/mlazos/67/orig 2025-12-04T09:33:40.6852739Z * [new branch] gh/mlazos/68/base -> origin/gh/mlazos/68/base 2025-12-04T09:33:40.6853943Z * [new branch] gh/mlazos/68/head -> origin/gh/mlazos/68/head 2025-12-04T09:33:40.6855324Z * [new branch] gh/mlazos/68/orig -> origin/gh/mlazos/68/orig 2025-12-04T09:33:40.6857135Z * [new branch] gh/mlazos/69/base -> origin/gh/mlazos/69/base 2025-12-04T09:33:40.6858411Z * [new branch] gh/mlazos/69/head -> origin/gh/mlazos/69/head 2025-12-04T09:33:40.6859646Z * [new branch] gh/mlazos/69/orig -> origin/gh/mlazos/69/orig 2025-12-04T09:33:40.6861408Z * [new branch] gh/mlazos/70/base -> origin/gh/mlazos/70/base 2025-12-04T09:33:40.6862599Z * [new branch] gh/mlazos/70/head -> origin/gh/mlazos/70/head 2025-12-04T09:33:40.6863975Z * [new branch] gh/mlazos/70/orig -> origin/gh/mlazos/70/orig 2025-12-04T09:33:40.6865719Z * [new branch] gh/mlazos/71/base -> origin/gh/mlazos/71/base 2025-12-04T09:33:40.6866906Z * [new branch] gh/mlazos/71/head -> origin/gh/mlazos/71/head 2025-12-04T09:33:40.6868347Z * [new branch] gh/mlazos/71/orig -> origin/gh/mlazos/71/orig 2025-12-04T09:33:40.6870018Z * [new branch] gh/mlazos/72/base -> origin/gh/mlazos/72/base 2025-12-04T09:33:40.6871320Z * [new branch] gh/mlazos/72/head -> origin/gh/mlazos/72/head 2025-12-04T09:33:40.6872470Z * [new branch] gh/mlazos/72/orig -> origin/gh/mlazos/72/orig 2025-12-04T09:33:40.6874388Z * [new branch] gh/mlazos/73/base -> origin/gh/mlazos/73/base 2025-12-04T09:33:40.6875630Z * [new branch] gh/mlazos/73/head -> origin/gh/mlazos/73/head 2025-12-04T09:33:40.6876908Z * [new branch] gh/mlazos/73/orig -> origin/gh/mlazos/73/orig 2025-12-04T09:33:40.6879003Z * [new branch] gh/mrmiywj/1/base -> origin/gh/mrmiywj/1/base 2025-12-04T09:33:40.6880254Z * [new branch] gh/mrmiywj/1/head -> origin/gh/mrmiywj/1/head 2025-12-04T09:33:40.6882421Z * [new branch] gh/muchulee8/73/base -> origin/gh/muchulee8/73/base 2025-12-04T09:33:40.6883998Z * [new branch] gh/muchulee8/73/head -> origin/gh/muchulee8/73/head 2025-12-04T09:33:40.6885269Z * [new branch] gh/muchulee8/73/orig -> origin/gh/muchulee8/73/orig 2025-12-04T09:33:40.6887512Z * [new branch] gh/naveenthangudu/1/base -> origin/gh/naveenthangudu/1/base 2025-12-04T09:33:40.6888730Z * [new branch] gh/naveenthangudu/1/head -> origin/gh/naveenthangudu/1/head 2025-12-04T09:33:40.6890106Z * [new branch] gh/naveenthangudu/1/orig -> origin/gh/naveenthangudu/1/orig 2025-12-04T09:33:40.6891904Z * [new branch] gh/naveenthangudu/2/base -> origin/gh/naveenthangudu/2/base 2025-12-04T09:33:40.6893121Z * [new branch] gh/naveenthangudu/2/head -> origin/gh/naveenthangudu/2/head 2025-12-04T09:33:40.6894968Z * [new branch] gh/naveenthangudu/2/orig -> origin/gh/naveenthangudu/2/orig 2025-12-04T09:33:40.6896706Z * [new branch] gh/naveenthangudu/3/base -> origin/gh/naveenthangudu/3/base 2025-12-04T09:33:40.6898072Z * [new branch] gh/naveenthangudu/3/head -> origin/gh/naveenthangudu/3/head 2025-12-04T09:33:40.6899369Z * [new branch] gh/naveenthangudu/3/orig -> origin/gh/naveenthangudu/3/orig 2025-12-04T09:33:40.6901185Z * [new branch] gh/naveenthangudu/4/base -> origin/gh/naveenthangudu/4/base 2025-12-04T09:33:40.6902368Z * [new branch] gh/naveenthangudu/4/head -> origin/gh/naveenthangudu/4/head 2025-12-04T09:33:40.6904027Z * [new branch] gh/naveenthangudu/4/orig -> origin/gh/naveenthangudu/4/orig 2025-12-04T09:33:40.6905709Z * [new branch] gh/naveenthangudu/5/base -> origin/gh/naveenthangudu/5/base 2025-12-04T09:33:40.6906898Z * [new branch] gh/naveenthangudu/5/head -> origin/gh/naveenthangudu/5/head 2025-12-04T09:33:40.6908408Z * [new branch] gh/naveenthangudu/5/orig -> origin/gh/naveenthangudu/5/orig 2025-12-04T09:33:40.6910162Z * [new branch] gh/naveenthangudu/6/base -> origin/gh/naveenthangudu/6/base 2025-12-04T09:33:40.6911374Z * [new branch] gh/naveenthangudu/6/head -> origin/gh/naveenthangudu/6/head 2025-12-04T09:33:40.6912633Z * [new branch] gh/naveenthangudu/6/orig -> origin/gh/naveenthangudu/6/orig 2025-12-04T09:33:40.6914395Z * [new branch] gh/naveenthangudu/7/base -> origin/gh/naveenthangudu/7/base 2025-12-04T09:33:40.6915575Z * [new branch] gh/naveenthangudu/7/head -> origin/gh/naveenthangudu/7/head 2025-12-04T09:33:40.6916744Z * [new branch] gh/naveenthangudu/7/orig -> origin/gh/naveenthangudu/7/orig 2025-12-04T09:33:40.6918528Z * [new branch] gh/naveenthangudu/8/base -> origin/gh/naveenthangudu/8/base 2025-12-04T09:33:40.6919845Z * [new branch] gh/naveenthangudu/8/head -> origin/gh/naveenthangudu/8/head 2025-12-04T09:33:40.6921106Z * [new branch] gh/naveenthangudu/8/orig -> origin/gh/naveenthangudu/8/orig 2025-12-04T09:33:40.6922935Z * [new branch] gh/naveenthangudu/9/base -> origin/gh/naveenthangudu/9/base 2025-12-04T09:33:40.6924031Z * [new branch] gh/naveenthangudu/9/head -> origin/gh/naveenthangudu/9/head 2025-12-04T09:33:40.6925369Z * [new branch] gh/naveenthangudu/9/orig -> origin/gh/naveenthangudu/9/orig 2025-12-04T09:33:40.6927364Z * [new branch] gh/nikitaved/1/base -> origin/gh/nikitaved/1/base 2025-12-04T09:33:40.6928562Z * [new branch] gh/nikitaved/1/head -> origin/gh/nikitaved/1/head 2025-12-04T09:33:40.6929843Z * [new branch] gh/nikitaved/1/orig -> origin/gh/nikitaved/1/orig 2025-12-04T09:33:40.6931720Z * [new branch] gh/nikitaved/10/base -> origin/gh/nikitaved/10/base 2025-12-04T09:33:40.6932869Z * [new branch] gh/nikitaved/10/head -> origin/gh/nikitaved/10/head 2025-12-04T09:33:40.6934124Z * [new branch] gh/nikitaved/10/orig -> origin/gh/nikitaved/10/orig 2025-12-04T09:33:40.6935910Z * [new branch] gh/nikitaved/11/base -> origin/gh/nikitaved/11/base 2025-12-04T09:33:40.6937240Z * [new branch] gh/nikitaved/11/head -> origin/gh/nikitaved/11/head 2025-12-04T09:33:40.6938983Z * [new branch] gh/nikitaved/11/orig -> origin/gh/nikitaved/11/orig 2025-12-04T09:33:40.6940684Z * [new branch] gh/nikitaved/12/base -> origin/gh/nikitaved/12/base 2025-12-04T09:33:40.6941907Z * [new branch] gh/nikitaved/12/head -> origin/gh/nikitaved/12/head 2025-12-04T09:33:40.6943179Z * [new branch] gh/nikitaved/12/orig -> origin/gh/nikitaved/12/orig 2025-12-04T09:33:40.6944968Z * [new branch] gh/nikitaved/13/base -> origin/gh/nikitaved/13/base 2025-12-04T09:33:40.6946163Z * [new branch] gh/nikitaved/13/head -> origin/gh/nikitaved/13/head 2025-12-04T09:33:40.6947440Z * [new branch] gh/nikitaved/13/orig -> origin/gh/nikitaved/13/orig 2025-12-04T09:33:40.6949235Z * [new branch] gh/nikitaved/14/base -> origin/gh/nikitaved/14/base 2025-12-04T09:33:40.6950399Z * [new branch] gh/nikitaved/14/head -> origin/gh/nikitaved/14/head 2025-12-04T09:33:40.6951655Z * [new branch] gh/nikitaved/14/orig -> origin/gh/nikitaved/14/orig 2025-12-04T09:33:40.6953441Z * [new branch] gh/nikitaved/15/base -> origin/gh/nikitaved/15/base 2025-12-04T09:33:40.6954753Z * [new branch] gh/nikitaved/15/head -> origin/gh/nikitaved/15/head 2025-12-04T09:33:40.6955997Z * [new branch] gh/nikitaved/15/orig -> origin/gh/nikitaved/15/orig 2025-12-04T09:33:40.6957751Z * [new branch] gh/nikitaved/16/base -> origin/gh/nikitaved/16/base 2025-12-04T09:33:40.6958931Z * [new branch] gh/nikitaved/16/head -> origin/gh/nikitaved/16/head 2025-12-04T09:33:40.6960163Z * [new branch] gh/nikitaved/16/orig -> origin/gh/nikitaved/16/orig 2025-12-04T09:33:40.6961985Z * [new branch] gh/nikitaved/2/base -> origin/gh/nikitaved/2/base 2025-12-04T09:33:40.6963147Z * [new branch] gh/nikitaved/2/head -> origin/gh/nikitaved/2/head 2025-12-04T09:33:40.6964924Z * [new branch] gh/nikitaved/2/orig -> origin/gh/nikitaved/2/orig 2025-12-04T09:33:40.6966648Z * [new branch] gh/nikitaved/4/base -> origin/gh/nikitaved/4/base 2025-12-04T09:33:40.6967852Z * [new branch] gh/nikitaved/4/head -> origin/gh/nikitaved/4/head 2025-12-04T09:33:40.6969245Z * [new branch] gh/nikitaved/4/orig -> origin/gh/nikitaved/4/orig 2025-12-04T09:33:40.6971057Z * [new branch] gh/nikitaved/5/base -> origin/gh/nikitaved/5/base 2025-12-04T09:33:40.6972303Z * [new branch] gh/nikitaved/5/head -> origin/gh/nikitaved/5/head 2025-12-04T09:33:40.6973679Z * [new branch] gh/nikitaved/5/orig -> origin/gh/nikitaved/5/orig 2025-12-04T09:33:40.6975235Z * [new branch] gh/nikitaved/6/base -> origin/gh/nikitaved/6/base 2025-12-04T09:33:40.6976505Z * [new branch] gh/nikitaved/6/head -> origin/gh/nikitaved/6/head 2025-12-04T09:33:40.6977888Z * [new branch] gh/nikitaved/6/orig -> origin/gh/nikitaved/6/orig 2025-12-04T09:33:40.6979574Z * [new branch] gh/nikitaved/8/base -> origin/gh/nikitaved/8/base 2025-12-04T09:33:40.6980861Z * [new branch] gh/nikitaved/8/head -> origin/gh/nikitaved/8/head 2025-12-04T09:33:40.6982114Z * [new branch] gh/nikitaved/8/orig -> origin/gh/nikitaved/8/orig 2025-12-04T09:33:40.6983740Z * [new branch] gh/nikitaved/9/base -> origin/gh/nikitaved/9/base 2025-12-04T09:33:40.6985004Z * [new branch] gh/nikitaved/9/head -> origin/gh/nikitaved/9/head 2025-12-04T09:33:40.6986288Z * [new branch] gh/nikitaved/9/orig -> origin/gh/nikitaved/9/orig 2025-12-04T09:33:40.6988862Z * [new branch] gh/oulgen/10/base -> origin/gh/oulgen/10/base 2025-12-04T09:33:40.6990092Z * [new branch] gh/oulgen/10/head -> origin/gh/oulgen/10/head 2025-12-04T09:33:40.6991384Z * [new branch] gh/oulgen/10/orig -> origin/gh/oulgen/10/orig 2025-12-04T09:33:40.6992990Z * [new branch] gh/oulgen/11/base -> origin/gh/oulgen/11/base 2025-12-04T09:33:40.6994255Z * [new branch] gh/oulgen/11/head -> origin/gh/oulgen/11/head 2025-12-04T09:33:40.6995515Z * [new branch] gh/oulgen/11/orig -> origin/gh/oulgen/11/orig 2025-12-04T09:33:40.6997414Z * [new branch] gh/oulgen/12/base -> origin/gh/oulgen/12/base 2025-12-04T09:33:40.6998637Z * [new branch] gh/oulgen/12/head -> origin/gh/oulgen/12/head 2025-12-04T09:33:40.6999928Z * [new branch] gh/oulgen/12/orig -> origin/gh/oulgen/12/orig 2025-12-04T09:33:40.7001569Z * [new branch] gh/oulgen/13/base -> origin/gh/oulgen/13/base 2025-12-04T09:33:40.7002820Z * [new branch] gh/oulgen/13/head -> origin/gh/oulgen/13/head 2025-12-04T09:33:40.7004160Z * [new branch] gh/oulgen/13/orig -> origin/gh/oulgen/13/orig 2025-12-04T09:33:40.7006340Z * [new branch] gh/oulgen/14/base -> origin/gh/oulgen/14/base 2025-12-04T09:33:40.7007545Z * [new branch] gh/oulgen/14/head -> origin/gh/oulgen/14/head 2025-12-04T09:33:40.7008891Z * [new branch] gh/oulgen/14/orig -> origin/gh/oulgen/14/orig 2025-12-04T09:33:40.7010503Z * [new branch] gh/oulgen/15/base -> origin/gh/oulgen/15/base 2025-12-04T09:33:40.7012086Z * [new branch] gh/oulgen/15/head -> origin/gh/oulgen/15/head 2025-12-04T09:33:40.7013093Z * [new branch] gh/oulgen/15/orig -> origin/gh/oulgen/15/orig 2025-12-04T09:33:40.7014665Z * [new branch] gh/oulgen/16/base -> origin/gh/oulgen/16/base 2025-12-04T09:33:40.7015920Z * [new branch] gh/oulgen/16/head -> origin/gh/oulgen/16/head 2025-12-04T09:33:40.7017299Z * [new branch] gh/oulgen/16/orig -> origin/gh/oulgen/16/orig 2025-12-04T09:33:40.7019131Z * [new branch] gh/oulgen/17/base -> origin/gh/oulgen/17/base 2025-12-04T09:33:40.7020805Z * [new branch] gh/oulgen/17/head -> origin/gh/oulgen/17/head 2025-12-04T09:33:40.7022200Z * [new branch] gh/oulgen/17/orig -> origin/gh/oulgen/17/orig 2025-12-04T09:33:40.7023949Z * [new branch] gh/oulgen/18/base -> origin/gh/oulgen/18/base 2025-12-04T09:33:40.7025104Z * [new branch] gh/oulgen/18/head -> origin/gh/oulgen/18/head 2025-12-04T09:33:40.7026542Z * [new branch] gh/oulgen/18/orig -> origin/gh/oulgen/18/orig 2025-12-04T09:33:40.7027972Z * [new branch] gh/oulgen/19/base -> origin/gh/oulgen/19/base 2025-12-04T09:33:40.7029283Z * [new branch] gh/oulgen/19/head -> origin/gh/oulgen/19/head 2025-12-04T09:33:40.7030453Z * [new branch] gh/oulgen/19/orig -> origin/gh/oulgen/19/orig 2025-12-04T09:33:40.7032161Z * [new branch] gh/oulgen/20/base -> origin/gh/oulgen/20/base 2025-12-04T09:33:40.7033415Z * [new branch] gh/oulgen/20/head -> origin/gh/oulgen/20/head 2025-12-04T09:33:40.7034654Z * [new branch] gh/oulgen/20/orig -> origin/gh/oulgen/20/orig 2025-12-04T09:33:40.7036755Z * [new branch] gh/oulgen/21/base -> origin/gh/oulgen/21/base 2025-12-04T09:33:40.7037349Z * [new branch] gh/oulgen/21/head -> origin/gh/oulgen/21/head 2025-12-04T09:33:40.7038874Z * [new branch] gh/oulgen/21/orig -> origin/gh/oulgen/21/orig 2025-12-04T09:33:40.7040532Z * [new branch] gh/oulgen/22/base -> origin/gh/oulgen/22/base 2025-12-04T09:33:40.7041759Z * [new branch] gh/oulgen/22/head -> origin/gh/oulgen/22/head 2025-12-04T09:33:40.7042999Z * [new branch] gh/oulgen/22/orig -> origin/gh/oulgen/22/orig 2025-12-04T09:33:40.7044674Z * [new branch] gh/oulgen/23/base -> origin/gh/oulgen/23/base 2025-12-04T09:33:40.7045936Z * [new branch] gh/oulgen/23/head -> origin/gh/oulgen/23/head 2025-12-04T09:33:40.7047220Z * [new branch] gh/oulgen/23/orig -> origin/gh/oulgen/23/orig 2025-12-04T09:33:40.7048783Z * [new branch] gh/oulgen/24/base -> origin/gh/oulgen/24/base 2025-12-04T09:33:40.7049988Z * [new branch] gh/oulgen/24/head -> origin/gh/oulgen/24/head 2025-12-04T09:33:40.7051220Z * [new branch] gh/oulgen/24/orig -> origin/gh/oulgen/24/orig 2025-12-04T09:33:40.7052904Z * [new branch] gh/oulgen/25/base -> origin/gh/oulgen/25/base 2025-12-04T09:33:40.7054149Z * [new branch] gh/oulgen/25/head -> origin/gh/oulgen/25/head 2025-12-04T09:33:40.7055493Z * [new branch] gh/oulgen/25/orig -> origin/gh/oulgen/25/orig 2025-12-04T09:33:40.7057194Z * [new branch] gh/oulgen/26/base -> origin/gh/oulgen/26/base 2025-12-04T09:33:40.7058530Z * [new branch] gh/oulgen/26/head -> origin/gh/oulgen/26/head 2025-12-04T09:33:40.7059939Z * [new branch] gh/oulgen/26/orig -> origin/gh/oulgen/26/orig 2025-12-04T09:33:40.7061535Z * [new branch] gh/oulgen/4/base -> origin/gh/oulgen/4/base 2025-12-04T09:33:40.7062793Z * [new branch] gh/oulgen/4/head -> origin/gh/oulgen/4/head 2025-12-04T09:33:40.7064025Z * [new branch] gh/oulgen/4/orig -> origin/gh/oulgen/4/orig 2025-12-04T09:33:40.7066205Z * [new branch] gh/oulgen/7/base -> origin/gh/oulgen/7/base 2025-12-04T09:33:40.7067456Z * [new branch] gh/oulgen/7/head -> origin/gh/oulgen/7/head 2025-12-04T09:33:40.7068714Z * [new branch] gh/oulgen/7/orig -> origin/gh/oulgen/7/orig 2025-12-04T09:33:40.7070465Z * [new branch] gh/oulgen/8/base -> origin/gh/oulgen/8/base 2025-12-04T09:33:40.7071734Z * [new branch] gh/oulgen/8/head -> origin/gh/oulgen/8/head 2025-12-04T09:33:40.7073150Z * [new branch] gh/oulgen/8/orig -> origin/gh/oulgen/8/orig 2025-12-04T09:33:40.7074778Z * [new branch] gh/oulgen/9/base -> origin/gh/oulgen/9/base 2025-12-04T09:33:40.7076036Z * [new branch] gh/oulgen/9/head -> origin/gh/oulgen/9/head 2025-12-04T09:33:40.7077374Z * [new branch] gh/oulgen/9/orig -> origin/gh/oulgen/9/orig 2025-12-04T09:33:40.7079075Z * [new branch] gh/patvig/mtia-serialization -> origin/gh/patvig/mtia-serialization 2025-12-04T09:33:40.7081160Z * [new branch] gh/pearu/108/base -> origin/gh/pearu/108/base 2025-12-04T09:33:40.7082544Z * [new branch] gh/pearu/108/head -> origin/gh/pearu/108/head 2025-12-04T09:33:40.7083955Z * [new branch] gh/pearu/108/orig -> origin/gh/pearu/108/orig 2025-12-04T09:33:40.7085566Z * [new branch] gh/pearu/109/base -> origin/gh/pearu/109/base 2025-12-04T09:33:40.7086832Z * [new branch] gh/pearu/109/head -> origin/gh/pearu/109/head 2025-12-04T09:33:40.7088149Z * [new branch] gh/pearu/109/orig -> origin/gh/pearu/109/orig 2025-12-04T09:33:40.7089932Z * [new branch] gh/pearu/110/base -> origin/gh/pearu/110/base 2025-12-04T09:33:40.7091193Z * [new branch] gh/pearu/110/head -> origin/gh/pearu/110/head 2025-12-04T09:33:40.7093006Z * [new branch] gh/pearu/110/orig -> origin/gh/pearu/110/orig 2025-12-04T09:33:40.7094753Z * [new branch] gh/pearu/111/base -> origin/gh/pearu/111/base 2025-12-04T09:33:40.7096111Z * [new branch] gh/pearu/111/head -> origin/gh/pearu/111/head 2025-12-04T09:33:40.7100387Z * [new branch] gh/pearu/111/orig -> origin/gh/pearu/111/orig 2025-12-04T09:33:40.7102221Z * [new branch] gh/pearu/112/base -> origin/gh/pearu/112/base 2025-12-04T09:33:40.7103511Z * [new branch] gh/pearu/112/head -> origin/gh/pearu/112/head 2025-12-04T09:33:40.7104817Z * [new branch] gh/pearu/112/orig -> origin/gh/pearu/112/orig 2025-12-04T09:33:40.7106402Z * [new branch] gh/pearu/115/base -> origin/gh/pearu/115/base 2025-12-04T09:33:40.7107670Z * [new branch] gh/pearu/115/head -> origin/gh/pearu/115/head 2025-12-04T09:33:40.7108936Z * [new branch] gh/pearu/115/orig -> origin/gh/pearu/115/orig 2025-12-04T09:33:40.7110565Z * [new branch] gh/pearu/116/base -> origin/gh/pearu/116/base 2025-12-04T09:33:40.7111833Z * [new branch] gh/pearu/116/head -> origin/gh/pearu/116/head 2025-12-04T09:33:40.7113516Z * [new branch] gh/pearu/116/orig -> origin/gh/pearu/116/orig 2025-12-04T09:33:40.7115250Z * [new branch] gh/pearu/117/base -> origin/gh/pearu/117/base 2025-12-04T09:33:40.7116478Z * [new branch] gh/pearu/117/head -> origin/gh/pearu/117/head 2025-12-04T09:33:40.7117799Z * [new branch] gh/pearu/117/orig -> origin/gh/pearu/117/orig 2025-12-04T09:33:40.7119459Z * [new branch] gh/pearu/118/base -> origin/gh/pearu/118/base 2025-12-04T09:33:40.7120723Z * [new branch] gh/pearu/118/head -> origin/gh/pearu/118/head 2025-12-04T09:33:40.7122006Z * [new branch] gh/pearu/118/orig -> origin/gh/pearu/118/orig 2025-12-04T09:33:40.7123643Z * [new branch] gh/pearu/119/base -> origin/gh/pearu/119/base 2025-12-04T09:33:40.7124865Z * [new branch] gh/pearu/119/head -> origin/gh/pearu/119/head 2025-12-04T09:33:40.7126182Z * [new branch] gh/pearu/119/orig -> origin/gh/pearu/119/orig 2025-12-04T09:33:40.7128004Z * [new branch] gh/pearu/139/base -> origin/gh/pearu/139/base 2025-12-04T09:33:40.7129279Z * [new branch] gh/pearu/139/head -> origin/gh/pearu/139/head 2025-12-04T09:33:40.7130980Z * [new branch] gh/pearu/139/orig -> origin/gh/pearu/139/orig 2025-12-04T09:33:40.7132753Z * [new branch] gh/pearu/140/base -> origin/gh/pearu/140/base 2025-12-04T09:33:40.7134139Z * [new branch] gh/pearu/140/head -> origin/gh/pearu/140/head 2025-12-04T09:33:40.7135281Z * [new branch] gh/pearu/140/orig -> origin/gh/pearu/140/orig 2025-12-04T09:33:40.7136999Z * [new branch] gh/pearu/142/base -> origin/gh/pearu/142/base 2025-12-04T09:33:40.7138428Z * [new branch] gh/pearu/142/head -> origin/gh/pearu/142/head 2025-12-04T09:33:40.7139715Z * [new branch] gh/pearu/142/orig -> origin/gh/pearu/142/orig 2025-12-04T09:33:40.7141360Z * [new branch] gh/pearu/143/base -> origin/gh/pearu/143/base 2025-12-04T09:33:40.7142642Z * [new branch] gh/pearu/143/head -> origin/gh/pearu/143/head 2025-12-04T09:33:40.7144000Z * [new branch] gh/pearu/143/orig -> origin/gh/pearu/143/orig 2025-12-04T09:33:40.7145764Z * [new branch] gh/pearu/147/base -> origin/gh/pearu/147/base 2025-12-04T09:33:40.7147016Z * [new branch] gh/pearu/147/head -> origin/gh/pearu/147/head 2025-12-04T09:33:40.7148351Z * [new branch] gh/pearu/147/orig -> origin/gh/pearu/147/orig 2025-12-04T09:33:40.7150007Z * [new branch] gh/pearu/149/base -> origin/gh/pearu/149/base 2025-12-04T09:33:40.7151229Z * [new branch] gh/pearu/149/head -> origin/gh/pearu/149/head 2025-12-04T09:33:40.7152478Z * [new branch] gh/pearu/149/orig -> origin/gh/pearu/149/orig 2025-12-04T09:33:40.7154578Z * [new branch] gh/pearu/150/base -> origin/gh/pearu/150/base 2025-12-04T09:33:40.7155848Z * [new branch] gh/pearu/150/head -> origin/gh/pearu/150/head 2025-12-04T09:33:40.7157103Z * [new branch] gh/pearu/150/orig -> origin/gh/pearu/150/orig 2025-12-04T09:33:40.7158855Z * [new branch] gh/pearu/151/base -> origin/gh/pearu/151/base 2025-12-04T09:33:40.7160125Z * [new branch] gh/pearu/151/head -> origin/gh/pearu/151/head 2025-12-04T09:33:40.7161372Z * [new branch] gh/pearu/151/orig -> origin/gh/pearu/151/orig 2025-12-04T09:33:40.7163299Z * [new branch] gh/pearu/152/base -> origin/gh/pearu/152/base 2025-12-04T09:33:40.7164960Z * [new branch] gh/pearu/152/head -> origin/gh/pearu/152/head 2025-12-04T09:33:40.7166089Z * [new branch] gh/pearu/152/orig -> origin/gh/pearu/152/orig 2025-12-04T09:33:40.7167838Z * [new branch] gh/pearu/153/base -> origin/gh/pearu/153/base 2025-12-04T09:33:40.7169113Z * [new branch] gh/pearu/153/head -> origin/gh/pearu/153/head 2025-12-04T09:33:40.7170384Z * [new branch] gh/pearu/153/orig -> origin/gh/pearu/153/orig 2025-12-04T09:33:40.7172068Z * [new branch] gh/pearu/154/base -> origin/gh/pearu/154/base 2025-12-04T09:33:40.7173403Z * [new branch] gh/pearu/154/head -> origin/gh/pearu/154/head 2025-12-04T09:33:40.7174612Z * [new branch] gh/pearu/154/orig -> origin/gh/pearu/154/orig 2025-12-04T09:33:40.7176349Z * [new branch] gh/pearu/155/base -> origin/gh/pearu/155/base 2025-12-04T09:33:40.7177786Z * [new branch] gh/pearu/155/head -> origin/gh/pearu/155/head 2025-12-04T09:33:40.7179049Z * [new branch] gh/pearu/155/orig -> origin/gh/pearu/155/orig 2025-12-04T09:33:40.7181396Z * [new branch] gh/pearu/156/base -> origin/gh/pearu/156/base 2025-12-04T09:33:40.7182680Z * [new branch] gh/pearu/156/head -> origin/gh/pearu/156/head 2025-12-04T09:33:40.7183978Z * [new branch] gh/pearu/156/orig -> origin/gh/pearu/156/orig 2025-12-04T09:33:40.7186182Z * [new branch] gh/pearu/56/base -> origin/gh/pearu/56/base 2025-12-04T09:33:40.7187745Z * [new branch] gh/pearu/56/head -> origin/gh/pearu/56/head 2025-12-04T09:33:40.7189003Z * [new branch] gh/pearu/56/orig -> origin/gh/pearu/56/orig 2025-12-04T09:33:40.7190924Z * [new branch] gh/pearu/97/base -> origin/gh/pearu/97/base 2025-12-04T09:33:40.7192221Z * [new branch] gh/pearu/97/head -> origin/gh/pearu/97/head 2025-12-04T09:33:40.7193987Z * [new branch] gh/pearu/97/orig -> origin/gh/pearu/97/orig 2025-12-04T09:33:40.7196128Z * [new branch] gh/pianpwk/21/base -> origin/gh/pianpwk/21/base 2025-12-04T09:33:40.7197469Z * [new branch] gh/pianpwk/21/head -> origin/gh/pianpwk/21/head 2025-12-04T09:33:40.7199329Z * [new branch] gh/pianpwk/28/base -> origin/gh/pianpwk/28/base 2025-12-04T09:33:40.7200578Z * [new branch] gh/pianpwk/28/head -> origin/gh/pianpwk/28/head 2025-12-04T09:33:40.7201863Z * [new branch] gh/pianpwk/28/orig -> origin/gh/pianpwk/28/orig 2025-12-04T09:33:40.7203563Z * [new branch] gh/pianpwk/29/base -> origin/gh/pianpwk/29/base 2025-12-04T09:33:40.7204915Z * [new branch] gh/pianpwk/29/head -> origin/gh/pianpwk/29/head 2025-12-04T09:33:40.7206185Z * [new branch] gh/pianpwk/29/orig -> origin/gh/pianpwk/29/orig 2025-12-04T09:33:40.7208080Z * [new branch] gh/pianpwk/30/base -> origin/gh/pianpwk/30/base 2025-12-04T09:33:40.7209373Z * [new branch] gh/pianpwk/30/head -> origin/gh/pianpwk/30/head 2025-12-04T09:33:40.7210717Z * [new branch] gh/pianpwk/30/orig -> origin/gh/pianpwk/30/orig 2025-12-04T09:33:40.7212405Z * [new branch] gh/pianpwk/31/base -> origin/gh/pianpwk/31/base 2025-12-04T09:33:40.7213668Z * [new branch] gh/pianpwk/31/head -> origin/gh/pianpwk/31/head 2025-12-04T09:33:40.7214953Z * [new branch] gh/pianpwk/31/orig -> origin/gh/pianpwk/31/orig 2025-12-04T09:33:40.7216528Z * [new branch] gh/pianpwk/32/base -> origin/gh/pianpwk/32/base 2025-12-04T09:33:40.7217895Z * [new branch] gh/pianpwk/32/head -> origin/gh/pianpwk/32/head 2025-12-04T09:33:40.7219643Z * [new branch] gh/pianpwk/32/orig -> origin/gh/pianpwk/32/orig 2025-12-04T09:33:40.7221246Z * [new branch] gh/pianpwk/33/base -> origin/gh/pianpwk/33/base 2025-12-04T09:33:40.7222389Z * [new branch] gh/pianpwk/33/head -> origin/gh/pianpwk/33/head 2025-12-04T09:33:40.7223644Z * [new branch] gh/pianpwk/33/orig -> origin/gh/pianpwk/33/orig 2025-12-04T09:33:40.7225636Z * [new branch] gh/pianpwk/34/base -> origin/gh/pianpwk/34/base 2025-12-04T09:33:40.7227284Z * [new branch] gh/pianpwk/34/head -> origin/gh/pianpwk/34/head 2025-12-04T09:33:40.7229451Z * [new branch] gh/pianpwk/34/orig -> origin/gh/pianpwk/34/orig 2025-12-04T09:33:40.7231118Z * [new branch] gh/pianpwk/35/base -> origin/gh/pianpwk/35/base 2025-12-04T09:33:40.7232926Z * [new branch] gh/pianpwk/35/head -> origin/gh/pianpwk/35/head 2025-12-04T09:33:40.7234285Z * [new branch] gh/pianpwk/35/orig -> origin/gh/pianpwk/35/orig 2025-12-04T09:33:40.7236333Z * [new branch] gh/rec/141/base -> origin/gh/rec/141/base 2025-12-04T09:33:40.7237626Z * [new branch] gh/rec/141/head -> origin/gh/rec/141/head 2025-12-04T09:33:40.7239298Z * [new branch] gh/rec/153/base -> origin/gh/rec/153/base 2025-12-04T09:33:40.7241005Z * [new branch] gh/rec/153/head -> origin/gh/rec/153/head 2025-12-04T09:33:40.7242225Z * [new branch] gh/rec/153/orig -> origin/gh/rec/153/orig 2025-12-04T09:33:40.7243975Z * [new branch] gh/rec/154/base -> origin/gh/rec/154/base 2025-12-04T09:33:40.7245115Z * [new branch] gh/rec/154/head -> origin/gh/rec/154/head 2025-12-04T09:33:40.7246422Z * [new branch] gh/rec/154/orig -> origin/gh/rec/154/orig 2025-12-04T09:33:40.7248046Z * [new branch] gh/rec/164/base -> origin/gh/rec/164/base 2025-12-04T09:33:40.7249348Z * [new branch] gh/rec/164/head -> origin/gh/rec/164/head 2025-12-04T09:33:40.7250593Z * [new branch] gh/rec/164/orig -> origin/gh/rec/164/orig 2025-12-04T09:33:40.7252410Z * [new branch] gh/rec/166/base -> origin/gh/rec/166/base 2025-12-04T09:33:40.7253644Z * [new branch] gh/rec/166/head -> origin/gh/rec/166/head 2025-12-04T09:33:40.7254894Z * [new branch] gh/rec/166/orig -> origin/gh/rec/166/orig 2025-12-04T09:33:40.7256548Z * [new branch] gh/rec/167/base -> origin/gh/rec/167/base 2025-12-04T09:33:40.7257938Z * [new branch] gh/rec/167/head -> origin/gh/rec/167/head 2025-12-04T09:33:40.7259269Z * [new branch] gh/rec/167/orig -> origin/gh/rec/167/orig 2025-12-04T09:33:40.7260991Z * [new branch] gh/rec/168/base -> origin/gh/rec/168/base 2025-12-04T09:33:40.7262263Z * [new branch] gh/rec/168/head -> origin/gh/rec/168/head 2025-12-04T09:33:40.7263506Z * [new branch] gh/rec/168/orig -> origin/gh/rec/168/orig 2025-12-04T09:33:40.7265119Z * [new branch] gh/rec/169/base -> origin/gh/rec/169/base 2025-12-04T09:33:40.7266324Z * [new branch] gh/rec/169/head -> origin/gh/rec/169/head 2025-12-04T09:33:40.7267630Z * [new branch] gh/rec/169/orig -> origin/gh/rec/169/orig 2025-12-04T09:33:40.7269378Z * [new branch] gh/rec/170/base -> origin/gh/rec/170/base 2025-12-04T09:33:40.7270625Z * [new branch] gh/rec/170/head -> origin/gh/rec/170/head 2025-12-04T09:33:40.7271857Z * [new branch] gh/rec/170/orig -> origin/gh/rec/170/orig 2025-12-04T09:33:40.7273562Z * [new branch] gh/rec/171/base -> origin/gh/rec/171/base 2025-12-04T09:33:40.7274770Z * [new branch] gh/rec/171/head -> origin/gh/rec/171/head 2025-12-04T09:33:40.7276097Z * [new branch] gh/rec/171/orig -> origin/gh/rec/171/orig 2025-12-04T09:33:40.7277765Z * [new branch] gh/rec/172/base -> origin/gh/rec/172/base 2025-12-04T09:33:40.7279053Z * [new branch] gh/rec/172/head -> origin/gh/rec/172/head 2025-12-04T09:33:40.7280239Z * [new branch] gh/rec/172/orig -> origin/gh/rec/172/orig 2025-12-04T09:33:40.7281941Z * [new branch] gh/rec/173/base -> origin/gh/rec/173/base 2025-12-04T09:33:40.7283174Z * [new branch] gh/rec/173/head -> origin/gh/rec/173/head 2025-12-04T09:33:40.7284443Z * [new branch] gh/rec/173/orig -> origin/gh/rec/173/orig 2025-12-04T09:33:40.7286200Z * [new branch] gh/rec/174/base -> origin/gh/rec/174/base 2025-12-04T09:33:40.7287437Z * [new branch] gh/rec/174/head -> origin/gh/rec/174/head 2025-12-04T09:33:40.7288747Z * [new branch] gh/rec/174/orig -> origin/gh/rec/174/orig 2025-12-04T09:33:40.7290859Z * [new branch] gh/rec/175/base -> origin/gh/rec/175/base 2025-12-04T09:33:40.7292082Z * [new branch] gh/rec/175/head -> origin/gh/rec/175/head 2025-12-04T09:33:40.7293400Z * [new branch] gh/rec/175/orig -> origin/gh/rec/175/orig 2025-12-04T09:33:40.7295093Z * [new branch] gh/rec/176/base -> origin/gh/rec/176/base 2025-12-04T09:33:40.7296377Z * [new branch] gh/rec/176/head -> origin/gh/rec/176/head 2025-12-04T09:33:40.7297776Z * [new branch] gh/rec/176/orig -> origin/gh/rec/176/orig 2025-12-04T09:33:40.7299457Z * [new branch] gh/rec/177/base -> origin/gh/rec/177/base 2025-12-04T09:33:40.7300732Z * [new branch] gh/rec/177/head -> origin/gh/rec/177/head 2025-12-04T09:33:40.7301976Z * [new branch] gh/rec/177/orig -> origin/gh/rec/177/orig 2025-12-04T09:33:40.7304161Z * [new branch] gh/robert-hardwick/3/base -> origin/gh/robert-hardwick/3/base 2025-12-04T09:33:40.7305464Z * [new branch] gh/robert-hardwick/3/head -> origin/gh/robert-hardwick/3/head 2025-12-04T09:33:40.7306706Z * [new branch] gh/robert-hardwick/3/orig -> origin/gh/robert-hardwick/3/orig 2025-12-04T09:33:40.7308464Z * [new branch] gh/robert-hardwick/4/base -> origin/gh/robert-hardwick/4/base 2025-12-04T09:33:40.7309625Z * [new branch] gh/robert-hardwick/4/head -> origin/gh/robert-hardwick/4/head 2025-12-04T09:33:40.7310919Z * [new branch] gh/robert-hardwick/4/orig -> origin/gh/robert-hardwick/4/orig 2025-12-04T09:33:40.7312603Z * [new branch] gh/robert-hardwick/5/base -> origin/gh/robert-hardwick/5/base 2025-12-04T09:33:40.7313873Z * [new branch] gh/robert-hardwick/5/head -> origin/gh/robert-hardwick/5/head 2025-12-04T09:33:40.7315192Z * [new branch] gh/robert-hardwick/5/orig -> origin/gh/robert-hardwick/5/orig 2025-12-04T09:33:40.7316832Z * [new branch] gh/robert-hardwick/6/base -> origin/gh/robert-hardwick/6/base 2025-12-04T09:33:40.7318125Z * [new branch] gh/robert-hardwick/6/head -> origin/gh/robert-hardwick/6/head 2025-12-04T09:33:40.7319496Z * [new branch] gh/robert-hardwick/6/orig -> origin/gh/robert-hardwick/6/orig 2025-12-04T09:33:40.7321125Z * [new branch] gh/robert-hardwick/7/base -> origin/gh/robert-hardwick/7/base 2025-12-04T09:33:40.7322362Z * [new branch] gh/robert-hardwick/7/head -> origin/gh/robert-hardwick/7/head 2025-12-04T09:33:40.7323600Z * [new branch] gh/robert-hardwick/7/orig -> origin/gh/robert-hardwick/7/orig 2025-12-04T09:33:40.7325270Z * [new branch] gh/robert-hardwick/8/base -> origin/gh/robert-hardwick/8/base 2025-12-04T09:33:40.7326499Z * [new branch] gh/robert-hardwick/8/head -> origin/gh/robert-hardwick/8/head 2025-12-04T09:33:40.7327770Z * [new branch] gh/robert-hardwick/8/orig -> origin/gh/robert-hardwick/8/orig 2025-12-04T09:33:40.7329448Z * [new branch] gh/robert-hardwick/9/base -> origin/gh/robert-hardwick/9/base 2025-12-04T09:33:40.7330699Z * [new branch] gh/robert-hardwick/9/head -> origin/gh/robert-hardwick/9/head 2025-12-04T09:33:40.7331957Z * [new branch] gh/robert-hardwick/9/orig -> origin/gh/robert-hardwick/9/orig 2025-12-04T09:33:40.7333933Z * [new branch] gh/rtimpe/1/base -> origin/gh/rtimpe/1/base 2025-12-04T09:33:40.7335389Z * [new branch] gh/rtimpe/1/head -> origin/gh/rtimpe/1/head 2025-12-04T09:33:40.7337036Z * [new branch] gh/rtimpe/2/base -> origin/gh/rtimpe/2/base 2025-12-04T09:33:40.7338739Z * [new branch] gh/rtimpe/2/head -> origin/gh/rtimpe/2/head 2025-12-04T09:33:40.7340468Z * [new branch] gh/rtimpe/22/base -> origin/gh/rtimpe/22/base 2025-12-04T09:33:40.7341757Z * [new branch] gh/rtimpe/22/head -> origin/gh/rtimpe/22/head 2025-12-04T09:33:40.7343061Z * [new branch] gh/rtimpe/22/orig -> origin/gh/rtimpe/22/orig 2025-12-04T09:33:40.7344633Z * [new branch] gh/rtimpe/23/base -> origin/gh/rtimpe/23/base 2025-12-04T09:33:40.7345997Z * [new branch] gh/rtimpe/23/head -> origin/gh/rtimpe/23/head 2025-12-04T09:33:40.7347145Z * [new branch] gh/rtimpe/23/orig -> origin/gh/rtimpe/23/orig 2025-12-04T09:33:40.7348778Z * [new branch] gh/rtimpe/24/base -> origin/gh/rtimpe/24/base 2025-12-04T09:33:40.7350017Z * [new branch] gh/rtimpe/24/head -> origin/gh/rtimpe/24/head 2025-12-04T09:33:40.7351247Z * [new branch] gh/rtimpe/24/orig -> origin/gh/rtimpe/24/orig 2025-12-04T09:33:40.7352973Z * [new branch] gh/rtimpe/25/base -> origin/gh/rtimpe/25/base 2025-12-04T09:33:40.7354216Z * [new branch] gh/rtimpe/25/head -> origin/gh/rtimpe/25/head 2025-12-04T09:33:40.7355546Z * [new branch] gh/rtimpe/25/orig -> origin/gh/rtimpe/25/orig 2025-12-04T09:33:40.7357153Z * [new branch] gh/rtimpe/26/base -> origin/gh/rtimpe/26/base 2025-12-04T09:33:40.7358431Z * [new branch] gh/rtimpe/26/head -> origin/gh/rtimpe/26/head 2025-12-04T09:33:40.7359683Z * [new branch] gh/rtimpe/26/orig -> origin/gh/rtimpe/26/orig 2025-12-04T09:33:40.7361251Z * [new branch] gh/rtimpe/27/base -> origin/gh/rtimpe/27/base 2025-12-04T09:33:40.7362535Z * [new branch] gh/rtimpe/27/head -> origin/gh/rtimpe/27/head 2025-12-04T09:33:40.7363799Z * [new branch] gh/rtimpe/27/orig -> origin/gh/rtimpe/27/orig 2025-12-04T09:33:40.7365461Z * [new branch] gh/rtimpe/28/base -> origin/gh/rtimpe/28/base 2025-12-04T09:33:40.7366680Z * [new branch] gh/rtimpe/28/head -> origin/gh/rtimpe/28/head 2025-12-04T09:33:40.7368438Z * [new branch] gh/rtimpe/28/orig -> origin/gh/rtimpe/28/orig 2025-12-04T09:33:40.7370250Z * [new branch] gh/rtimpe/29/base -> origin/gh/rtimpe/29/base 2025-12-04T09:33:40.7371467Z * [new branch] gh/rtimpe/29/head -> origin/gh/rtimpe/29/head 2025-12-04T09:33:40.7372733Z * [new branch] gh/rtimpe/29/orig -> origin/gh/rtimpe/29/orig 2025-12-04T09:33:40.7374306Z * [new branch] gh/rtimpe/3/base -> origin/gh/rtimpe/3/base 2025-12-04T09:33:40.7375489Z * [new branch] gh/rtimpe/3/head -> origin/gh/rtimpe/3/head 2025-12-04T09:33:40.7377176Z * [new branch] gh/rtimpe/30/base -> origin/gh/rtimpe/30/base 2025-12-04T09:33:40.7378535Z * [new branch] gh/rtimpe/30/head -> origin/gh/rtimpe/30/head 2025-12-04T09:33:40.7379778Z * [new branch] gh/rtimpe/30/orig -> origin/gh/rtimpe/30/orig 2025-12-04T09:33:40.7381420Z * [new branch] gh/rtimpe/31/base -> origin/gh/rtimpe/31/base 2025-12-04T09:33:40.7382685Z * [new branch] gh/rtimpe/31/head -> origin/gh/rtimpe/31/head 2025-12-04T09:33:40.7384027Z * [new branch] gh/rtimpe/31/orig -> origin/gh/rtimpe/31/orig 2025-12-04T09:33:40.7385790Z * [new branch] gh/rtimpe/32/base -> origin/gh/rtimpe/32/base 2025-12-04T09:33:40.7387034Z * [new branch] gh/rtimpe/32/head -> origin/gh/rtimpe/32/head 2025-12-04T09:33:40.7388284Z * [new branch] gh/rtimpe/32/orig -> origin/gh/rtimpe/32/orig 2025-12-04T09:33:40.7390254Z * [new branch] gh/rtimpe/33/base -> origin/gh/rtimpe/33/base 2025-12-04T09:33:40.7391536Z * [new branch] gh/rtimpe/33/head -> origin/gh/rtimpe/33/head 2025-12-04T09:33:40.7392818Z * [new branch] gh/rtimpe/33/orig -> origin/gh/rtimpe/33/orig 2025-12-04T09:33:40.7394335Z * [new branch] gh/rtimpe/34/base -> origin/gh/rtimpe/34/base 2025-12-04T09:33:40.7395586Z * [new branch] gh/rtimpe/34/head -> origin/gh/rtimpe/34/head 2025-12-04T09:33:40.7397156Z * [new branch] gh/rtimpe/34/orig -> origin/gh/rtimpe/34/orig 2025-12-04T09:33:40.7399233Z * [new branch] gh/rtimpe/35/base -> origin/gh/rtimpe/35/base 2025-12-04T09:33:40.7400603Z * [new branch] gh/rtimpe/35/head -> origin/gh/rtimpe/35/head 2025-12-04T09:33:40.7401863Z * [new branch] gh/rtimpe/35/orig -> origin/gh/rtimpe/35/orig 2025-12-04T09:33:40.7403595Z * [new branch] gh/rtimpe/4/base -> origin/gh/rtimpe/4/base 2025-12-04T09:33:40.7404861Z * [new branch] gh/rtimpe/4/head -> origin/gh/rtimpe/4/head 2025-12-04T09:33:40.7406993Z * [new branch] gh/ruisizhang123/1/base -> origin/gh/ruisizhang123/1/base 2025-12-04T09:33:40.7408411Z * [new branch] gh/ruisizhang123/1/head -> origin/gh/ruisizhang123/1/head 2025-12-04T09:33:40.7409692Z * [new branch] gh/ruisizhang123/1/orig -> origin/gh/ruisizhang123/1/orig 2025-12-04T09:33:40.7411402Z * [new branch] gh/ruisizhang123/4/base -> origin/gh/ruisizhang123/4/base 2025-12-04T09:33:40.7412597Z * [new branch] gh/ruisizhang123/4/head -> origin/gh/ruisizhang123/4/head 2025-12-04T09:33:40.7413859Z * [new branch] gh/ruisizhang123/4/orig -> origin/gh/ruisizhang123/4/orig 2025-12-04T09:33:40.7415495Z * [new branch] gh/ruisizhang123/5/base -> origin/gh/ruisizhang123/5/base 2025-12-04T09:33:40.7416750Z * [new branch] gh/ruisizhang123/5/head -> origin/gh/ruisizhang123/5/head 2025-12-04T09:33:40.7418125Z * [new branch] gh/ruisizhang123/5/orig -> origin/gh/ruisizhang123/5/orig 2025-12-04T09:33:40.7419860Z * [new branch] gh/ruisizhang123/6/base -> origin/gh/ruisizhang123/6/base 2025-12-04T09:33:40.7421252Z * [new branch] gh/ruisizhang123/6/head -> origin/gh/ruisizhang123/6/head 2025-12-04T09:33:40.7422492Z * [new branch] gh/ruisizhang123/6/orig -> origin/gh/ruisizhang123/6/orig 2025-12-04T09:33:40.7424217Z * [new branch] gh/ruisizhang123/7/base -> origin/gh/ruisizhang123/7/base 2025-12-04T09:33:40.7425503Z * [new branch] gh/ruisizhang123/7/head -> origin/gh/ruisizhang123/7/head 2025-12-04T09:33:40.7426768Z * [new branch] gh/ruisizhang123/7/orig -> origin/gh/ruisizhang123/7/orig 2025-12-04T09:33:40.7428352Z * [new branch] gh/ruisizhang123/8/base -> origin/gh/ruisizhang123/8/base 2025-12-04T09:33:40.7429635Z * [new branch] gh/ruisizhang123/8/head -> origin/gh/ruisizhang123/8/head 2025-12-04T09:33:40.7430882Z * [new branch] gh/ruisizhang123/8/orig -> origin/gh/ruisizhang123/8/orig 2025-12-04T09:33:40.7432532Z * [new branch] gh/ruisizhang123/9/base -> origin/gh/ruisizhang123/9/base 2025-12-04T09:33:40.7433776Z * [new branch] gh/ruisizhang123/9/head -> origin/gh/ruisizhang123/9/head 2025-12-04T09:33:40.7435019Z * [new branch] gh/ruisizhang123/9/orig -> origin/gh/ruisizhang123/9/orig 2025-12-04T09:33:40.7437191Z * [new branch] gh/seemethere/52/base -> origin/gh/seemethere/52/base 2025-12-04T09:33:40.7438436Z * [new branch] gh/seemethere/52/head -> origin/gh/seemethere/52/head 2025-12-04T09:33:40.7439752Z * [new branch] gh/seemethere/52/orig -> origin/gh/seemethere/52/orig 2025-12-04T09:33:40.7441396Z * [new branch] gh/seemethere/53/base -> origin/gh/seemethere/53/base 2025-12-04T09:33:40.7442660Z * [new branch] gh/seemethere/53/head -> origin/gh/seemethere/53/head 2025-12-04T09:33:40.7443882Z * [new branch] gh/seemethere/53/orig -> origin/gh/seemethere/53/orig 2025-12-04T09:33:40.7445509Z * [new branch] gh/seemethere/54/base -> origin/gh/seemethere/54/base 2025-12-04T09:33:40.7446800Z * [new branch] gh/seemethere/54/head -> origin/gh/seemethere/54/head 2025-12-04T09:33:40.7448134Z * [new branch] gh/seemethere/54/orig -> origin/gh/seemethere/54/orig 2025-12-04T09:33:40.7449652Z * [new branch] gh/seemethere/55/base -> origin/gh/seemethere/55/base 2025-12-04T09:33:40.7450817Z * [new branch] gh/seemethere/55/head -> origin/gh/seemethere/55/head 2025-12-04T09:33:40.7452143Z * [new branch] gh/seemethere/55/orig -> origin/gh/seemethere/55/orig 2025-12-04T09:33:40.7453821Z * [new branch] gh/seemethere/59/base -> origin/gh/seemethere/59/base 2025-12-04T09:33:40.7455168Z * [new branch] gh/seemethere/59/head -> origin/gh/seemethere/59/head 2025-12-04T09:33:40.7456381Z * [new branch] gh/seemethere/59/orig -> origin/gh/seemethere/59/orig 2025-12-04T09:33:40.7458162Z * [new branch] gh/seemethere/62/base -> origin/gh/seemethere/62/base 2025-12-04T09:33:40.7459437Z * [new branch] gh/seemethere/62/head -> origin/gh/seemethere/62/head 2025-12-04T09:33:40.7460705Z * [new branch] gh/seemethere/62/orig -> origin/gh/seemethere/62/orig 2025-12-04T09:33:40.7462345Z * [new branch] gh/seemethere/63/base -> origin/gh/seemethere/63/base 2025-12-04T09:33:40.7463592Z * [new branch] gh/seemethere/63/head -> origin/gh/seemethere/63/head 2025-12-04T09:33:40.7464883Z * [new branch] gh/seemethere/63/orig -> origin/gh/seemethere/63/orig 2025-12-04T09:33:40.7466537Z * [new branch] gh/seemethere/71/base -> origin/gh/seemethere/71/base 2025-12-04T09:33:40.7467738Z * [new branch] gh/seemethere/71/head -> origin/gh/seemethere/71/head 2025-12-04T09:33:40.7469019Z * [new branch] gh/seemethere/71/orig -> origin/gh/seemethere/71/orig 2025-12-04T09:33:40.7470895Z * [new branch] gh/seemethere/72/base -> origin/gh/seemethere/72/base 2025-12-04T09:33:40.7472188Z * [new branch] gh/seemethere/72/head -> origin/gh/seemethere/72/head 2025-12-04T09:33:40.7473445Z * [new branch] gh/seemethere/72/orig -> origin/gh/seemethere/72/orig 2025-12-04T09:33:40.7475055Z * [new branch] gh/seemethere/73/base -> origin/gh/seemethere/73/base 2025-12-04T09:33:40.7476305Z * [new branch] gh/seemethere/73/head -> origin/gh/seemethere/73/head 2025-12-04T09:33:40.7477550Z * [new branch] gh/seemethere/73/orig -> origin/gh/seemethere/73/orig 2025-12-04T09:33:40.7479313Z * [new branch] gh/seemethere/74/base -> origin/gh/seemethere/74/base 2025-12-04T09:33:40.7480564Z * [new branch] gh/seemethere/74/head -> origin/gh/seemethere/74/head 2025-12-04T09:33:40.7481880Z * [new branch] gh/seemethere/74/orig -> origin/gh/seemethere/74/orig 2025-12-04T09:33:40.7484076Z * [new branch] gh/seemethere/75/base -> origin/gh/seemethere/75/base 2025-12-04T09:33:40.7485357Z * [new branch] gh/seemethere/75/head -> origin/gh/seemethere/75/head 2025-12-04T09:33:40.7486656Z * [new branch] gh/seemethere/75/orig -> origin/gh/seemethere/75/orig 2025-12-04T09:33:40.7488487Z * [new branch] gh/seemethere/76/base -> origin/gh/seemethere/76/base 2025-12-04T09:33:40.7489795Z * [new branch] gh/seemethere/76/head -> origin/gh/seemethere/76/head 2025-12-04T09:33:40.7491072Z * [new branch] gh/seemethere/76/orig -> origin/gh/seemethere/76/orig 2025-12-04T09:33:40.7493790Z * [new branch] gh/shunting314/145/base -> origin/gh/shunting314/145/base 2025-12-04T09:33:40.7495275Z * [new branch] gh/shunting314/145/head -> origin/gh/shunting314/145/head 2025-12-04T09:33:40.7496818Z * [new branch] gh/shunting314/145/orig -> origin/gh/shunting314/145/orig 2025-12-04T09:33:40.7498990Z * [new branch] gh/shunting314/176/base -> origin/gh/shunting314/176/base 2025-12-04T09:33:40.7500493Z * [new branch] gh/shunting314/176/head -> origin/gh/shunting314/176/head 2025-12-04T09:33:40.7501800Z * [new branch] gh/shunting314/176/orig -> origin/gh/shunting314/176/orig 2025-12-04T09:33:40.7504109Z * [new branch] gh/shunting314/249/base -> origin/gh/shunting314/249/base 2025-12-04T09:33:40.7505484Z * [new branch] gh/shunting314/249/head -> origin/gh/shunting314/249/head 2025-12-04T09:33:40.7507017Z * [new branch] gh/shunting314/249/orig -> origin/gh/shunting314/249/orig 2025-12-04T09:33:40.7508776Z * [new branch] gh/shunting314/253/base -> origin/gh/shunting314/253/base 2025-12-04T09:33:40.7509963Z * [new branch] gh/shunting314/253/head -> origin/gh/shunting314/253/head 2025-12-04T09:33:40.7511206Z * [new branch] gh/shunting314/253/orig -> origin/gh/shunting314/253/orig 2025-12-04T09:33:40.7512915Z * [new branch] gh/shunting314/256/base -> origin/gh/shunting314/256/base 2025-12-04T09:33:40.7514208Z * [new branch] gh/shunting314/256/head -> origin/gh/shunting314/256/head 2025-12-04T09:33:40.7515379Z * [new branch] gh/shunting314/256/orig -> origin/gh/shunting314/256/orig 2025-12-04T09:33:40.7517451Z * [new branch] gh/shunting314/257/base -> origin/gh/shunting314/257/base 2025-12-04T09:33:40.7518798Z * [new branch] gh/shunting314/257/head -> origin/gh/shunting314/257/head 2025-12-04T09:33:40.7520028Z * [new branch] gh/shunting314/257/orig -> origin/gh/shunting314/257/orig 2025-12-04T09:33:40.7522421Z * [new branch] gh/shunting314/258/base -> origin/gh/shunting314/258/base 2025-12-04T09:33:40.7523631Z * [new branch] gh/shunting314/258/head -> origin/gh/shunting314/258/head 2025-12-04T09:33:40.7525062Z * [new branch] gh/shunting314/258/orig -> origin/gh/shunting314/258/orig 2025-12-04T09:33:40.7526585Z * [new branch] gh/shunting314/259/base -> origin/gh/shunting314/259/base 2025-12-04T09:33:40.7527904Z * [new branch] gh/shunting314/259/head -> origin/gh/shunting314/259/head 2025-12-04T09:33:40.7529163Z * [new branch] gh/shunting314/259/orig -> origin/gh/shunting314/259/orig 2025-12-04T09:33:40.7530956Z * [new branch] gh/shunting314/260/base -> origin/gh/shunting314/260/base 2025-12-04T09:33:40.7532307Z * [new branch] gh/shunting314/260/head -> origin/gh/shunting314/260/head 2025-12-04T09:33:40.7533602Z * [new branch] gh/shunting314/260/orig -> origin/gh/shunting314/260/orig 2025-12-04T09:33:40.7535347Z * [new branch] gh/shunting314/261/base -> origin/gh/shunting314/261/base 2025-12-04T09:33:40.7536716Z * [new branch] gh/shunting314/261/head -> origin/gh/shunting314/261/head 2025-12-04T09:33:40.7538166Z * [new branch] gh/shunting314/261/orig -> origin/gh/shunting314/261/orig 2025-12-04T09:33:40.7539936Z * [new branch] gh/shunting314/262/base -> origin/gh/shunting314/262/base 2025-12-04T09:33:40.7541263Z * [new branch] gh/shunting314/262/head -> origin/gh/shunting314/262/head 2025-12-04T09:33:40.7542700Z * [new branch] gh/shunting314/262/orig -> origin/gh/shunting314/262/orig 2025-12-04T09:33:40.7544491Z * [new branch] gh/shunting314/263/base -> origin/gh/shunting314/263/base 2025-12-04T09:33:40.7545942Z * [new branch] gh/shunting314/263/head -> origin/gh/shunting314/263/head 2025-12-04T09:33:40.7547198Z * [new branch] gh/shunting314/263/orig -> origin/gh/shunting314/263/orig 2025-12-04T09:33:40.7548933Z * [new branch] gh/shunting314/264/base -> origin/gh/shunting314/264/base 2025-12-04T09:33:40.7550318Z * [new branch] gh/shunting314/264/head -> origin/gh/shunting314/264/head 2025-12-04T09:33:40.7551467Z * [new branch] gh/shunting314/264/orig -> origin/gh/shunting314/264/orig 2025-12-04T09:33:40.7553240Z * [new branch] gh/shunting314/265/base -> origin/gh/shunting314/265/base 2025-12-04T09:33:40.7554423Z * [new branch] gh/shunting314/265/head -> origin/gh/shunting314/265/head 2025-12-04T09:33:40.7555742Z * [new branch] gh/shunting314/265/orig -> origin/gh/shunting314/265/orig 2025-12-04T09:33:40.7557392Z * [new branch] gh/shunting314/266/base -> origin/gh/shunting314/266/base 2025-12-04T09:33:40.7558914Z * [new branch] gh/shunting314/266/head -> origin/gh/shunting314/266/head 2025-12-04T09:33:40.7560307Z * [new branch] gh/shunting314/266/orig -> origin/gh/shunting314/266/orig 2025-12-04T09:33:40.7562257Z * [new branch] gh/shunting314/267/base -> origin/gh/shunting314/267/base 2025-12-04T09:33:40.7563701Z * [new branch] gh/shunting314/267/head -> origin/gh/shunting314/267/head 2025-12-04T09:33:40.7565005Z * [new branch] gh/shunting314/267/orig -> origin/gh/shunting314/267/orig 2025-12-04T09:33:40.7567252Z * [new branch] gh/shunting314/268/base -> origin/gh/shunting314/268/base 2025-12-04T09:33:40.7568584Z * [new branch] gh/shunting314/268/head -> origin/gh/shunting314/268/head 2025-12-04T09:33:40.7569915Z * [new branch] gh/shunting314/268/orig -> origin/gh/shunting314/268/orig 2025-12-04T09:33:40.7571659Z * [new branch] gh/shunting314/269/base -> origin/gh/shunting314/269/base 2025-12-04T09:33:40.7572942Z * [new branch] gh/shunting314/269/head -> origin/gh/shunting314/269/head 2025-12-04T09:33:40.7574215Z * [new branch] gh/shunting314/269/orig -> origin/gh/shunting314/269/orig 2025-12-04T09:33:40.7576209Z * [new branch] gh/silverguo/1/base -> origin/gh/silverguo/1/base 2025-12-04T09:33:40.7577705Z * [new branch] gh/silverguo/1/head -> origin/gh/silverguo/1/head 2025-12-04T09:33:40.7579258Z * [new branch] gh/silverguo/2/base -> origin/gh/silverguo/2/base 2025-12-04T09:33:40.7580413Z * [new branch] gh/silverguo/2/head -> origin/gh/silverguo/2/head 2025-12-04T09:33:40.7581961Z * [new branch] gh/silverguo/3/base -> origin/gh/silverguo/3/base 2025-12-04T09:33:40.7583192Z * [new branch] gh/silverguo/3/head -> origin/gh/silverguo/3/head 2025-12-04T09:33:40.7584729Z * [new branch] gh/silverguo/4/base -> origin/gh/silverguo/4/base 2025-12-04T09:33:40.7586023Z * [new branch] gh/silverguo/4/head -> origin/gh/silverguo/4/head 2025-12-04T09:33:40.7588001Z * [new branch] gh/slayton58/39/base -> origin/gh/slayton58/39/base 2025-12-04T09:33:40.7589265Z * [new branch] gh/slayton58/39/head -> origin/gh/slayton58/39/head 2025-12-04T09:33:40.7590608Z * [new branch] gh/slayton58/39/orig -> origin/gh/slayton58/39/orig 2025-12-04T09:33:40.7592358Z * [new branch] gh/slayton58/42/base -> origin/gh/slayton58/42/base 2025-12-04T09:33:40.7593612Z * [new branch] gh/slayton58/42/head -> origin/gh/slayton58/42/head 2025-12-04T09:33:40.7595004Z * [new branch] gh/slayton58/42/orig -> origin/gh/slayton58/42/orig 2025-12-04T09:33:40.7600186Z * [new branch] gh/slayton58/43/base -> origin/gh/slayton58/43/base 2025-12-04T09:33:40.7601700Z * [new branch] gh/slayton58/43/head -> origin/gh/slayton58/43/head 2025-12-04T09:33:40.7602994Z * [new branch] gh/slayton58/43/orig -> origin/gh/slayton58/43/orig 2025-12-04T09:33:40.7604870Z * [new branch] gh/slayton58/44/base -> origin/gh/slayton58/44/base 2025-12-04T09:33:40.7606244Z * [new branch] gh/slayton58/44/head -> origin/gh/slayton58/44/head 2025-12-04T09:33:40.7607440Z * [new branch] gh/slayton58/44/orig -> origin/gh/slayton58/44/orig 2025-12-04T09:33:40.7609107Z * [new branch] gh/slayton58/45/base -> origin/gh/slayton58/45/base 2025-12-04T09:33:40.7610337Z * [new branch] gh/slayton58/45/head -> origin/gh/slayton58/45/head 2025-12-04T09:33:40.7611698Z * [new branch] gh/slayton58/45/orig -> origin/gh/slayton58/45/orig 2025-12-04T09:33:40.7613440Z * [new branch] gh/slayton58/46/base -> origin/gh/slayton58/46/base 2025-12-04T09:33:40.7614828Z * [new branch] gh/slayton58/46/head -> origin/gh/slayton58/46/head 2025-12-04T09:33:40.7616144Z * [new branch] gh/slayton58/46/orig -> origin/gh/slayton58/46/orig 2025-12-04T09:33:40.7617991Z * [new branch] gh/slayton58/6/base -> origin/gh/slayton58/6/base 2025-12-04T09:33:40.7619376Z * [new branch] gh/slayton58/6/head -> origin/gh/slayton58/6/head 2025-12-04T09:33:40.7621153Z * [new branch] gh/slayton58/7/base -> origin/gh/slayton58/7/base 2025-12-04T09:33:40.7622376Z * [new branch] gh/slayton58/7/head -> origin/gh/slayton58/7/head 2025-12-04T09:33:40.7624650Z * [new branch] gh/soulitzer/269/base -> origin/gh/soulitzer/269/base 2025-12-04T09:33:40.7625874Z * [new branch] gh/soulitzer/269/head -> origin/gh/soulitzer/269/head 2025-12-04T09:33:40.7627173Z * [new branch] gh/soulitzer/269/orig -> origin/gh/soulitzer/269/orig 2025-12-04T09:33:40.7629010Z * [new branch] gh/soulitzer/276/base -> origin/gh/soulitzer/276/base 2025-12-04T09:33:40.7630487Z * [new branch] gh/soulitzer/276/head -> origin/gh/soulitzer/276/head 2025-12-04T09:33:40.7631760Z * [new branch] gh/soulitzer/276/orig -> origin/gh/soulitzer/276/orig 2025-12-04T09:33:40.7633844Z * [new branch] gh/soulitzer/287/base -> origin/gh/soulitzer/287/base 2025-12-04T09:33:40.7635067Z * [new branch] gh/soulitzer/287/head -> origin/gh/soulitzer/287/head 2025-12-04T09:33:40.7636486Z * [new branch] gh/soulitzer/287/orig -> origin/gh/soulitzer/287/orig 2025-12-04T09:33:40.7638352Z * [new branch] gh/soulitzer/296/base -> origin/gh/soulitzer/296/base 2025-12-04T09:33:40.7639678Z * [new branch] gh/soulitzer/296/head -> origin/gh/soulitzer/296/head 2025-12-04T09:33:40.7640966Z * [new branch] gh/soulitzer/296/orig -> origin/gh/soulitzer/296/orig 2025-12-04T09:33:40.7642644Z * [new branch] gh/soulitzer/299/base -> origin/gh/soulitzer/299/base 2025-12-04T09:33:40.7644051Z * [new branch] gh/soulitzer/299/head -> origin/gh/soulitzer/299/head 2025-12-04T09:33:40.7645315Z * [new branch] gh/soulitzer/299/orig -> origin/gh/soulitzer/299/orig 2025-12-04T09:33:40.7647092Z * [new branch] gh/soulitzer/300/base -> origin/gh/soulitzer/300/base 2025-12-04T09:33:40.7648494Z * [new branch] gh/soulitzer/300/head -> origin/gh/soulitzer/300/head 2025-12-04T09:33:40.7649753Z * [new branch] gh/soulitzer/300/orig -> origin/gh/soulitzer/300/orig 2025-12-04T09:33:40.7651604Z * [new branch] gh/soulitzer/301/base -> origin/gh/soulitzer/301/base 2025-12-04T09:33:40.7652948Z * [new branch] gh/soulitzer/301/head -> origin/gh/soulitzer/301/head 2025-12-04T09:33:40.7654258Z * [new branch] gh/soulitzer/301/orig -> origin/gh/soulitzer/301/orig 2025-12-04T09:33:40.7655944Z * [new branch] gh/soulitzer/313/base -> origin/gh/soulitzer/313/base 2025-12-04T09:33:40.7657262Z * [new branch] gh/soulitzer/313/head -> origin/gh/soulitzer/313/head 2025-12-04T09:33:40.7658711Z * [new branch] gh/soulitzer/313/orig -> origin/gh/soulitzer/313/orig 2025-12-04T09:33:40.7660318Z * [new branch] gh/soulitzer/319/base -> origin/gh/soulitzer/319/base 2025-12-04T09:33:40.7661583Z * [new branch] gh/soulitzer/319/head -> origin/gh/soulitzer/319/head 2025-12-04T09:33:40.7662827Z * [new branch] gh/soulitzer/319/orig -> origin/gh/soulitzer/319/orig 2025-12-04T09:33:40.7664734Z * [new branch] gh/soulitzer/320/base -> origin/gh/soulitzer/320/base 2025-12-04T09:33:40.7666013Z * [new branch] gh/soulitzer/320/head -> origin/gh/soulitzer/320/head 2025-12-04T09:33:40.7667268Z * [new branch] gh/soulitzer/320/orig -> origin/gh/soulitzer/320/orig 2025-12-04T09:33:40.7669104Z * [new branch] gh/soulitzer/336/base -> origin/gh/soulitzer/336/base 2025-12-04T09:33:40.7670375Z * [new branch] gh/soulitzer/336/head -> origin/gh/soulitzer/336/head 2025-12-04T09:33:40.7671636Z * [new branch] gh/soulitzer/336/orig -> origin/gh/soulitzer/336/orig 2025-12-04T09:33:40.7673365Z * [new branch] gh/soulitzer/347/base -> origin/gh/soulitzer/347/base 2025-12-04T09:33:40.7674546Z * [new branch] gh/soulitzer/347/head -> origin/gh/soulitzer/347/head 2025-12-04T09:33:40.7675811Z * [new branch] gh/soulitzer/347/orig -> origin/gh/soulitzer/347/orig 2025-12-04T09:33:40.7677789Z * [new branch] gh/soulitzer/349/base -> origin/gh/soulitzer/349/base 2025-12-04T09:33:40.7679082Z * [new branch] gh/soulitzer/349/head -> origin/gh/soulitzer/349/head 2025-12-04T09:33:40.7680371Z * [new branch] gh/soulitzer/349/orig -> origin/gh/soulitzer/349/orig 2025-12-04T09:33:40.7681965Z * [new branch] gh/soulitzer/350/base -> origin/gh/soulitzer/350/base 2025-12-04T09:33:40.7683272Z * [new branch] gh/soulitzer/350/head -> origin/gh/soulitzer/350/head 2025-12-04T09:33:40.7684516Z * [new branch] gh/soulitzer/350/orig -> origin/gh/soulitzer/350/orig 2025-12-04T09:33:40.7686300Z * [new branch] gh/soulitzer/351/base -> origin/gh/soulitzer/351/base 2025-12-04T09:33:40.7687540Z * [new branch] gh/soulitzer/351/head -> origin/gh/soulitzer/351/head 2025-12-04T09:33:40.7688830Z * [new branch] gh/soulitzer/351/orig -> origin/gh/soulitzer/351/orig 2025-12-04T09:33:40.7690515Z * [new branch] gh/soulitzer/353/base -> origin/gh/soulitzer/353/base 2025-12-04T09:33:40.7691921Z * [new branch] gh/soulitzer/353/head -> origin/gh/soulitzer/353/head 2025-12-04T09:33:40.7693654Z * [new branch] gh/soulitzer/353/orig -> origin/gh/soulitzer/353/orig 2025-12-04T09:33:40.7696248Z * [new branch] gh/soulitzer/358/base -> origin/gh/soulitzer/358/base 2025-12-04T09:33:40.7697766Z * [new branch] gh/soulitzer/358/head -> origin/gh/soulitzer/358/head 2025-12-04T09:33:40.7699082Z * [new branch] gh/soulitzer/358/orig -> origin/gh/soulitzer/358/orig 2025-12-04T09:33:40.7701420Z * [new branch] gh/soulitzer/359/base -> origin/gh/soulitzer/359/base 2025-12-04T09:33:40.7702784Z * [new branch] gh/soulitzer/359/head -> origin/gh/soulitzer/359/head 2025-12-04T09:33:40.7704094Z * [new branch] gh/soulitzer/359/orig -> origin/gh/soulitzer/359/orig 2025-12-04T09:33:40.7705853Z * [new branch] gh/soulitzer/374/base -> origin/gh/soulitzer/374/base 2025-12-04T09:33:40.7707104Z * [new branch] gh/soulitzer/374/head -> origin/gh/soulitzer/374/head 2025-12-04T09:33:40.7708396Z * [new branch] gh/soulitzer/374/orig -> origin/gh/soulitzer/374/orig 2025-12-04T09:33:40.7710193Z * [new branch] gh/soulitzer/375/base -> origin/gh/soulitzer/375/base 2025-12-04T09:33:40.7711521Z * [new branch] gh/soulitzer/375/head -> origin/gh/soulitzer/375/head 2025-12-04T09:33:40.7712648Z * [new branch] gh/soulitzer/375/orig -> origin/gh/soulitzer/375/orig 2025-12-04T09:33:40.7714363Z * [new branch] gh/soulitzer/380/base -> origin/gh/soulitzer/380/base 2025-12-04T09:33:40.7715628Z * [new branch] gh/soulitzer/380/head -> origin/gh/soulitzer/380/head 2025-12-04T09:33:40.7716907Z * [new branch] gh/soulitzer/380/orig -> origin/gh/soulitzer/380/orig 2025-12-04T09:33:40.7718624Z * [new branch] gh/soulitzer/385/base -> origin/gh/soulitzer/385/base 2025-12-04T09:33:40.7719982Z * [new branch] gh/soulitzer/385/head -> origin/gh/soulitzer/385/head 2025-12-04T09:33:40.7721741Z * [new branch] gh/soulitzer/385/orig -> origin/gh/soulitzer/385/orig 2025-12-04T09:33:40.7723464Z * [new branch] gh/soulitzer/386/base -> origin/gh/soulitzer/386/base 2025-12-04T09:33:40.7724720Z * [new branch] gh/soulitzer/386/head -> origin/gh/soulitzer/386/head 2025-12-04T09:33:40.7726059Z * [new branch] gh/soulitzer/386/orig -> origin/gh/soulitzer/386/orig 2025-12-04T09:33:40.7727772Z * [new branch] gh/soulitzer/387/base -> origin/gh/soulitzer/387/base 2025-12-04T09:33:40.7729038Z * [new branch] gh/soulitzer/387/head -> origin/gh/soulitzer/387/head 2025-12-04T09:33:40.7730252Z * [new branch] gh/soulitzer/387/orig -> origin/gh/soulitzer/387/orig 2025-12-04T09:33:40.7731976Z * [new branch] gh/soulitzer/388/base -> origin/gh/soulitzer/388/base 2025-12-04T09:33:40.7733222Z * [new branch] gh/soulitzer/388/head -> origin/gh/soulitzer/388/head 2025-12-04T09:33:40.7734456Z * [new branch] gh/soulitzer/388/orig -> origin/gh/soulitzer/388/orig 2025-12-04T09:33:40.7736216Z * [new branch] gh/soulitzer/389/base -> origin/gh/soulitzer/389/base 2025-12-04T09:33:40.7737655Z * [new branch] gh/soulitzer/389/head -> origin/gh/soulitzer/389/head 2025-12-04T09:33:40.7738979Z * [new branch] gh/soulitzer/389/orig -> origin/gh/soulitzer/389/orig 2025-12-04T09:33:40.7740628Z * [new branch] gh/soulitzer/390/base -> origin/gh/soulitzer/390/base 2025-12-04T09:33:40.7741893Z * [new branch] gh/soulitzer/390/head -> origin/gh/soulitzer/390/head 2025-12-04T09:33:40.7743153Z * [new branch] gh/soulitzer/390/orig -> origin/gh/soulitzer/390/orig 2025-12-04T09:33:40.7744866Z * [new branch] gh/soulitzer/391/base -> origin/gh/soulitzer/391/base 2025-12-04T09:33:40.7746114Z * [new branch] gh/soulitzer/391/head -> origin/gh/soulitzer/391/head 2025-12-04T09:33:40.7747343Z * [new branch] gh/soulitzer/391/orig -> origin/gh/soulitzer/391/orig 2025-12-04T09:33:40.7749052Z * [new branch] gh/soulitzer/392/base -> origin/gh/soulitzer/392/base 2025-12-04T09:33:40.7750257Z * [new branch] gh/soulitzer/392/head -> origin/gh/soulitzer/392/head 2025-12-04T09:33:40.7751509Z * [new branch] gh/soulitzer/392/orig -> origin/gh/soulitzer/392/orig 2025-12-04T09:33:40.7753711Z * [new branch] gh/swolchok/728/next -> origin/gh/swolchok/728/next 2025-12-04T09:33:40.7755711Z * [new branch] gh/swolchok/819/base -> origin/gh/swolchok/819/base 2025-12-04T09:33:40.7757001Z * [new branch] gh/swolchok/819/head -> origin/gh/swolchok/819/head 2025-12-04T09:33:40.7758238Z * [new branch] gh/swolchok/819/orig -> origin/gh/swolchok/819/orig 2025-12-04T09:33:40.7759915Z * [new branch] gh/swolchok/824/base -> origin/gh/swolchok/824/base 2025-12-04T09:33:40.7761266Z * [new branch] gh/swolchok/824/head -> origin/gh/swolchok/824/head 2025-12-04T09:33:40.7762456Z * [new branch] gh/swolchok/824/orig -> origin/gh/swolchok/824/orig 2025-12-04T09:33:40.7764154Z * [new branch] gh/swolchok/829/base -> origin/gh/swolchok/829/base 2025-12-04T09:33:40.7765348Z * [new branch] gh/swolchok/829/head -> origin/gh/swolchok/829/head 2025-12-04T09:33:40.7766678Z * [new branch] gh/swolchok/829/orig -> origin/gh/swolchok/829/orig 2025-12-04T09:33:40.7768426Z * [new branch] gh/swolchok/839/base -> origin/gh/swolchok/839/base 2025-12-04T09:33:40.7769590Z * [new branch] gh/swolchok/839/head -> origin/gh/swolchok/839/head 2025-12-04T09:33:40.7770939Z * [new branch] gh/swolchok/839/orig -> origin/gh/swolchok/839/orig 2025-12-04T09:33:40.7773051Z * [new branch] gh/swolchok/841/base -> origin/gh/swolchok/841/base 2025-12-04T09:33:40.7774318Z * [new branch] gh/swolchok/841/head -> origin/gh/swolchok/841/head 2025-12-04T09:33:40.7775650Z * [new branch] gh/swolchok/841/orig -> origin/gh/swolchok/841/orig 2025-12-04T09:33:40.7777446Z * [new branch] gh/swolchok/842/base -> origin/gh/swolchok/842/base 2025-12-04T09:33:40.7778788Z * [new branch] gh/swolchok/842/head -> origin/gh/swolchok/842/head 2025-12-04T09:33:40.7780072Z * [new branch] gh/swolchok/842/orig -> origin/gh/swolchok/842/orig 2025-12-04T09:33:40.7781686Z * [new branch] gh/swolchok/845/base -> origin/gh/swolchok/845/base 2025-12-04T09:33:40.7782938Z * [new branch] gh/swolchok/845/head -> origin/gh/swolchok/845/head 2025-12-04T09:33:40.7784297Z * [new branch] gh/swolchok/845/orig -> origin/gh/swolchok/845/orig 2025-12-04T09:33:40.7786095Z * [new branch] gh/swolchok/848/base -> origin/gh/swolchok/848/base 2025-12-04T09:33:40.7787446Z * [new branch] gh/swolchok/848/head -> origin/gh/swolchok/848/head 2025-12-04T09:33:40.7788818Z * [new branch] gh/swolchok/848/orig -> origin/gh/swolchok/848/orig 2025-12-04T09:33:40.7790528Z * [new branch] gh/swolchok/856/base -> origin/gh/swolchok/856/base 2025-12-04T09:33:40.7791837Z * [new branch] gh/swolchok/856/head -> origin/gh/swolchok/856/head 2025-12-04T09:33:40.7793114Z * [new branch] gh/swolchok/856/orig -> origin/gh/swolchok/856/orig 2025-12-04T09:33:40.7794989Z * [new branch] gh/swolchok/860/base -> origin/gh/swolchok/860/base 2025-12-04T09:33:40.7796401Z * [new branch] gh/swolchok/860/head -> origin/gh/swolchok/860/head 2025-12-04T09:33:40.7797722Z * [new branch] gh/swolchok/860/orig -> origin/gh/swolchok/860/orig 2025-12-04T09:33:40.7799585Z * [new branch] gh/swolchok/861/base -> origin/gh/swolchok/861/base 2025-12-04T09:33:40.7800949Z * [new branch] gh/swolchok/861/head -> origin/gh/swolchok/861/head 2025-12-04T09:33:40.7802193Z * [new branch] gh/swolchok/861/orig -> origin/gh/swolchok/861/orig 2025-12-04T09:33:40.7804044Z * [new branch] gh/swolchok/862/base -> origin/gh/swolchok/862/base 2025-12-04T09:33:40.7805248Z * [new branch] gh/swolchok/862/head -> origin/gh/swolchok/862/head 2025-12-04T09:33:40.7806599Z * [new branch] gh/swolchok/862/orig -> origin/gh/swolchok/862/orig 2025-12-04T09:33:40.7809060Z * [new branch] gh/swolchok/863/base -> origin/gh/swolchok/863/base 2025-12-04T09:33:40.7810286Z * [new branch] gh/swolchok/863/head -> origin/gh/swolchok/863/head 2025-12-04T09:33:40.7811726Z * [new branch] gh/swolchok/863/orig -> origin/gh/swolchok/863/orig 2025-12-04T09:33:40.7813524Z * [new branch] gh/swolchok/864/base -> origin/gh/swolchok/864/base 2025-12-04T09:33:40.7814736Z * [new branch] gh/swolchok/864/head -> origin/gh/swolchok/864/head 2025-12-04T09:33:40.7815997Z * [new branch] gh/swolchok/864/orig -> origin/gh/swolchok/864/orig 2025-12-04T09:33:40.7817822Z * [new branch] gh/swolchok/865/base -> origin/gh/swolchok/865/base 2025-12-04T09:33:40.7819315Z * [new branch] gh/swolchok/865/head -> origin/gh/swolchok/865/head 2025-12-04T09:33:40.7820552Z * [new branch] gh/swolchok/865/orig -> origin/gh/swolchok/865/orig 2025-12-04T09:33:40.7822833Z * [new branch] gh/swolchok/866/base -> origin/gh/swolchok/866/base 2025-12-04T09:33:40.7824171Z * [new branch] gh/swolchok/866/head -> origin/gh/swolchok/866/head 2025-12-04T09:33:40.7825562Z * [new branch] gh/swolchok/866/orig -> origin/gh/swolchok/866/orig 2025-12-04T09:33:40.7827277Z * [new branch] gh/swolchok/867/base -> origin/gh/swolchok/867/base 2025-12-04T09:33:40.7828585Z * [new branch] gh/swolchok/867/head -> origin/gh/swolchok/867/head 2025-12-04T09:33:40.7829897Z * [new branch] gh/swolchok/867/orig -> origin/gh/swolchok/867/orig 2025-12-04T09:33:40.7831598Z * [new branch] gh/swolchok/868/base -> origin/gh/swolchok/868/base 2025-12-04T09:33:40.7832915Z * [new branch] gh/swolchok/868/head -> origin/gh/swolchok/868/head 2025-12-04T09:33:40.7834240Z * [new branch] gh/swolchok/868/orig -> origin/gh/swolchok/868/orig 2025-12-04T09:33:40.7835988Z * [new branch] gh/swolchok/869/base -> origin/gh/swolchok/869/base 2025-12-04T09:33:40.7837302Z * [new branch] gh/swolchok/869/head -> origin/gh/swolchok/869/head 2025-12-04T09:33:40.7838569Z * [new branch] gh/swolchok/869/orig -> origin/gh/swolchok/869/orig 2025-12-04T09:33:40.7840411Z * [new branch] gh/swolchok/870/base -> origin/gh/swolchok/870/base 2025-12-04T09:33:40.7841596Z * [new branch] gh/swolchok/870/head -> origin/gh/swolchok/870/head 2025-12-04T09:33:40.7842938Z * [new branch] gh/swolchok/870/orig -> origin/gh/swolchok/870/orig 2025-12-04T09:33:40.7844679Z * [new branch] gh/swolchok/871/base -> origin/gh/swolchok/871/base 2025-12-04T09:33:40.7846075Z * [new branch] gh/swolchok/871/head -> origin/gh/swolchok/871/head 2025-12-04T09:33:40.7847560Z * [new branch] gh/swolchok/871/orig -> origin/gh/swolchok/871/orig 2025-12-04T09:33:40.7849881Z * [new branch] gh/teja-rao/4/base -> origin/gh/teja-rao/4/base 2025-12-04T09:33:40.7851197Z * [new branch] gh/teja-rao/4/head -> origin/gh/teja-rao/4/head 2025-12-04T09:33:40.7852517Z * [new branch] gh/teja-rao/4/orig -> origin/gh/teja-rao/4/orig 2025-12-04T09:33:40.7854542Z * [new branch] gh/tianyu-l/2/base -> origin/gh/tianyu-l/2/base 2025-12-04T09:33:40.7855845Z * [new branch] gh/tianyu-l/2/head -> origin/gh/tianyu-l/2/head 2025-12-04T09:33:40.7857158Z * [new branch] gh/tianyu-l/2/orig -> origin/gh/tianyu-l/2/orig 2025-12-04T09:33:40.7859086Z * [new branch] gh/tianyu-l/3/base -> origin/gh/tianyu-l/3/base 2025-12-04T09:33:40.7860370Z * [new branch] gh/tianyu-l/3/orig -> origin/gh/tianyu-l/3/orig 2025-12-04T09:33:40.7862078Z * [new branch] gh/tianyu-l/4/base -> origin/gh/tianyu-l/4/base 2025-12-04T09:33:40.7863293Z * [new branch] gh/tianyu-l/4/head -> origin/gh/tianyu-l/4/head 2025-12-04T09:33:40.7865046Z * [new branch] gh/tianyu-l/4/orig -> origin/gh/tianyu-l/4/orig 2025-12-04T09:33:40.7867612Z * [new branch] gh/tugsbayasgalan/10/base -> origin/gh/tugsbayasgalan/10/base 2025-12-04T09:33:40.7868878Z * [new branch] gh/tugsbayasgalan/10/head -> origin/gh/tugsbayasgalan/10/head 2025-12-04T09:33:40.7870177Z * [new branch] gh/tugsbayasgalan/10/orig -> origin/gh/tugsbayasgalan/10/orig 2025-12-04T09:33:40.7871830Z * [new branch] gh/tugsbayasgalan/13/base -> origin/gh/tugsbayasgalan/13/base 2025-12-04T09:33:40.7873077Z * [new branch] gh/tugsbayasgalan/13/head -> origin/gh/tugsbayasgalan/13/head 2025-12-04T09:33:40.7874327Z * [new branch] gh/tugsbayasgalan/13/orig -> origin/gh/tugsbayasgalan/13/orig 2025-12-04T09:33:40.7876343Z * [new branch] gh/tugsbayasgalan/17/base -> origin/gh/tugsbayasgalan/17/base 2025-12-04T09:33:40.7877601Z * [new branch] gh/tugsbayasgalan/17/head -> origin/gh/tugsbayasgalan/17/head 2025-12-04T09:33:40.7878856Z * [new branch] gh/tugsbayasgalan/17/orig -> origin/gh/tugsbayasgalan/17/orig 2025-12-04T09:33:40.7880654Z * [new branch] gh/tugsbayasgalan/2/base -> origin/gh/tugsbayasgalan/2/base 2025-12-04T09:33:40.7881886Z * [new branch] gh/tugsbayasgalan/2/head -> origin/gh/tugsbayasgalan/2/head 2025-12-04T09:33:40.7883146Z * [new branch] gh/tugsbayasgalan/2/orig -> origin/gh/tugsbayasgalan/2/orig 2025-12-04T09:33:40.7885217Z * [new branch] gh/tugsbayasgalan/28/base -> origin/gh/tugsbayasgalan/28/base 2025-12-04T09:33:40.7886457Z * [new branch] gh/tugsbayasgalan/28/head -> origin/gh/tugsbayasgalan/28/head 2025-12-04T09:33:40.7887745Z * [new branch] gh/tugsbayasgalan/28/orig -> origin/gh/tugsbayasgalan/28/orig 2025-12-04T09:33:40.7889503Z * [new branch] gh/tugsbayasgalan/32/base -> origin/gh/tugsbayasgalan/32/base 2025-12-04T09:33:40.7890811Z * [new branch] gh/tugsbayasgalan/32/head -> origin/gh/tugsbayasgalan/32/head 2025-12-04T09:33:40.7892012Z * [new branch] gh/tugsbayasgalan/32/orig -> origin/gh/tugsbayasgalan/32/orig 2025-12-04T09:33:40.7893940Z * [new branch] gh/tugsbayasgalan/35/base -> origin/gh/tugsbayasgalan/35/base 2025-12-04T09:33:40.7895357Z * [new branch] gh/tugsbayasgalan/35/head -> origin/gh/tugsbayasgalan/35/head 2025-12-04T09:33:40.7896762Z * [new branch] gh/tugsbayasgalan/35/orig -> origin/gh/tugsbayasgalan/35/orig 2025-12-04T09:33:40.7899082Z * [new branch] gh/tugsbayasgalan/36/base -> origin/gh/tugsbayasgalan/36/base 2025-12-04T09:33:40.7900319Z * [new branch] gh/tugsbayasgalan/36/head -> origin/gh/tugsbayasgalan/36/head 2025-12-04T09:33:40.7901595Z * [new branch] gh/tugsbayasgalan/36/orig -> origin/gh/tugsbayasgalan/36/orig 2025-12-04T09:33:40.7903340Z * [new branch] gh/tugsbayasgalan/37/base -> origin/gh/tugsbayasgalan/37/base 2025-12-04T09:33:40.7904625Z * [new branch] gh/tugsbayasgalan/37/head -> origin/gh/tugsbayasgalan/37/head 2025-12-04T09:33:40.7905817Z * [new branch] gh/tugsbayasgalan/37/orig -> origin/gh/tugsbayasgalan/37/orig 2025-12-04T09:33:40.7907490Z * [new branch] gh/tugsbayasgalan/43/base -> origin/gh/tugsbayasgalan/43/base 2025-12-04T09:33:40.7908727Z * [new branch] gh/tugsbayasgalan/43/head -> origin/gh/tugsbayasgalan/43/head 2025-12-04T09:33:40.7909964Z * [new branch] gh/tugsbayasgalan/43/orig -> origin/gh/tugsbayasgalan/43/orig 2025-12-04T09:33:40.7911637Z * [new branch] gh/tugsbayasgalan/48/base -> origin/gh/tugsbayasgalan/48/base 2025-12-04T09:33:40.7912923Z * [new branch] gh/tugsbayasgalan/48/head -> origin/gh/tugsbayasgalan/48/head 2025-12-04T09:33:40.7914193Z * [new branch] gh/tugsbayasgalan/48/orig -> origin/gh/tugsbayasgalan/48/orig 2025-12-04T09:33:40.7915907Z * [new branch] gh/tugsbayasgalan/51/base -> origin/gh/tugsbayasgalan/51/base 2025-12-04T09:33:40.7917302Z * [new branch] gh/tugsbayasgalan/51/head -> origin/gh/tugsbayasgalan/51/head 2025-12-04T09:33:40.7918521Z * [new branch] gh/tugsbayasgalan/51/orig -> origin/gh/tugsbayasgalan/51/orig 2025-12-04T09:33:40.7920004Z * [new branch] gh/tugsbayasgalan/52/base -> origin/gh/tugsbayasgalan/52/base 2025-12-04T09:33:40.7921348Z * [new branch] gh/tugsbayasgalan/52/head -> origin/gh/tugsbayasgalan/52/head 2025-12-04T09:33:40.7922551Z * [new branch] gh/tugsbayasgalan/52/orig -> origin/gh/tugsbayasgalan/52/orig 2025-12-04T09:33:40.7924313Z * [new branch] gh/tugsbayasgalan/53/base -> origin/gh/tugsbayasgalan/53/base 2025-12-04T09:33:40.7925559Z * [new branch] gh/tugsbayasgalan/53/head -> origin/gh/tugsbayasgalan/53/head 2025-12-04T09:33:40.7926838Z * [new branch] gh/tugsbayasgalan/53/orig -> origin/gh/tugsbayasgalan/53/orig 2025-12-04T09:33:40.7928764Z * [new branch] gh/tugsbayasgalan/55/base -> origin/gh/tugsbayasgalan/55/base 2025-12-04T09:33:40.7930190Z * [new branch] gh/tugsbayasgalan/55/head -> origin/gh/tugsbayasgalan/55/head 2025-12-04T09:33:40.7931461Z * [new branch] gh/tugsbayasgalan/55/orig -> origin/gh/tugsbayasgalan/55/orig 2025-12-04T09:33:40.7933377Z * [new branch] gh/tugsbayasgalan/59/base -> origin/gh/tugsbayasgalan/59/base 2025-12-04T09:33:40.7934747Z * [new branch] gh/tugsbayasgalan/59/head -> origin/gh/tugsbayasgalan/59/head 2025-12-04T09:33:40.7935964Z * [new branch] gh/tugsbayasgalan/59/orig -> origin/gh/tugsbayasgalan/59/orig 2025-12-04T09:33:40.7937725Z * [new branch] gh/tugsbayasgalan/6/base -> origin/gh/tugsbayasgalan/6/base 2025-12-04T09:33:40.7938953Z * [new branch] gh/tugsbayasgalan/6/head -> origin/gh/tugsbayasgalan/6/head 2025-12-04T09:33:40.7940223Z * [new branch] gh/tugsbayasgalan/6/orig -> origin/gh/tugsbayasgalan/6/orig 2025-12-04T09:33:40.7941793Z * [new branch] gh/tugsbayasgalan/60/base -> origin/gh/tugsbayasgalan/60/base 2025-12-04T09:33:40.7943051Z * [new branch] gh/tugsbayasgalan/60/head -> origin/gh/tugsbayasgalan/60/head 2025-12-04T09:33:40.7944313Z * [new branch] gh/tugsbayasgalan/60/orig -> origin/gh/tugsbayasgalan/60/orig 2025-12-04T09:33:40.7946696Z * [new branch] gh/tugsbayasgalan/61/base -> origin/gh/tugsbayasgalan/61/base 2025-12-04T09:33:40.7947942Z * [new branch] gh/tugsbayasgalan/61/head -> origin/gh/tugsbayasgalan/61/head 2025-12-04T09:33:40.7949225Z * [new branch] gh/tugsbayasgalan/61/orig -> origin/gh/tugsbayasgalan/61/orig 2025-12-04T09:33:40.7951041Z * [new branch] gh/tugsbayasgalan/63/base -> origin/gh/tugsbayasgalan/63/base 2025-12-04T09:33:40.7952355Z * [new branch] gh/tugsbayasgalan/63/head -> origin/gh/tugsbayasgalan/63/head 2025-12-04T09:33:40.7953622Z * [new branch] gh/tugsbayasgalan/63/orig -> origin/gh/tugsbayasgalan/63/orig 2025-12-04T09:33:40.7955361Z * [new branch] gh/tugsbayasgalan/67/base -> origin/gh/tugsbayasgalan/67/base 2025-12-04T09:33:40.7956595Z * [new branch] gh/tugsbayasgalan/67/head -> origin/gh/tugsbayasgalan/67/head 2025-12-04T09:33:40.7957864Z * [new branch] gh/tugsbayasgalan/67/orig -> origin/gh/tugsbayasgalan/67/orig 2025-12-04T09:33:40.7959847Z * [new branch] gh/tugsbayasgalan/68/base -> origin/gh/tugsbayasgalan/68/base 2025-12-04T09:33:40.7961172Z * [new branch] gh/tugsbayasgalan/68/head -> origin/gh/tugsbayasgalan/68/head 2025-12-04T09:33:40.7962448Z * [new branch] gh/tugsbayasgalan/68/orig -> origin/gh/tugsbayasgalan/68/orig 2025-12-04T09:33:40.7964294Z * [new branch] gh/tugsbayasgalan/7/base -> origin/gh/tugsbayasgalan/7/base 2025-12-04T09:33:40.7965581Z * [new branch] gh/tugsbayasgalan/7/head -> origin/gh/tugsbayasgalan/7/head 2025-12-04T09:33:40.7966950Z * [new branch] gh/tugsbayasgalan/7/orig -> origin/gh/tugsbayasgalan/7/orig 2025-12-04T09:33:40.7968905Z * [new branch] gh/tugsbayasgalan/70/base -> origin/gh/tugsbayasgalan/70/base 2025-12-04T09:33:40.7970448Z * [new branch] gh/tugsbayasgalan/70/head -> origin/gh/tugsbayasgalan/70/head 2025-12-04T09:33:40.7971748Z * [new branch] gh/tugsbayasgalan/70/orig -> origin/gh/tugsbayasgalan/70/orig 2025-12-04T09:33:40.7973695Z * [new branch] gh/tugsbayasgalan/71/base -> origin/gh/tugsbayasgalan/71/base 2025-12-04T09:33:40.7975095Z * [new branch] gh/tugsbayasgalan/71/head -> origin/gh/tugsbayasgalan/71/head 2025-12-04T09:33:40.7976475Z * [new branch] gh/tugsbayasgalan/71/orig -> origin/gh/tugsbayasgalan/71/orig 2025-12-04T09:33:40.7978591Z * [new branch] gh/tugsbayasgalan/72/base -> origin/gh/tugsbayasgalan/72/base 2025-12-04T09:33:40.7979868Z * [new branch] gh/tugsbayasgalan/72/head -> origin/gh/tugsbayasgalan/72/head 2025-12-04T09:33:40.7981144Z * [new branch] gh/tugsbayasgalan/72/orig -> origin/gh/tugsbayasgalan/72/orig 2025-12-04T09:33:40.7983012Z * [new branch] gh/tugsbayasgalan/73/base -> origin/gh/tugsbayasgalan/73/base 2025-12-04T09:33:40.7984460Z * [new branch] gh/tugsbayasgalan/73/head -> origin/gh/tugsbayasgalan/73/head 2025-12-04T09:33:40.7985739Z * [new branch] gh/tugsbayasgalan/73/orig -> origin/gh/tugsbayasgalan/73/orig 2025-12-04T09:33:40.7987678Z * [new branch] gh/tugsbayasgalan/74/base -> origin/gh/tugsbayasgalan/74/base 2025-12-04T09:33:40.7988997Z * [new branch] gh/tugsbayasgalan/74/head -> origin/gh/tugsbayasgalan/74/head 2025-12-04T09:33:40.7990256Z * [new branch] gh/tugsbayasgalan/74/orig -> origin/gh/tugsbayasgalan/74/orig 2025-12-04T09:33:40.7992131Z * [new branch] gh/tugsbayasgalan/75/base -> origin/gh/tugsbayasgalan/75/base 2025-12-04T09:33:40.7993283Z * [new branch] gh/tugsbayasgalan/75/head -> origin/gh/tugsbayasgalan/75/head 2025-12-04T09:33:40.7995021Z * [new branch] gh/tugsbayasgalan/75/orig -> origin/gh/tugsbayasgalan/75/orig 2025-12-04T09:33:40.7996728Z * [new branch] gh/tugsbayasgalan/76/base -> origin/gh/tugsbayasgalan/76/base 2025-12-04T09:33:40.8001719Z * [new branch] gh/tugsbayasgalan/76/head -> origin/gh/tugsbayasgalan/76/head 2025-12-04T09:33:40.8002948Z * [new branch] gh/tugsbayasgalan/76/orig -> origin/gh/tugsbayasgalan/76/orig 2025-12-04T09:33:40.8005056Z * [new branch] gh/tugsbayasgalan/77/base -> origin/gh/tugsbayasgalan/77/base 2025-12-04T09:33:40.8006325Z * [new branch] gh/tugsbayasgalan/77/head -> origin/gh/tugsbayasgalan/77/head 2025-12-04T09:33:40.8007528Z * [new branch] gh/tugsbayasgalan/77/orig -> origin/gh/tugsbayasgalan/77/orig 2025-12-04T09:33:40.8009382Z * [new branch] gh/tugsbayasgalan/78/base -> origin/gh/tugsbayasgalan/78/base 2025-12-04T09:33:40.8010800Z * [new branch] gh/tugsbayasgalan/78/head -> origin/gh/tugsbayasgalan/78/head 2025-12-04T09:33:40.8012073Z * [new branch] gh/tugsbayasgalan/78/orig -> origin/gh/tugsbayasgalan/78/orig 2025-12-04T09:33:40.8013828Z * [new branch] gh/tugsbayasgalan/79/base -> origin/gh/tugsbayasgalan/79/base 2025-12-04T09:33:40.8015150Z * [new branch] gh/tugsbayasgalan/79/head -> origin/gh/tugsbayasgalan/79/head 2025-12-04T09:33:40.8016444Z * [new branch] gh/tugsbayasgalan/79/orig -> origin/gh/tugsbayasgalan/79/orig 2025-12-04T09:33:40.8018336Z * [new branch] gh/tugsbayasgalan/8/base -> origin/gh/tugsbayasgalan/8/base 2025-12-04T09:33:40.8019527Z * [new branch] gh/tugsbayasgalan/8/head -> origin/gh/tugsbayasgalan/8/head 2025-12-04T09:33:40.8020897Z * [new branch] gh/tugsbayasgalan/8/orig -> origin/gh/tugsbayasgalan/8/orig 2025-12-04T09:33:40.8022579Z * [new branch] gh/tugsbayasgalan/80/base -> origin/gh/tugsbayasgalan/80/base 2025-12-04T09:33:40.8023770Z * [new branch] gh/tugsbayasgalan/80/head -> origin/gh/tugsbayasgalan/80/head 2025-12-04T09:33:40.8025057Z * [new branch] gh/tugsbayasgalan/80/orig -> origin/gh/tugsbayasgalan/80/orig 2025-12-04T09:33:40.8026910Z * [new branch] gh/tugsbayasgalan/81/base -> origin/gh/tugsbayasgalan/81/base 2025-12-04T09:33:40.8028079Z * [new branch] gh/tugsbayasgalan/81/head -> origin/gh/tugsbayasgalan/81/head 2025-12-04T09:33:40.8029357Z * [new branch] gh/tugsbayasgalan/81/orig -> origin/gh/tugsbayasgalan/81/orig 2025-12-04T09:33:40.8031813Z * [new branch] gh/tugsbayasgalan/82/base -> origin/gh/tugsbayasgalan/82/base 2025-12-04T09:33:40.8033283Z * [new branch] gh/tugsbayasgalan/82/head -> origin/gh/tugsbayasgalan/82/head 2025-12-04T09:33:40.8034590Z * [new branch] gh/tugsbayasgalan/82/orig -> origin/gh/tugsbayasgalan/82/orig 2025-12-04T09:33:40.8036221Z * [new branch] gh/tugsbayasgalan/83/base -> origin/gh/tugsbayasgalan/83/base 2025-12-04T09:33:40.8037508Z * [new branch] gh/tugsbayasgalan/83/head -> origin/gh/tugsbayasgalan/83/head 2025-12-04T09:33:40.8038786Z * [new branch] gh/tugsbayasgalan/83/orig -> origin/gh/tugsbayasgalan/83/orig 2025-12-04T09:33:40.8040486Z * [new branch] gh/tugsbayasgalan/84/base -> origin/gh/tugsbayasgalan/84/base 2025-12-04T09:33:40.8041988Z * [new branch] gh/tugsbayasgalan/84/head -> origin/gh/tugsbayasgalan/84/head 2025-12-04T09:33:40.8043255Z * [new branch] gh/tugsbayasgalan/84/orig -> origin/gh/tugsbayasgalan/84/orig 2025-12-04T09:33:40.8044845Z * [new branch] gh/tugsbayasgalan/85/base -> origin/gh/tugsbayasgalan/85/base 2025-12-04T09:33:40.8046145Z * [new branch] gh/tugsbayasgalan/85/head -> origin/gh/tugsbayasgalan/85/head 2025-12-04T09:33:40.8047452Z * [new branch] gh/tugsbayasgalan/85/orig -> origin/gh/tugsbayasgalan/85/orig 2025-12-04T09:33:40.8049212Z * [new branch] gh/tugsbayasgalan/86/base -> origin/gh/tugsbayasgalan/86/base 2025-12-04T09:33:40.8050571Z * [new branch] gh/tugsbayasgalan/86/head -> origin/gh/tugsbayasgalan/86/head 2025-12-04T09:33:40.8051845Z * [new branch] gh/tugsbayasgalan/86/orig -> origin/gh/tugsbayasgalan/86/orig 2025-12-04T09:33:40.8054054Z * [new branch] gh/tugsbayasgalan/87/base -> origin/gh/tugsbayasgalan/87/base 2025-12-04T09:33:40.8055264Z * [new branch] gh/tugsbayasgalan/87/head -> origin/gh/tugsbayasgalan/87/head 2025-12-04T09:33:40.8056524Z * [new branch] gh/tugsbayasgalan/87/orig -> origin/gh/tugsbayasgalan/87/orig 2025-12-04T09:33:40.8058664Z * [new branch] gh/tugsbayasgalan/88/base -> origin/gh/tugsbayasgalan/88/base 2025-12-04T09:33:40.8059890Z * [new branch] gh/tugsbayasgalan/88/head -> origin/gh/tugsbayasgalan/88/head 2025-12-04T09:33:40.8061626Z * [new branch] gh/tugsbayasgalan/88/orig -> origin/gh/tugsbayasgalan/88/orig 2025-12-04T09:33:40.8063430Z * [new branch] gh/tugsbayasgalan/89/base -> origin/gh/tugsbayasgalan/89/base 2025-12-04T09:33:40.8065197Z * [new branch] gh/tugsbayasgalan/89/head -> origin/gh/tugsbayasgalan/89/head 2025-12-04T09:33:40.8066507Z * [new branch] gh/tugsbayasgalan/89/orig -> origin/gh/tugsbayasgalan/89/orig 2025-12-04T09:33:40.8068190Z * [new branch] gh/tugsbayasgalan/9/base -> origin/gh/tugsbayasgalan/9/base 2025-12-04T09:33:40.8069366Z * [new branch] gh/tugsbayasgalan/9/head -> origin/gh/tugsbayasgalan/9/head 2025-12-04T09:33:40.8070635Z * [new branch] gh/tugsbayasgalan/9/orig -> origin/gh/tugsbayasgalan/9/orig 2025-12-04T09:33:40.8073371Z * [new branch] gh/tugsbayasgalan/90/base -> origin/gh/tugsbayasgalan/90/base 2025-12-04T09:33:40.8074540Z * [new branch] gh/tugsbayasgalan/90/head -> origin/gh/tugsbayasgalan/90/head 2025-12-04T09:33:40.8075799Z * [new branch] gh/tugsbayasgalan/90/orig -> origin/gh/tugsbayasgalan/90/orig 2025-12-04T09:33:40.8077912Z * [new branch] gh/tugsbayasgalan/91/base -> origin/gh/tugsbayasgalan/91/base 2025-12-04T09:33:40.8079114Z * [new branch] gh/tugsbayasgalan/91/head -> origin/gh/tugsbayasgalan/91/head 2025-12-04T09:33:40.8080428Z * [new branch] gh/tugsbayasgalan/91/orig -> origin/gh/tugsbayasgalan/91/orig 2025-12-04T09:33:40.8082242Z * [new branch] gh/tugsbayasgalan/92/base -> origin/gh/tugsbayasgalan/92/base 2025-12-04T09:33:40.8083844Z * [new branch] gh/tugsbayasgalan/92/head -> origin/gh/tugsbayasgalan/92/head 2025-12-04T09:33:40.8085089Z * [new branch] gh/tugsbayasgalan/92/orig -> origin/gh/tugsbayasgalan/92/orig 2025-12-04T09:33:40.8086971Z * [new branch] gh/tugsbayasgalan/93/base -> origin/gh/tugsbayasgalan/93/base 2025-12-04T09:33:40.8088314Z * [new branch] gh/tugsbayasgalan/93/head -> origin/gh/tugsbayasgalan/93/head 2025-12-04T09:33:40.8089661Z * [new branch] gh/tugsbayasgalan/93/orig -> origin/gh/tugsbayasgalan/93/orig 2025-12-04T09:33:40.8092187Z * [new branch] gh/v0i0/14/base -> origin/gh/v0i0/14/base 2025-12-04T09:33:40.8093393Z * [new branch] gh/v0i0/14/head -> origin/gh/v0i0/14/head 2025-12-04T09:33:40.8095186Z * [new branch] gh/v0i0/14/orig -> origin/gh/v0i0/14/orig 2025-12-04T09:33:40.8097546Z * [new branch] gh/v0i0/15/base -> origin/gh/v0i0/15/base 2025-12-04T09:33:40.8098985Z * [new branch] gh/v0i0/15/head -> origin/gh/v0i0/15/head 2025-12-04T09:33:40.8100280Z * [new branch] gh/v0i0/15/orig -> origin/gh/v0i0/15/orig 2025-12-04T09:33:40.8101970Z * [new branch] gh/v0i0/16/base -> origin/gh/v0i0/16/base 2025-12-04T09:33:40.8103236Z * [new branch] gh/v0i0/16/head -> origin/gh/v0i0/16/head 2025-12-04T09:33:40.8104489Z * [new branch] gh/v0i0/16/orig -> origin/gh/v0i0/16/orig 2025-12-04T09:33:40.8106194Z * [new branch] gh/v0i0/17/base -> origin/gh/v0i0/17/base 2025-12-04T09:33:40.8107485Z * [new branch] gh/v0i0/17/head -> origin/gh/v0i0/17/head 2025-12-04T09:33:40.8109289Z * [new branch] gh/v0i0/17/orig -> origin/gh/v0i0/17/orig 2025-12-04T09:33:40.8111025Z * [new branch] gh/v0i0/18/base -> origin/gh/v0i0/18/base 2025-12-04T09:33:40.8112803Z * [new branch] gh/v0i0/18/head -> origin/gh/v0i0/18/head 2025-12-04T09:33:40.8114069Z * [new branch] gh/v0i0/18/orig -> origin/gh/v0i0/18/orig 2025-12-04T09:33:40.8115906Z * [new branch] gh/v0i0/19/base -> origin/gh/v0i0/19/base 2025-12-04T09:33:40.8117223Z * [new branch] gh/v0i0/19/head -> origin/gh/v0i0/19/head 2025-12-04T09:33:40.8118467Z * [new branch] gh/v0i0/19/orig -> origin/gh/v0i0/19/orig 2025-12-04T09:33:40.8120580Z * [new branch] gh/vishal9-team/1/base -> origin/gh/vishal9-team/1/base 2025-12-04T09:33:40.8121884Z * [new branch] gh/vishal9-team/1/head -> origin/gh/vishal9-team/1/head 2025-12-04T09:33:40.8123373Z * [new branch] gh/vishal9-team/2/base -> origin/gh/vishal9-team/2/base 2025-12-04T09:33:40.8124722Z * [new branch] gh/vishal9-team/2/head -> origin/gh/vishal9-team/2/head 2025-12-04T09:33:40.8125973Z * [new branch] gh/vishal9-team/2/orig -> origin/gh/vishal9-team/2/orig 2025-12-04T09:33:40.8128303Z * [new branch] gh/vishal9-team/3/base -> origin/gh/vishal9-team/3/base 2025-12-04T09:33:40.8129545Z * [new branch] gh/vishal9-team/3/head -> origin/gh/vishal9-team/3/head 2025-12-04T09:33:40.8130802Z * [new branch] gh/vishal9-team/3/orig -> origin/gh/vishal9-team/3/orig 2025-12-04T09:33:40.8132444Z * [new branch] gh/vishal9-team/4/base -> origin/gh/vishal9-team/4/base 2025-12-04T09:33:40.8133711Z * [new branch] gh/vishal9-team/4/head -> origin/gh/vishal9-team/4/head 2025-12-04T09:33:40.8134924Z * [new branch] gh/vishal9-team/4/orig -> origin/gh/vishal9-team/4/orig 2025-12-04T09:33:40.8136823Z * [new branch] gh/vkuzo/1/next -> origin/gh/vkuzo/1/next 2025-12-04T09:33:40.8138650Z * [new branch] gh/vkuzo/2/next -> origin/gh/vkuzo/2/next 2025-12-04T09:33:40.8140295Z * [new branch] gh/vkuzo/3/next -> origin/gh/vkuzo/3/next 2025-12-04T09:33:40.8142347Z * [new branch] gh/wconstab/424/base -> origin/gh/wconstab/424/base 2025-12-04T09:33:40.8143739Z * [new branch] gh/wconstab/424/head -> origin/gh/wconstab/424/head 2025-12-04T09:33:40.8145052Z * [new branch] gh/wconstab/424/orig -> origin/gh/wconstab/424/orig 2025-12-04T09:33:40.8146879Z * [new branch] gh/wconstab/435/base -> origin/gh/wconstab/435/base 2025-12-04T09:33:40.8148136Z * [new branch] gh/wconstab/435/head -> origin/gh/wconstab/435/head 2025-12-04T09:33:40.8149415Z * [new branch] gh/wconstab/435/orig -> origin/gh/wconstab/435/orig 2025-12-04T09:33:40.8151148Z * [new branch] gh/wconstab/444/base -> origin/gh/wconstab/444/base 2025-12-04T09:33:40.8152465Z * [new branch] gh/wconstab/444/head -> origin/gh/wconstab/444/head 2025-12-04T09:33:40.8153763Z * [new branch] gh/wconstab/444/orig -> origin/gh/wconstab/444/orig 2025-12-04T09:33:40.8155479Z * [new branch] gh/wconstab/447/base -> origin/gh/wconstab/447/base 2025-12-04T09:33:40.8156736Z * [new branch] gh/wconstab/447/head -> origin/gh/wconstab/447/head 2025-12-04T09:33:40.8157947Z * [new branch] gh/wconstab/447/orig -> origin/gh/wconstab/447/orig 2025-12-04T09:33:40.8159666Z * [new branch] gh/wconstab/448/base -> origin/gh/wconstab/448/base 2025-12-04T09:33:40.8160937Z * [new branch] gh/wconstab/448/head -> origin/gh/wconstab/448/head 2025-12-04T09:33:40.8162218Z * [new branch] gh/wconstab/448/orig -> origin/gh/wconstab/448/orig 2025-12-04T09:33:40.8163838Z * [new branch] gh/wconstab/449/base -> origin/gh/wconstab/449/base 2025-12-04T09:33:40.8165129Z * [new branch] gh/wconstab/449/head -> origin/gh/wconstab/449/head 2025-12-04T09:33:40.8166430Z * [new branch] gh/wconstab/449/orig -> origin/gh/wconstab/449/orig 2025-12-04T09:33:40.8167993Z * [new branch] gh/wconstab/450/base -> origin/gh/wconstab/450/base 2025-12-04T09:33:40.8169354Z * [new branch] gh/wconstab/450/head -> origin/gh/wconstab/450/head 2025-12-04T09:33:40.8170672Z * [new branch] gh/wconstab/450/orig -> origin/gh/wconstab/450/orig 2025-12-04T09:33:40.8172209Z * [new branch] gh/wconstab/451/base -> origin/gh/wconstab/451/base 2025-12-04T09:33:40.8173593Z * [new branch] gh/wconstab/451/head -> origin/gh/wconstab/451/head 2025-12-04T09:33:40.8174803Z * [new branch] gh/wconstab/451/orig -> origin/gh/wconstab/451/orig 2025-12-04T09:33:40.8176597Z * [new branch] gh/wconstab/452/base -> origin/gh/wconstab/452/base 2025-12-04T09:33:40.8177911Z * [new branch] gh/wconstab/452/head -> origin/gh/wconstab/452/head 2025-12-04T09:33:40.8179198Z * [new branch] gh/wconstab/452/orig -> origin/gh/wconstab/452/orig 2025-12-04T09:33:40.8180848Z * [new branch] gh/wconstab/453/base -> origin/gh/wconstab/453/base 2025-12-04T09:33:40.8182238Z * [new branch] gh/wconstab/453/head -> origin/gh/wconstab/453/head 2025-12-04T09:33:40.8183599Z * [new branch] gh/wconstab/453/orig -> origin/gh/wconstab/453/orig 2025-12-04T09:33:40.8185128Z * [new branch] gh/wconstab/454/base -> origin/gh/wconstab/454/base 2025-12-04T09:33:40.8186420Z * [new branch] gh/wconstab/454/head -> origin/gh/wconstab/454/head 2025-12-04T09:33:40.8187704Z * [new branch] gh/wconstab/454/orig -> origin/gh/wconstab/454/orig 2025-12-04T09:33:40.8189395Z * [new branch] gh/wconstab/455/base -> origin/gh/wconstab/455/base 2025-12-04T09:33:40.8190674Z * [new branch] gh/wconstab/455/head -> origin/gh/wconstab/455/head 2025-12-04T09:33:40.8191895Z * [new branch] gh/wconstab/455/orig -> origin/gh/wconstab/455/orig 2025-12-04T09:33:40.8193876Z * [new branch] gh/wconstab/456/base -> origin/gh/wconstab/456/base 2025-12-04T09:33:40.8195561Z * [new branch] gh/wconstab/456/head -> origin/gh/wconstab/456/head 2025-12-04T09:33:40.8199522Z * [new branch] gh/wconstab/456/orig -> origin/gh/wconstab/456/orig 2025-12-04T09:33:40.8199783Z * [new branch] gh/wconstab/457/base -> origin/gh/wconstab/457/base 2025-12-04T09:33:40.8200688Z * [new branch] gh/wconstab/457/head -> origin/gh/wconstab/457/head 2025-12-04T09:33:40.8201448Z * [new branch] gh/wconstab/457/orig -> origin/gh/wconstab/457/orig 2025-12-04T09:33:40.8203317Z * [new branch] gh/wconstab/458/base -> origin/gh/wconstab/458/base 2025-12-04T09:33:40.8204688Z * [new branch] gh/wconstab/458/head -> origin/gh/wconstab/458/head 2025-12-04T09:33:40.8205992Z * [new branch] gh/wconstab/458/orig -> origin/gh/wconstab/458/orig 2025-12-04T09:33:40.8207571Z * [new branch] gh/wconstab/459/base -> origin/gh/wconstab/459/base 2025-12-04T09:33:40.8208895Z * [new branch] gh/wconstab/459/head -> origin/gh/wconstab/459/head 2025-12-04T09:33:40.8210069Z * [new branch] gh/wconstab/459/orig -> origin/gh/wconstab/459/orig 2025-12-04T09:33:40.8212449Z * [new branch] gh/wconstab/460/base -> origin/gh/wconstab/460/base 2025-12-04T09:33:40.8214464Z * [new branch] gh/wconstab/460/head -> origin/gh/wconstab/460/head 2025-12-04T09:33:40.8215860Z * [new branch] gh/wconstab/460/orig -> origin/gh/wconstab/460/orig 2025-12-04T09:33:40.8218121Z * [new branch] gh/wconstab/461/base -> origin/gh/wconstab/461/base 2025-12-04T09:33:40.8219398Z * [new branch] gh/wconstab/461/head -> origin/gh/wconstab/461/head 2025-12-04T09:33:40.8220616Z * [new branch] gh/wconstab/461/orig -> origin/gh/wconstab/461/orig 2025-12-04T09:33:40.8222234Z * [new branch] gh/wconstab/462/base -> origin/gh/wconstab/462/base 2025-12-04T09:33:40.8223588Z * [new branch] gh/wconstab/462/head -> origin/gh/wconstab/462/head 2025-12-04T09:33:40.8224974Z * [new branch] gh/wconstab/462/orig -> origin/gh/wconstab/462/orig 2025-12-04T09:33:40.8226787Z * [new branch] gh/wconstab/463/base -> origin/gh/wconstab/463/base 2025-12-04T09:33:40.8228137Z * [new branch] gh/wconstab/463/head -> origin/gh/wconstab/463/head 2025-12-04T09:33:40.8229401Z * [new branch] gh/wconstab/463/orig -> origin/gh/wconstab/463/orig 2025-12-04T09:33:40.8231639Z * [new branch] gh/wconstab/464/base -> origin/gh/wconstab/464/base 2025-12-04T09:33:40.8233196Z * [new branch] gh/wconstab/464/head -> origin/gh/wconstab/464/head 2025-12-04T09:33:40.8234318Z * [new branch] gh/wconstab/464/orig -> origin/gh/wconstab/464/orig 2025-12-04T09:33:40.8236054Z * [new branch] gh/wconstab/465/base -> origin/gh/wconstab/465/base 2025-12-04T09:33:40.8237381Z * [new branch] gh/wconstab/465/head -> origin/gh/wconstab/465/head 2025-12-04T09:33:40.8238570Z * [new branch] gh/wconstab/465/orig -> origin/gh/wconstab/465/orig 2025-12-04T09:33:40.8240422Z * [new branch] gh/wconstab/466/base -> origin/gh/wconstab/466/base 2025-12-04T09:33:40.8241593Z * [new branch] gh/wconstab/466/head -> origin/gh/wconstab/466/head 2025-12-04T09:33:40.8243265Z * [new branch] gh/wconstab/466/orig -> origin/gh/wconstab/466/orig 2025-12-04T09:33:40.8245372Z * [new branch] gh/wconstab/467/base -> origin/gh/wconstab/467/base 2025-12-04T09:33:40.8246722Z * [new branch] gh/wconstab/467/head -> origin/gh/wconstab/467/head 2025-12-04T09:33:40.8248051Z * [new branch] gh/wconstab/467/orig -> origin/gh/wconstab/467/orig 2025-12-04T09:33:40.8249714Z * [new branch] gh/wconstab/468/base -> origin/gh/wconstab/468/base 2025-12-04T09:33:40.8250960Z * [new branch] gh/wconstab/468/head -> origin/gh/wconstab/468/head 2025-12-04T09:33:40.8252244Z * [new branch] gh/wconstab/468/orig -> origin/gh/wconstab/468/orig 2025-12-04T09:33:40.8254534Z * [new branch] gh/weifengpy/39/base -> origin/gh/weifengpy/39/base 2025-12-04T09:33:40.8256272Z * [new branch] gh/weifengpy/39/head -> origin/gh/weifengpy/39/head 2025-12-04T09:33:40.8257680Z * [new branch] gh/weifengpy/39/orig -> origin/gh/weifengpy/39/orig 2025-12-04T09:33:40.8259567Z * [new branch] gh/weifengpy/40/base -> origin/gh/weifengpy/40/base 2025-12-04T09:33:40.8260851Z * [new branch] gh/weifengpy/40/head -> origin/gh/weifengpy/40/head 2025-12-04T09:33:40.8262129Z * [new branch] gh/weifengpy/40/orig -> origin/gh/weifengpy/40/orig 2025-12-04T09:33:40.8263961Z * [new branch] gh/weifengpy/41/base -> origin/gh/weifengpy/41/base 2025-12-04T09:33:40.8265345Z * [new branch] gh/weifengpy/41/head -> origin/gh/weifengpy/41/head 2025-12-04T09:33:40.8266731Z * [new branch] gh/weifengpy/41/orig -> origin/gh/weifengpy/41/orig 2025-12-04T09:33:40.8268951Z * [new branch] gh/williamwen42/250/base -> origin/gh/williamwen42/250/base 2025-12-04T09:33:40.8270282Z * [new branch] gh/williamwen42/250/head -> origin/gh/williamwen42/250/head 2025-12-04T09:33:40.8271532Z * [new branch] gh/williamwen42/250/orig -> origin/gh/williamwen42/250/orig 2025-12-04T09:33:40.8273295Z * [new branch] gh/williamwen42/279/base -> origin/gh/williamwen42/279/base 2025-12-04T09:33:40.8274760Z * [new branch] gh/williamwen42/279/head -> origin/gh/williamwen42/279/head 2025-12-04T09:33:40.8275989Z * [new branch] gh/williamwen42/279/orig -> origin/gh/williamwen42/279/orig 2025-12-04T09:33:40.8277691Z * [new branch] gh/williamwen42/282/base -> origin/gh/williamwen42/282/base 2025-12-04T09:33:40.8278976Z * [new branch] gh/williamwen42/282/head -> origin/gh/williamwen42/282/head 2025-12-04T09:33:40.8280193Z * [new branch] gh/williamwen42/282/orig -> origin/gh/williamwen42/282/orig 2025-12-04T09:33:40.8281979Z * [new branch] gh/williamwen42/287/base -> origin/gh/williamwen42/287/base 2025-12-04T09:33:40.8283732Z * [new branch] gh/williamwen42/287/head -> origin/gh/williamwen42/287/head 2025-12-04T09:33:40.8285058Z * [new branch] gh/williamwen42/287/orig -> origin/gh/williamwen42/287/orig 2025-12-04T09:33:40.8286859Z * [new branch] gh/williamwen42/288/base -> origin/gh/williamwen42/288/base 2025-12-04T09:33:40.8288148Z * [new branch] gh/williamwen42/288/head -> origin/gh/williamwen42/288/head 2025-12-04T09:33:40.8289442Z * [new branch] gh/williamwen42/288/orig -> origin/gh/williamwen42/288/orig 2025-12-04T09:33:40.8291385Z * [new branch] gh/williamwen42/296/base -> origin/gh/williamwen42/296/base 2025-12-04T09:33:40.8292765Z * [new branch] gh/williamwen42/296/head -> origin/gh/williamwen42/296/head 2025-12-04T09:33:40.8294081Z * [new branch] gh/williamwen42/296/orig -> origin/gh/williamwen42/296/orig 2025-12-04T09:33:40.8295663Z * [new branch] gh/williamwen42/297/base -> origin/gh/williamwen42/297/base 2025-12-04T09:33:40.8297286Z * [new branch] gh/williamwen42/297/head -> origin/gh/williamwen42/297/head 2025-12-04T09:33:40.8298627Z * [new branch] gh/williamwen42/297/orig -> origin/gh/williamwen42/297/orig 2025-12-04T09:33:40.8300339Z * [new branch] gh/williamwen42/306/base -> origin/gh/williamwen42/306/base 2025-12-04T09:33:40.8301680Z * [new branch] gh/williamwen42/306/head -> origin/gh/williamwen42/306/head 2025-12-04T09:33:40.8302947Z * [new branch] gh/williamwen42/306/orig -> origin/gh/williamwen42/306/orig 2025-12-04T09:33:40.8304698Z * [new branch] gh/williamwen42/309/base -> origin/gh/williamwen42/309/base 2025-12-04T09:33:40.8306125Z * [new branch] gh/williamwen42/309/head -> origin/gh/williamwen42/309/head 2025-12-04T09:33:40.8307468Z * [new branch] gh/williamwen42/309/orig -> origin/gh/williamwen42/309/orig 2025-12-04T09:33:40.8309149Z * [new branch] gh/williamwen42/310/base -> origin/gh/williamwen42/310/base 2025-12-04T09:33:40.8310424Z * [new branch] gh/williamwen42/310/head -> origin/gh/williamwen42/310/head 2025-12-04T09:33:40.8311743Z * [new branch] gh/williamwen42/310/orig -> origin/gh/williamwen42/310/orig 2025-12-04T09:33:40.8314962Z * [new branch] gh/williamwen42/311/base -> origin/gh/williamwen42/311/base 2025-12-04T09:33:40.8316655Z * [new branch] gh/williamwen42/311/head -> origin/gh/williamwen42/311/head 2025-12-04T09:33:40.8317921Z * [new branch] gh/williamwen42/311/orig -> origin/gh/williamwen42/311/orig 2025-12-04T09:33:40.8319504Z * [new branch] gh/williamwen42/319/base -> origin/gh/williamwen42/319/base 2025-12-04T09:33:40.8320805Z * [new branch] gh/williamwen42/319/head -> origin/gh/williamwen42/319/head 2025-12-04T09:33:40.8322088Z * [new branch] gh/williamwen42/319/orig -> origin/gh/williamwen42/319/orig 2025-12-04T09:33:40.8323812Z * [new branch] gh/williamwen42/325/base -> origin/gh/williamwen42/325/base 2025-12-04T09:33:40.8325250Z * [new branch] gh/williamwen42/325/head -> origin/gh/williamwen42/325/head 2025-12-04T09:33:40.8326577Z * [new branch] gh/williamwen42/325/orig -> origin/gh/williamwen42/325/orig 2025-12-04T09:33:40.8328401Z * [new branch] gh/williamwen42/326/base -> origin/gh/williamwen42/326/base 2025-12-04T09:33:40.8329763Z * [new branch] gh/williamwen42/326/head -> origin/gh/williamwen42/326/head 2025-12-04T09:33:40.8331047Z * [new branch] gh/williamwen42/326/orig -> origin/gh/williamwen42/326/orig 2025-12-04T09:33:40.8332797Z * [new branch] gh/williamwen42/327/base -> origin/gh/williamwen42/327/base 2025-12-04T09:33:40.8334067Z * [new branch] gh/williamwen42/327/head -> origin/gh/williamwen42/327/head 2025-12-04T09:33:40.8335337Z * [new branch] gh/williamwen42/327/orig -> origin/gh/williamwen42/327/orig 2025-12-04T09:33:40.8337215Z * [new branch] gh/williamwen42/328/base -> origin/gh/williamwen42/328/base 2025-12-04T09:33:40.8338701Z * [new branch] gh/williamwen42/328/head -> origin/gh/williamwen42/328/head 2025-12-04T09:33:40.8339877Z * [new branch] gh/williamwen42/328/orig -> origin/gh/williamwen42/328/orig 2025-12-04T09:33:40.8342047Z * [new branch] gh/williamwen42/329/base -> origin/gh/williamwen42/329/base 2025-12-04T09:33:40.8343513Z * [new branch] gh/williamwen42/329/head -> origin/gh/williamwen42/329/head 2025-12-04T09:33:40.8344827Z * [new branch] gh/williamwen42/329/orig -> origin/gh/williamwen42/329/orig 2025-12-04T09:33:40.8347123Z * [new branch] gh/williamwen42/330/base -> origin/gh/williamwen42/330/base 2025-12-04T09:33:40.8348520Z * [new branch] gh/williamwen42/330/head -> origin/gh/williamwen42/330/head 2025-12-04T09:33:40.8349780Z * [new branch] gh/williamwen42/330/orig -> origin/gh/williamwen42/330/orig 2025-12-04T09:33:40.8351464Z * [new branch] gh/williamwen42/331/base -> origin/gh/williamwen42/331/base 2025-12-04T09:33:40.8352724Z * [new branch] gh/williamwen42/331/head -> origin/gh/williamwen42/331/head 2025-12-04T09:33:40.8353996Z * [new branch] gh/williamwen42/331/orig -> origin/gh/williamwen42/331/orig 2025-12-04T09:33:40.8355577Z * [new branch] gh/williamwen42/332/base -> origin/gh/williamwen42/332/base 2025-12-04T09:33:40.8356837Z * [new branch] gh/williamwen42/332/head -> origin/gh/williamwen42/332/head 2025-12-04T09:33:40.8358100Z * [new branch] gh/williamwen42/332/orig -> origin/gh/williamwen42/332/orig 2025-12-04T09:33:40.8360110Z * [new branch] gh/williamwen42/333/base -> origin/gh/williamwen42/333/base 2025-12-04T09:33:40.8361484Z * [new branch] gh/williamwen42/333/head -> origin/gh/williamwen42/333/head 2025-12-04T09:33:40.8362767Z * [new branch] gh/williamwen42/333/orig -> origin/gh/williamwen42/333/orig 2025-12-04T09:33:40.8364535Z * [new branch] gh/williamwen42/334/base -> origin/gh/williamwen42/334/base 2025-12-04T09:33:40.8365777Z * [new branch] gh/williamwen42/334/head -> origin/gh/williamwen42/334/head 2025-12-04T09:33:40.8367067Z * [new branch] gh/williamwen42/334/orig -> origin/gh/williamwen42/334/orig 2025-12-04T09:33:40.8372993Z * [new branch] gh/williamwen42/335/base -> origin/gh/williamwen42/335/base 2025-12-04T09:33:40.8374349Z * [new branch] gh/williamwen42/335/head -> origin/gh/williamwen42/335/head 2025-12-04T09:33:40.8375723Z * [new branch] gh/williamwen42/335/orig -> origin/gh/williamwen42/335/orig 2025-12-04T09:33:40.8377669Z * [new branch] gh/williamwen42/336/base -> origin/gh/williamwen42/336/base 2025-12-04T09:33:40.8378907Z * [new branch] gh/williamwen42/336/head -> origin/gh/williamwen42/336/head 2025-12-04T09:33:40.8380116Z * [new branch] gh/williamwen42/336/orig -> origin/gh/williamwen42/336/orig 2025-12-04T09:33:40.8381873Z * [new branch] gh/williamwen42/337/base -> origin/gh/williamwen42/337/base 2025-12-04T09:33:40.8383283Z * [new branch] gh/williamwen42/337/head -> origin/gh/williamwen42/337/head 2025-12-04T09:33:40.8384582Z * [new branch] gh/williamwen42/337/orig -> origin/gh/williamwen42/337/orig 2025-12-04T09:33:40.8386400Z * [new branch] gh/williamwen42/338/base -> origin/gh/williamwen42/338/base 2025-12-04T09:33:40.8387687Z * [new branch] gh/williamwen42/338/head -> origin/gh/williamwen42/338/head 2025-12-04T09:33:40.8388974Z * [new branch] gh/williamwen42/338/orig -> origin/gh/williamwen42/338/orig 2025-12-04T09:33:40.8390732Z * [new branch] gh/williamwen42/339/base -> origin/gh/williamwen42/339/base 2025-12-04T09:33:40.8391974Z * [new branch] gh/williamwen42/339/head -> origin/gh/williamwen42/339/head 2025-12-04T09:33:40.8393159Z * [new branch] gh/williamwen42/339/orig -> origin/gh/williamwen42/339/orig 2025-12-04T09:33:40.8394962Z * [new branch] gh/williamwen42/340/base -> origin/gh/williamwen42/340/base 2025-12-04T09:33:40.8396324Z * [new branch] gh/williamwen42/340/head -> origin/gh/williamwen42/340/head 2025-12-04T09:33:40.8400964Z * [new branch] gh/williamwen42/340/orig -> origin/gh/williamwen42/340/orig 2025-12-04T09:33:40.8403069Z * [new branch] gh/williamwen42/341/base -> origin/gh/williamwen42/341/base 2025-12-04T09:33:40.8404506Z * [new branch] gh/williamwen42/341/head -> origin/gh/williamwen42/341/head 2025-12-04T09:33:40.8405825Z * [new branch] gh/williamwen42/341/orig -> origin/gh/williamwen42/341/orig 2025-12-04T09:33:40.8407543Z * [new branch] gh/williamwen42/342/base -> origin/gh/williamwen42/342/base 2025-12-04T09:33:40.8408834Z * [new branch] gh/williamwen42/342/head -> origin/gh/williamwen42/342/head 2025-12-04T09:33:40.8410121Z * [new branch] gh/williamwen42/342/orig -> origin/gh/williamwen42/342/orig 2025-12-04T09:33:40.8411955Z * [new branch] gh/williamwen42/343/base -> origin/gh/williamwen42/343/base 2025-12-04T09:33:40.8413268Z * [new branch] gh/williamwen42/343/head -> origin/gh/williamwen42/343/head 2025-12-04T09:33:40.8414533Z * [new branch] gh/williamwen42/343/orig -> origin/gh/williamwen42/343/orig 2025-12-04T09:33:40.8416321Z * [new branch] gh/williamwen42/344/base -> origin/gh/williamwen42/344/base 2025-12-04T09:33:40.8417696Z * [new branch] gh/williamwen42/344/head -> origin/gh/williamwen42/344/head 2025-12-04T09:33:40.8418992Z * [new branch] gh/williamwen42/344/orig -> origin/gh/williamwen42/344/orig 2025-12-04T09:33:40.8420835Z * [new branch] gh/williamwen42/345/base -> origin/gh/williamwen42/345/base 2025-12-04T09:33:40.8422239Z * [new branch] gh/williamwen42/345/head -> origin/gh/williamwen42/345/head 2025-12-04T09:33:40.8423520Z * [new branch] gh/williamwen42/345/orig -> origin/gh/williamwen42/345/orig 2025-12-04T09:33:40.8425346Z * [new branch] gh/williamwen42/346/base -> origin/gh/williamwen42/346/base 2025-12-04T09:33:40.8426704Z * [new branch] gh/williamwen42/346/head -> origin/gh/williamwen42/346/head 2025-12-04T09:33:40.8428046Z * [new branch] gh/williamwen42/346/orig -> origin/gh/williamwen42/346/orig 2025-12-04T09:33:40.8429837Z * [new branch] gh/williamwen42/347/base -> origin/gh/williamwen42/347/base 2025-12-04T09:33:40.8431045Z * [new branch] gh/williamwen42/347/head -> origin/gh/williamwen42/347/head 2025-12-04T09:33:40.8432335Z * [new branch] gh/williamwen42/347/orig -> origin/gh/williamwen42/347/orig 2025-12-04T09:33:40.8433995Z * [new branch] gh/williamwen42/348/base -> origin/gh/williamwen42/348/base 2025-12-04T09:33:40.8435240Z * [new branch] gh/williamwen42/348/head -> origin/gh/williamwen42/348/head 2025-12-04T09:33:40.8455873Z * [new branch] gh/williamwen42/348/orig -> origin/gh/williamwen42/348/orig 2025-12-04T09:33:40.8456370Z * [new branch] gh/williamwen42/349/base -> origin/gh/williamwen42/349/base 2025-12-04T09:33:40.8456659Z * [new branch] gh/williamwen42/349/head -> origin/gh/williamwen42/349/head 2025-12-04T09:33:40.8457011Z * [new branch] gh/williamwen42/349/orig -> origin/gh/williamwen42/349/orig 2025-12-04T09:33:40.8457306Z * [new branch] gh/williamwen42/350/base -> origin/gh/williamwen42/350/base 2025-12-04T09:33:40.8457579Z * [new branch] gh/williamwen42/350/head -> origin/gh/williamwen42/350/head 2025-12-04T09:33:40.8458076Z * [new branch] gh/williamwen42/350/orig -> origin/gh/williamwen42/350/orig 2025-12-04T09:33:40.8458353Z * [new branch] gh/williamwen42/351/base -> origin/gh/williamwen42/351/base 2025-12-04T09:33:40.8458623Z * [new branch] gh/williamwen42/351/head -> origin/gh/williamwen42/351/head 2025-12-04T09:33:40.8458916Z * [new branch] gh/williamwen42/351/orig -> origin/gh/williamwen42/351/orig 2025-12-04T09:33:40.8459190Z * [new branch] gh/williamwen42/352/base -> origin/gh/williamwen42/352/base 2025-12-04T09:33:40.8459478Z * [new branch] gh/williamwen42/352/head -> origin/gh/williamwen42/352/head 2025-12-04T09:33:40.8459749Z * [new branch] gh/williamwen42/352/orig -> origin/gh/williamwen42/352/orig 2025-12-04T09:33:40.8460022Z * [new branch] gh/williamwen42/353/base -> origin/gh/williamwen42/353/base 2025-12-04T09:33:40.8460315Z * [new branch] gh/williamwen42/353/head -> origin/gh/williamwen42/353/head 2025-12-04T09:33:40.8460590Z * [new branch] gh/williamwen42/353/orig -> origin/gh/williamwen42/353/orig 2025-12-04T09:33:40.8461117Z * [new branch] gh/williamwen42/354/base -> origin/gh/williamwen42/354/base 2025-12-04T09:33:40.8462668Z * [new branch] gh/williamwen42/354/head -> origin/gh/williamwen42/354/head 2025-12-04T09:33:40.8463973Z * [new branch] gh/williamwen42/354/orig -> origin/gh/williamwen42/354/orig 2025-12-04T09:33:40.8465707Z * [new branch] gh/williamwen42/355/base -> origin/gh/williamwen42/355/base 2025-12-04T09:33:40.8466977Z * [new branch] gh/williamwen42/355/head -> origin/gh/williamwen42/355/head 2025-12-04T09:33:40.8468268Z * [new branch] gh/williamwen42/355/orig -> origin/gh/williamwen42/355/orig 2025-12-04T09:33:40.8470146Z * [new branch] gh/williamwen42/356/base -> origin/gh/williamwen42/356/base 2025-12-04T09:33:40.8471860Z * [new branch] gh/williamwen42/356/head -> origin/gh/williamwen42/356/head 2025-12-04T09:33:40.8473133Z * [new branch] gh/williamwen42/356/orig -> origin/gh/williamwen42/356/orig 2025-12-04T09:33:40.8474892Z * [new branch] gh/williamwen42/357/base -> origin/gh/williamwen42/357/base 2025-12-04T09:33:40.8476338Z * [new branch] gh/williamwen42/357/head -> origin/gh/williamwen42/357/head 2025-12-04T09:33:40.8477563Z * [new branch] gh/williamwen42/357/orig -> origin/gh/williamwen42/357/orig 2025-12-04T09:33:40.8479311Z * [new branch] gh/williamwen42/358/base -> origin/gh/williamwen42/358/base 2025-12-04T09:33:40.8480534Z * [new branch] gh/williamwen42/358/head -> origin/gh/williamwen42/358/head 2025-12-04T09:33:40.8481914Z * [new branch] gh/williamwen42/358/orig -> origin/gh/williamwen42/358/orig 2025-12-04T09:33:40.8483832Z * [new branch] gh/xmfan/169/base -> origin/gh/xmfan/169/base 2025-12-04T09:33:40.8485126Z * [new branch] gh/xmfan/169/head -> origin/gh/xmfan/169/head 2025-12-04T09:33:40.8486720Z * [new branch] gh/xmfan/170/base -> origin/gh/xmfan/170/base 2025-12-04T09:33:40.8487893Z * [new branch] gh/xmfan/170/head -> origin/gh/xmfan/170/head 2025-12-04T09:33:40.8489539Z * [new branch] gh/xmfan/274/base -> origin/gh/xmfan/274/base 2025-12-04T09:33:40.8490758Z * [new branch] gh/xmfan/274/head -> origin/gh/xmfan/274/head 2025-12-04T09:33:40.8492157Z * [new branch] gh/xmfan/274/orig -> origin/gh/xmfan/274/orig 2025-12-04T09:33:40.8493825Z * [new branch] gh/xmfan/277/base -> origin/gh/xmfan/277/base 2025-12-04T09:33:40.8495052Z * [new branch] gh/xmfan/277/head -> origin/gh/xmfan/277/head 2025-12-04T09:33:40.8496461Z * [new branch] gh/xmfan/277/orig -> origin/gh/xmfan/277/orig 2025-12-04T09:33:40.8498330Z * [new branch] gh/xmfan/301/base -> origin/gh/xmfan/301/base 2025-12-04T09:33:40.8499488Z * [new branch] gh/xmfan/301/head -> origin/gh/xmfan/301/head 2025-12-04T09:33:40.8500740Z * [new branch] gh/xmfan/301/orig -> origin/gh/xmfan/301/orig 2025-12-04T09:33:40.8502382Z * [new branch] gh/xmfan/304/base -> origin/gh/xmfan/304/base 2025-12-04T09:33:40.8503616Z * [new branch] gh/xmfan/304/head -> origin/gh/xmfan/304/head 2025-12-04T09:33:40.8504853Z * [new branch] gh/xmfan/304/orig -> origin/gh/xmfan/304/orig 2025-12-04T09:33:40.8506500Z * [new branch] gh/xmfan/309/base -> origin/gh/xmfan/309/base 2025-12-04T09:33:40.8507708Z * [new branch] gh/xmfan/309/head -> origin/gh/xmfan/309/head 2025-12-04T09:33:40.8509078Z * [new branch] gh/xmfan/309/orig -> origin/gh/xmfan/309/orig 2025-12-04T09:33:40.8510714Z * [new branch] gh/xmfan/310/base -> origin/gh/xmfan/310/base 2025-12-04T09:33:40.8511940Z * [new branch] gh/xmfan/310/head -> origin/gh/xmfan/310/head 2025-12-04T09:33:40.8513253Z * [new branch] gh/xmfan/310/orig -> origin/gh/xmfan/310/orig 2025-12-04T09:33:40.8514911Z * [new branch] gh/xmfan/311/base -> origin/gh/xmfan/311/base 2025-12-04T09:33:40.8516157Z * [new branch] gh/xmfan/311/head -> origin/gh/xmfan/311/head 2025-12-04T09:33:40.8517392Z * [new branch] gh/xmfan/311/orig -> origin/gh/xmfan/311/orig 2025-12-04T09:33:40.8519056Z * [new branch] gh/xmfan/312/base -> origin/gh/xmfan/312/base 2025-12-04T09:33:40.8520323Z * [new branch] gh/xmfan/312/head -> origin/gh/xmfan/312/head 2025-12-04T09:33:40.8521575Z * [new branch] gh/xmfan/312/orig -> origin/gh/xmfan/312/orig 2025-12-04T09:33:40.8523228Z * [new branch] gh/xmfan/313/base -> origin/gh/xmfan/313/base 2025-12-04T09:33:40.8524430Z * [new branch] gh/xmfan/313/head -> origin/gh/xmfan/313/head 2025-12-04T09:33:40.8526279Z * [new branch] gh/xmfan/313/orig -> origin/gh/xmfan/313/orig 2025-12-04T09:33:40.8528313Z * [new branch] gh/xuanzhang816/27/base -> origin/gh/xuanzhang816/27/base 2025-12-04T09:33:40.8529597Z * [new branch] gh/xuanzhang816/27/head -> origin/gh/xuanzhang816/27/head 2025-12-04T09:33:40.8530864Z * [new branch] gh/xuanzhang816/27/orig -> origin/gh/xuanzhang816/27/orig 2025-12-04T09:33:40.8532676Z * [new branch] gh/xuanzhang816/32/base -> origin/gh/xuanzhang816/32/base 2025-12-04T09:33:40.8533910Z * [new branch] gh/xuanzhang816/32/head -> origin/gh/xuanzhang816/32/head 2025-12-04T09:33:40.8535154Z * [new branch] gh/xuanzhang816/32/orig -> origin/gh/xuanzhang816/32/orig 2025-12-04T09:33:40.8537004Z * [new branch] gh/xuanzhang816/33/base -> origin/gh/xuanzhang816/33/base 2025-12-04T09:33:40.8538333Z * [new branch] gh/xuanzhang816/33/head -> origin/gh/xuanzhang816/33/head 2025-12-04T09:33:40.8539549Z * [new branch] gh/xuanzhang816/33/orig -> origin/gh/xuanzhang816/33/orig 2025-12-04T09:33:40.8541601Z * [new branch] gh/xuanzhang816/34/base -> origin/gh/xuanzhang816/34/base 2025-12-04T09:33:40.8542992Z * [new branch] gh/xuanzhang816/34/head -> origin/gh/xuanzhang816/34/head 2025-12-04T09:33:40.8544287Z * [new branch] gh/xuanzhang816/34/orig -> origin/gh/xuanzhang816/34/orig 2025-12-04T09:33:40.8546188Z * [new branch] gh/xuanzhang816/35/base -> origin/gh/xuanzhang816/35/base 2025-12-04T09:33:40.8547464Z * [new branch] gh/xuanzhang816/35/head -> origin/gh/xuanzhang816/35/head 2025-12-04T09:33:40.8549262Z * [new branch] gh/xuanzhang816/35/orig -> origin/gh/xuanzhang816/35/orig 2025-12-04T09:33:40.8551665Z * [new branch] gh/yanbing-j/11/base -> origin/gh/yanbing-j/11/base 2025-12-04T09:33:40.8553003Z * [new branch] gh/yanbing-j/11/head -> origin/gh/yanbing-j/11/head 2025-12-04T09:33:40.8554343Z * [new branch] gh/yanbing-j/11/orig -> origin/gh/yanbing-j/11/orig 2025-12-04T09:33:40.8556031Z * [new branch] gh/yanbing-j/12/base -> origin/gh/yanbing-j/12/base 2025-12-04T09:33:40.8557268Z * [new branch] gh/yanbing-j/12/head -> origin/gh/yanbing-j/12/head 2025-12-04T09:33:40.8558516Z * [new branch] gh/yanbing-j/12/orig -> origin/gh/yanbing-j/12/orig 2025-12-04T09:33:40.8560305Z * [new branch] gh/yanbing-j/13/base -> origin/gh/yanbing-j/13/base 2025-12-04T09:33:40.8561538Z * [new branch] gh/yanbing-j/13/head -> origin/gh/yanbing-j/13/head 2025-12-04T09:33:40.8562880Z * [new branch] gh/yanbing-j/13/orig -> origin/gh/yanbing-j/13/orig 2025-12-04T09:33:40.8564546Z * [new branch] gh/yanbing-j/14/base -> origin/gh/yanbing-j/14/base 2025-12-04T09:33:40.8565807Z * [new branch] gh/yanbing-j/14/head -> origin/gh/yanbing-j/14/head 2025-12-04T09:33:40.8567037Z * [new branch] gh/yanbing-j/14/orig -> origin/gh/yanbing-j/14/orig 2025-12-04T09:33:40.8568625Z * [new branch] gh/yanbing-j/15/base -> origin/gh/yanbing-j/15/base 2025-12-04T09:33:40.8569925Z * [new branch] gh/yanbing-j/15/head -> origin/gh/yanbing-j/15/head 2025-12-04T09:33:40.8571185Z * [new branch] gh/yanbing-j/15/orig -> origin/gh/yanbing-j/15/orig 2025-12-04T09:33:40.8572786Z * [new branch] gh/yanbing-j/18/base -> origin/gh/yanbing-j/18/base 2025-12-04T09:33:40.8574455Z * [new branch] gh/yanbing-j/18/head -> origin/gh/yanbing-j/18/head 2025-12-04T09:33:40.8575721Z * [new branch] gh/yanbing-j/18/orig -> origin/gh/yanbing-j/18/orig 2025-12-04T09:33:40.8577553Z * [new branch] gh/yanbing-j/19/base -> origin/gh/yanbing-j/19/base 2025-12-04T09:33:40.8578847Z * [new branch] gh/yanbing-j/19/head -> origin/gh/yanbing-j/19/head 2025-12-04T09:33:40.8580081Z * [new branch] gh/yanbing-j/19/orig -> origin/gh/yanbing-j/19/orig 2025-12-04T09:33:40.8581731Z * [new branch] gh/yanbing-j/20/base -> origin/gh/yanbing-j/20/base 2025-12-04T09:33:40.8583175Z * [new branch] gh/yanbing-j/20/head -> origin/gh/yanbing-j/20/head 2025-12-04T09:33:40.8584480Z * [new branch] gh/yanbing-j/20/orig -> origin/gh/yanbing-j/20/orig 2025-12-04T09:33:40.8586213Z * [new branch] gh/yanbing-j/21/base -> origin/gh/yanbing-j/21/base 2025-12-04T09:33:40.8587492Z * [new branch] gh/yanbing-j/21/head -> origin/gh/yanbing-j/21/head 2025-12-04T09:33:40.8589210Z * [new branch] gh/yanbing-j/22/base -> origin/gh/yanbing-j/22/base 2025-12-04T09:33:40.8590509Z * [new branch] gh/yanbing-j/22/head -> origin/gh/yanbing-j/22/head 2025-12-04T09:33:40.8591778Z * [new branch] gh/yanbing-j/22/orig -> origin/gh/yanbing-j/22/orig 2025-12-04T09:33:40.8593732Z * [new branch] gh/yanbing-j/23/base -> origin/gh/yanbing-j/23/base 2025-12-04T09:33:40.8595154Z * [new branch] gh/yanbing-j/23/head -> origin/gh/yanbing-j/23/head 2025-12-04T09:33:40.8596664Z * [new branch] gh/yanbing-j/23/orig -> origin/gh/yanbing-j/23/orig 2025-12-04T09:33:40.8598331Z * [new branch] gh/yanbing-j/24/base -> origin/gh/yanbing-j/24/base 2025-12-04T09:33:40.8599597Z * [new branch] gh/yanbing-j/24/head -> origin/gh/yanbing-j/24/head 2025-12-04T09:33:40.8601415Z * [new branch] gh/yanbing-j/24/orig -> origin/gh/yanbing-j/24/orig 2025-12-04T09:33:40.8603024Z * [new branch] gh/yanbing-j/25/base -> origin/gh/yanbing-j/25/base 2025-12-04T09:33:40.8604271Z * [new branch] gh/yanbing-j/25/head -> origin/gh/yanbing-j/25/head 2025-12-04T09:33:40.8606034Z * [new branch] gh/yanbing-j/25/orig -> origin/gh/yanbing-j/25/orig 2025-12-04T09:33:40.8607735Z * [new branch] gh/yanbing-j/26/base -> origin/gh/yanbing-j/26/base 2025-12-04T09:33:40.8609056Z * [new branch] gh/yanbing-j/26/head -> origin/gh/yanbing-j/26/head 2025-12-04T09:33:40.8610303Z * [new branch] gh/yanbing-j/26/orig -> origin/gh/yanbing-j/26/orig 2025-12-04T09:33:40.8612551Z * [new branch] gh/yang-yu-hang/1/base -> origin/gh/yang-yu-hang/1/base 2025-12-04T09:33:40.8614058Z * [new branch] gh/yang-yu-hang/1/head -> origin/gh/yang-yu-hang/1/head 2025-12-04T09:33:40.8615552Z * [new branch] gh/yang-yu-hang/1/orig -> origin/gh/yang-yu-hang/1/orig 2025-12-04T09:33:40.8617243Z * [new branch] gh/yang-yu-hang/2/base -> origin/gh/yang-yu-hang/2/base 2025-12-04T09:33:40.8618833Z * [new branch] gh/yang-yu-hang/2/head -> origin/gh/yang-yu-hang/2/head 2025-12-04T09:33:40.8620410Z * [new branch] gh/yang-yu-hang/2/orig -> origin/gh/yang-yu-hang/2/orig 2025-12-04T09:33:40.8622152Z * [new branch] gh/yang-yu-hang/3/base -> origin/gh/yang-yu-hang/3/base 2025-12-04T09:33:40.8623490Z * [new branch] gh/yang-yu-hang/3/head -> origin/gh/yang-yu-hang/3/head 2025-12-04T09:33:40.8624833Z * [new branch] gh/yang-yu-hang/3/orig -> origin/gh/yang-yu-hang/3/orig 2025-12-04T09:33:40.8626862Z * [new branch] gh/yangw-dev/12/base -> origin/gh/yangw-dev/12/base 2025-12-04T09:33:40.8628097Z * [new branch] gh/yangw-dev/12/head -> origin/gh/yangw-dev/12/head 2025-12-04T09:33:40.8629422Z * [new branch] gh/yangw-dev/12/orig -> origin/gh/yangw-dev/12/orig 2025-12-04T09:33:40.8631228Z * [new branch] gh/yangw-dev/13/base -> origin/gh/yangw-dev/13/base 2025-12-04T09:33:40.8633032Z * [new branch] gh/yangw-dev/13/head -> origin/gh/yangw-dev/13/head 2025-12-04T09:33:40.8634322Z * [new branch] gh/yangw-dev/13/orig -> origin/gh/yangw-dev/13/orig 2025-12-04T09:33:40.8636414Z * [new branch] gh/yangw-dev/14/base -> origin/gh/yangw-dev/14/base 2025-12-04T09:33:40.8637720Z * [new branch] gh/yangw-dev/14/head -> origin/gh/yangw-dev/14/head 2025-12-04T09:33:40.8638975Z * [new branch] gh/yangw-dev/14/orig -> origin/gh/yangw-dev/14/orig 2025-12-04T09:33:40.8640722Z * [new branch] gh/yangw-dev/15/base -> origin/gh/yangw-dev/15/base 2025-12-04T09:33:40.8642060Z * [new branch] gh/yangw-dev/15/head -> origin/gh/yangw-dev/15/head 2025-12-04T09:33:40.8643309Z * [new branch] gh/yangw-dev/15/orig -> origin/gh/yangw-dev/15/orig 2025-12-04T09:33:40.8644987Z * [new branch] gh/yangw-dev/19/base -> origin/gh/yangw-dev/19/base 2025-12-04T09:33:40.8646252Z * [new branch] gh/yangw-dev/19/head -> origin/gh/yangw-dev/19/head 2025-12-04T09:33:40.8647544Z * [new branch] gh/yangw-dev/19/orig -> origin/gh/yangw-dev/19/orig 2025-12-04T09:33:40.8649360Z * [new branch] gh/yangw-dev/26/base -> origin/gh/yangw-dev/26/base 2025-12-04T09:33:40.8650655Z * [new branch] gh/yangw-dev/26/head -> origin/gh/yangw-dev/26/head 2025-12-04T09:33:40.8651911Z * [new branch] gh/yangw-dev/26/orig -> origin/gh/yangw-dev/26/orig 2025-12-04T09:33:40.8653551Z * [new branch] gh/yangw-dev/27/base -> origin/gh/yangw-dev/27/base 2025-12-04T09:33:40.8654909Z * [new branch] gh/yangw-dev/27/head -> origin/gh/yangw-dev/27/head 2025-12-04T09:33:40.8656065Z * [new branch] gh/yangw-dev/27/orig -> origin/gh/yangw-dev/27/orig 2025-12-04T09:33:40.8658307Z * [new branch] gh/ydwu4/292/base -> origin/gh/ydwu4/292/base 2025-12-04T09:33:40.8659928Z * [new branch] gh/ydwu4/292/head -> origin/gh/ydwu4/292/head 2025-12-04T09:33:40.8661267Z * [new branch] gh/ydwu4/292/orig -> origin/gh/ydwu4/292/orig 2025-12-04T09:33:40.8663496Z * [new branch] gh/ydwu4/294/base -> origin/gh/ydwu4/294/base 2025-12-04T09:33:40.8664810Z * [new branch] gh/ydwu4/294/head -> origin/gh/ydwu4/294/head 2025-12-04T09:33:40.8666181Z * [new branch] gh/ydwu4/294/orig -> origin/gh/ydwu4/294/orig 2025-12-04T09:33:40.8668120Z * [new branch] gh/ydwu4/295/base -> origin/gh/ydwu4/295/base 2025-12-04T09:33:40.8669464Z * [new branch] gh/ydwu4/295/head -> origin/gh/ydwu4/295/head 2025-12-04T09:33:40.8670711Z * [new branch] gh/ydwu4/295/orig -> origin/gh/ydwu4/295/orig 2025-12-04T09:33:40.8672369Z * [new branch] gh/ydwu4/296/base -> origin/gh/ydwu4/296/base 2025-12-04T09:33:40.8673559Z * [new branch] gh/ydwu4/296/head -> origin/gh/ydwu4/296/head 2025-12-04T09:33:40.8674815Z * [new branch] gh/ydwu4/296/orig -> origin/gh/ydwu4/296/orig 2025-12-04T09:33:40.8676634Z * [new branch] gh/ydwu4/306/base -> origin/gh/ydwu4/306/base 2025-12-04T09:33:40.8677942Z * [new branch] gh/ydwu4/306/head -> origin/gh/ydwu4/306/head 2025-12-04T09:33:40.8679284Z * [new branch] gh/ydwu4/306/orig -> origin/gh/ydwu4/306/orig 2025-12-04T09:33:40.8680960Z * [new branch] gh/ydwu4/312/base -> origin/gh/ydwu4/312/base 2025-12-04T09:33:40.8682244Z * [new branch] gh/ydwu4/312/head -> origin/gh/ydwu4/312/head 2025-12-04T09:33:40.8683700Z * [new branch] gh/ydwu4/312/orig -> origin/gh/ydwu4/312/orig 2025-12-04T09:33:40.8685347Z * [new branch] gh/ydwu4/322/base -> origin/gh/ydwu4/322/base 2025-12-04T09:33:40.8686664Z * [new branch] gh/ydwu4/322/head -> origin/gh/ydwu4/322/head 2025-12-04T09:33:40.8687835Z * [new branch] gh/ydwu4/322/orig -> origin/gh/ydwu4/322/orig 2025-12-04T09:33:40.8689514Z * [new branch] gh/ydwu4/327/base -> origin/gh/ydwu4/327/base 2025-12-04T09:33:40.8690793Z * [new branch] gh/ydwu4/327/head -> origin/gh/ydwu4/327/head 2025-12-04T09:33:40.8692089Z * [new branch] gh/ydwu4/327/orig -> origin/gh/ydwu4/327/orig 2025-12-04T09:33:40.8693907Z * [new branch] gh/ydwu4/328/base -> origin/gh/ydwu4/328/base 2025-12-04T09:33:40.8695150Z * [new branch] gh/ydwu4/328/head -> origin/gh/ydwu4/328/head 2025-12-04T09:33:40.8696459Z * [new branch] gh/ydwu4/328/orig -> origin/gh/ydwu4/328/orig 2025-12-04T09:33:40.8698200Z * [new branch] gh/ydwu4/329/base -> origin/gh/ydwu4/329/base 2025-12-04T09:33:40.8699426Z * [new branch] gh/ydwu4/329/head -> origin/gh/ydwu4/329/head 2025-12-04T09:33:40.8700758Z * [new branch] gh/ydwu4/329/orig -> origin/gh/ydwu4/329/orig 2025-12-04T09:33:40.8703032Z * [new branch] gh/ydwu4/330/base -> origin/gh/ydwu4/330/base 2025-12-04T09:33:40.8704247Z * [new branch] gh/ydwu4/330/head -> origin/gh/ydwu4/330/head 2025-12-04T09:33:40.8705549Z * [new branch] gh/ydwu4/330/orig -> origin/gh/ydwu4/330/orig 2025-12-04T09:33:40.8707189Z * [new branch] gh/ydwu4/331/base -> origin/gh/ydwu4/331/base 2025-12-04T09:33:40.8708531Z * [new branch] gh/ydwu4/331/head -> origin/gh/ydwu4/331/head 2025-12-04T09:33:40.8709708Z * [new branch] gh/ydwu4/331/orig -> origin/gh/ydwu4/331/orig 2025-12-04T09:33:40.8711222Z * [new branch] gh/ydwu4/332/base -> origin/gh/ydwu4/332/base 2025-12-04T09:33:40.8712441Z * [new branch] gh/ydwu4/332/head -> origin/gh/ydwu4/332/head 2025-12-04T09:33:40.8713710Z * [new branch] gh/ydwu4/332/orig -> origin/gh/ydwu4/332/orig 2025-12-04T09:33:40.8715218Z * [new branch] gh/ydwu4/333/base -> origin/gh/ydwu4/333/base 2025-12-04T09:33:40.8716510Z * [new branch] gh/ydwu4/333/head -> origin/gh/ydwu4/333/head 2025-12-04T09:33:40.8717826Z * [new branch] gh/ydwu4/333/orig -> origin/gh/ydwu4/333/orig 2025-12-04T09:33:40.8719344Z * [new branch] gh/ydwu4/334/base -> origin/gh/ydwu4/334/base 2025-12-04T09:33:40.8720722Z * [new branch] gh/ydwu4/334/head -> origin/gh/ydwu4/334/head 2025-12-04T09:33:40.8721972Z * [new branch] gh/ydwu4/334/orig -> origin/gh/ydwu4/334/orig 2025-12-04T09:33:40.8723485Z * [new branch] gh/ydwu4/335/base -> origin/gh/ydwu4/335/base 2025-12-04T09:33:40.8724728Z * [new branch] gh/ydwu4/335/head -> origin/gh/ydwu4/335/head 2025-12-04T09:33:40.8726040Z * [new branch] gh/ydwu4/335/orig -> origin/gh/ydwu4/335/orig 2025-12-04T09:33:40.8728113Z * [new branch] gh/ydwu4/337/base -> origin/gh/ydwu4/337/base 2025-12-04T09:33:40.8729372Z * [new branch] gh/ydwu4/337/head -> origin/gh/ydwu4/337/head 2025-12-04T09:33:40.8730614Z * [new branch] gh/ydwu4/337/orig -> origin/gh/ydwu4/337/orig 2025-12-04T09:33:40.8732364Z * [new branch] gh/ydwu4/339/base -> origin/gh/ydwu4/339/base 2025-12-04T09:33:40.8733693Z * [new branch] gh/ydwu4/339/head -> origin/gh/ydwu4/339/head 2025-12-04T09:33:40.8735062Z * [new branch] gh/ydwu4/339/orig -> origin/gh/ydwu4/339/orig 2025-12-04T09:33:40.8737296Z * [new branch] gh/yf225/133/base -> origin/gh/yf225/133/base 2025-12-04T09:33:40.8738581Z * [new branch] gh/yf225/133/head -> origin/gh/yf225/133/head 2025-12-04T09:33:40.8740276Z * [new branch] gh/yf225/93/base -> origin/gh/yf225/93/base 2025-12-04T09:33:40.8741555Z * [new branch] gh/yf225/93/head -> origin/gh/yf225/93/head 2025-12-04T09:33:40.8744202Z * [new branch] gh/yifuwang/152/base -> origin/gh/yifuwang/152/base 2025-12-04T09:33:40.8745901Z * [new branch] gh/yifuwang/152/head -> origin/gh/yifuwang/152/head 2025-12-04T09:33:40.8747209Z * [new branch] gh/yifuwang/152/orig -> origin/gh/yifuwang/152/orig 2025-12-04T09:33:40.8748929Z * [new branch] gh/yifuwang/195/base -> origin/gh/yifuwang/195/base 2025-12-04T09:33:40.8750279Z * [new branch] gh/yifuwang/195/head -> origin/gh/yifuwang/195/head 2025-12-04T09:33:40.8751654Z * [new branch] gh/yifuwang/195/orig -> origin/gh/yifuwang/195/orig 2025-12-04T09:33:40.8754249Z * [new branch] gh/yiming0416/1/base -> origin/gh/yiming0416/1/base 2025-12-04T09:33:40.8755480Z * [new branch] gh/yiming0416/1/head -> origin/gh/yiming0416/1/head 2025-12-04T09:33:40.8757048Z * [new branch] gh/yiming0416/2/base -> origin/gh/yiming0416/2/base 2025-12-04T09:33:40.8758226Z * [new branch] gh/yiming0416/2/head -> origin/gh/yiming0416/2/head 2025-12-04T09:33:40.8760318Z * [new branch] gh/yushangdi/1/base -> origin/gh/yushangdi/1/base 2025-12-04T09:33:40.8761647Z * [new branch] gh/yushangdi/1/head -> origin/gh/yushangdi/1/head 2025-12-04T09:33:40.8763288Z * [new branch] gh/yushangdi/10/base -> origin/gh/yushangdi/10/base 2025-12-04T09:33:40.8764581Z * [new branch] gh/yushangdi/10/head -> origin/gh/yushangdi/10/head 2025-12-04T09:33:40.8765835Z * [new branch] gh/yushangdi/10/orig -> origin/gh/yushangdi/10/orig 2025-12-04T09:33:40.8767556Z * [new branch] gh/yushangdi/11/base -> origin/gh/yushangdi/11/base 2025-12-04T09:33:40.8768873Z * [new branch] gh/yushangdi/11/head -> origin/gh/yushangdi/11/head 2025-12-04T09:33:40.8770188Z * [new branch] gh/yushangdi/11/orig -> origin/gh/yushangdi/11/orig 2025-12-04T09:33:40.8771702Z * [new branch] gh/yushangdi/2/base -> origin/gh/yushangdi/2/base 2025-12-04T09:33:40.8772896Z * [new branch] gh/yushangdi/2/head -> origin/gh/yushangdi/2/head 2025-12-04T09:33:40.8774620Z * [new branch] gh/yushangdi/7/base -> origin/gh/yushangdi/7/base 2025-12-04T09:33:40.8775789Z * [new branch] gh/yushangdi/7/head -> origin/gh/yushangdi/7/head 2025-12-04T09:33:40.8777116Z * [new branch] gh/yushangdi/7/orig -> origin/gh/yushangdi/7/orig 2025-12-04T09:33:40.8779252Z * [new branch] gh/yushangdi/8/base -> origin/gh/yushangdi/8/base 2025-12-04T09:33:40.8781109Z * [new branch] gh/yushangdi/8/head -> origin/gh/yushangdi/8/head 2025-12-04T09:33:40.8782382Z * [new branch] gh/yushangdi/8/orig -> origin/gh/yushangdi/8/orig 2025-12-04T09:33:40.8784047Z * [new branch] gh/yushangdi/9/base -> origin/gh/yushangdi/9/base 2025-12-04T09:33:40.8785418Z * [new branch] gh/yushangdi/9/head -> origin/gh/yushangdi/9/head 2025-12-04T09:33:40.8786734Z * [new branch] gh/yushangdi/9/orig -> origin/gh/yushangdi/9/orig 2025-12-04T09:33:40.8788774Z * [new branch] gh/zklaus/19/base -> origin/gh/zklaus/19/base 2025-12-04T09:33:40.8790007Z * [new branch] gh/zklaus/19/head -> origin/gh/zklaus/19/head 2025-12-04T09:33:40.8791314Z * [new branch] gh/zklaus/19/orig -> origin/gh/zklaus/19/orig 2025-12-04T09:33:40.8793085Z * [new branch] gh/zklaus/20/base -> origin/gh/zklaus/20/base 2025-12-04T09:33:40.8794348Z * [new branch] gh/zklaus/20/head -> origin/gh/zklaus/20/head 2025-12-04T09:33:40.8795607Z * [new branch] gh/zklaus/20/orig -> origin/gh/zklaus/20/orig 2025-12-04T09:33:40.8797504Z * [new branch] gh/zklaus/21/base -> origin/gh/zklaus/21/base 2025-12-04T09:33:40.8798806Z * [new branch] gh/zklaus/21/head -> origin/gh/zklaus/21/head 2025-12-04T09:33:40.8800007Z * [new branch] gh/zklaus/21/orig -> origin/gh/zklaus/21/orig 2025-12-04T09:33:40.8801822Z * [new branch] gh/zklaus/22/base -> origin/gh/zklaus/22/base 2025-12-04T09:33:40.8803106Z * [new branch] gh/zklaus/22/head -> origin/gh/zklaus/22/head 2025-12-04T09:33:40.8804378Z * [new branch] gh/zklaus/22/orig -> origin/gh/zklaus/22/orig 2025-12-04T09:33:40.8806077Z * [new branch] gh/zklaus/23/base -> origin/gh/zklaus/23/base 2025-12-04T09:33:40.8807339Z * [new branch] gh/zklaus/23/head -> origin/gh/zklaus/23/head 2025-12-04T09:33:40.8808649Z * [new branch] gh/zklaus/23/orig -> origin/gh/zklaus/23/orig 2025-12-04T09:33:40.8810263Z * [new branch] gh/zklaus/24/base -> origin/gh/zklaus/24/base 2025-12-04T09:33:40.8811536Z * [new branch] gh/zklaus/24/head -> origin/gh/zklaus/24/head 2025-12-04T09:33:40.8812793Z * [new branch] gh/zklaus/24/orig -> origin/gh/zklaus/24/orig 2025-12-04T09:33:40.8815039Z * [new branch] gh/zou3519/1197/base -> origin/gh/zou3519/1197/base 2025-12-04T09:33:40.8816168Z * [new branch] gh/zou3519/1197/head -> origin/gh/zou3519/1197/head 2025-12-04T09:33:40.8817686Z * [new branch] gh/zou3519/1197/orig -> origin/gh/zou3519/1197/orig 2025-12-04T09:33:40.8819781Z * [new branch] gh/zou3519/1199/base -> origin/gh/zou3519/1199/base 2025-12-04T09:33:40.8821134Z * [new branch] gh/zou3519/1199/head -> origin/gh/zou3519/1199/head 2025-12-04T09:33:40.8822424Z * [new branch] gh/zou3519/1199/orig -> origin/gh/zou3519/1199/orig 2025-12-04T09:33:40.8824350Z * [new branch] gh/zou3519/1200/base -> origin/gh/zou3519/1200/base 2025-12-04T09:33:40.8825668Z * [new branch] gh/zou3519/1200/head -> origin/gh/zou3519/1200/head 2025-12-04T09:33:40.8826944Z * [new branch] gh/zou3519/1200/orig -> origin/gh/zou3519/1200/orig 2025-12-04T09:33:40.8828674Z * [new branch] gh/zou3519/1201/base -> origin/gh/zou3519/1201/base 2025-12-04T09:33:40.8829878Z * [new branch] gh/zou3519/1201/head -> origin/gh/zou3519/1201/head 2025-12-04T09:33:40.8831158Z * [new branch] gh/zou3519/1201/orig -> origin/gh/zou3519/1201/orig 2025-12-04T09:33:40.8832697Z * [new branch] gh/zou3519/1202/base -> origin/gh/zou3519/1202/base 2025-12-04T09:33:40.8833975Z * [new branch] gh/zou3519/1202/head -> origin/gh/zou3519/1202/head 2025-12-04T09:33:40.8835599Z * [new branch] gh/zou3519/1202/orig -> origin/gh/zou3519/1202/orig 2025-12-04T09:33:40.8837706Z * [new branch] gh/zpcore/1/base -> origin/gh/zpcore/1/base 2025-12-04T09:33:40.8838962Z * [new branch] gh/zpcore/1/head -> origin/gh/zpcore/1/head 2025-12-04T09:33:40.8840720Z * [new branch] gh/zpcore/11/base -> origin/gh/zpcore/11/base 2025-12-04T09:33:40.8842084Z * [new branch] gh/zpcore/11/head -> origin/gh/zpcore/11/head 2025-12-04T09:33:40.8843357Z * [new branch] gh/zpcore/11/orig -> origin/gh/zpcore/11/orig 2025-12-04T09:33:40.8845519Z * [new branch] gh/zpcore/12/base -> origin/gh/zpcore/12/base 2025-12-04T09:33:40.8846807Z * [new branch] gh/zpcore/12/head -> origin/gh/zpcore/12/head 2025-12-04T09:33:40.8848138Z * [new branch] gh/zpcore/12/orig -> origin/gh/zpcore/12/orig 2025-12-04T09:33:40.8849934Z * [new branch] gh/zpcore/13/base -> origin/gh/zpcore/13/base 2025-12-04T09:33:40.8851189Z * [new branch] gh/zpcore/13/head -> origin/gh/zpcore/13/head 2025-12-04T09:33:40.8852570Z * [new branch] gh/zpcore/13/orig -> origin/gh/zpcore/13/orig 2025-12-04T09:33:40.8854328Z * [new branch] gh/zpcore/14/base -> origin/gh/zpcore/14/base 2025-12-04T09:33:40.8855632Z * [new branch] gh/zpcore/14/head -> origin/gh/zpcore/14/head 2025-12-04T09:33:40.8856856Z * [new branch] gh/zpcore/14/orig -> origin/gh/zpcore/14/orig 2025-12-04T09:33:40.8858954Z * [new branch] gh/zpcore/15/base -> origin/gh/zpcore/15/base 2025-12-04T09:33:40.8860165Z * [new branch] gh/zpcore/15/head -> origin/gh/zpcore/15/head 2025-12-04T09:33:40.8861436Z * [new branch] gh/zpcore/15/orig -> origin/gh/zpcore/15/orig 2025-12-04T09:33:40.8863134Z * [new branch] gh/zpcore/2/base -> origin/gh/zpcore/2/base 2025-12-04T09:33:40.8864450Z * [new branch] gh/zpcore/2/head -> origin/gh/zpcore/2/head 2025-12-04T09:33:40.8866726Z * [new branch] gh/zpcore/21/base -> origin/gh/zpcore/21/base 2025-12-04T09:33:40.8868189Z * [new branch] gh/zpcore/21/head -> origin/gh/zpcore/21/head 2025-12-04T09:33:40.8869481Z * [new branch] gh/zpcore/21/orig -> origin/gh/zpcore/21/orig 2025-12-04T09:33:40.8871440Z * [new branch] gh/zpcore/22/base -> origin/gh/zpcore/22/base 2025-12-04T09:33:40.8872717Z * [new branch] gh/zpcore/22/head -> origin/gh/zpcore/22/head 2025-12-04T09:33:40.8874019Z * [new branch] gh/zpcore/22/orig -> origin/gh/zpcore/22/orig 2025-12-04T09:33:40.8875783Z * [new branch] gh/zpcore/23/base -> origin/gh/zpcore/23/base 2025-12-04T09:33:40.8877064Z * [new branch] gh/zpcore/23/head -> origin/gh/zpcore/23/head 2025-12-04T09:33:40.8878371Z * [new branch] gh/zpcore/23/orig -> origin/gh/zpcore/23/orig 2025-12-04T09:33:40.8879883Z * [new branch] gh/zpcore/24/base -> origin/gh/zpcore/24/base 2025-12-04T09:33:40.8881175Z * [new branch] gh/zpcore/24/head -> origin/gh/zpcore/24/head 2025-12-04T09:33:40.8882409Z * [new branch] gh/zpcore/24/orig -> origin/gh/zpcore/24/orig 2025-12-04T09:33:40.8884372Z * [new branch] gh/zpcore/25/base -> origin/gh/zpcore/25/base 2025-12-04T09:33:40.8885656Z * [new branch] gh/zpcore/25/head -> origin/gh/zpcore/25/head 2025-12-04T09:33:40.8887024Z * [new branch] gh/zpcore/25/orig -> origin/gh/zpcore/25/orig 2025-12-04T09:33:40.8888981Z * [new branch] gh/zpcore/26/base -> origin/gh/zpcore/26/base 2025-12-04T09:33:40.8890240Z * [new branch] gh/zpcore/26/head -> origin/gh/zpcore/26/head 2025-12-04T09:33:40.8891464Z * [new branch] gh/zpcore/26/orig -> origin/gh/zpcore/26/orig 2025-12-04T09:33:40.8893278Z * [new branch] gh/zpcore/27/base -> origin/gh/zpcore/27/base 2025-12-04T09:33:40.8894599Z * [new branch] gh/zpcore/27/head -> origin/gh/zpcore/27/head 2025-12-04T09:33:40.8895839Z * [new branch] gh/zpcore/27/orig -> origin/gh/zpcore/27/orig 2025-12-04T09:33:40.8901290Z * [new branch] gh/zpcore/28/base -> origin/gh/zpcore/28/base 2025-12-04T09:33:40.8903075Z * [new branch] gh/zpcore/28/head -> origin/gh/zpcore/28/head 2025-12-04T09:33:40.8904330Z * [new branch] gh/zpcore/28/orig -> origin/gh/zpcore/28/orig 2025-12-04T09:33:40.8906400Z * [new branch] gh/zpcore/3/base -> origin/gh/zpcore/3/base 2025-12-04T09:33:40.8907609Z * [new branch] gh/zpcore/3/head -> origin/gh/zpcore/3/head 2025-12-04T09:33:40.8909279Z * [new branch] gh/zpcore/4/base -> origin/gh/zpcore/4/base 2025-12-04T09:33:40.8910502Z * [new branch] gh/zpcore/4/head -> origin/gh/zpcore/4/head 2025-12-04T09:33:40.8912041Z * [new branch] gh/zpcore/5/base -> origin/gh/zpcore/5/base 2025-12-04T09:33:40.8913220Z * [new branch] gh/zpcore/5/head -> origin/gh/zpcore/5/head 2025-12-04T09:33:40.8915230Z * [new branch] gh/zpcore/6/base -> origin/gh/zpcore/6/base 2025-12-04T09:33:40.8916449Z * [new branch] gh/zpcore/6/head -> origin/gh/zpcore/6/head 2025-12-04T09:33:40.8918438Z * [new branch] gh/zpcore/7/base -> origin/gh/zpcore/7/base 2025-12-04T09:33:40.8919642Z * [new branch] gh/zpcore/7/head -> origin/gh/zpcore/7/head 2025-12-04T09:33:40.8921233Z * [new branch] gh/zpcore/8/base -> origin/gh/zpcore/8/base 2025-12-04T09:33:40.8922525Z * [new branch] gh/zpcore/8/head -> origin/gh/zpcore/8/head 2025-12-04T09:33:40.8924008Z * [new branch] google-main -> origin/google-main 2025-12-04T09:33:40.8925991Z * [new branch] guangyey/external_stream -> origin/guangyey/external_stream 2025-12-04T09:33:40.8927099Z * [new branch] guangyey/test_2025 -> origin/guangyey/test_2025 2025-12-04T09:33:40.8929101Z * [new branch] guilhermeleobas/cherry-pick-55d87d9dfd9 -> origin/guilhermeleobas/cherry-pick-55d87d9dfd9 2025-12-04T09:33:40.8930728Z * [new branch] hameerabbasi/complex_tensor_subclass -> origin/hameerabbasi/complex_tensor_subclass 2025-12-04T09:33:40.8932168Z * [new branch] hameerabbasi/fix-ctensor-gradcheck-tests -> origin/hameerabbasi/fix-ctensor-gradcheck-tests 2025-12-04T09:33:40.8933887Z * [new branch] hameerabbasi/gradcheck-allclose -> origin/hameerabbasi/gradcheck-allclose 2025-12-04T09:33:40.8935099Z * [new branch] hc_baseline -> origin/hc_baseline 2025-12-04T09:33:40.8936488Z * [new branch] hhh_rand -> origin/hhh_rand 2025-12-04T09:33:40.8938687Z * [new branch] huba/f1 -> origin/huba/f1 2025-12-04T09:33:40.8940717Z * [new branch] increase-timeout-linux-jammy-cuda12_8-py3_10-gcc11-test -> origin/increase-timeout-linux-jammy-cuda12_8-py3_10-gcc11-test 2025-12-04T09:33:40.8941491Z * [new branch] inlining -> origin/inlining 2025-12-04T09:33:40.8942945Z * [new branch] inlining-ezyang -> origin/inlining-ezyang 2025-12-04T09:33:40.8944282Z * [new branch] install-torchao-0.13.0 -> origin/install-torchao-0.13.0 2025-12-04T09:33:40.8946013Z * [new branch] instrument-trunk-pull-linux-with-job-test-filters -> origin/instrument-trunk-pull-linux-with-job-test-filters 2025-12-04T09:33:40.8946905Z * [new branch] invoke-subgraph -> origin/invoke-subgraph 2025-12-04T09:33:40.8948291Z * [new branch] issue#58739 -> origin/issue#58739 2025-12-04T09:33:40.8949835Z * [new branch] jainapurva-patch-1 -> origin/jainapurva-patch-1 2025-12-04T09:33:40.8951412Z * [new branch] jathu/o3 -> origin/jathu/o3 2025-12-04T09:33:40.8952622Z * [new branch] jathu/sve -> origin/jathu/sve 2025-12-04T09:33:40.8954604Z * [new branch] jcaip/test-cusparselt-version-0.6.2 -> origin/jcaip/test-cusparselt-version-0.6.2 2025-12-04T09:33:40.8955780Z * [new branch] jcaip/update-cusparselt-0.6.2 -> origin/jcaip/update-cusparselt-0.6.2 2025-12-04T09:33:40.8957504Z * [new branch] jiannanWang/memorysnapshot_filter -> origin/jiannanWang/memorysnapshot_filter 2025-12-04T09:33:40.8958758Z * [new branch] jiannanWang/profilerstepwarning -> origin/jiannanWang/profilerstepwarning 2025-12-04T09:33:40.8960119Z * [new branch] jithunnair-amd-patch-1 -> origin/jithunnair-amd-patch-1 2025-12-04T09:33:40.8961470Z * [new branch] jithunnair-amd-patch-10 -> origin/jithunnair-amd-patch-10 2025-12-04T09:33:40.8962823Z * [new branch] jithunnair-amd-patch-2 -> origin/jithunnair-amd-patch-2 2025-12-04T09:33:40.8964270Z * [new branch] jithunnair-amd-patch-3 -> origin/jithunnair-amd-patch-3 2025-12-04T09:33:40.8965650Z * [new branch] jithunnair-amd-patch-4 -> origin/jithunnair-amd-patch-4 2025-12-04T09:33:40.8966957Z * [new branch] jithunnair-amd-patch-5 -> origin/jithunnair-amd-patch-5 2025-12-04T09:33:40.8968308Z * [new branch] jithunnair-amd-patch-6 -> origin/jithunnair-amd-patch-6 2025-12-04T09:33:40.8969621Z * [new branch] jithunnair-amd-patch-7 -> origin/jithunnair-amd-patch-7 2025-12-04T09:33:40.8971505Z * [new branch] jithunnair-amd-patch-8 -> origin/jithunnair-amd-patch-8 2025-12-04T09:33:40.8972863Z * [new branch] jithunnair-amd-patch-9 -> origin/jithunnair-amd-patch-9 2025-12-04T09:33:40.8974658Z * [new branch] justinchu/native-qdq -> origin/justinchu/native-qdq 2025-12-04T09:33:40.8977004Z * [new branch] kainan666/xlf_debug -> origin/kainan666/xlf_debug 2025-12-04T09:33:40.8978341Z * [new branch] kainan_test -> origin/kainan_test 2025-12-04T09:33:40.8979665Z * [new branch] larryliu0820-patch-1 -> origin/larryliu0820-patch-1 2025-12-04T09:33:40.8981493Z * [new branch] leslie/test_group_gemm_epilogues -> origin/leslie/test_group_gemm_epilogues 2025-12-04T09:33:40.8983300Z * [new branch] lessw2020/fix_cutlass_cache_error -> origin/lessw2020/fix_cutlass_cache_error 2025-12-04T09:33:40.8984906Z * [new branch] liaoxuan/shm_all_reduce -> origin/liaoxuan/shm_all_reduce 2025-12-04T09:33:40.8986185Z * [new branch] liaoxuan/test_fa_disable_softmax -> origin/liaoxuan/test_fa_disable_softmax 2025-12-04T09:33:40.8987350Z * [new branch] liaoxuan/test_int8_sdpa -> origin/liaoxuan/test_int8_sdpa 2025-12-04T09:33:40.8988573Z * [new branch] llama4-stable -> origin/llama4-stable 2025-12-04T09:33:40.8990719Z * [new branch] lts/release/1.8 -> origin/lts/release/1.8 2025-12-04T09:33:40.8992444Z * [new branch] lucaskabela/#94773 -> origin/lucaskabela/#94773 2025-12-04T09:33:40.8993693Z * [new branch] lucaskabela/fix_164876 -> origin/lucaskabela/fix_164876 2025-12-04T09:33:40.8994912Z * [new branch] lucaskabela/flop_counter -> origin/lucaskabela/flop_counter 2025-12-04T09:33:40.8996322Z * [new branch] lucaskabela/func_under_decomp -> origin/lucaskabela/func_under_decomp 2025-12-04T09:33:40.8997631Z * [new branch] lucaskabela/functional_in_dynamo -> origin/lucaskabela/functional_in_dynamo 2025-12-04T09:33:40.8998945Z * [new branch] lucaskabela/install_params_as_graph_attr -> origin/lucaskabela/install_params_as_graph_attr 2025-12-04T09:33:40.9000548Z * [new branch] lucaskabela/parameters_as_graph_attr -> origin/lucaskabela/parameters_as_graph_attr 2025-12-04T09:33:40.9002312Z * [new branch] lucaskabela/remove_aot_dispatcher_metadata -> origin/lucaskabela/remove_aot_dispatcher_metadata 2025-12-04T09:33:40.9003411Z * [new branch] lucaskabela/rnn_decomp -> origin/lucaskabela/rnn_decomp 2025-12-04T09:33:40.9004720Z * [new branch] lucaskabela/typing_backends -> origin/lucaskabela/typing_backends 2025-12-04T09:33:40.9006003Z * [new branch] lucaskabela/typing_ctx_manager -> origin/lucaskabela/typing_ctx_manager 2025-12-04T09:33:40.9007273Z * [new branch] lucaskabela/typing_nn_module -> origin/lucaskabela/typing_nn_module 2025-12-04T09:33:40.9008607Z * [new branch] lucaskabela/typing_user_defined -> origin/lucaskabela/typing_user_defined 2025-12-04T09:33:40.9009826Z * [new branch] lucaskabela/typing_variables -> origin/lucaskabela/typing_variables 2025-12-04T09:33:40.9011158Z * [new branch] lucaskabela/typing_variables_dicts -> origin/lucaskabela/typing_variables_dicts 2025-12-04T09:33:40.9012817Z * [new branch] lucaskabela/typing_variables_functions -> origin/lucaskabela/typing_variables_functions 2025-12-04T09:33:40.9014047Z * [new branch] lucaskabela/typing_variables_lists -> origin/lucaskabela/typing_variables_lists 2025-12-04T09:33:40.9015754Z * [new branch] lw/torch_box_by_ref -> origin/lw/torch_box_by_ref 2025-12-04T09:33:40.9017155Z * [new branch] main -> origin/main 2025-12-04T09:33:40.9018714Z * [new branch] malfet-patch-1 -> origin/malfet-patch-1 2025-12-04T09:33:40.9020261Z * [new branch] malfet-patch-2 -> origin/malfet-patch-2 2025-12-04T09:33:40.9021777Z * [new branch] malfet-patch-3 -> origin/malfet-patch-3 2025-12-04T09:33:40.9023317Z * [new branch] malfet-patch-4 -> origin/malfet-patch-4 2025-12-04T09:33:40.9024618Z * [new branch] malfet-patch-5 -> origin/malfet-patch-5 2025-12-04T09:33:40.9025913Z * [new branch] malfet-patch-6 -> origin/malfet-patch-6 2025-12-04T09:33:40.9027298Z * [new branch] malfet-patch-7 -> origin/malfet-patch-7 2025-12-04T09:33:40.9028699Z * [new branch] malfet-patch-8 -> origin/malfet-patch-8 2025-12-04T09:33:40.9030457Z * [new branch] malfet/add-3.14-ci -> origin/malfet/add-3.14-ci 2025-12-04T09:33:40.9032020Z * [new branch] malfet/be-do-not-make-typos-in-build-artifacts -> origin/malfet/be-do-not-make-typos-in-build-artifacts 2025-12-04T09:33:40.9033215Z * [new branch] malfet/be-move-more-settings-to-checkout-pytorch -> origin/malfet/be-move-more-settings-to-checkout-pytorch 2025-12-04T09:33:40.9034692Z * [new branch] malfet/be-remove-misisng-neon-headers -> origin/malfet/be-remove-misisng-neon-headers 2025-12-04T09:33:40.9036081Z * [new branch] malfet/mps-implement-col2im -> origin/malfet/mps-implement-col2im 2025-12-04T09:33:40.9037917Z * [new branch] manuel/aoti_metal_shimify-thread_safe -> origin/manuel/aoti_metal_shimify-thread_safe 2025-12-04T09:33:40.9038946Z * [new branch] manuel/inductor_link_openmp -> origin/manuel/inductor_link_openmp 2025-12-04T09:33:40.9040780Z * [new branch] masnesral/metaconda -> origin/masnesral/metaconda 2025-12-04T09:33:40.9042175Z * [new branch] mem_profiler_flaky_fix -> origin/mem_profiler_flaky_fix 2025-12-04T09:33:40.9043478Z * [new branch] mem_profiler_stack_trace -> origin/mem_profiler_stack_trace 2025-12-04T09:33:40.9044787Z * [new branch] memory_profiler_stack -> origin/memory_profiler_stack 2025-12-04T09:33:40.9046230Z * [new branch] metascroy-patch-1 -> origin/metascroy-patch-1 2025-12-04T09:33:40.9047487Z * [new branch] mingw_posix -> origin/mingw_posix 2025-12-04T09:33:40.9049257Z * [new branch] mlazos/S429861-debug -> origin/mlazos/S429861-debug 2025-12-04T09:33:40.9050411Z * [new branch] mlazos/aa -> origin/mlazos/aa 2025-12-04T09:33:40.9051692Z * [new branch] mlazos/acts -> origin/mlazos/acts 2025-12-04T09:33:40.9052898Z * [new branch] mlazos/arg-renames -> origin/mlazos/arg-renames 2025-12-04T09:33:40.9054145Z * [new branch] mlazos/bad-cudagraphs -> origin/mlazos/bad-cudagraphs 2025-12-04T09:33:40.9055394Z * [new branch] mlazos/baseline-graph-breaks -> origin/mlazos/baseline-graph-breaks 2025-12-04T09:33:40.9056539Z * [new branch] mlazos/beta-tensor -> origin/mlazos/beta-tensor 2025-12-04T09:33:40.9057807Z * [new branch] mlazos/buffers -> origin/mlazos/buffers 2025-12-04T09:33:40.9058891Z * [new branch] mlazos/buffers2 -> origin/mlazos/buffers2 2025-12-04T09:33:40.9060574Z * [new branch] mlazos/buffers3 -> origin/mlazos/buffers3 2025-12-04T09:33:40.9062071Z * [new branch] mlazos/bwd -> origin/mlazos/bwd 2025-12-04T09:33:40.9063227Z * [new branch] mlazos/combo-test -> origin/mlazos/combo-test 2025-12-04T09:33:40.9064514Z * [new branch] mlazos/ctx-cleanup -> origin/mlazos/ctx-cleanup 2025-12-04T09:33:40.9065785Z * [new branch] mlazos/cuda-cmd-log -> origin/mlazos/cuda-cmd-log 2025-12-04T09:33:40.9067208Z * [new branch] mlazos/cudagraph-tests -> origin/mlazos/cudagraph-tests 2025-12-04T09:33:40.9068516Z * [new branch] mlazos/cudagraphs-measurement -> origin/mlazos/cudagraphs-measurement 2025-12-04T09:33:40.9069823Z * [new branch] mlazos/cutlass-test -> origin/mlazos/cutlass-test 2025-12-04T09:33:40.9071161Z * [new branch] mlazos/cutlass-topo-bug -> origin/mlazos/cutlass-topo-bug 2025-12-04T09:33:40.9072349Z * [new branch] mlazos/dataclass-proxy -> origin/mlazos/dataclass-proxy 2025-12-04T09:33:40.9073578Z * [new branch] mlazos/dc-attrs -> origin/mlazos/dc-attrs 2025-12-04T09:33:40.9075342Z * [new branch] mlazos/dc-helion -> origin/mlazos/dc-helion 2025-12-04T09:33:40.9076622Z * [new branch] mlazos/dict-fix -> origin/mlazos/dict-fix 2025-12-04T09:33:40.9077888Z * [new branch] mlazos/disable-tf -> origin/mlazos/disable-tf 2025-12-04T09:33:40.9079113Z * [new branch] mlazos/dupe-fix -> origin/mlazos/dupe-fix 2025-12-04T09:33:40.9080450Z * [new branch] mlazos/dyn-batch -> origin/mlazos/dyn-batch 2025-12-04T09:33:40.9081777Z * [new branch] mlazos/evt -> origin/mlazos/evt 2025-12-04T09:33:40.9083096Z * [new branch] mlazos/extract-examples -> origin/mlazos/extract-examples 2025-12-04T09:33:40.9084316Z * [new branch] mlazos/foreach-op -> origin/mlazos/foreach-op 2025-12-04T09:33:40.9085541Z * [new branch] mlazos/fp8 -> origin/mlazos/fp8 2025-12-04T09:33:40.9086799Z * [new branch] mlazos/fp8-bias -> origin/mlazos/fp8-bias 2025-12-04T09:33:40.9088192Z * [new branch] mlazos/fp8-bias-fusion -> origin/mlazos/fp8-bias-fusion 2025-12-04T09:33:40.9089425Z * [new branch] mlazos/fp8-fixes -> origin/mlazos/fp8-fixes 2025-12-04T09:33:40.9090671Z * [new branch] mlazos/freezing -> origin/mlazos/freezing 2025-12-04T09:33:40.9091936Z * [new branch] mlazos/h-comp -> origin/mlazos/h-comp 2025-12-04T09:33:40.9093254Z * [new branch] mlazos/h-comp2 -> origin/mlazos/h-comp2 2025-12-04T09:33:40.9094539Z * [new branch] mlazos/hash-hop -> origin/mlazos/hash-hop 2025-12-04T09:33:40.9095816Z * [new branch] mlazos/hc -> origin/mlazos/hc 2025-12-04T09:33:40.9097347Z * [new branch] mlazos/hc-cycles -> origin/mlazos/hc-cycles 2025-12-04T09:33:40.9098735Z * [new branch] mlazos/hc-fixes -> origin/mlazos/hc-fixes 2025-12-04T09:33:40.9099968Z * [new branch] mlazos/hc-fixes3 -> origin/mlazos/hc-fixes3 2025-12-04T09:33:40.9101209Z * [new branch] mlazos/hc-fixes4 -> origin/mlazos/hc-fixes4 2025-12-04T09:33:40.9102536Z * [new branch] mlazos/hc-hf -> origin/mlazos/hc-hf 2025-12-04T09:33:40.9103779Z * [new branch] mlazos/hc-mut -> origin/mlazos/hc-mut 2025-12-04T09:33:40.9105037Z * [new branch] mlazos/hc10 -> origin/mlazos/hc10 2025-12-04T09:33:40.9106299Z * [new branch] mlazos/hc11 -> origin/mlazos/hc11 2025-12-04T09:33:40.9108026Z * [new branch] mlazos/hc12 -> origin/mlazos/hc12 2025-12-04T09:33:40.9109264Z * [new branch] mlazos/hc13 -> origin/mlazos/hc13 2025-12-04T09:33:40.9110508Z * [new branch] mlazos/hc14 -> origin/mlazos/hc14 2025-12-04T09:33:40.9111771Z * [new branch] mlazos/hc15 -> origin/mlazos/hc15 2025-12-04T09:33:40.9113013Z * [new branch] mlazos/hc2 -> origin/mlazos/hc2 2025-12-04T09:33:40.9114258Z * [new branch] mlazos/hc4 -> origin/mlazos/hc4 2025-12-04T09:33:40.9115500Z * [new branch] mlazos/hc5 -> origin/mlazos/hc5 2025-12-04T09:33:40.9116792Z * [new branch] mlazos/hc6 -> origin/mlazos/hc6 2025-12-04T09:33:40.9118054Z * [new branch] mlazos/hc7 -> origin/mlazos/hc7 2025-12-04T09:33:40.9119309Z * [new branch] mlazos/hc8 -> origin/mlazos/hc8 2025-12-04T09:33:40.9120458Z * [new branch] mlazos/hc9 -> origin/mlazos/hc9 2025-12-04T09:33:40.9121755Z * [new branch] mlazos/hc_baseline2 -> origin/mlazos/hc_baseline2 2025-12-04T09:33:40.9123097Z * [new branch] mlazos/inductor-streams -> origin/mlazos/inductor-streams 2025-12-04T09:33:40.9124121Z * [new branch] mlazos/main -> origin/mlazos/main 2025-12-04T09:33:40.9125416Z * [new branch] mlazos/mcg2 -> origin/mlazos/mcg2 2025-12-04T09:33:40.9126686Z * [new branch] mlazos/meta-guards -> origin/mlazos/meta-guards 2025-12-04T09:33:40.9128541Z * [new branch] mlazos/mlazos/foreach-map-adam -> origin/mlazos/mlazos/foreach-map-adam 2025-12-04T09:33:40.9129771Z * [new branch] mlazos/mlazos/tf-mode-backup -> origin/mlazos/mlazos/tf-mode-backup 2025-12-04T09:33:40.9131013Z * [new branch] mlazos/mod-fix -> origin/mlazos/mod-fix 2025-12-04T09:33:40.9132387Z * [new branch] mlazos/mode-fix -> origin/mlazos/mode-fix 2025-12-04T09:33:40.9133714Z * [new branch] mlazos/offsets -> origin/mlazos/offsets 2025-12-04T09:33:40.9134742Z * [new branch] mlazos/overguarding -> origin/mlazos/overguarding 2025-12-04T09:33:40.9136057Z * [new branch] mlazos/proxy-ctors -> origin/mlazos/proxy-ctors 2025-12-04T09:33:40.9137421Z * [new branch] mlazos/quant-fix -> origin/mlazos/quant-fix 2025-12-04T09:33:40.9138720Z * [new branch] mlazos/resnet-fix -> origin/mlazos/resnet-fix 2025-12-04T09:33:40.9139999Z * [new branch] mlazos/rm-buf-names -> origin/mlazos/rm-buf-names 2025-12-04T09:33:40.9141257Z * [new branch] mlazos/rm-code -> origin/mlazos/rm-code 2025-12-04T09:33:40.9142622Z * [new branch] mlazos/rm-spam -> origin/mlazos/rm-spam 2025-12-04T09:33:40.9144006Z * [new branch] mlazos/rtp -> origin/mlazos/rtp 2025-12-04T09:33:40.9145267Z * [new branch] mlazos/static-idx-dbg -> origin/mlazos/static-idx-dbg 2025-12-04T09:33:40.9146556Z * [new branch] mlazos/static-inputs-log -> origin/mlazos/static-inputs-log 2025-12-04T09:33:40.9147619Z * [new branch] mlazos/stests -> origin/mlazos/stests 2025-12-04T09:33:40.9148926Z * [new branch] mlazos/stream-ops -> origin/mlazos/stream-ops 2025-12-04T09:33:40.9150147Z * [new branch] mlazos/td-fix2 -> origin/mlazos/td-fix2 2025-12-04T09:33:40.9151511Z * [new branch] mlazos/tensor-hasattr2 -> origin/mlazos/tensor-hasattr2 2025-12-04T09:33:40.9152716Z * [new branch] mlazos/test -> origin/mlazos/test 2025-12-04T09:33:40.9153984Z * [new branch] mlazos/tf-mode -> origin/mlazos/tf-mode 2025-12-04T09:33:40.9155286Z * [new branch] mlazos/tf-mode-backup2 -> origin/mlazos/tf-mode-backup2 2025-12-04T09:33:40.9156577Z * [new branch] mlazos/tf-mode-reland -> origin/mlazos/tf-mode-reland 2025-12-04T09:33:40.9157960Z * [new branch] mlazos/tf-mode-reland2 -> origin/mlazos/tf-mode-reland2 2025-12-04T09:33:40.9159211Z * [new branch] mlazos/tf-mode-reland3 -> origin/mlazos/tf-mode-reland3 2025-12-04T09:33:40.9160940Z * [new branch] mlazos/triton-no-epi -> origin/mlazos/triton-no-epi 2025-12-04T09:33:40.9162271Z * [new branch] mlazos/tune-proto -> origin/mlazos/tune-proto 2025-12-04T09:33:40.9163609Z * [new branch] mlazos/tuple-fixes -> origin/mlazos/tuple-fixes 2025-12-04T09:33:40.9164947Z * [new branch] mlazos/tuple-fixes2 -> origin/mlazos/tuple-fixes2 2025-12-04T09:33:40.9166234Z * [new branch] mlazos/tuple-handling -> origin/mlazos/tuple-handling 2025-12-04T09:33:40.9167542Z * [new branch] mlazos/user-stream-base -> origin/mlazos/user-stream-base 2025-12-04T09:33:40.9168795Z * [new branch] mlazos/user-streams -> origin/mlazos/user-streams 2025-12-04T09:33:40.9170108Z * [new branch] mlazos/user-streams-backup -> origin/mlazos/user-streams-backup 2025-12-04T09:33:40.9171407Z * [new branch] mlazos/user-streams-backup2 -> origin/mlazos/user-streams-backup2 2025-12-04T09:33:40.9172594Z * [new branch] mlazos/vary-beta -> origin/mlazos/vary-beta 2025-12-04T09:33:40.9173880Z * [new branch] mlazos/vary-beta2 -> origin/mlazos/vary-beta2 2025-12-04T09:33:40.9175147Z * [new branch] mlazos/weird-perf1 -> origin/mlazos/weird-perf1 2025-12-04T09:33:40.9176594Z * [new branch] mm_out_dtype_compile -> origin/mm_out_dtype_compile 2025-12-04T09:33:40.9178084Z * [new branch] module-shim -> origin/module-shim 2025-12-04T09:33:40.9179437Z * [new branch] move_config -> origin/move_config 2025-12-04T09:33:40.9181227Z * [new branch] msaroufim/reduce -> origin/msaroufim/reduce 2025-12-04T09:33:40.9182991Z * [new branch] mtia/basic-cmake -> origin/mtia/basic-cmake 2025-12-04T09:33:40.9184791Z * [new branch] mwizak/fix-triton-block-shape -> origin/mwizak/fix-triton-block-shape 2025-12-04T09:33:40.9185988Z * [new branch] my_varlen_backup -> origin/my_varlen_backup 2025-12-04T09:33:40.9187335Z * [new branch] nativert_num_outputs -> origin/nativert_num_outputs 2025-12-04T09:33:40.9188640Z * [new branch] new-codegen -> origin/new-codegen 2025-12-04T09:33:40.9190015Z * [new branch] newtest-base -> origin/newtest-base 2025-12-04T09:33:40.9191717Z * [new branch] ngimel/addmm_dtype -> origin/ngimel/addmm_dtype 2025-12-04T09:33:40.9192888Z * [new branch] ngimel/div_inv -> origin/ngimel/div_inv 2025-12-04T09:33:40.9194124Z * [new branch] ngimel/error_index_list -> origin/ngimel/error_index_list 2025-12-04T09:33:40.9195353Z * [new branch] ngimel/gather_grid -> origin/ngimel/gather_grid 2025-12-04T09:33:40.9196835Z * [new branch] ngimel/gather_grid_release -> origin/ngimel/gather_grid_release 2025-12-04T09:33:40.9197956Z * [new branch] ngimel/gg_new -> origin/ngimel/gg_new 2025-12-04T09:33:40.9199178Z * [new branch] ngimel/hostalloc -> origin/ngimel/hostalloc 2025-12-04T09:33:40.9200383Z * [new branch] ngimel/storage_id -> origin/ngimel/storage_id 2025-12-04T09:33:40.9201901Z * [new branch] nightly -> origin/nightly 2025-12-04T09:33:40.9203799Z * [new branch] nikitaved/addmm_1_rowcol_lt_path_check -> origin/nikitaved/addmm_1_rowcol_lt_path_check 2025-12-04T09:33:40.9205096Z * [new branch] nikitaved/addmm_epilogue_fusions_2d_bias -> origin/nikitaved/addmm_epilogue_fusions_2d_bias 2025-12-04T09:33:40.9206316Z * [new branch] nikitaved/addmm_epilogue_fusions_inductor -> origin/nikitaved/addmm_epilogue_fusions_inductor 2025-12-04T09:33:40.9207828Z * [new branch] nikitaved/addmm_epilogue_fusions_scratch -> origin/nikitaved/addmm_epilogue_fusions_scratch 2025-12-04T09:33:40.9209275Z * [new branch] nikitaved/grad_addmm_epilogue_fusions -> origin/nikitaved/grad_addmm_epilogue_fusions 2025-12-04T09:33:40.9210881Z * [new branch] nikitaved/simpler_can_use_32bit_index -> origin/nikitaved/simpler_can_use_32bit_index 2025-12-04T09:33:40.9212126Z * [new branch] nikitaved/test -> origin/nikitaved/test 2025-12-04T09:33:40.9213774Z * [new branch] nmacchioni-perf-test-async-autotune -> origin/nmacchioni-perf-test-async-autotune 2025-12-04T09:33:40.9215032Z * [new branch] no_distributed_log_spew -> origin/no_distributed_log_spew 2025-12-04T09:33:40.9216351Z * [new branch] nofun-hack -> origin/nofun-hack 2025-12-04T09:33:40.9217833Z * [new branch] norm_bench -> origin/norm_bench 2025-12-04T09:33:40.9219858Z * [new branch] nullplay/fuse_matmul -> origin/nullplay/fuse_matmul 2025-12-04T09:33:40.9221206Z * [new branch] nullplay_fuse_matmul -> origin/nullplay_fuse_matmul 2025-12-04T09:33:40.9222605Z * [new branch] optimizer_test -> origin/optimizer_test 2025-12-04T09:33:40.9224729Z * [new branch] orig/release/1.10 -> origin/orig/release/1.10 2025-12-04T09:33:40.9226563Z * [new branch] orig/release/1.11 -> origin/orig/release/1.11 2025-12-04T09:33:40.9227882Z * [new branch] orig/release/1.12 -> origin/orig/release/1.12 2025-12-04T09:33:40.9229392Z * [new branch] orig/release/1.13 -> origin/orig/release/1.13 2025-12-04T09:33:40.9230787Z * [new branch] orig/release/1.6 -> origin/orig/release/1.6 2025-12-04T09:33:40.9232244Z * [new branch] orig/release/1.7 -> origin/orig/release/1.7 2025-12-04T09:33:40.9233575Z * [new branch] orig/release/1.8 -> origin/orig/release/1.8 2025-12-04T09:33:40.9234900Z * [new branch] orig/release/1.9 -> origin/orig/release/1.9 2025-12-04T09:33:40.9236200Z * [new branch] orig/release/2.0 -> origin/orig/release/2.0 2025-12-04T09:33:40.9237492Z * [new branch] orig/release/2.1 -> origin/orig/release/2.1 2025-12-04T09:33:40.9238815Z * [new branch] orig/release/2.2 -> origin/orig/release/2.2 2025-12-04T09:33:40.9240037Z * [new branch] orig/release/2.3 -> origin/orig/release/2.3 2025-12-04T09:33:40.9241285Z * [new branch] orig/release/2.4 -> origin/orig/release/2.4 2025-12-04T09:33:40.9242609Z * [new branch] orig/release/2.5 -> origin/orig/release/2.5 2025-12-04T09:33:40.9243884Z * [new branch] orig/release/2.6 -> origin/orig/release/2.6 2025-12-04T09:33:40.9245500Z * [new branch] orig/release/2.7 -> origin/orig/release/2.7 2025-12-04T09:33:40.9247280Z * [new branch] orig/release/2.8 -> origin/orig/release/2.8 2025-12-04T09:33:40.9248465Z * [new branch] orig/release/2.9 -> origin/orig/release/2.9 2025-12-04T09:33:40.9251199Z * [new branch] origin/gh/fxdawnn/1/base -> origin/origin/gh/fxdawnn/1/base 2025-12-04T09:33:40.9252373Z * [new branch] origin/gh/fxdawnn/1/orig -> origin/origin/gh/fxdawnn/1/orig 2025-12-04T09:33:40.9254479Z * [new branch] origin/gh/zpcore/14/orig -> origin/origin/gh/zpcore/14/orig 2025-12-04T09:33:40.9255891Z * [new branch] oulgen-patch-1 -> origin/oulgen-patch-1 2025-12-04T09:33:40.9257392Z * [new branch] oulgen-patch-2 -> origin/oulgen-patch-2 2025-12-04T09:33:40.9258993Z * [new branch] oulgen-patch-3 -> origin/oulgen-patch-3 2025-12-04T09:33:40.9260460Z * [new branch] oulgen-patch-4 -> origin/oulgen-patch-4 2025-12-04T09:33:40.9261780Z * [new branch] padded-tensor -> origin/padded-tensor 2025-12-04T09:33:40.9263150Z * [new branch] pca2 -> origin/pca2 2025-12-04T09:33:40.9264616Z * [new branch] per_channel_backup -> origin/per_channel_backup 2025-12-04T09:33:40.9266006Z * [new branch] perf_ops -> origin/perf_ops 2025-12-04T09:33:40.9267304Z * [new branch] perf_ops_2_9 -> origin/perf_ops_2_9 2025-12-04T09:33:40.9268772Z * [new branch] pianpwk-patch-1 -> origin/pianpwk-patch-1 2025-12-04T09:33:40.9270534Z * [new branch] pianpwk/__draft_debug_mode -> origin/pianpwk/__draft_debug_mode 2025-12-04T09:33:40.9271823Z * [new branch] pianpwk/_debug_mode_for_triton_draft -> origin/pianpwk/_debug_mode_for_triton_draft 2025-12-04T09:33:40.9273095Z * [new branch] pianpwk/_debug_nn_module_compile -> origin/pianpwk/_debug_nn_module_compile 2025-12-04T09:33:40.9274245Z * [new branch] pianpwk/_draft_triton_11_3 -> origin/pianpwk/_draft_triton_11_3 2025-12-04T09:33:40.9275438Z * [new branch] pianpwk/_manual_bucket_draft -> origin/pianpwk/_manual_bucket_draft 2025-12-04T09:33:40.9276972Z * [new branch] pianpwk/_profile_w_dispatch_keys -> origin/pianpwk/_profile_w_dispatch_keys 2025-12-04T09:33:40.9278508Z * [new branch] pianpwk/_super_draft_debug_mode -> origin/pianpwk/_super_draft_debug_mode 2025-12-04T09:33:40.9280038Z * [new branch] pianpwk/_unbacked_local_shard_size -> origin/pianpwk/_unbacked_local_shard_size 2025-12-04T09:33:40.9281202Z * [new branch] pianpwk/anomaly_tb -> origin/pianpwk/anomaly_tb 2025-12-04T09:33:40.9282468Z * [new branch] pianpwk/auto_fx_annotate -> origin/pianpwk/auto_fx_annotate 2025-12-04T09:33:40.9283874Z * [new branch] pianpwk/backed_size_oblivious_export -> origin/pianpwk/backed_size_oblivious_export 2025-12-04T09:33:40.9284989Z * [new branch] pianpwk/bert_dynamic_perf -> origin/pianpwk/bert_dynamic_perf 2025-12-04T09:33:40.9286358Z * [new branch] pianpwk/debug_fwd_stack_traces -> origin/pianpwk/debug_fwd_stack_traces 2025-12-04T09:33:40.9287716Z * [new branch] pianpwk/debug_hash_tensor -> origin/pianpwk/debug_hash_tensor 2025-12-04T09:33:40.9288992Z * [new branch] pianpwk/debug_mode_annotate -> origin/pianpwk/debug_mode_annotate 2025-12-04T09:33:40.9290195Z * [new branch] pianpwk/debug_mode_defaults -> origin/pianpwk/debug_mode_defaults 2025-12-04T09:33:40.9291423Z * [new branch] pianpwk/debug_mode_hacks -> origin/pianpwk/debug_mode_hacks 2025-12-04T09:33:40.9292744Z * [new branch] pianpwk/debug_mode_opcall_refactor -> origin/pianpwk/debug_mode_opcall_refactor 2025-12-04T09:33:40.9293930Z * [new branch] pianpwk/debug_mode_show_ids -> origin/pianpwk/debug_mode_show_ids 2025-12-04T09:33:40.9295200Z * [new branch] pianpwk/debug_mode_triton -> origin/pianpwk/debug_mode_triton 2025-12-04T09:33:40.9296761Z * [new branch] pianpwk/debug_show_stack_trace -> origin/pianpwk/debug_show_stack_trace 2025-12-04T09:33:40.9298254Z * [new branch] pianpwk/debug_wait_on_collective -> origin/pianpwk/debug_wait_on_collective 2025-12-04T09:33:40.9299573Z * [new branch] pianpwk/debugmode_compile_tf -> origin/pianpwk/debugmode_compile_tf 2025-12-04T09:33:40.9301220Z * [new branch] pianpwk/dispatch_key_debugging_for_debug -> origin/pianpwk/dispatch_key_debugging_for_debug 2025-12-04T09:33:40.9302220Z * [new branch] pianpwk/draft_debug_mode_tfcompile -> origin/pianpwk/draft_debug_mode_tfcompile 2025-12-04T09:33:40.9303497Z * [new branch] pianpwk/draft_multikernel_nn -> origin/pianpwk/draft_multikernel_nn 2025-12-04T09:33:40.9304888Z * [new branch] pianpwk/draft_multikernel_status_10_5 -> origin/pianpwk/draft_multikernel_status_10_5 2025-12-04T09:33:40.9306315Z * [new branch] pianpwk/dtensor_custom_chunk -> origin/pianpwk/dtensor_custom_chunk 2025-12-04T09:33:40.9308141Z * [new branch] pianpwk/dtensor_unbacked_keypath -> origin/pianpwk/dtensor_unbacked_keypath 2025-12-04T09:33:40.9309477Z * [new branch] pianpwk/event_list_tree -> origin/pianpwk/event_list_tree 2025-12-04T09:33:40.9310719Z * [new branch] pianpwk/false_numel_refs -> origin/pianpwk/false_numel_refs 2025-12-04T09:33:40.9311940Z * [new branch] pianpwk/maybe_guard_rel -> origin/pianpwk/maybe_guard_rel 2025-12-04T09:33:40.9313264Z * [new branch] pianpwk/multikernel_hints_draft -> origin/pianpwk/multikernel_hints_draft 2025-12-04T09:33:40.9314647Z * [new branch] pianpwk/no_size_oblivious_slice_scat -> origin/pianpwk/no_size_oblivious_slice_scat 2025-12-04T09:33:40.9315883Z * [new branch] pianpwk/oblivious_reshape_view_better -> origin/pianpwk/oblivious_reshape_view_better 2025-12-04T09:33:40.9317003Z * [new branch] pianpwk/pre_forward_hook -> origin/pianpwk/pre_forward_hook 2025-12-04T09:33:40.9318308Z * [new branch] pianpwk/skip_python_keys_alternate -> origin/pianpwk/skip_python_keys_alternate 2025-12-04T09:33:40.9319579Z * [new branch] pianpwk/skip_python_keys_in_guards -> origin/pianpwk/skip_python_keys_in_guards 2025-12-04T09:33:40.9320732Z * [new branch] pianpwk/sym_tokens_draft -> origin/pianpwk/sym_tokens_draft 2025-12-04T09:33:40.9322093Z * [new branch] pianpwk/symint_one_hot -> origin/pianpwk/symint_one_hot 2025-12-04T09:33:40.9323626Z * [new branch] pianpwk/test_pointwise_guard_or_false -> origin/pianpwk/test_pointwise_guard_or_false 2025-12-04T09:33:40.9324852Z * [new branch] pianpwk/totally_draft_sym_wrap -> origin/pianpwk/totally_draft_sym_wrap 2025-12-04T09:33:40.9326039Z * [new branch] pianpwk/try_dumb_stuff -> origin/pianpwk/try_dumb_stuff 2025-12-04T09:33:40.9327283Z * [new branch] pianpwk/try_dumb_stuff_2 -> origin/pianpwk/try_dumb_stuff_2 2025-12-04T09:33:40.9329092Z * [new branch] pianpwk/unbacked_dtensor_mm -> origin/pianpwk/unbacked_dtensor_mm 2025-12-04T09:33:40.9330328Z * [new branch] pianpwk/unbacked_tracing_12_2 -> origin/pianpwk/unbacked_tracing_12_2 2025-12-04T09:33:40.9331525Z * [new branch] pianpwk/user_symints -> origin/pianpwk/user_symints 2025-12-04T09:33:40.9332795Z * [new branch] pianpwk/wan21_reshape -> origin/pianpwk/wan21_reshape 2025-12-04T09:33:40.9334569Z * [new branch] piz/fix_partial_backward_1112 -> origin/piz/fix_partial_backward_1112 2025-12-04T09:33:40.9335802Z * [new branch] piz/prop_cache_clean -> origin/piz/prop_cache_clean 2025-12-04T09:33:40.9337262Z * [new branch] pool-separate -> origin/pool-separate 2025-12-04T09:33:40.9338633Z * [new branch] pr-156087 -> origin/pr-156087 2025-12-04T09:33:40.9340462Z * [new branch] pr/131860 -> origin/pr/131860 2025-12-04T09:33:40.9342445Z * [new branch] predispatch_to -> origin/predispatch_to 2025-12-04T09:33:40.9343830Z * [new branch] protect-c17 -> origin/protect-c17 2025-12-04T09:33:40.9345192Z * [new branch] pt-opt-cuda3 -> origin/pt-opt-cuda3 2025-12-04T09:33:40.9347118Z * [new branch] python_compiled_autograd -> origin/python_compiled_autograd 2025-12-04T09:33:40.9349037Z * [new branch] q1l1/fix_device_moved_constant_type_unknown -> origin/q1l1/fix_device_moved_constant_type_unknown 2025-12-04T09:33:40.9350348Z * [new branch] q1l1/fix_wrong_default_type_for_kernel_call_args -> origin/q1l1/fix_wrong_default_type_for_kernel_call_args 2025-12-04T09:33:40.9352437Z * [new branch] qchip/export-D54134695 -> origin/qchip/export-D54134695 2025-12-04T09:33:40.9353854Z * [new branch] quote-pytest_cache -> origin/quote-pytest_cache 2025-12-04T09:33:40.9355526Z * [new branch] reland-accgrad-stream-warn -> origin/reland-accgrad-stream-warn 2025-12-04T09:33:40.9357374Z * [new branch] release/1.10 -> origin/release/1.10 2025-12-04T09:33:40.9358692Z * [new branch] release/1.11 -> origin/release/1.11 2025-12-04T09:33:40.9360020Z * [new branch] release/1.12 -> origin/release/1.12 2025-12-04T09:33:40.9361333Z * [new branch] release/1.13 -> origin/release/1.13 2025-12-04T09:33:40.9362683Z * [new branch] release/1.4 -> origin/release/1.4 2025-12-04T09:33:40.9363733Z * [new branch] release/1.4.1 -> origin/release/1.4.1 2025-12-04T09:33:40.9364995Z * [new branch] release/1.5 -> origin/release/1.5 2025-12-04T09:33:40.9366346Z * [new branch] release/1.6 -> origin/release/1.6 2025-12-04T09:33:40.9367676Z * [new branch] release/1.7 -> origin/release/1.7 2025-12-04T09:33:40.9369174Z * [new branch] release/1.8 -> origin/release/1.8 2025-12-04T09:33:40.9370394Z * [new branch] release/1.9 -> origin/release/1.9 2025-12-04T09:33:40.9371704Z * [new branch] release/2.0 -> origin/release/2.0 2025-12-04T09:33:40.9373061Z * [new branch] release/2.1 -> origin/release/2.1 2025-12-04T09:33:40.9374891Z * [new branch] release/2.2 -> origin/release/2.2 2025-12-04T09:33:40.9376428Z * [new branch] release/2.3 -> origin/release/2.3 2025-12-04T09:33:40.9378319Z * [new branch] release/2.4 -> origin/release/2.4 2025-12-04T09:33:40.9380187Z * [new branch] release/2.5 -> origin/release/2.5 2025-12-04T09:33:40.9381726Z * [new branch] release/2.6 -> origin/release/2.6 2025-12-04T09:33:40.9384405Z * [new branch] release/2.7 -> origin/release/2.7 2025-12-04T09:33:40.9385607Z * [new branch] release/2.8 -> origin/release/2.8 2025-12-04T09:33:40.9386235Z * [new branch] release/2.9 -> origin/release/2.9 2025-12-04T09:33:40.9387498Z * [new branch] release_notes -> origin/release_notes 2025-12-04T09:33:40.9388837Z * [new branch] remove_pyinterpreter -> origin/remove_pyinterpreter 2025-12-04T09:33:40.9390516Z * [new branch] replace-pytorch-labs-20250812-195836 -> origin/replace-pytorch-labs-20250812-195836 2025-12-04T09:33:40.9391571Z * [new branch] replace-pytorch-labs-20250812-200248 -> origin/replace-pytorch-labs-20250812-200248 2025-12-04T09:33:40.9392848Z * [new branch] replace-pytorch-labs-20250812-200324 -> origin/replace-pytorch-labs-20250812-200324 2025-12-04T09:33:40.9394072Z * [new branch] replace-pytorch-labs-20250812-204020 -> origin/replace-pytorch-labs-20250812-204020 2025-12-04T09:33:40.9396725Z * [new branch] revert-131069-gh/krzysztofjordan/1/head -> origin/revert-131069-gh/krzysztofjordan/1/head 2025-12-04T09:33:40.9401964Z * [new branch] revert-131469-gh/andrewor14/51/head -> origin/revert-131469-gh/andrewor14/51/head 2025-12-04T09:33:40.9404493Z * [new branch] revert-152361-gh/fadara01/1/head -> origin/revert-152361-gh/fadara01/1/head 2025-12-04T09:33:40.9406960Z * [new branch] revert-156870-gh/skarjala/3/head -> origin/revert-156870-gh/skarjala/3/head 2025-12-04T09:33:40.9408694Z * [new branch] revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ -> origin/revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ 2025-12-04T09:33:40.9409864Z * [new branch] revert-hoo-invoke-subgraph -> origin/revert-hoo-invoke-subgraph 2025-12-04T09:33:40.9411666Z * [new branch] revert_always_build_distributed -> origin/revert_always_build_distributed 2025-12-04T09:33:40.9412944Z * [new branch] rms_norm_patch -> origin/rms_norm_patch 2025-12-04T09:33:40.9414944Z * [new branch] ruisi/fix_all_to_all_estimation -> origin/ruisi/fix_all_to_all_estimation 2025-12-04T09:33:40.9415844Z * [new branch] ruisi/fix_comm_estimation -> origin/ruisi/fix_comm_estimation 2025-12-04T09:33:40.9417307Z * [new branch] ruisi/fix_dynamic_shape_estimation -> origin/ruisi/fix_dynamic_shape_estimation 2025-12-04T09:33:40.9418577Z * [new branch] ruisi/fix_llama3_autobucketing -> origin/ruisi/fix_llama3_autobucketing 2025-12-04T09:33:40.9420168Z * [new branch] ruisi/fix_manual_bucketing_ep_pass -> origin/ruisi/fix_manual_bucketing_ep_pass 2025-12-04T09:33:40.9421659Z * [new branch] ruisi/manual_bucket_pass -> origin/ruisi/manual_bucket_pass 2025-12-04T09:33:40.9423712Z * [new branch] ryanguo99/cleanup-dynamo-expected-failures -> origin/ryanguo99/cleanup-dynamo-expected-failures 2025-12-04T09:33:40.9424640Z * [new branch] ryanguo99/fix-closure-var -> origin/ryanguo99/fix-closure-var 2025-12-04T09:33:40.9426899Z * [new branch] rzou/faketensor_bench -> origin/rzou/faketensor_bench 2025-12-04T09:33:40.9428189Z * [new branch] rzou/njt -> origin/rzou/njt 2025-12-04T09:33:40.9429500Z * [new branch] rzou/pca -> origin/rzou/pca 2025-12-04T09:33:40.9430710Z * [new branch] rzou/realprop -> origin/rzou/realprop 2025-12-04T09:33:40.9432092Z * [new branch] samplevllm -> origin/samplevllm 2025-12-04T09:33:40.9434336Z * [new branch] sanchitintel/weird_thing_with_test_cpu_select_algorithm -> origin/sanchitintel/weird_thing_with_test_cpu_select_algorithm 2025-12-04T09:33:40.9435543Z * [new branch] sapling-pr-archive-SS-JIA -> origin/sapling-pr-archive-SS-JIA 2025-12-04T09:33:40.9437026Z * [new branch] sapling-pr-archive-tushar00jain -> origin/sapling-pr-archive-tushar00jain 2025-12-04T09:33:40.9438236Z * [new branch] save -> origin/save 2025-12-04T09:33:40.9439731Z * [new branch] scaled_mm -> origin/scaled_mm 2025-12-04T09:33:40.9441028Z * [new branch] scan_attempt -> origin/scan_attempt 2025-12-04T09:33:40.9442709Z * [new branch] sdym/2.5.1 -> origin/sdym/2.5.1 2025-12-04T09:33:40.9444157Z * [new branch] sekyondaMeta-dynamoconfig-fix -> origin/sekyondaMeta-dynamoconfig-fix 2025-12-04T09:33:40.9445825Z * [new branch] shengf/fx-xform-perf -> origin/shengf/fx-xform-perf 2025-12-04T09:33:40.9447623Z * [new branch] shoumikhin-patch-1 -> origin/shoumikhin-patch-1 2025-12-04T09:33:40.9449012Z * [new branch] solve-accuracy-fix -> origin/solve-accuracy-fix 2025-12-04T09:33:40.9450314Z * [new branch] some_rocm_inductor_skips -> origin/some_rocm_inductor_skips 2025-12-04T09:33:40.9452179Z * [new branch] soulitzer/stash-tls-ac -> origin/soulitzer/stash-tls-ac 2025-12-04T09:33:40.9453478Z * [new branch] sparse-mm-bf16-support -> origin/sparse-mm-bf16-support 2025-12-04T09:33:40.9454809Z * [new branch] starterTaskUpdate -> origin/starterTaskUpdate 2025-12-04T09:33:40.9456156Z * [new branch] suo -> origin/suo 2025-12-04T09:33:40.9457682Z * [new branch] sve-poc -> origin/sve-poc 2025-12-04T09:33:40.9459184Z * [new branch] switch-bn -> origin/switch-bn 2025-12-04T09:33:40.9460569Z * [new branch] sy_annotation_in_autograd_hop -> origin/sy_annotation_in_autograd_hop 2025-12-04T09:33:40.9461829Z * [new branch] sy_aot_eager_record -> origin/sy_aot_eager_record 2025-12-04T09:33:40.9463196Z * [new branch] sy_custom_bucketing -> origin/sy_custom_bucketing 2025-12-04T09:33:40.9464603Z * [new branch] sy_debug_mode_test -> origin/sy_debug_mode_test 2025-12-04T09:33:40.9465885Z * [new branch] sy_deserialize -> origin/sy_deserialize 2025-12-04T09:33:40.9467158Z * [new branch] sy_dump_gm_code -> origin/sy_dump_gm_code 2025-12-04T09:33:40.9468550Z * [new branch] sy_exp -> origin/sy_exp 2025-12-04T09:33:40.9469889Z * [new branch] sy_export_annotation -> origin/sy_export_annotation 2025-12-04T09:33:40.9471256Z * [new branch] sy_invoke_subgraph -> origin/sy_invoke_subgraph 2025-12-04T09:33:40.9472608Z * [new branch] sy_kernel_bw_name -> origin/sy_kernel_bw_name 2025-12-04T09:33:40.9473928Z * [new branch] sy_multi_arch -> origin/sy_multi_arch 2025-12-04T09:33:40.9475265Z * [new branch] sy_nn_module_stack -> origin/sy_nn_module_stack 2025-12-04T09:33:40.9476612Z * [new branch] sy_original_dtensor -> origin/sy_original_dtensor 2025-12-04T09:33:40.9477929Z * [new branch] sy_profiler_cia -> origin/sy_profiler_cia 2025-12-04T09:33:40.9479219Z * [new branch] symm_mem_sync -> origin/symm_mem_sync 2025-12-04T09:33:40.9480755Z * [new branch] sympy-bottleneck-repro -> origin/sympy-bottleneck-repro 2025-12-04T09:33:40.9482105Z * [new branch] tensordict_integration -> origin/tensordict_integration 2025-12-04T09:33:40.9483531Z * [new branch] test-move-conda-builds -> origin/test-move-conda-builds 2025-12-04T09:33:40.9484798Z * [new branch] test-old -> origin/test-old 2025-12-04T09:33:40.9486506Z * [new branch] test/bmm_heur -> origin/test/bmm_heur 2025-12-04T09:33:40.9488354Z * [new branch] tianren/customOp_autotune_fix -> origin/tianren/customOp_autotune_fix 2025-12-04T09:33:40.9489645Z * [new branch] tianren/customOp_enable_max_autotune -> origin/tianren/customOp_enable_max_autotune 2025-12-04T09:33:40.9490800Z * [new branch] tianren/customOp_fusion -> origin/tianren/customOp_fusion 2025-12-04T09:33:40.9492162Z * [new branch] tianren/customop_collectiveop_benchmark -> origin/tianren/customop_collectiveop_benchmark 2025-12-04T09:33:40.9493734Z * [new branch] tianren/customop_collectiveop_benchmark_fix -> origin/tianren/customop_collectiveop_benchmark_fix 2025-12-04T09:33:40.9495292Z * [new branch] tianren/customop_dynamic_config -> origin/tianren/customop_dynamic_config 2025-12-04T09:33:40.9496702Z * [new branch] tianren/dynamic_range_input -> origin/tianren/dynamic_range_input 2025-12-04T09:33:40.9498235Z * [new branch] tianren/dynamic_range_input_fix -> origin/tianren/dynamic_range_input_fix 2025-12-04T09:33:40.9499464Z * [new branch] tianren/dynamic_range_input_merge -> origin/tianren/dynamic_range_input_merge 2025-12-04T09:33:40.9500800Z * [new branch] tianren/flex_paged_attn_fix_temp -> origin/tianren/flex_paged_attn_fix_temp 2025-12-04T09:33:40.9502121Z * [new branch] tianren/fx_codegen_dump -> origin/tianren/fx_codegen_dump 2025-12-04T09:33:40.9503358Z * [new branch] tianren/symmetric_memory -> origin/tianren/symmetric_memory 2025-12-04T09:33:40.9504559Z * [new branch] tianren/test -> origin/tianren/test 2025-12-04T09:33:40.9505965Z * [new branch] tidy_performance_cyy -> origin/tidy_performance_cyy 2025-12-04T09:33:40.9507292Z * [new branch] tmp -> origin/tmp 2025-12-04T09:33:40.9508664Z * [new branch] torchtitan_ep -> origin/torchtitan_ep 2025-12-04T09:33:40.9510113Z * [new branch] torchtitan_integration -> origin/torchtitan_integration 2025-12-04T09:33:40.9511621Z * [new branch] trace_fsdp_torchtune_lora -> origin/trace_fsdp_torchtune_lora 2025-12-04T09:33:40.9512781Z * [new branch] traceable_fsdp_unit_tests -> origin/traceable_fsdp_unit_tests 2025-12-04T09:33:40.9514155Z * [new branch] tree_loop_vec_base -> origin/tree_loop_vec_base 2025-12-04T09:33:40.9515521Z * [new branch] triton_kernel -> origin/triton_kernel 2025-12-04T09:33:40.9516920Z * [new branch] tt_pkg_1908 -> origin/tt_pkg_1908 2025-12-04T09:33:40.9518216Z * [new branch] type_dec -> origin/type_dec 2025-12-04T09:33:40.9519631Z * [new branch] udate-sphinx-dependancies -> origin/udate-sphinx-dependancies 2025-12-04T09:33:40.9521613Z * [new branch] update-audio-commit-hash/17630256502-1803-1 -> origin/update-audio-commit-hash/17630256502-1803-1 2025-12-04T09:33:40.9522834Z * [new branch] update-audio-commit-hash/19087141161-1916-1 -> origin/update-audio-commit-hash/19087141161-1916-1 2025-12-04T09:33:40.9523954Z * [new branch] update-audio-commit-hash/19250643381-1929-1 -> origin/update-audio-commit-hash/19250643381-1929-1 2025-12-04T09:33:40.9525287Z * [new branch] update-audio-commit-hash/19397724337-1935-1 -> origin/update-audio-commit-hash/19397724337-1935-1 2025-12-04T09:33:40.9526370Z * [new branch] update-audio-commit-hash/19555670148-1941-1 -> origin/update-audio-commit-hash/19555670148-1941-1 2025-12-04T09:33:40.9527986Z * [new branch] update-audio-commit-hash/19750627930-1946-1 -> origin/update-audio-commit-hash/19750627930-1946-1 2025-12-04T09:33:40.9530210Z * [new branch] update-triton-commit-hash/13663274526-1487-2 -> origin/update-triton-commit-hash/13663274526-1487-2 2025-12-04T09:33:40.9531948Z * [new branch] update-vision-commit-hash/19087141161-1916-1 -> origin/update-vision-commit-hash/19087141161-1916-1 2025-12-04T09:33:40.9533073Z * [new branch] update-vision-commit-hash/19184897099-1925-1 -> origin/update-vision-commit-hash/19184897099-1925-1 2025-12-04T09:33:40.9534216Z * [new branch] update-vision-commit-hash/19250643381-1929-1 -> origin/update-vision-commit-hash/19250643381-1929-1 2025-12-04T09:33:40.9535601Z * [new branch] update-vision-commit-hash/19381328640-1934-1 -> origin/update-vision-commit-hash/19381328640-1934-1 2025-12-04T09:33:40.9536727Z * [new branch] update-vision-commit-hash/19485237164-1938-1 -> origin/update-vision-commit-hash/19485237164-1938-1 2025-12-04T09:33:40.9538704Z * [new branch] update-vllm-commit-hash/18451675449-1879-1 -> origin/update-vllm-commit-hash/18451675449-1879-1 2025-12-04T09:33:40.9540513Z * [new branch] update-vllm-dockerfile -> origin/update-vllm-dockerfile 2025-12-04T09:33:40.9542355Z * [new branch] update-xla-commit-hash/19224287370-211-1 -> origin/update-xla-commit-hash/19224287370-211-1 2025-12-04T09:33:40.9543536Z * [new branch] update-xla-commit-hash/19422028566-212-1 -> origin/update-xla-commit-hash/19422028566-212-1 2025-12-04T09:33:40.9544747Z * [new branch] update-xla-commit-hash/19626841311-213-1 -> origin/update-xla-commit-hash/19626841311-213-1 2025-12-04T09:33:40.9546144Z * [new branch] update_docs_torch_multinomial_issue#125388 -> origin/update_docs_torch_multinomial_issue#125388 2025-12-04T09:33:40.9547385Z * [new branch] update_operator_readme -> origin/update_operator_readme 2025-12-04T09:33:40.9548787Z * [new branch] update_slow_tests_1722488736 -> origin/update_slow_tests_1722488736 2025-12-04T09:33:40.9550244Z * [new branch] update_slow_tests_1722879173 -> origin/update_slow_tests_1722879173 2025-12-04T09:33:40.9551607Z * [new branch] update_slow_tests_1762155677 -> origin/update_slow_tests_1762155677 2025-12-04T09:33:40.9553104Z * [new branch] update_slow_tests_1763365283 -> origin/update_slow_tests_1763365283 2025-12-04T09:33:40.9554380Z * [new branch] update_submodule_FBGEMM -> origin/update_submodule_FBGEMM 2025-12-04T09:33:40.9555738Z * [new branch] update_submodule_kineto -> origin/update_submodule_kineto 2025-12-04T09:33:40.9557176Z * [new branch] update_submodule_tensorpipe -> origin/update_submodule_tensorpipe 2025-12-04T09:33:40.9558685Z * [new branch] upload-tests-for-autorevert -> origin/upload-tests-for-autorevert 2025-12-04T09:33:40.9560027Z * [new branch] v0.1.2 -> origin/v0.1.2 2025-12-04T09:33:40.9561597Z * [new branch] v1.0.1 -> origin/v1.0.1 2025-12-04T09:33:40.9563042Z * [new branch] v1.0.3 -> origin/v1.0.3 2025-12-04T09:33:40.9564656Z * [new branch] v1.1.0 -> origin/v1.1.0 2025-12-04T09:33:40.9566534Z * [new branch] v1.2.0 -> origin/v1.2.0 2025-12-04T09:33:40.9567978Z * [new branch] v1.3.0 -> origin/v1.3.0 2025-12-04T09:33:40.9569389Z * [new branch] v1.3.1 -> origin/v1.3.1 2025-12-04T09:33:40.9570804Z * [new branch] validate_fn -> origin/validate_fn 2025-12-04T09:33:40.9572307Z * [new branch] validations_2.6 -> origin/validations_2.6 2025-12-04T09:33:40.9573742Z * [new branch] validations_2.8 -> origin/validations_2.8 2025-12-04T09:33:40.9575095Z * [new branch] varlen-api -> origin/varlen-api 2025-12-04T09:33:40.9576458Z * [new branch] varlen-api-backup -> origin/varlen-api-backup 2025-12-04T09:33:40.9577893Z * [new branch] varlen_batch_invariance -> origin/varlen_batch_invariance 2025-12-04T09:33:40.9579432Z * [new branch] viable/strict -> origin/viable/strict 2025-12-04T09:33:40.9581395Z * [new branch] vishal9-team/dtensor_parallelism_toy -> origin/vishal9-team/dtensor_parallelism_toy 2025-12-04T09:33:40.9582628Z * [new branch] vllmbuildci -> origin/vllmbuildci 2025-12-04T09:33:40.9584025Z * [new branch] vllmpin -> origin/vllmpin 2025-12-04T09:33:40.9585597Z * [new branch] vscode-recommend-pyrefly -> origin/vscode-recommend-pyrefly 2025-12-04T09:33:40.9586983Z * [new branch] wdvr-patch-1 -> origin/wdvr-patch-1 2025-12-04T09:33:40.9588696Z * [new branch] wdvr/iss_145259 -> origin/wdvr/iss_145259 2025-12-04T09:33:40.9590346Z * [new branch] whc/pei -> origin/whc/pei 2025-12-04T09:33:40.9591699Z * [new branch] whc/pp_fix -> origin/whc/pp_fix 2025-12-04T09:33:40.9593016Z * [new branch] whc/sharding -> origin/whc/sharding 2025-12-04T09:33:40.9594194Z * [new branch] whc/sharding2 -> origin/whc/sharding2 2025-12-04T09:33:40.9595383Z * [new branch] whc/uneven -> origin/whc/uneven 2025-12-04T09:33:40.9597193Z * [new branch] whc/uneven-merge -> origin/whc/uneven-merge 2025-12-04T09:33:40.9598525Z * [new branch] win_warnings -> origin/win_warnings 2025-12-04T09:33:40.9599841Z * [new branch] windows_libtorch_free -> origin/windows_libtorch_free 2025-12-04T09:33:40.9601133Z * [new branch] xmfan-war -> origin/xmfan-war 2025-12-04T09:33:40.9602931Z * [new branch] xmfan/ca_0516 -> origin/xmfan/ca_0516 2025-12-04T09:33:40.9604265Z * [new branch] xmfan/ca_1051b93192 -> origin/xmfan/ca_1051b93192 2025-12-04T09:33:40.9605788Z * [new branch] xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 -> origin/xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 2025-12-04T09:33:40.9606465Z * [new branch] xmfan/ca_5a2be192d1 -> origin/xmfan/ca_5a2be192d1 2025-12-04T09:33:40.9607752Z * [new branch] xmfan/ca_9d59b516e9 -> origin/xmfan/ca_9d59b516e9 2025-12-04T09:33:40.9608863Z * [new branch] xmfan/ca_apr8 -> origin/xmfan/ca_apr8 2025-12-04T09:33:40.9610036Z * [new branch] xmfan/ca_base -> origin/xmfan/ca_base 2025-12-04T09:33:40.9611501Z * [new branch] xmfan/ca_dynamic -> origin/xmfan/ca_dynamic 2025-12-04T09:33:40.9613220Z * [new branch] xmfan/ca_fix_dyn -> origin/xmfan/ca_fix_dyn 2025-12-04T09:33:40.9615013Z * [new branch] xmfan/ca_fix_lowering -> origin/xmfan/ca_fix_lowering 2025-12-04T09:33:40.9616268Z * [new branch] xmfan/ca_fix_polyfills -> origin/xmfan/ca_fix_polyfills 2025-12-04T09:33:40.9617448Z * [new branch] xmfan/ca_jan3 -> origin/xmfan/ca_jan3 2025-12-04T09:33:40.9618704Z * [new branch] xmfan/ca_jun18 -> origin/xmfan/ca_jun18 2025-12-04T09:33:40.9619970Z * [new branch] xmfan/ca_jun24 -> origin/xmfan/ca_jun24 2025-12-04T09:33:40.9621206Z * [new branch] xmfan/ca_nested -> origin/xmfan/ca_nested 2025-12-04T09:33:40.9622634Z * [new branch] xmfan/ca_overhead -> origin/xmfan/ca_overhead 2025-12-04T09:33:40.9623929Z * [new branch] xmfan/ca_overhead_0eba7e5451 -> origin/xmfan/ca_overhead_0eba7e5451 2025-12-04T09:33:40.9625491Z * [new branch] xmfan/cacu_jun18 -> origin/xmfan/cacu_jun18 2025-12-04T09:33:40.9626726Z * [new branch] xmfan/cacu_jun19 -> origin/xmfan/cacu_jun19 2025-12-04T09:33:40.9627964Z * [new branch] xmfan/cacu_jun4 -> origin/xmfan/cacu_jun4 2025-12-04T09:33:40.9629230Z * [new branch] xmfan/disable_duck_shape -> origin/xmfan/disable_duck_shape 2025-12-04T09:33:40.9630603Z * [new branch] xmfan/fca_cpp_node_passthrough -> origin/xmfan/fca_cpp_node_passthrough 2025-12-04T09:33:40.9632101Z * [new branch] xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 2025-12-04T09:33:40.9633297Z * [new branch] xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 2025-12-04T09:33:40.9634380Z * [new branch] xmfan/single_step -> origin/xmfan/single_step 2025-12-04T09:33:40.9635586Z * [new branch] xmfan/sth_0829 -> origin/xmfan/sth_0829 2025-12-04T09:33:40.9636877Z * [new branch] xmfan/test -> origin/xmfan/test 2025-12-04T09:33:40.9638705Z * [new branch] yguo/debug-0226-constexpr -> origin/yguo/debug-0226-constexpr 2025-12-04T09:33:40.9639902Z * [new branch] yguo/new_latest_changes -> origin/yguo/new_latest_changes 2025-12-04T09:33:40.9641103Z * [new branch] yguo/patch_constexpr_changes -> origin/yguo/patch_constexpr_changes 2025-12-04T09:33:40.9642826Z * [new branch] yiming/bootcamp -> origin/yiming/bootcamp 2025-12-04T09:33:40.9644151Z * [new branch] yiming/run_with_start_end_rng_hop -> origin/yiming/run_with_start_end_rng_hop 2025-12-04T09:33:40.9645402Z * [new branch] yolo-llama3 -> origin/yolo-llama3 2025-12-04T09:33:40.9647129Z * [new branch] zainr/canary-test -> origin/zainr/canary-test 2025-12-04T09:33:40.9648558Z * [new branch] zainr/cleanup-gh-runners -> origin/zainr/cleanup-gh-runners 2025-12-04T09:33:40.9649742Z * [new branch] zainr/pull-migration-c -> origin/zainr/pull-migration-c 2025-12-04T09:33:40.9650829Z * [new branch] zainr/test2 -> origin/zainr/test2 2025-12-04T09:33:40.9652432Z * [new branch] zasdfgbnm-patch-3 -> origin/zasdfgbnm-patch-3 2025-12-04T09:33:40.9653678Z * [new branch] zb2p -> origin/zb2p 2025-12-04T09:33:40.9655123Z * [new branch] zeros-and-scatter-part2 -> origin/zeros-and-scatter-part2 2025-12-04T09:33:40.9657206Z * [new branch] zhxchen17/ci/vllm_lora_oom -> origin/zhxchen17/ci/vllm_lora_oom 2025-12-04T09:33:40.9658579Z * [new branch] zhxchen17/ci/vllm_multimodal_oom -> origin/zhxchen17/ci/vllm_multimodal_oom 2025-12-04T09:33:40.9659908Z * [new branch] zhxchen17/ci/vllm_pin -> origin/zhxchen17/ci/vllm_pin 2025-12-04T09:33:40.9661758Z * [new branch] zhxchen17/dynamo/unsafe_drop_all_guards -> origin/zhxchen17/dynamo/unsafe_drop_all_guards 2025-12-04T09:33:40.9663569Z * [new branch] zhxchen17/export/call_override -> origin/zhxchen17/export/call_override 2025-12-04T09:33:40.9664770Z * [new branch] zhxchen17/export/codemod1 -> origin/zhxchen17/export/codemod1 2025-12-04T09:33:40.9666066Z * [new branch] zhxchen17/export/ctx_return -> origin/zhxchen17/export/ctx_return 2025-12-04T09:33:40.9667493Z * [new branch] zhxchen17/export/disable_side_effect_warn -> origin/zhxchen17/export/disable_side_effect_warn 2025-12-04T09:33:40.9668542Z * [new branch] zhxchen17/export/pytree_check -> origin/zhxchen17/export/pytree_check 2025-12-04T09:33:40.9670354Z * [new branch] zhxchen17/precompile/aoti -> origin/zhxchen17/precompile/aoti 2025-12-04T09:33:40.9671642Z * [new branch] zhxchen17/precompile/globals -> origin/zhxchen17/precompile/globals 2025-12-04T09:33:40.9672990Z * [new branch] zhxchen17/precompile/inductor_guards -> origin/zhxchen17/precompile/inductor_guards 2025-12-04T09:33:40.9674437Z * [new branch] zhxchen17/scratch/0 -> origin/zhxchen17/scratch/0 2025-12-04T09:33:40.9675861Z * [new branch] zhxchen17/torch_export_api_update -> origin/zhxchen17/torch_export_api_update 2025-12-04T09:33:40.9677664Z * [new branch] zhxhcen17/moodycamel -> origin/zhxhcen17/moodycamel 2025-12-04T09:33:40.9679386Z * [new branch] zxiiro/build-times -> origin/zxiiro/build-times 2025-12-04T09:33:40.9680688Z * [new branch] zxiiro/c7i.2xlarge -> origin/zxiiro/c7i.2xlarge 2025-12-04T09:33:40.9682032Z * [new branch] zxiiro/c7i.2xlarge.h100 -> origin/zxiiro/c7i.2xlarge.h100 2025-12-04T09:33:40.9683311Z * [new branch] zxiiro/main -> origin/zxiiro/main 2025-12-04T09:33:40.9684531Z * [new branch] zxiiro/risc64 -> origin/zxiiro/risc64 2025-12-04T09:33:40.9685948Z * [new branch] zxiiro/test-multicloud-arc -> origin/zxiiro/test-multicloud-arc 2025-12-04T09:33:40.9687171Z * [new tag] bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug -> bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug 2025-12-04T09:33:40.9688183Z * [new tag] ci/binaries/77164 -> ci/binaries/77164 2025-12-04T09:33:40.9689275Z * [new tag] ciflow/b200/115316 -> ciflow/b200/115316 2025-12-04T09:33:40.9690006Z * [new tag] ciflow/b200/160685 -> ciflow/b200/160685 2025-12-04T09:33:40.9690840Z * [new tag] ciflow/b200/161607 -> ciflow/b200/161607 2025-12-04T09:33:40.9691811Z * [new tag] ciflow/b200/161938 -> ciflow/b200/161938 2025-12-04T09:33:40.9692853Z * [new tag] ciflow/b200/167207 -> ciflow/b200/167207 2025-12-04T09:33:40.9693431Z * [new tag] ciflow/b200/167989 -> ciflow/b200/167989 2025-12-04T09:33:40.9694515Z * [new tag] ciflow/b200/168096 -> ciflow/b200/168096 2025-12-04T09:33:40.9695514Z * [new tag] ciflow/b200/168175 -> ciflow/b200/168175 2025-12-04T09:33:40.9696444Z * [new tag] ciflow/b200/168195 -> ciflow/b200/168195 2025-12-04T09:33:40.9697728Z * [new tag] ciflow/b200/169200 -> ciflow/b200/169200 2025-12-04T09:33:40.9698538Z * [new tag] ciflow/b200/169216 -> ciflow/b200/169216 2025-12-04T09:33:40.9699954Z * [new tag] ciflow/b200/169380 -> ciflow/b200/169380 2025-12-04T09:33:40.9701141Z * [new tag] ciflow/b200/169412 -> ciflow/b200/169412 2025-12-04T09:33:40.9702190Z * [new tag] ciflow/b200/169470 -> ciflow/b200/169470 2025-12-04T09:33:40.9702987Z * [new tag] ciflow/b200/169471 -> ciflow/b200/169471 2025-12-04T09:33:40.9704044Z * [new tag] ciflow/b200/169472 -> ciflow/b200/169472 2025-12-04T09:33:40.9705008Z * [new tag] ciflow/b200/169514 -> ciflow/b200/169514 2025-12-04T09:33:40.9705799Z * [new tag] ciflow/b200/169517 -> ciflow/b200/169517 2025-12-04T09:33:40.9707074Z * [new tag] ciflow/binaries/165922 -> ciflow/binaries/165922 2025-12-04T09:33:40.9707813Z * [new tag] ciflow/binaries/169510 -> ciflow/binaries/169510 2025-12-04T09:33:40.9709100Z * [new tag] ciflow/binaries_wheel/157994 -> ciflow/binaries_wheel/157994 2025-12-04T09:33:40.9710001Z * [new tag] ciflow/binaries_wheel/166829 -> ciflow/binaries_wheel/166829 2025-12-04T09:33:40.9710821Z * [new tag] ciflow/binaries_wheel/167972 -> ciflow/binaries_wheel/167972 2025-12-04T09:33:40.9711935Z * [new tag] ciflow/binaries_wheel/167981 -> ciflow/binaries_wheel/167981 2025-12-04T09:33:40.9712744Z * [new tag] ciflow/dynamo/167695 -> ciflow/dynamo/167695 2025-12-04T09:33:40.9713580Z * [new tag] ciflow/dynamo/168096 -> ciflow/dynamo/168096 2025-12-04T09:33:40.9714691Z * [new tag] ciflow/dynamo/169525 -> ciflow/dynamo/169525 2025-12-04T09:33:40.9715669Z * [new tag] ciflow/h100-cutlass-backend/161938 -> ciflow/h100-cutlass-backend/161938 2025-12-04T09:33:40.9716527Z * [new tag] ciflow/h100-cutlass-backend/161940 -> ciflow/h100-cutlass-backend/161940 2025-12-04T09:33:40.9717675Z * [new tag] ciflow/h100-distributed/168923 -> ciflow/h100-distributed/168923 2025-12-04T09:33:40.9718530Z * [new tag] ciflow/h100-symm-mem/167552 -> ciflow/h100-symm-mem/167552 2025-12-04T09:33:40.9719321Z * [new tag] ciflow/h100-symm-mem/168129 -> ciflow/h100-symm-mem/168129 2025-12-04T09:33:40.9720160Z * [new tag] ciflow/h100-symm-mem/168917 -> ciflow/h100-symm-mem/168917 2025-12-04T09:33:40.9721475Z * [new tag] ciflow/h100-symm-mem/169156 -> ciflow/h100-symm-mem/169156 2025-12-04T09:33:40.9722176Z * [new tag] ciflow/h100-symm-mem/169200 -> ciflow/h100-symm-mem/169200 2025-12-04T09:33:40.9723101Z * [new tag] ciflow/h100-symm-mem/169216 -> ciflow/h100-symm-mem/169216 2025-12-04T09:33:40.9723903Z * [new tag] ciflow/h100-symm-mem/169338 -> ciflow/h100-symm-mem/169338 2025-12-04T09:33:40.9724986Z * [new tag] ciflow/h100-symm-mem/169355 -> ciflow/h100-symm-mem/169355 2025-12-04T09:33:40.9725701Z * [new tag] ciflow/h100-symm-mem/169543 -> ciflow/h100-symm-mem/169543 2025-12-04T09:33:40.9726715Z * [new tag] ciflow/h100/115316 -> ciflow/h100/115316 2025-12-04T09:33:40.9727460Z * [new tag] ciflow/h100/160685 -> ciflow/h100/160685 2025-12-04T09:33:40.9728267Z * [new tag] ciflow/h100/160729 -> ciflow/h100/160729 2025-12-04T09:33:40.9729108Z * [new tag] ciflow/h100/161607 -> ciflow/h100/161607 2025-12-04T09:33:40.9729875Z * [new tag] ciflow/h100/161938 -> ciflow/h100/161938 2025-12-04T09:33:40.9730773Z * [new tag] ciflow/h100/167207 -> ciflow/h100/167207 2025-12-04T09:33:40.9731507Z * [new tag] ciflow/h100/167989 -> ciflow/h100/167989 2025-12-04T09:33:40.9732325Z * [new tag] ciflow/h100/168096 -> ciflow/h100/168096 2025-12-04T09:33:40.9733137Z * [new tag] ciflow/h100/168175 -> ciflow/h100/168175 2025-12-04T09:33:40.9733947Z * [new tag] ciflow/h100/168195 -> ciflow/h100/168195 2025-12-04T09:33:40.9734790Z * [new tag] ciflow/h100/168980 -> ciflow/h100/168980 2025-12-04T09:33:40.9736031Z * [new tag] ciflow/h100/169200 -> ciflow/h100/169200 2025-12-04T09:33:40.9737170Z * [new tag] ciflow/h100/169216 -> ciflow/h100/169216 2025-12-04T09:33:40.9738237Z * [new tag] ciflow/h100/169380 -> ciflow/h100/169380 2025-12-04T09:33:40.9739002Z * [new tag] ciflow/h100/169412 -> ciflow/h100/169412 2025-12-04T09:33:40.9739980Z * [new tag] ciflow/h100/169470 -> ciflow/h100/169470 2025-12-04T09:33:40.9740720Z * [new tag] ciflow/h100/169471 -> ciflow/h100/169471 2025-12-04T09:33:40.9741582Z * [new tag] ciflow/h100/169472 -> ciflow/h100/169472 2025-12-04T09:33:40.9742424Z * [new tag] ciflow/h100/169514 -> ciflow/h100/169514 2025-12-04T09:33:40.9743590Z * [new tag] ciflow/inductor-cu126/168096 -> ciflow/inductor-cu126/168096 2025-12-04T09:33:40.9745069Z * [new tag] ciflow/inductor-micro-benchmark-cpu-x86/168096 -> ciflow/inductor-micro-benchmark-cpu-x86/168096 2025-12-04T09:33:40.9745867Z * [new tag] ciflow/inductor-micro-benchmark/166165 -> ciflow/inductor-micro-benchmark/166165 2025-12-04T09:33:40.9746731Z * [new tag] ciflow/inductor-micro-benchmark/168096 -> ciflow/inductor-micro-benchmark/168096 2025-12-04T09:33:40.9747857Z * [new tag] ciflow/inductor-perf-compare/168096 -> ciflow/inductor-perf-compare/168096 2025-12-04T09:33:40.9749336Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi300/168073 -> ciflow/inductor-perf-test-nightly-rocm-mi300/168073 2025-12-04T09:33:40.9750051Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi300/168096 -> ciflow/inductor-perf-test-nightly-rocm-mi300/168096 2025-12-04T09:33:40.9750997Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi300/169024 -> ciflow/inductor-perf-test-nightly-rocm-mi300/169024 2025-12-04T09:33:40.9752580Z * [new tag] ciflow/inductor-perf-test-nightly-rocm-mi355/169024 -> ciflow/inductor-perf-test-nightly-rocm-mi355/169024 2025-12-04T09:33:40.9753351Z * [new tag] ciflow/inductor-perf-test-nightly/168096 -> ciflow/inductor-perf-test-nightly/168096 2025-12-04T09:33:40.9754838Z * [new tag] ciflow/inductor-periodic/168096 -> ciflow/inductor-periodic/168096 2025-12-04T09:33:40.9755516Z * [new tag] ciflow/inductor-periodic/169024 -> ciflow/inductor-periodic/169024 2025-12-04T09:33:40.9756672Z * [new tag] ciflow/inductor-periodic/169425 -> ciflow/inductor-periodic/169425 2025-12-04T09:33:40.9757771Z * [new tag] ciflow/inductor-rocm-mi200/165545 -> ciflow/inductor-rocm-mi200/165545 2025-12-04T09:33:40.9758653Z * [new tag] ciflow/inductor-rocm-mi200/165997 -> ciflow/inductor-rocm-mi200/165997 2025-12-04T09:33:40.9759505Z * [new tag] ciflow/inductor-rocm-mi200/168096 -> ciflow/inductor-rocm-mi200/168096 2025-12-04T09:33:40.9760407Z * [new tag] ciflow/inductor-rocm-mi200/169063 -> ciflow/inductor-rocm-mi200/169063 2025-12-04T09:33:40.9761262Z * [new tag] ciflow/inductor-rocm-mi200/169425 -> ciflow/inductor-rocm-mi200/169425 2025-12-04T09:33:40.9762472Z * [new tag] ciflow/inductor-rocm-mi300/165545 -> ciflow/inductor-rocm-mi300/165545 2025-12-04T09:33:40.9763135Z * [new tag] ciflow/inductor-rocm-mi300/168096 -> ciflow/inductor-rocm-mi300/168096 2025-12-04T09:33:40.9763914Z * [new tag] ciflow/inductor-rocm-mi300/169063 -> ciflow/inductor-rocm-mi300/169063 2025-12-04T09:33:40.9764747Z * [new tag] ciflow/inductor-rocm-mi300/169425 -> ciflow/inductor-rocm-mi300/169425 2025-12-04T09:33:40.9765948Z * [new tag] ciflow/inductor-rocm/162052 -> ciflow/inductor-rocm/162052 2025-12-04T09:33:40.9767327Z * [new tag] ciflow/inductor-rocm/168971 -> ciflow/inductor-rocm/168971 2025-12-04T09:33:40.9768173Z * [new tag] ciflow/inductor-windows/168096 -> ciflow/inductor-windows/168096 2025-12-04T09:33:40.9769218Z * [new tag] ciflow/inductor/144542 -> ciflow/inductor/144542 2025-12-04T09:33:40.9769925Z * [new tag] ciflow/inductor/146506 -> ciflow/inductor/146506 2025-12-04T09:33:40.9771032Z * [new tag] ciflow/inductor/147990 -> ciflow/inductor/147990 2025-12-04T09:33:40.9771836Z * [new tag] ciflow/inductor/148294 -> ciflow/inductor/148294 2025-12-04T09:33:40.9772696Z * [new tag] ciflow/inductor/148492 -> ciflow/inductor/148492 2025-12-04T09:33:40.9773513Z * [new tag] ciflow/inductor/157149 -> ciflow/inductor/157149 2025-12-04T09:33:40.9774335Z * [new tag] ciflow/inductor/157994 -> ciflow/inductor/157994 2025-12-04T09:33:40.9775150Z * [new tag] ciflow/inductor/160685 -> ciflow/inductor/160685 2025-12-04T09:33:40.9775971Z * [new tag] ciflow/inductor/160686 -> ciflow/inductor/160686 2025-12-04T09:33:40.9776775Z * [new tag] ciflow/inductor/160687 -> ciflow/inductor/160687 2025-12-04T09:33:40.9777720Z * [new tag] ciflow/inductor/160688 -> ciflow/inductor/160688 2025-12-04T09:33:40.9779001Z * [new tag] ciflow/inductor/160706 -> ciflow/inductor/160706 2025-12-04T09:33:40.9780169Z * [new tag] ciflow/inductor/160729 -> ciflow/inductor/160729 2025-12-04T09:33:40.9781253Z * [new tag] ciflow/inductor/161938 -> ciflow/inductor/161938 2025-12-04T09:33:40.9781999Z * [new tag] ciflow/inductor/161939 -> ciflow/inductor/161939 2025-12-04T09:33:40.9782901Z * [new tag] ciflow/inductor/161940 -> ciflow/inductor/161940 2025-12-04T09:33:40.9783732Z * [new tag] ciflow/inductor/162052 -> ciflow/inductor/162052 2025-12-04T09:33:40.9785145Z * [new tag] ciflow/inductor/162275 -> ciflow/inductor/162275 2025-12-04T09:33:40.9785913Z * [new tag] ciflow/inductor/162795 -> ciflow/inductor/162795 2025-12-04T09:33:40.9787109Z * [new tag] ciflow/inductor/163245 -> ciflow/inductor/163245 2025-12-04T09:33:40.9787890Z * [new tag] ciflow/inductor/163335 -> ciflow/inductor/163335 2025-12-04T09:33:40.9788745Z * [new tag] ciflow/inductor/163503 -> ciflow/inductor/163503 2025-12-04T09:33:40.9789609Z * [new tag] ciflow/inductor/163942 -> ciflow/inductor/163942 2025-12-04T09:33:40.9790752Z * [new tag] ciflow/inductor/165270 -> ciflow/inductor/165270 2025-12-04T09:33:40.9791498Z * [new tag] ciflow/inductor/165274 -> ciflow/inductor/165274 2025-12-04T09:33:40.9792377Z * [new tag] ciflow/inductor/165322 -> ciflow/inductor/165322 2025-12-04T09:33:40.9793225Z * [new tag] ciflow/inductor/165597 -> ciflow/inductor/165597 2025-12-04T09:33:40.9794099Z * [new tag] ciflow/inductor/166063 -> ciflow/inductor/166063 2025-12-04T09:33:40.9794947Z * [new tag] ciflow/inductor/166075 -> ciflow/inductor/166075 2025-12-04T09:33:40.9795907Z * [new tag] ciflow/inductor/166165 -> ciflow/inductor/166165 2025-12-04T09:33:40.9797355Z * [new tag] ciflow/inductor/166254 -> ciflow/inductor/166254 2025-12-04T09:33:40.9798076Z * [new tag] ciflow/inductor/166483 -> ciflow/inductor/166483 2025-12-04T09:33:40.9798979Z * [new tag] ciflow/inductor/166494 -> ciflow/inductor/166494 2025-12-04T09:33:40.9799848Z * [new tag] ciflow/inductor/166545 -> ciflow/inductor/166545 2025-12-04T09:33:40.9800941Z * [new tag] ciflow/inductor/166788 -> ciflow/inductor/166788 2025-12-04T09:33:40.9801916Z * [new tag] ciflow/inductor/166846 -> ciflow/inductor/166846 2025-12-04T09:33:40.9802726Z * [new tag] ciflow/inductor/167300 -> ciflow/inductor/167300 2025-12-04T09:33:40.9803576Z * [new tag] ciflow/inductor/167407 -> ciflow/inductor/167407 2025-12-04T09:33:40.9804660Z * [new tag] ciflow/inductor/167536 -> ciflow/inductor/167536 2025-12-04T09:33:40.9805434Z * [new tag] ciflow/inductor/167552 -> ciflow/inductor/167552 2025-12-04T09:33:40.9806288Z * [new tag] ciflow/inductor/167555 -> ciflow/inductor/167555 2025-12-04T09:33:40.9807392Z * [new tag] ciflow/inductor/167583 -> ciflow/inductor/167583 2025-12-04T09:33:40.9808345Z * [new tag] ciflow/inductor/167599 -> ciflow/inductor/167599 2025-12-04T09:33:40.9809367Z * [new tag] ciflow/inductor/167647 -> ciflow/inductor/167647 2025-12-04T09:33:40.9810118Z * [new tag] ciflow/inductor/167677 -> ciflow/inductor/167677 2025-12-04T09:33:40.9811000Z * [new tag] ciflow/inductor/167680 -> ciflow/inductor/167680 2025-12-04T09:33:40.9811854Z * [new tag] ciflow/inductor/167695 -> ciflow/inductor/167695 2025-12-04T09:33:40.9812740Z * [new tag] ciflow/inductor/167742 -> ciflow/inductor/167742 2025-12-04T09:33:40.9813614Z * [new tag] ciflow/inductor/167768 -> ciflow/inductor/167768 2025-12-04T09:33:40.9814873Z * [new tag] ciflow/inductor/167773 -> ciflow/inductor/167773 2025-12-04T09:33:40.9815631Z * [new tag] ciflow/inductor/167781 -> ciflow/inductor/167781 2025-12-04T09:33:40.9816511Z * [new tag] ciflow/inductor/167880 -> ciflow/inductor/167880 2025-12-04T09:33:40.9817558Z * [new tag] ciflow/inductor/167887 -> ciflow/inductor/167887 2025-12-04T09:33:40.9818376Z * [new tag] ciflow/inductor/167972 -> ciflow/inductor/167972 2025-12-04T09:33:40.9819192Z * [new tag] ciflow/inductor/167989 -> ciflow/inductor/167989 2025-12-04T09:33:40.9820174Z * [new tag] ciflow/inductor/168002 -> ciflow/inductor/168002 2025-12-04T09:33:40.9820941Z * [new tag] ciflow/inductor/168050 -> ciflow/inductor/168050 2025-12-04T09:33:40.9821975Z * [new tag] ciflow/inductor/168051 -> ciflow/inductor/168051 2025-12-04T09:33:40.9822677Z * [new tag] ciflow/inductor/168052 -> ciflow/inductor/168052 2025-12-04T09:33:40.9823688Z * [new tag] ciflow/inductor/168073 -> ciflow/inductor/168073 2025-12-04T09:33:40.9824399Z * [new tag] ciflow/inductor/168096 -> ciflow/inductor/168096 2025-12-04T09:33:40.9825375Z * [new tag] ciflow/inductor/168114 -> ciflow/inductor/168114 2025-12-04T09:33:40.9826140Z * [new tag] ciflow/inductor/168115 -> ciflow/inductor/168115 2025-12-04T09:33:40.9826987Z * [new tag] ciflow/inductor/168127 -> ciflow/inductor/168127 2025-12-04T09:33:40.9827864Z * [new tag] ciflow/inductor/168129 -> ciflow/inductor/168129 2025-12-04T09:33:40.9828735Z * [new tag] ciflow/inductor/168157 -> ciflow/inductor/168157 2025-12-04T09:33:40.9829869Z * [new tag] ciflow/inductor/168175 -> ciflow/inductor/168175 2025-12-04T09:33:40.9830553Z * [new tag] ciflow/inductor/168185 -> ciflow/inductor/168185 2025-12-04T09:33:40.9831388Z * [new tag] ciflow/inductor/168195 -> ciflow/inductor/168195 2025-12-04T09:33:40.9832452Z * [new tag] ciflow/inductor/168209 -> ciflow/inductor/168209 2025-12-04T09:33:40.9833239Z * [new tag] ciflow/inductor/168266 -> ciflow/inductor/168266 2025-12-04T09:33:40.9834328Z * [new tag] ciflow/inductor/168316 -> ciflow/inductor/168316 2025-12-04T09:33:40.9835376Z * [new tag] ciflow/inductor/168326 -> ciflow/inductor/168326 2025-12-04T09:33:40.9836123Z * [new tag] ciflow/inductor/168368 -> ciflow/inductor/168368 2025-12-04T09:33:40.9837006Z * [new tag] ciflow/inductor/168894 -> ciflow/inductor/168894 2025-12-04T09:33:40.9837874Z * [new tag] ciflow/inductor/168934 -> ciflow/inductor/168934 2025-12-04T09:33:40.9838758Z * [new tag] ciflow/inductor/168939 -> ciflow/inductor/168939 2025-12-04T09:33:40.9839625Z * [new tag] ciflow/inductor/168946 -> ciflow/inductor/168946 2025-12-04T09:33:40.9840482Z * [new tag] ciflow/inductor/168950 -> ciflow/inductor/168950 2025-12-04T09:33:40.9841372Z * [new tag] ciflow/inductor/168951 -> ciflow/inductor/168951 2025-12-04T09:33:40.9842222Z * [new tag] ciflow/inductor/168952 -> ciflow/inductor/168952 2025-12-04T09:33:40.9843639Z * [new tag] ciflow/inductor/168955 -> ciflow/inductor/168955 2025-12-04T09:33:40.9844373Z * [new tag] ciflow/inductor/168971 -> ciflow/inductor/168971 2025-12-04T09:33:40.9845220Z * [new tag] ciflow/inductor/168979 -> ciflow/inductor/168979 2025-12-04T09:33:40.9846089Z * [new tag] ciflow/inductor/168980 -> ciflow/inductor/168980 2025-12-04T09:33:40.9847178Z * [new tag] ciflow/inductor/168983 -> ciflow/inductor/168983 2025-12-04T09:33:40.9847955Z * [new tag] ciflow/inductor/169006 -> ciflow/inductor/169006 2025-12-04T09:33:40.9848948Z * [new tag] ciflow/inductor/169023 -> ciflow/inductor/169023 2025-12-04T09:33:40.9849740Z * [new tag] ciflow/inductor/169024 -> ciflow/inductor/169024 2025-12-04T09:33:40.9850590Z * [new tag] ciflow/inductor/169025 -> ciflow/inductor/169025 2025-12-04T09:33:40.9851475Z * [new tag] ciflow/inductor/169066 -> ciflow/inductor/169066 2025-12-04T09:33:40.9852351Z * [new tag] ciflow/inductor/169091 -> ciflow/inductor/169091 2025-12-04T09:33:40.9853225Z * [new tag] ciflow/inductor/169102 -> ciflow/inductor/169102 2025-12-04T09:33:40.9854091Z * [new tag] ciflow/inductor/169103 -> ciflow/inductor/169103 2025-12-04T09:33:40.9854954Z * [new tag] ciflow/inductor/169121 -> ciflow/inductor/169121 2025-12-04T09:33:40.9855833Z * [new tag] ciflow/inductor/169134 -> ciflow/inductor/169134 2025-12-04T09:33:40.9856664Z * [new tag] ciflow/inductor/169135 -> ciflow/inductor/169135 2025-12-04T09:33:40.9857781Z * [new tag] ciflow/inductor/169141 -> ciflow/inductor/169141 2025-12-04T09:33:40.9858627Z * [new tag] ciflow/inductor/169151 -> ciflow/inductor/169151 2025-12-04T09:33:40.9859721Z * [new tag] ciflow/inductor/169161 -> ciflow/inductor/169161 2025-12-04T09:33:40.9860530Z * [new tag] ciflow/inductor/169167 -> ciflow/inductor/169167 2025-12-04T09:33:40.9861670Z * [new tag] ciflow/inductor/169177 -> ciflow/inductor/169177 2025-12-04T09:33:40.9862724Z * [new tag] ciflow/inductor/169185 -> ciflow/inductor/169185 2025-12-04T09:33:40.9863438Z * [new tag] ciflow/inductor/169196 -> ciflow/inductor/169196 2025-12-04T09:33:40.9864294Z * [new tag] ciflow/inductor/169200 -> ciflow/inductor/169200 2025-12-04T09:33:40.9865367Z * [new tag] ciflow/inductor/169204 -> ciflow/inductor/169204 2025-12-04T09:33:40.9866189Z * [new tag] ciflow/inductor/169216 -> ciflow/inductor/169216 2025-12-04T09:33:40.9867232Z * [new tag] ciflow/inductor/169219 -> ciflow/inductor/169219 2025-12-04T09:33:40.9867964Z * [new tag] ciflow/inductor/169220 -> ciflow/inductor/169220 2025-12-04T09:33:40.9869129Z * [new tag] ciflow/inductor/169230 -> ciflow/inductor/169230 2025-12-04T09:33:40.9869897Z * [new tag] ciflow/inductor/169242 -> ciflow/inductor/169242 2025-12-04T09:33:40.9870786Z * [new tag] ciflow/inductor/169245 -> ciflow/inductor/169245 2025-12-04T09:33:40.9871941Z * [new tag] ciflow/inductor/169260 -> ciflow/inductor/169260 2025-12-04T09:33:40.9872657Z * [new tag] ciflow/inductor/169282 -> ciflow/inductor/169282 2025-12-04T09:33:40.9873536Z * [new tag] ciflow/inductor/169286 -> ciflow/inductor/169286 2025-12-04T09:33:40.9874406Z * [new tag] ciflow/inductor/169299 -> ciflow/inductor/169299 2025-12-04T09:33:40.9875558Z * [new tag] ciflow/inductor/169304 -> ciflow/inductor/169304 2025-12-04T09:33:40.9876824Z * [new tag] ciflow/inductor/169305 -> ciflow/inductor/169305 2025-12-04T09:33:40.9877586Z * [new tag] ciflow/inductor/169308 -> ciflow/inductor/169308 2025-12-04T09:33:40.9878488Z * [new tag] ciflow/inductor/169319 -> ciflow/inductor/169319 2025-12-04T09:33:40.9879493Z * [new tag] ciflow/inductor/169326 -> ciflow/inductor/169326 2025-12-04T09:33:40.9880262Z * [new tag] ciflow/inductor/169332 -> ciflow/inductor/169332 2025-12-04T09:33:40.9881122Z * [new tag] ciflow/inductor/169333 -> ciflow/inductor/169333 2025-12-04T09:33:40.9882364Z * [new tag] ciflow/inductor/169336 -> ciflow/inductor/169336 2025-12-04T09:33:40.9883177Z * [new tag] ciflow/inductor/169340 -> ciflow/inductor/169340 2025-12-04T09:33:40.9884053Z * [new tag] ciflow/inductor/169341 -> ciflow/inductor/169341 2025-12-04T09:33:40.9884923Z * [new tag] ciflow/inductor/169343 -> ciflow/inductor/169343 2025-12-04T09:33:40.9885791Z * [new tag] ciflow/inductor/169346 -> ciflow/inductor/169346 2025-12-04T09:33:40.9886939Z * [new tag] ciflow/inductor/169348 -> ciflow/inductor/169348 2025-12-04T09:33:40.9887789Z * [new tag] ciflow/inductor/169350 -> ciflow/inductor/169350 2025-12-04T09:33:40.9888945Z * [new tag] ciflow/inductor/169355 -> ciflow/inductor/169355 2025-12-04T09:33:40.9889747Z * [new tag] ciflow/inductor/169370 -> ciflow/inductor/169370 2025-12-04T09:33:40.9891121Z * [new tag] ciflow/inductor/169375 -> ciflow/inductor/169375 2025-12-04T09:33:40.9891899Z * [new tag] ciflow/inductor/169389 -> ciflow/inductor/169389 2025-12-04T09:33:40.9892766Z * [new tag] ciflow/inductor/169391 -> ciflow/inductor/169391 2025-12-04T09:33:40.9893760Z * [new tag] ciflow/inductor/169393 -> ciflow/inductor/169393 2025-12-04T09:33:40.9894521Z * [new tag] ciflow/inductor/169399 -> ciflow/inductor/169399 2025-12-04T09:33:40.9895685Z * [new tag] ciflow/inductor/169400 -> ciflow/inductor/169400 2025-12-04T09:33:40.9896558Z * [new tag] ciflow/inductor/169415 -> ciflow/inductor/169415 2025-12-04T09:33:40.9901162Z * [new tag] ciflow/inductor/169417 -> ciflow/inductor/169417 2025-12-04T09:33:40.9901846Z * [new tag] ciflow/inductor/169418 -> ciflow/inductor/169418 2025-12-04T09:33:40.9903090Z * [new tag] ciflow/inductor/169430 -> ciflow/inductor/169430 2025-12-04T09:33:40.9903862Z * [new tag] ciflow/inductor/169432 -> ciflow/inductor/169432 2025-12-04T09:33:40.9904754Z * [new tag] ciflow/inductor/169436 -> ciflow/inductor/169436 2025-12-04T09:33:40.9905874Z * [new tag] ciflow/inductor/169437 -> ciflow/inductor/169437 2025-12-04T09:33:40.9907218Z * [new tag] ciflow/inductor/169438 -> ciflow/inductor/169438 2025-12-04T09:33:40.9907973Z * [new tag] ciflow/inductor/169441 -> ciflow/inductor/169441 2025-12-04T09:33:40.9908854Z * [new tag] ciflow/inductor/169446 -> ciflow/inductor/169446 2025-12-04T09:33:40.9909961Z * [new tag] ciflow/inductor/169447 -> ciflow/inductor/169447 2025-12-04T09:33:40.9910729Z * [new tag] ciflow/inductor/169452 -> ciflow/inductor/169452 2025-12-04T09:33:40.9912041Z * [new tag] ciflow/inductor/169455 -> ciflow/inductor/169455 2025-12-04T09:33:40.9912839Z * [new tag] ciflow/inductor/169459 -> ciflow/inductor/169459 2025-12-04T09:33:40.9913998Z * [new tag] ciflow/inductor/169463 -> ciflow/inductor/169463 2025-12-04T09:33:40.9914965Z * [new tag] ciflow/inductor/169476 -> ciflow/inductor/169476 2025-12-04T09:33:40.9915795Z * [new tag] ciflow/inductor/169485 -> ciflow/inductor/169485 2025-12-04T09:33:40.9916663Z * [new tag] ciflow/inductor/169493 -> ciflow/inductor/169493 2025-12-04T09:33:40.9917699Z * [new tag] ciflow/inductor/169496 -> ciflow/inductor/169496 2025-12-04T09:33:40.9918473Z * [new tag] ciflow/inductor/169497 -> ciflow/inductor/169497 2025-12-04T09:33:40.9919325Z * [new tag] ciflow/inductor/169503 -> ciflow/inductor/169503 2025-12-04T09:33:40.9920226Z * [new tag] ciflow/inductor/169504 -> ciflow/inductor/169504 2025-12-04T09:33:40.9921463Z * [new tag] ciflow/inductor/169505 -> ciflow/inductor/169505 2025-12-04T09:33:40.9922899Z * [new tag] ciflow/inductor/169508 -> ciflow/inductor/169508 2025-12-04T09:33:40.9923784Z * [new tag] ciflow/inductor/169509 -> ciflow/inductor/169509 2025-12-04T09:33:40.9924823Z * [new tag] ciflow/inductor/169513 -> ciflow/inductor/169513 2025-12-04T09:33:40.9925611Z * [new tag] ciflow/inductor/169514 -> ciflow/inductor/169514 2025-12-04T09:33:40.9926483Z * [new tag] ciflow/inductor/169515 -> ciflow/inductor/169515 2025-12-04T09:33:40.9927373Z * [new tag] ciflow/inductor/169517 -> ciflow/inductor/169517 2025-12-04T09:33:40.9928235Z * [new tag] ciflow/inductor/169519 -> ciflow/inductor/169519 2025-12-04T09:33:40.9929121Z * [new tag] ciflow/inductor/169520 -> ciflow/inductor/169520 2025-12-04T09:33:40.9930021Z * [new tag] ciflow/inductor/169521 -> ciflow/inductor/169521 2025-12-04T09:33:40.9930900Z * [new tag] ciflow/inductor/169524 -> ciflow/inductor/169524 2025-12-04T09:33:40.9931937Z * [new tag] ciflow/inductor/169527 -> ciflow/inductor/169527 2025-12-04T09:33:40.9932686Z * [new tag] ciflow/inductor/169528 -> ciflow/inductor/169528 2025-12-04T09:33:40.9933820Z * [new tag] ciflow/inductor/169532 -> ciflow/inductor/169532 2025-12-04T09:33:40.9934578Z * [new tag] ciflow/inductor/169535 -> ciflow/inductor/169535 2025-12-04T09:33:40.9935568Z * [new tag] ciflow/inductor/169536 -> ciflow/inductor/169536 2025-12-04T09:33:40.9936443Z * [new tag] ciflow/inductor/169547 -> ciflow/inductor/169547 2025-12-04T09:33:40.9937327Z * [new tag] ciflow/inductor/169548 -> ciflow/inductor/169548 2025-12-04T09:33:40.9938327Z * [new tag] ciflow/inductor/169549 -> ciflow/inductor/169549 2025-12-04T09:33:40.9939141Z * [new tag] ciflow/inductor/169551 -> ciflow/inductor/169551 2025-12-04T09:33:40.9940013Z * [new tag] ciflow/inductor/169552 -> ciflow/inductor/169552 2025-12-04T09:33:40.9940877Z * [new tag] ciflow/inductor/169553 -> ciflow/inductor/169553 2025-12-04T09:33:40.9941751Z * [new tag] ciflow/inductor/169557 -> ciflow/inductor/169557 2025-12-04T09:33:40.9943309Z * [new tag] ciflow/inductor/3b9a386 -> ciflow/inductor/3b9a386 2025-12-04T09:33:40.9944379Z * [new tag] ciflow/inductor/3d4b92b -> ciflow/inductor/3d4b92b 2025-12-04T09:33:40.9945457Z * [new tag] ciflow/inductor/d224ac7 -> ciflow/inductor/d224ac7 2025-12-04T09:33:40.9946546Z * [new tag] ciflow/linux-aarch64/157994 -> ciflow/linux-aarch64/157994 2025-12-04T09:33:40.9947400Z * [new tag] ciflow/linux-aarch64/166075 -> ciflow/linux-aarch64/166075 2025-12-04T09:33:40.9948149Z * [new tag] ciflow/linux-aarch64/166876 -> ciflow/linux-aarch64/166876 2025-12-04T09:33:40.9948985Z * [new tag] ciflow/linux-aarch64/167981 -> ciflow/linux-aarch64/167981 2025-12-04T09:33:40.9950131Z * [new tag] ciflow/mps/166254 -> ciflow/mps/166254 2025-12-04T09:33:40.9950841Z * [new tag] ciflow/mps/169017 -> ciflow/mps/169017 2025-12-04T09:33:40.9951975Z * [new tag] ciflow/mps/169372 -> ciflow/mps/169372 2025-12-04T09:33:40.9952696Z * [new tag] ciflow/mps/169478 -> ciflow/mps/169478 2025-12-04T09:33:40.9953867Z * [new tag] ciflow/op-benchmark/157994 -> ciflow/op-benchmark/157994 2025-12-04T09:33:40.9954592Z * [new tag] ciflow/op-benchmark/166075 -> ciflow/op-benchmark/166075 2025-12-04T09:33:40.9955378Z * [new tag] ciflow/op-benchmark/169544 -> ciflow/op-benchmark/169544 2025-12-04T09:33:40.9956603Z * [new tag] ciflow/periodic-rocm-mi200/165997 -> ciflow/periodic-rocm-mi200/165997 2025-12-04T09:33:40.9957423Z * [new tag] ciflow/periodic-rocm-mi200/166517 -> ciflow/periodic-rocm-mi200/166517 2025-12-04T09:33:40.9958278Z * [new tag] ciflow/periodic-rocm-mi200/169063 -> ciflow/periodic-rocm-mi200/169063 2025-12-04T09:33:40.9959143Z * [new tag] ciflow/periodic-rocm-mi200/169425 -> ciflow/periodic-rocm-mi200/169425 2025-12-04T09:33:40.9960229Z * [new tag] ciflow/periodic-rocm-mi300/166517 -> ciflow/periodic-rocm-mi300/166517 2025-12-04T09:33:40.9960967Z * [new tag] ciflow/periodic-rocm-mi300/169063 -> ciflow/periodic-rocm-mi300/169063 2025-12-04T09:33:40.9961785Z * [new tag] ciflow/periodic-rocm-mi300/169425 -> ciflow/periodic-rocm-mi300/169425 2025-12-04T09:33:40.9963118Z * [new tag] ciflow/periodic/054a2fd -> ciflow/periodic/054a2fd 2025-12-04T09:33:40.9963829Z * [new tag] ciflow/periodic/167207 -> ciflow/periodic/167207 2025-12-04T09:33:40.9964934Z * [new tag] ciflow/periodic/167978 -> ciflow/periodic/167978 2025-12-04T09:33:40.9965608Z * [new tag] ciflow/periodic/168096 -> ciflow/periodic/168096 2025-12-04T09:33:40.9966454Z * [new tag] ciflow/periodic/169286 -> ciflow/periodic/169286 2025-12-04T09:33:40.9967660Z * [new tag] ciflow/periodic/2a6d37d -> ciflow/periodic/2a6d37d 2025-12-04T09:33:40.9968688Z * [new tag] ciflow/periodic/317eeb8 -> ciflow/periodic/317eeb8 2025-12-04T09:33:40.9969604Z * [new tag] ciflow/periodic/3c32 -> ciflow/periodic/3c32 2025-12-04T09:33:40.9970709Z * [new tag] ciflow/periodic/3e98831 -> ciflow/periodic/3e98831 2025-12-04T09:33:40.9972395Z * [new tag] ciflow/periodic/7c648509a7470ace9fb2bae960dd4790f7e943e9 -> ciflow/periodic/7c648509a7470ace9fb2bae960dd4790f7e943e9 2025-12-04T09:33:40.9973412Z * [new tag] ciflow/periodic/94512-point -> ciflow/periodic/94512-point 2025-12-04T09:33:40.9974791Z * [new tag] ciflow/periodic/csl/test87519 -> ciflow/periodic/csl/test87519 2025-12-04T09:33:40.9975907Z * [new tag] ciflow/periodic/csltest88275 -> ciflow/periodic/csltest88275 2025-12-04T09:33:40.9976795Z * [new tag] ciflow/periodic/csltest88761 -> ciflow/periodic/csltest88761 2025-12-04T09:33:40.9978170Z * [new tag] ciflow/periodic/release_1.12 -> ciflow/periodic/release_1.12 2025-12-04T09:33:40.9979440Z * [new tag] ciflow/periodic/release_1.12.0 -> ciflow/periodic/release_1.12.0 2025-12-04T09:33:40.9980656Z * [new tag] ciflow/periodic/sha-ec5b83 -> ciflow/periodic/sha-ec5b83 2025-12-04T09:33:40.9981564Z * [new tag] ciflow/pull/167207 -> ciflow/pull/167207 2025-12-04T09:33:40.9982977Z * [new tag] ciflow/quantization-periodic/169207 -> ciflow/quantization-periodic/169207 2025-12-04T09:33:40.9983759Z * [new tag] ciflow/rocm-mi200/165545 -> ciflow/rocm-mi200/165545 2025-12-04T09:33:40.9984577Z * [new tag] ciflow/rocm-mi200/165997 -> ciflow/rocm-mi200/165997 2025-12-04T09:33:40.9985410Z * [new tag] ciflow/rocm-mi200/168096 -> ciflow/rocm-mi200/168096 2025-12-04T09:33:40.9986507Z * [new tag] ciflow/rocm-mi200/168275 -> ciflow/rocm-mi200/168275 2025-12-04T09:33:40.9987206Z * [new tag] ciflow/rocm-mi200/169063 -> ciflow/rocm-mi200/169063 2025-12-04T09:33:40.9988306Z * [new tag] ciflow/rocm-mi200/169356 -> ciflow/rocm-mi200/169356 2025-12-04T09:33:40.9989017Z * [new tag] ciflow/rocm-mi200/169425 -> ciflow/rocm-mi200/169425 2025-12-04T09:33:40.9990155Z * [new tag] ciflow/rocm-mi300/165545 -> ciflow/rocm-mi300/165545 2025-12-04T09:33:40.9991203Z * [new tag] ciflow/rocm-mi300/167157 -> ciflow/rocm-mi300/167157 2025-12-04T09:33:40.9991905Z * [new tag] ciflow/rocm-mi300/168096 -> ciflow/rocm-mi300/168096 2025-12-04T09:33:40.9992734Z * [new tag] ciflow/rocm-mi300/169063 -> ciflow/rocm-mi300/169063 2025-12-04T09:33:40.9993560Z * [new tag] ciflow/rocm-mi300/169425 -> ciflow/rocm-mi300/169425 2025-12-04T09:33:40.9994691Z * [new tag] ciflow/rocm-mi355/167157 -> ciflow/rocm-mi355/167157 2025-12-04T09:33:40.9995434Z * [new tag] ciflow/rocm-mi355/168275 -> ciflow/rocm-mi355/168275 2025-12-04T09:33:40.9996386Z * [new tag] ciflow/rocm-mi355/169425 -> ciflow/rocm-mi355/169425 2025-12-04T09:33:40.9997629Z * [new tag] ciflow/rocm-navi31/168275 -> ciflow/rocm-navi31/168275 2025-12-04T09:33:40.9998296Z * [new tag] ciflow/rocm-navi31/169425 -> ciflow/rocm-navi31/169425 2025-12-04T09:33:40.9999406Z * [new tag] ciflow/rocm/115316 -> ciflow/rocm/115316 2025-12-04T09:33:41.0000148Z * [new tag] ciflow/rocm/148492 -> ciflow/rocm/148492 2025-12-04T09:33:41.0000951Z * [new tag] ciflow/rocm/160685 -> ciflow/rocm/160685 2025-12-04T09:33:41.0001792Z * [new tag] ciflow/rocm/161607 -> ciflow/rocm/161607 2025-12-04T09:33:41.0002868Z * [new tag] ciflow/rocm/162052 -> ciflow/rocm/162052 2025-12-04T09:33:41.0003537Z * [new tag] ciflow/rocm/165997 -> ciflow/rocm/165997 2025-12-04T09:33:41.0004498Z * [new tag] ciflow/rocm/166165 -> ciflow/rocm/166165 2025-12-04T09:33:41.0005218Z * [new tag] ciflow/rocm/166517 -> ciflow/rocm/166517 2025-12-04T09:33:41.0005994Z * [new tag] ciflow/rocm/167207 -> ciflow/rocm/167207 2025-12-04T09:33:41.0006811Z * [new tag] ciflow/rocm/167536 -> ciflow/rocm/167536 2025-12-04T09:33:41.0007617Z * [new tag] ciflow/rocm/167781 -> ciflow/rocm/167781 2025-12-04T09:33:41.0008885Z * [new tag] ciflow/rocm/167989 -> ciflow/rocm/167989 2025-12-04T09:33:41.0010009Z * [new tag] ciflow/rocm/168073 -> ciflow/rocm/168073 2025-12-04T09:33:41.0011061Z * [new tag] ciflow/rocm/168195 -> ciflow/rocm/168195 2025-12-04T09:33:41.0011835Z * [new tag] ciflow/rocm/168939 -> ciflow/rocm/168939 2025-12-04T09:33:41.0012894Z * [new tag] ciflow/rocm/168971 -> ciflow/rocm/168971 2025-12-04T09:33:41.0013813Z * [new tag] ciflow/rocm/169024 -> ciflow/rocm/169024 2025-12-04T09:33:41.0014446Z * [new tag] ciflow/rocm/169200 -> ciflow/rocm/169200 2025-12-04T09:33:41.0015301Z * [new tag] ciflow/rocm/169216 -> ciflow/rocm/169216 2025-12-04T09:33:41.0016156Z * [new tag] ciflow/rocm/169312 -> ciflow/rocm/169312 2025-12-04T09:33:41.0017072Z * [new tag] ciflow/rocm/169380 -> ciflow/rocm/169380 2025-12-04T09:33:41.0018156Z * [new tag] ciflow/rocm/169427 -> ciflow/rocm/169427 2025-12-04T09:33:41.0018972Z * [new tag] ciflow/rocm/169455 -> ciflow/rocm/169455 2025-12-04T09:33:41.0020043Z * [new tag] ciflow/rocm/169470 -> ciflow/rocm/169470 2025-12-04T09:33:41.0020804Z * [new tag] ciflow/rocm/169471 -> ciflow/rocm/169471 2025-12-04T09:33:41.0021815Z * [new tag] ciflow/rocm/169472 -> ciflow/rocm/169472 2025-12-04T09:33:41.0022536Z * [new tag] ciflow/rocm/169514 -> ciflow/rocm/169514 2025-12-04T09:33:41.0023873Z * [new tag] ciflow/slow/01c7106 -> ciflow/slow/01c7106 2025-12-04T09:33:41.0024830Z * [new tag] ciflow/slow/0577043 -> ciflow/slow/0577043 2025-12-04T09:33:41.0026468Z * [new tag] ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym -> ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym 2025-12-04T09:33:41.0026944Z * [new tag] ciflow/slow/0e81104 -> ciflow/slow/0e81104 2025-12-04T09:33:41.0027777Z * [new tag] ciflow/slow/167207 -> ciflow/slow/167207 2025-12-04T09:33:41.0028604Z * [new tag] ciflow/slow/168050 -> ciflow/slow/168050 2025-12-04T09:33:41.0029696Z * [new tag] ciflow/slow/1732077 -> ciflow/slow/1732077 2025-12-04T09:33:41.0030831Z * [new tag] ciflow/slow/187eb7c -> ciflow/slow/187eb7c 2025-12-04T09:33:41.0032246Z * [new tag] ciflow/slow/1faef89 -> ciflow/slow/1faef89 2025-12-04T09:33:41.0033528Z * [new tag] ciflow/slow/3920ec1 -> ciflow/slow/3920ec1 2025-12-04T09:33:41.0034741Z * [new tag] ciflow/slow/3b7c6b2 -> ciflow/slow/3b7c6b2 2025-12-04T09:33:41.0035806Z * [new tag] ciflow/slow/59a3759 -> ciflow/slow/59a3759 2025-12-04T09:33:41.0036883Z * [new tag] ciflow/slow/70ef0bb -> ciflow/slow/70ef0bb 2025-12-04T09:33:41.0037865Z * [new tag] ciflow/slow/788ff06 -> ciflow/slow/788ff06 2025-12-04T09:33:41.0039525Z * [new tag] ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym -> ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym 2025-12-04T09:33:41.0040047Z * [new tag] ciflow/slow/9d85864 -> ciflow/slow/9d85864 2025-12-04T09:33:41.0041247Z * [new tag] ciflow/slow/9ffad5b -> ciflow/slow/9ffad5b 2025-12-04T09:33:41.0042086Z * [new tag] ciflow/slow/a206e8b -> ciflow/slow/a206e8b 2025-12-04T09:33:41.0043288Z * [new tag] ciflow/slow/a837609 -> ciflow/slow/a837609 2025-12-04T09:33:41.0044356Z * [new tag] ciflow/slow/af841f3 -> ciflow/slow/af841f3 2025-12-04T09:33:41.0045990Z * [new tag] ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym -> ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym 2025-12-04T09:33:41.0046668Z * [new tag] ciflow/torchbench/168175 -> ciflow/torchbench/168175 2025-12-04T09:33:41.0047809Z * [new tag] ciflow/trunk/148492 -> ciflow/trunk/148492 2025-12-04T09:33:41.0048462Z * [new tag] ciflow/trunk/157149 -> ciflow/trunk/157149 2025-12-04T09:33:41.0049457Z * [new tag] ciflow/trunk/157994 -> ciflow/trunk/157994 2025-12-04T09:33:41.0050181Z * [new tag] ciflow/trunk/159718 -> ciflow/trunk/159718 2025-12-04T09:33:41.0051031Z * [new tag] ciflow/trunk/160685 -> ciflow/trunk/160685 2025-12-04T09:33:41.0051816Z * [new tag] ciflow/trunk/160729 -> ciflow/trunk/160729 2025-12-04T09:33:41.0052709Z * [new tag] ciflow/trunk/162275 -> ciflow/trunk/162275 2025-12-04T09:33:41.0053519Z * [new tag] ciflow/trunk/162795 -> ciflow/trunk/162795 2025-12-04T09:33:41.0054367Z * [new tag] ciflow/trunk/163245 -> ciflow/trunk/163245 2025-12-04T09:33:41.0055172Z * [new tag] ciflow/trunk/163942 -> ciflow/trunk/163942 2025-12-04T09:33:41.0055995Z * [new tag] ciflow/trunk/165274 -> ciflow/trunk/165274 2025-12-04T09:33:41.0057487Z * [new tag] ciflow/trunk/165483 -> ciflow/trunk/165483 2025-12-04T09:33:41.0058702Z * [new tag] ciflow/trunk/165728 -> ciflow/trunk/165728 2025-12-04T09:33:41.0059783Z * [new tag] ciflow/trunk/165922 -> ciflow/trunk/165922 2025-12-04T09:33:41.0060559Z * [new tag] ciflow/trunk/166075 -> ciflow/trunk/166075 2025-12-04T09:33:41.0061520Z * [new tag] ciflow/trunk/166165 -> ciflow/trunk/166165 2025-12-04T09:33:41.0062559Z * [new tag] ciflow/trunk/166829 -> ciflow/trunk/166829 2025-12-04T09:33:41.0063427Z * [new tag] ciflow/trunk/166843 -> ciflow/trunk/166843 2025-12-04T09:33:41.0064416Z * [new tag] ciflow/trunk/166876 -> ciflow/trunk/166876 2025-12-04T09:33:41.0065204Z * [new tag] ciflow/trunk/167207 -> ciflow/trunk/167207 2025-12-04T09:33:41.0066234Z * [new tag] ciflow/trunk/167536 -> ciflow/trunk/167536 2025-12-04T09:33:41.0067003Z * [new tag] ciflow/trunk/167552 -> ciflow/trunk/167552 2025-12-04T09:33:41.0067957Z * [new tag] ciflow/trunk/167555 -> ciflow/trunk/167555 2025-12-04T09:33:41.0068801Z * [new tag] ciflow/trunk/167599 -> ciflow/trunk/167599 2025-12-04T09:33:41.0069793Z * [new tag] ciflow/trunk/167659 -> ciflow/trunk/167659 2025-12-04T09:33:41.0070670Z * [new tag] ciflow/trunk/167672 -> ciflow/trunk/167672 2025-12-04T09:33:41.0071639Z * [new tag] ciflow/trunk/167742 -> ciflow/trunk/167742 2025-12-04T09:33:41.0072430Z * [new tag] ciflow/trunk/167781 -> ciflow/trunk/167781 2025-12-04T09:33:41.0073638Z * [new tag] ciflow/trunk/167837 -> ciflow/trunk/167837 2025-12-04T09:33:41.0074403Z * [new tag] ciflow/trunk/167887 -> ciflow/trunk/167887 2025-12-04T09:33:41.0075376Z * [new tag] ciflow/trunk/167978 -> ciflow/trunk/167978 2025-12-04T09:33:41.0076225Z * [new tag] ciflow/trunk/168050 -> ciflow/trunk/168050 2025-12-04T09:33:41.0077028Z * [new tag] ciflow/trunk/168051 -> ciflow/trunk/168051 2025-12-04T09:33:41.0078025Z * [new tag] ciflow/trunk/168096 -> ciflow/trunk/168096 2025-12-04T09:33:41.0078751Z * [new tag] ciflow/trunk/168127 -> ciflow/trunk/168127 2025-12-04T09:33:41.0079767Z * [new tag] ciflow/trunk/168157 -> ciflow/trunk/168157 2025-12-04T09:33:41.0080521Z * [new tag] ciflow/trunk/168175 -> ciflow/trunk/168175 2025-12-04T09:33:41.0081525Z * [new tag] ciflow/trunk/168209 -> ciflow/trunk/168209 2025-12-04T09:33:41.0082511Z * [new tag] ciflow/trunk/168213 -> ciflow/trunk/168213 2025-12-04T09:33:41.0083487Z * [new tag] ciflow/trunk/168226 -> ciflow/trunk/168226 2025-12-04T09:33:41.0084283Z * [new tag] ciflow/trunk/168262 -> ciflow/trunk/168262 2025-12-04T09:33:41.0085146Z * [new tag] ciflow/trunk/168275 -> ciflow/trunk/168275 2025-12-04T09:33:41.0086275Z * [new tag] ciflow/trunk/168328 -> ciflow/trunk/168328 2025-12-04T09:33:41.0087024Z * [new tag] ciflow/trunk/168368 -> ciflow/trunk/168368 2025-12-04T09:33:41.0087999Z * [new tag] ciflow/trunk/168917 -> ciflow/trunk/168917 2025-12-04T09:33:41.0088786Z * [new tag] ciflow/trunk/168933 -> ciflow/trunk/168933 2025-12-04T09:33:41.0089904Z * [new tag] ciflow/trunk/168941 -> ciflow/trunk/168941 2025-12-04T09:33:41.0090682Z * [new tag] ciflow/trunk/168955 -> ciflow/trunk/168955 2025-12-04T09:33:41.0091747Z * [new tag] ciflow/trunk/168980 -> ciflow/trunk/168980 2025-12-04T09:33:41.0092905Z * [new tag] ciflow/trunk/169004 -> ciflow/trunk/169004 2025-12-04T09:33:41.0093659Z * [new tag] ciflow/trunk/169006 -> ciflow/trunk/169006 2025-12-04T09:33:41.0094685Z * [new tag] ciflow/trunk/169023 -> ciflow/trunk/169023 2025-12-04T09:33:41.0095410Z * [new tag] ciflow/trunk/169025 -> ciflow/trunk/169025 2025-12-04T09:33:41.0096756Z * [new tag] ciflow/trunk/169048 -> ciflow/trunk/169048 2025-12-04T09:33:41.0097483Z * [new tag] ciflow/trunk/169066 -> ciflow/trunk/169066 2025-12-04T09:33:41.0098300Z * [new tag] ciflow/trunk/169091 -> ciflow/trunk/169091 2025-12-04T09:33:41.0099284Z * [new tag] ciflow/trunk/169102 -> ciflow/trunk/169102 2025-12-04T09:33:41.0100635Z * [new tag] ciflow/trunk/169103 -> ciflow/trunk/169103 2025-12-04T09:33:41.0101653Z * [new tag] ciflow/trunk/169125 -> ciflow/trunk/169125 2025-12-04T09:33:41.0102697Z * [new tag] ciflow/trunk/169139 -> ciflow/trunk/169139 2025-12-04T09:33:41.0103703Z * [new tag] ciflow/trunk/169148 -> ciflow/trunk/169148 2025-12-04T09:33:41.0104501Z * [new tag] ciflow/trunk/169151 -> ciflow/trunk/169151 2025-12-04T09:33:41.0105518Z * [new tag] ciflow/trunk/169156 -> ciflow/trunk/169156 2025-12-04T09:33:41.0106575Z * [new tag] ciflow/trunk/169176 -> ciflow/trunk/169176 2025-12-04T09:33:41.0107322Z * [new tag] ciflow/trunk/169204 -> ciflow/trunk/169204 2025-12-04T09:33:41.0108270Z * [new tag] ciflow/trunk/169207 -> ciflow/trunk/169207 2025-12-04T09:33:41.0109090Z * [new tag] ciflow/trunk/169211 -> ciflow/trunk/169211 2025-12-04T09:33:41.0110299Z * [new tag] ciflow/trunk/169231 -> ciflow/trunk/169231 2025-12-04T09:33:41.0111188Z * [new tag] ciflow/trunk/169260 -> ciflow/trunk/169260 2025-12-04T09:33:41.0112391Z * [new tag] ciflow/trunk/169271 -> ciflow/trunk/169271 2025-12-04T09:33:41.0113162Z * [new tag] ciflow/trunk/169280 -> ciflow/trunk/169280 2025-12-04T09:33:41.0114227Z * [new tag] ciflow/trunk/169281 -> ciflow/trunk/169281 2025-12-04T09:33:41.0114981Z * [new tag] ciflow/trunk/169286 -> ciflow/trunk/169286 2025-12-04T09:33:41.0116143Z * [new tag] ciflow/trunk/169293 -> ciflow/trunk/169293 2025-12-04T09:33:41.0116906Z * [new tag] ciflow/trunk/169296 -> ciflow/trunk/169296 2025-12-04T09:33:41.0117896Z * [new tag] ciflow/trunk/169304 -> ciflow/trunk/169304 2025-12-04T09:33:41.0118660Z * [new tag] ciflow/trunk/169305 -> ciflow/trunk/169305 2025-12-04T09:33:41.0119668Z * [new tag] ciflow/trunk/169312 -> ciflow/trunk/169312 2025-12-04T09:33:41.0120846Z * [new tag] ciflow/trunk/169328 -> ciflow/trunk/169328 2025-12-04T09:33:41.0121616Z * [new tag] ciflow/trunk/169343 -> ciflow/trunk/169343 2025-12-04T09:33:41.0122717Z * [new tag] ciflow/trunk/169355 -> ciflow/trunk/169355 2025-12-04T09:33:41.0123495Z * [new tag] ciflow/trunk/169370 -> ciflow/trunk/169370 2025-12-04T09:33:41.0124652Z * [new tag] ciflow/trunk/169379 -> ciflow/trunk/169379 2025-12-04T09:33:41.0125401Z * [new tag] ciflow/trunk/169380 -> ciflow/trunk/169380 2025-12-04T09:33:41.0126376Z * [new tag] ciflow/trunk/169385 -> ciflow/trunk/169385 2025-12-04T09:33:41.0127165Z * [new tag] ciflow/trunk/169387 -> ciflow/trunk/169387 2025-12-04T09:33:41.0128349Z * [new tag] ciflow/trunk/169410 -> ciflow/trunk/169410 2025-12-04T09:33:41.0129138Z * [new tag] ciflow/trunk/169412 -> ciflow/trunk/169412 2025-12-04T09:33:41.0130000Z * [new tag] ciflow/trunk/169418 -> ciflow/trunk/169418 2025-12-04T09:33:41.0131072Z * [new tag] ciflow/trunk/169423 -> ciflow/trunk/169423 2025-12-04T09:33:41.0131738Z * [new tag] ciflow/trunk/169427 -> ciflow/trunk/169427 2025-12-04T09:33:41.0132769Z * [new tag] ciflow/trunk/169430 -> ciflow/trunk/169430 2025-12-04T09:33:41.0133559Z * [new tag] ciflow/trunk/169437 -> ciflow/trunk/169437 2025-12-04T09:33:41.0134539Z * [new tag] ciflow/trunk/169442 -> ciflow/trunk/169442 2025-12-04T09:33:41.0135299Z * [new tag] ciflow/trunk/169452 -> ciflow/trunk/169452 2025-12-04T09:33:41.0136365Z * [new tag] ciflow/trunk/169454 -> ciflow/trunk/169454 2025-12-04T09:33:41.0137164Z * [new tag] ciflow/trunk/169459 -> ciflow/trunk/169459 2025-12-04T09:33:41.0138380Z * [new tag] ciflow/trunk/169474 -> ciflow/trunk/169474 2025-12-04T09:33:41.0139149Z * [new tag] ciflow/trunk/169475 -> ciflow/trunk/169475 2025-12-04T09:33:41.0140138Z * [new tag] ciflow/trunk/169476 -> ciflow/trunk/169476 2025-12-04T09:33:41.0141175Z * [new tag] ciflow/trunk/169487 -> ciflow/trunk/169487 2025-12-04T09:33:41.0141984Z * [new tag] ciflow/trunk/169497 -> ciflow/trunk/169497 2025-12-04T09:33:41.0142856Z * [new tag] ciflow/trunk/169503 -> ciflow/trunk/169503 2025-12-04T09:33:41.0143718Z * [new tag] ciflow/trunk/169505 -> ciflow/trunk/169505 2025-12-04T09:33:41.0144605Z * [new tag] ciflow/trunk/169507 -> ciflow/trunk/169507 2025-12-04T09:33:41.0145603Z * [new tag] ciflow/trunk/169514 -> ciflow/trunk/169514 2025-12-04T09:33:41.0146467Z * [new tag] ciflow/trunk/169517 -> ciflow/trunk/169517 2025-12-04T09:33:41.0147283Z * [new tag] ciflow/trunk/169519 -> ciflow/trunk/169519 2025-12-04T09:33:41.0148145Z * [new tag] ciflow/trunk/169528 -> ciflow/trunk/169528 2025-12-04T09:33:41.0149010Z * [new tag] ciflow/trunk/169541 -> ciflow/trunk/169541 2025-12-04T09:33:41.0150157Z * [new tag] ciflow/trunk/169555 -> ciflow/trunk/169555 2025-12-04T09:33:41.0151566Z * [new tag] ciflow/unstable/123 -> ciflow/unstable/123 2025-12-04T09:33:41.0152569Z * [new tag] ciflow/vllm/165270 -> ciflow/vllm/165270 2025-12-04T09:33:41.0153303Z * [new tag] ciflow/vllm/165274 -> ciflow/vllm/165274 2025-12-04T09:33:41.0154108Z * [new tag] ciflow/vllm/166494 -> ciflow/vllm/166494 2025-12-04T09:33:41.0154980Z * [new tag] ciflow/vllm/169219 -> ciflow/vllm/169219 2025-12-04T09:33:41.0155771Z * [new tag] ciflow/vllm/169220 -> ciflow/vllm/169220 2025-12-04T09:33:41.0156877Z * [new tag] ciflow/xpu/157994 -> ciflow/xpu/157994 2025-12-04T09:33:41.0157605Z * [new tag] ciflow/xpu/159718 -> ciflow/xpu/159718 2025-12-04T09:33:41.0158459Z * [new tag] ciflow/xpu/161940 -> ciflow/xpu/161940 2025-12-04T09:33:41.0159513Z * [new tag] ciflow/xpu/163251 -> ciflow/xpu/163251 2025-12-04T09:33:41.0160190Z * [new tag] ciflow/xpu/166829 -> ciflow/xpu/166829 2025-12-04T09:33:41.0161029Z * [new tag] ciflow/xpu/166843 -> ciflow/xpu/166843 2025-12-04T09:33:41.0161882Z * [new tag] ciflow/xpu/167972 -> ciflow/xpu/167972 2025-12-04T09:33:41.0162671Z * [new tag] ciflow/xpu/167981 -> ciflow/xpu/167981 2025-12-04T09:33:41.0163504Z * [new tag] ciflow/xpu/168213 -> ciflow/xpu/168213 2025-12-04T09:33:41.0164303Z * [new tag] ciflow/xpu/168262 -> ciflow/xpu/168262 2025-12-04T09:33:41.0165281Z * [new tag] ciflow/xpu/168328 -> ciflow/xpu/168328 2025-12-04T09:33:41.0166369Z * [new tag] ciflow/xpu/168950 -> ciflow/xpu/168950 2025-12-04T09:33:41.0167671Z * [new tag] ciflow/xpu/169039 -> ciflow/xpu/169039 2025-12-04T09:33:41.0168720Z * [new tag] ciflow/xpu/169200 -> ciflow/xpu/169200 2025-12-04T09:33:41.0169503Z * [new tag] ciflow/xpu/169203 -> ciflow/xpu/169203 2025-12-04T09:33:41.0171009Z * [new tag] ciflow/xpu/169230 -> ciflow/xpu/169230 2025-12-04T09:33:41.0171776Z * [new tag] ciflow/xpu/169231 -> ciflow/xpu/169231 2025-12-04T09:33:41.0172917Z * [new tag] ciflow/xpu/169241 -> ciflow/xpu/169241 2025-12-04T09:33:41.0173683Z * [new tag] ciflow/xpu/169280 -> ciflow/xpu/169280 2025-12-04T09:33:41.0174520Z * [new tag] ciflow/xpu/169296 -> ciflow/xpu/169296 2025-12-04T09:33:41.0175649Z * [new tag] ciflow/xpu/169353 -> ciflow/xpu/169353 2025-12-04T09:33:41.0176439Z * [new tag] ciflow/xpu/169410 -> ciflow/xpu/169410 2025-12-04T09:33:41.0177530Z * [new tag] ciflow/xpu/169442 -> ciflow/xpu/169442 2025-12-04T09:33:41.0178335Z * [new tag] ciflow/xpu/169555 -> ciflow/xpu/169555 2025-12-04T09:33:41.0179568Z * [new tag] cslpull75 -> cslpull75 2025-12-04T09:33:41.0180391Z * [new tag] cslpull76 -> cslpull76 2025-12-04T09:33:41.0181530Z * [new tag] cslpull77 -> cslpull77 2025-12-04T09:33:41.0182414Z * [new tag] cslpull78 -> cslpull78 2025-12-04T09:33:41.0183688Z * [new tag] cslpull79 -> cslpull79 2025-12-04T09:33:41.0184923Z * [new tag] cslpull80 -> cslpull80 2025-12-04T09:33:41.0185950Z * [new tag] cslpull81 -> cslpull81 2025-12-04T09:33:41.0186766Z * [new tag] cslpull82 -> cslpull82 2025-12-04T09:33:41.0187924Z * [new tag] cslpull83 -> cslpull83 2025-12-04T09:33:41.0188935Z * [new tag] cslpull84 -> cslpull84 2025-12-04T09:33:41.0189770Z * [new tag] cslpull85 -> cslpull85 2025-12-04T09:33:41.0190873Z * [new tag] cslpull86 -> cslpull86 2025-12-04T09:33:41.0191864Z * [new tag] cslpull87 -> cslpull87 2025-12-04T09:33:41.0192866Z * [new tag] cslpull88 -> cslpull88 2025-12-04T09:33:41.0193852Z * [new tag] cslpull89 -> cslpull89 2025-12-04T09:33:41.0194472Z * [new tag] cslpull90 -> cslpull90 2025-12-04T09:33:41.0196170Z * [new tag] cslpull91 -> cslpull91 2025-12-04T09:33:41.0197105Z * [new tag] cslpull92 -> cslpull92 2025-12-04T09:33:41.0198265Z * [new tag] flight_5 -> flight_5 2025-12-04T09:33:41.0199432Z * [new tag] flight_5.1 -> flight_5.1 2025-12-04T09:33:41.0200434Z * [new tag] flight_5.2 -> flight_5.2 2025-12-04T09:33:41.0201481Z * [new tag] flight_5.3 -> flight_5.3 2025-12-04T09:33:41.0202539Z * [new tag] forpull1 -> forpull1 2025-12-04T09:33:41.0203794Z * [new tag] malfet/tag-2ef5611 -> malfet/tag-2ef5611 2025-12-04T09:33:41.0204822Z * [new tag] malfet/tag-317b1a0 -> malfet/tag-317b1a0 2025-12-04T09:33:41.0205667Z * [new tag] malfet/tag-ec6f767 -> malfet/tag-ec6f767 2025-12-04T09:33:41.0206994Z * [new tag] nightly-binary -> nightly-binary 2025-12-04T09:33:41.0208101Z * [new tag] sqzhang_flight4_plus -> sqzhang_flight4_plus 2025-12-04T09:33:41.0209217Z * [new tag] sqzhang_flight_3 -> sqzhang_flight_3 2025-12-04T09:33:41.0210653Z * [new tag] trunk/02d8bd6974cf84b721680d773dbdb1b6f40ce272 -> trunk/02d8bd6974cf84b721680d773dbdb1b6f40ce272 2025-12-04T09:33:41.0211633Z * [new tag] trunk/066997fb38ade71e00d78e9d572e380b5f02bd3e -> trunk/066997fb38ade71e00d78e9d572e380b5f02bd3e 2025-12-04T09:33:41.0213064Z * [new tag] trunk/076e7b19fa1d481ad778d06d2b49ba57d3ce8c88 -> trunk/076e7b19fa1d481ad778d06d2b49ba57d3ce8c88 2025-12-04T09:33:41.0214283Z * [new tag] trunk/07dcc0b83db3211653a38565a24e15acdba75654 -> trunk/07dcc0b83db3211653a38565a24e15acdba75654 2025-12-04T09:33:41.0215233Z * [new tag] trunk/082e96b68dfcd16cab7cfafc4d3d055767dab3eb -> trunk/082e96b68dfcd16cab7cfafc4d3d055767dab3eb 2025-12-04T09:33:41.0216236Z * [new tag] trunk/088048f2fea28ff7d450f65c72419ca45780d30b -> trunk/088048f2fea28ff7d450f65c72419ca45780d30b 2025-12-04T09:33:41.0217306Z * [new tag] trunk/09076941a95c76f4d9ad189d064dfd8baa39e672 -> trunk/09076941a95c76f4d9ad189d064dfd8baa39e672 2025-12-04T09:33:41.0218354Z * [new tag] trunk/0b80a4c62b94402844bf221791c096b0035c6d75 -> trunk/0b80a4c62b94402844bf221791c096b0035c6d75 2025-12-04T09:33:41.0219760Z * [new tag] trunk/0bbbdf1750567a980634ad907a325357ba8ba8f2 -> trunk/0bbbdf1750567a980634ad907a325357ba8ba8f2 2025-12-04T09:33:41.0220802Z * [new tag] trunk/0c281dd78773b2bc17c58ead0e4cd4ac46e775c5 -> trunk/0c281dd78773b2bc17c58ead0e4cd4ac46e775c5 2025-12-04T09:33:41.0222333Z * [new tag] trunk/135f3753c418a6879b1954904184937b67e61688 -> trunk/135f3753c418a6879b1954904184937b67e61688 2025-12-04T09:33:41.0223323Z * [new tag] trunk/15da21026cb13cd20257dc9e96830db108743c10 -> trunk/15da21026cb13cd20257dc9e96830db108743c10 2025-12-04T09:33:41.0224414Z * [new tag] trunk/166efdad2ac827f30fb02504c6017520257f88ec -> trunk/166efdad2ac827f30fb02504c6017520257f88ec 2025-12-04T09:33:41.0225382Z * [new tag] trunk/174272c15fae553d8488140af931f7d8050a313f -> trunk/174272c15fae553d8488140af931f7d8050a313f 2025-12-04T09:33:41.0226724Z * [new tag] trunk/18f3ca08f13b8de61307f5e8cd7d4cccb67e9d11 -> trunk/18f3ca08f13b8de61307f5e8cd7d4cccb67e9d11 2025-12-04T09:33:41.0227672Z * [new tag] trunk/1902eddfe655a15ebcf2c72bd81ade110fdeef63 -> trunk/1902eddfe655a15ebcf2c72bd81ade110fdeef63 2025-12-04T09:33:41.0228672Z * [new tag] trunk/195f92e98d3d66738577f11f22c4b5c8a1c76dd5 -> trunk/195f92e98d3d66738577f11f22c4b5c8a1c76dd5 2025-12-04T09:33:41.0229693Z * [new tag] trunk/1aa13e17de39e3c768ea7aebaad166ce72a06676 -> trunk/1aa13e17de39e3c768ea7aebaad166ce72a06676 2025-12-04T09:33:41.0230711Z * [new tag] trunk/1afe2832f58e24e54a5bfda5a5afa9b96fdea40e -> trunk/1afe2832f58e24e54a5bfda5a5afa9b96fdea40e 2025-12-04T09:33:41.0231686Z * [new tag] trunk/1c87554d74140eaee964ca8b1832cede67f5f520 -> trunk/1c87554d74140eaee964ca8b1832cede67f5f520 2025-12-04T09:33:41.0232738Z * [new tag] trunk/1ccb743b7b5be955f49736c162c4f5004b8a0dd8 -> trunk/1ccb743b7b5be955f49736c162c4f5004b8a0dd8 2025-12-04T09:33:41.0233985Z * [new tag] trunk/1cee47d6ce0a02227185b566593f002dd639ca0c -> trunk/1cee47d6ce0a02227185b566593f002dd639ca0c 2025-12-04T09:33:41.0234776Z * [new tag] trunk/1d21b4df2babe322e5d085ceb6de884eb260a62d -> trunk/1d21b4df2babe322e5d085ceb6de884eb260a62d 2025-12-04T09:33:41.0235836Z * [new tag] trunk/1e34fb2550e4aa650314f7a6d9f6daf4da7478a8 -> trunk/1e34fb2550e4aa650314f7a6d9f6daf4da7478a8 2025-12-04T09:33:41.0236920Z * [new tag] trunk/1e526fb5b1d93bfc70691c5c3955fdffc1b7b7de -> trunk/1e526fb5b1d93bfc70691c5c3955fdffc1b7b7de 2025-12-04T09:33:41.0237956Z * [new tag] trunk/1ee32a8b1f554a312d79bad01ded24f38cd95543 -> trunk/1ee32a8b1f554a312d79bad01ded24f38cd95543 2025-12-04T09:33:41.0239415Z * [new tag] trunk/201e2c4117eb9744594dad6a5c18213d7b4705d7 -> trunk/201e2c4117eb9744594dad6a5c18213d7b4705d7 2025-12-04T09:33:41.0240380Z * [new tag] trunk/2353a0f60eb4b4cb6675907a7fa9fbedc1c02e7f -> trunk/2353a0f60eb4b4cb6675907a7fa9fbedc1c02e7f 2025-12-04T09:33:41.0241597Z * [new tag] trunk/285779b1621cf9f073a062b0889a642d200308d9 -> trunk/285779b1621cf9f073a062b0889a642d200308d9 2025-12-04T09:33:41.0242426Z * [new tag] trunk/2887faaec6295d081580d09fce161201826c6d87 -> trunk/2887faaec6295d081580d09fce161201826c6d87 2025-12-04T09:33:41.0243402Z * [new tag] trunk/296e67c92635443c67b11c0ae1bd045f03ebb7bc -> trunk/296e67c92635443c67b11c0ae1bd045f03ebb7bc 2025-12-04T09:33:41.0244487Z * [new tag] trunk/29856679769b3dede478767e2fe6cfb51197cb25 -> trunk/29856679769b3dede478767e2fe6cfb51197cb25 2025-12-04T09:33:41.0245561Z * [new tag] trunk/29e5455a4740c326ab187c7aa7b5ef98034ea563 -> trunk/29e5455a4740c326ab187c7aa7b5ef98034ea563 2025-12-04T09:33:41.0246798Z * [new tag] trunk/2ac3ef882afb23136adc188975f0a8802fc68adf -> trunk/2ac3ef882afb23136adc188975f0a8802fc68adf 2025-12-04T09:33:41.0247572Z * [new tag] trunk/2bec68e73b64715354af076ad309335f943e36cd -> trunk/2bec68e73b64715354af076ad309335f943e36cd 2025-12-04T09:33:41.0248579Z * [new tag] trunk/2c87367e6f88662cd5cedbd1537748b7948c38e1 -> trunk/2c87367e6f88662cd5cedbd1537748b7948c38e1 2025-12-04T09:33:41.0249696Z * [new tag] trunk/2d1f78fe3ec13820f136a2e0336da12a25f41708 -> trunk/2d1f78fe3ec13820f136a2e0336da12a25f41708 2025-12-04T09:33:41.0250950Z * [new tag] trunk/2df6058f116a65722a0e03073402feb242572d35 -> trunk/2df6058f116a65722a0e03073402feb242572d35 2025-12-04T09:33:41.0252030Z * [new tag] trunk/2e0c2e170fe658c440775c8e5c44228aafcc47ec -> trunk/2e0c2e170fe658c440775c8e5c44228aafcc47ec 2025-12-04T09:33:41.0253422Z * [new tag] trunk/2f9b7dad7b5419b063bd0f2e204de192720ebb94 -> trunk/2f9b7dad7b5419b063bd0f2e204de192720ebb94 2025-12-04T09:33:41.0254332Z * [new tag] trunk/305168768a95d69c444df5cd334bb774edfe06f1 -> trunk/305168768a95d69c444df5cd334bb774edfe06f1 2025-12-04T09:33:41.0255454Z * [new tag] trunk/31fc12773026e8e00f054dd79ad9b2491e693b48 -> trunk/31fc12773026e8e00f054dd79ad9b2491e693b48 2025-12-04T09:33:41.0256455Z * [new tag] trunk/320de0c6b0a3e7c6d2693ea5c28d5d0156ba7991 -> trunk/320de0c6b0a3e7c6d2693ea5c28d5d0156ba7991 2025-12-04T09:33:41.0257585Z * [new tag] trunk/3418bd29475dff06695045fcdf93e7d0dac67da8 -> trunk/3418bd29475dff06695045fcdf93e7d0dac67da8 2025-12-04T09:33:41.0258603Z * [new tag] trunk/34a98608afa0cb5b48f0d6d30432fdd0a2614ddf -> trunk/34a98608afa0cb5b48f0d6d30432fdd0a2614ddf 2025-12-04T09:33:41.0259569Z * [new tag] trunk/35b7a9a26c5923d98aebaa41a031dae21788a9ee -> trunk/35b7a9a26c5923d98aebaa41a031dae21788a9ee 2025-12-04T09:33:41.0260668Z * [new tag] trunk/39d07dbf03a911bdd45d1af78d8638dc92074938 -> trunk/39d07dbf03a911bdd45d1af78d8638dc92074938 2025-12-04T09:33:41.0261522Z * [new tag] trunk/3cd98b4205ada151042cc7ff097a82d4a4b18725 -> trunk/3cd98b4205ada151042cc7ff097a82d4a4b18725 2025-12-04T09:33:41.0262579Z * [new tag] trunk/3d35fd20a78ff4d016fa80f4e5fad37191d7bcae -> trunk/3d35fd20a78ff4d016fa80f4e5fad37191d7bcae 2025-12-04T09:33:41.0263591Z * [new tag] trunk/409a5fee945c46a3edaf5df162812f201bfd7b2f -> trunk/409a5fee945c46a3edaf5df162812f201bfd7b2f 2025-12-04T09:33:41.0264585Z * [new tag] trunk/42e9005cda22da3f1c559c3649218cebd671027c -> trunk/42e9005cda22da3f1c559c3649218cebd671027c 2025-12-04T09:33:41.0267751Z * [new tag] trunk/43b94713bbf340d3c124fde02d0f73add4021247 -> trunk/43b94713bbf340d3c124fde02d0f73add4021247 2025-12-04T09:33:41.0268232Z * [new tag] trunk/44ac69388a4a5eb463dbd2a13f00d1e3b924566c -> trunk/44ac69388a4a5eb463dbd2a13f00d1e3b924566c 2025-12-04T09:33:41.0268695Z * [new tag] trunk/45d14e2497292be06ad36eaa1aaaf7c630a2586a -> trunk/45d14e2497292be06ad36eaa1aaaf7c630a2586a 2025-12-04T09:33:41.0269168Z * [new tag] trunk/45d310ad84854dff730c0b12e577d7998d978686 -> trunk/45d310ad84854dff730c0b12e577d7998d978686 2025-12-04T09:33:41.0270117Z * [new tag] trunk/47b28ddf7bd74b50fa93b307a7d3b183a6d77f54 -> trunk/47b28ddf7bd74b50fa93b307a7d3b183a6d77f54 2025-12-04T09:33:41.0270938Z * [new tag] trunk/481e5ab336275bd3acd5fa8a611b05b4469012af -> trunk/481e5ab336275bd3acd5fa8a611b05b4469012af 2025-12-04T09:33:41.0272007Z * [new tag] trunk/491731647f6b8a9345dcfb3bc9416aea254a7d96 -> trunk/491731647f6b8a9345dcfb3bc9416aea254a7d96 2025-12-04T09:33:41.0273281Z * [new tag] trunk/49a04d26088acc17d948ddd66920f3e16371e873 -> trunk/49a04d26088acc17d948ddd66920f3e16371e873 2025-12-04T09:33:41.0274225Z * [new tag] trunk/4bebc827c47d2f1f0fa1a417a5201a97aef3d985 -> trunk/4bebc827c47d2f1f0fa1a417a5201a97aef3d985 2025-12-04T09:33:41.0275091Z * [new tag] trunk/4c246677784c6a14bc2dbb9ff8773ef0a3a3222f -> trunk/4c246677784c6a14bc2dbb9ff8773ef0a3a3222f 2025-12-04T09:33:41.0276469Z * [new tag] trunk/4cfb47ff548b6d996641058cf04a70e311a4c3aa -> trunk/4cfb47ff548b6d996641058cf04a70e311a4c3aa 2025-12-04T09:33:41.0277458Z * [new tag] trunk/4e0061c1aa52f606dda8cfab0bd7591e588faf2c -> trunk/4e0061c1aa52f606dda8cfab0bd7591e588faf2c 2025-12-04T09:33:41.0279067Z * [new tag] trunk/4fefb8e7e942386ffac764a41b232241f82bea3a -> trunk/4fefb8e7e942386ffac764a41b232241f82bea3a 2025-12-04T09:33:41.0280040Z * [new tag] trunk/503b2640023521f5a35cd9a52fc8033d73a95d0d -> trunk/503b2640023521f5a35cd9a52fc8033d73a95d0d 2025-12-04T09:33:41.0281303Z * [new tag] trunk/518c2b1b3dab9a2ef2849e04b3bc2f20c1c41db9 -> trunk/518c2b1b3dab9a2ef2849e04b3bc2f20c1c41db9 2025-12-04T09:33:41.0282227Z * [new tag] trunk/5191b2fa68ba19960912bfd7fd721c79d76bb1f3 -> trunk/5191b2fa68ba19960912bfd7fd721c79d76bb1f3 2025-12-04T09:33:41.0283506Z * [new tag] trunk/52ac0f0dc4acacd219f1317fbc28ec631c01e07a -> trunk/52ac0f0dc4acacd219f1317fbc28ec631c01e07a 2025-12-04T09:33:41.0284542Z * [new tag] trunk/539ba711b029de9f191070f4f0d12f18f5b7f292 -> trunk/539ba711b029de9f191070f4f0d12f18f5b7f292 2025-12-04T09:33:41.0285571Z * [new tag] trunk/556375b55deebebbc56cb7aef81f4d52f031ba28 -> trunk/556375b55deebebbc56cb7aef81f4d52f031ba28 2025-12-04T09:33:41.0286681Z * [new tag] trunk/55c4ab554845481d0a69a3811937575fe8bb1a66 -> trunk/55c4ab554845481d0a69a3811937575fe8bb1a66 2025-12-04T09:33:41.0287609Z * [new tag] trunk/5634469fda9e5d98869c82c7d03bb08914245f96 -> trunk/5634469fda9e5d98869c82c7d03bb08914245f96 2025-12-04T09:33:41.0288497Z * [new tag] trunk/5778f6ff894686a975a9a23645178ae4c87ad5dc -> trunk/5778f6ff894686a975a9a23645178ae4c87ad5dc 2025-12-04T09:33:41.0289558Z * [new tag] trunk/587d63a3e07de5dc91065f9ef70bcacda9989068 -> trunk/587d63a3e07de5dc91065f9ef70bcacda9989068 2025-12-04T09:33:41.0291189Z * [new tag] trunk/597930f6b568852356ca9795dac76f9e4653adbd -> trunk/597930f6b568852356ca9795dac76f9e4653adbd 2025-12-04T09:33:41.0291924Z * [new tag] trunk/597df3a4e2a67b9fdbe1a89b2f4d74f822274db6 -> trunk/597df3a4e2a67b9fdbe1a89b2f4d74f822274db6 2025-12-04T09:33:41.0293242Z * [new tag] trunk/59abd50e931f4efb21b053f7a2911f5d8a49d883 -> trunk/59abd50e931f4efb21b053f7a2911f5d8a49d883 2025-12-04T09:33:41.0294187Z * [new tag] trunk/5a607febc04c3a2b5824c75f3f60307867439a2c -> trunk/5a607febc04c3a2b5824c75f3f60307867439a2c 2025-12-04T09:33:41.0295285Z * [new tag] trunk/5bf1cdf4755c54ef462b44cb8041b0a57311556b -> trunk/5bf1cdf4755c54ef462b44cb8041b0a57311556b 2025-12-04T09:33:41.0296348Z * [new tag] trunk/5f0030ba63d334d7e8c93a09e41403b89e4c573c -> trunk/5f0030ba63d334d7e8c93a09e41403b89e4c573c 2025-12-04T09:33:41.0297506Z * [new tag] trunk/5f21d27e71268464d362a96c9ac09ea475f7f202 -> trunk/5f21d27e71268464d362a96c9ac09ea475f7f202 2025-12-04T09:33:41.0298856Z * [new tag] trunk/5fafc13038c9988d9ac21fa793fbd5890604b447 -> trunk/5fafc13038c9988d9ac21fa793fbd5890604b447 2025-12-04T09:33:41.0299723Z * [new tag] trunk/61be54a31dc09b59d99b62176fb935aee0b924ef -> trunk/61be54a31dc09b59d99b62176fb935aee0b924ef 2025-12-04T09:33:41.0300755Z * [new tag] trunk/62d3ccd71484ed6a760d909b41487101bbc65719 -> trunk/62d3ccd71484ed6a760d909b41487101bbc65719 2025-12-04T09:33:41.0301769Z * [new tag] trunk/641cdb68ae27668eb441d0e49c87a0602c120c2b -> trunk/641cdb68ae27668eb441d0e49c87a0602c120c2b 2025-12-04T09:33:41.0302851Z * [new tag] trunk/65c4620d6bb0c6029f69762c22b91dda2294da9a -> trunk/65c4620d6bb0c6029f69762c22b91dda2294da9a 2025-12-04T09:33:41.0303921Z * [new tag] trunk/66004b993744b4106bf8afaba71f3c228a804206 -> trunk/66004b993744b4106bf8afaba71f3c228a804206 2025-12-04T09:33:41.0304957Z * [new tag] trunk/6658a04c7ca67acb64512341342e7b3ee13ee386 -> trunk/6658a04c7ca67acb64512341342e7b3ee13ee386 2025-12-04T09:33:41.0305981Z * [new tag] trunk/6864e309092a71f8ab0ca6a4dc7f8a4073fd31c4 -> trunk/6864e309092a71f8ab0ca6a4dc7f8a4073fd31c4 2025-12-04T09:33:41.0307171Z * [new tag] trunk/6c261c6cb07892c90ca19ed51c9705b1659a3f7d -> trunk/6c261c6cb07892c90ca19ed51c9705b1659a3f7d 2025-12-04T09:33:41.0308085Z * [new tag] trunk/6c8b6a043f1628188b6396b3a2a6e000ca68362b -> trunk/6c8b6a043f1628188b6396b3a2a6e000ca68362b 2025-12-04T09:33:41.0309097Z * [new tag] trunk/6ceb4a32f92ae67ce5d7d97931d17401ebf5ffa5 -> trunk/6ceb4a32f92ae67ce5d7d97931d17401ebf5ffa5 2025-12-04T09:33:41.0310164Z * [new tag] trunk/6e404e9b7d6f5fb0de86aa73888c3038248c17f8 -> trunk/6e404e9b7d6f5fb0de86aa73888c3038248c17f8 2025-12-04T09:33:41.0311196Z * [new tag] trunk/6ec30b490aee1db6bcdc7340abddef25784f08ec -> trunk/6ec30b490aee1db6bcdc7340abddef25784f08ec 2025-12-04T09:33:41.0312216Z * [new tag] trunk/6f2783a6c08e1db34275ff25176ffe9aebc30a71 -> trunk/6f2783a6c08e1db34275ff25176ffe9aebc30a71 2025-12-04T09:33:41.0313275Z * [new tag] trunk/6f53fefeb90ad3281119b5cfc4aa9ffd8a066e3d -> trunk/6f53fefeb90ad3281119b5cfc4aa9ffd8a066e3d 2025-12-04T09:33:41.0314316Z * [new tag] trunk/6f7dcf51e46d0c880db1a2f5c70de57adb576f4a -> trunk/6f7dcf51e46d0c880db1a2f5c70de57adb576f4a 2025-12-04T09:33:41.0315682Z * [new tag] trunk/6ff831180d2fa436c7f1c1af3adac641fce9d60e -> trunk/6ff831180d2fa436c7f1c1af3adac641fce9d60e 2025-12-04T09:33:41.0316591Z * [new tag] trunk/70076464a63ab218a7ceefb0e76ccd7131deb8f8 -> trunk/70076464a63ab218a7ceefb0e76ccd7131deb8f8 2025-12-04T09:33:41.0317605Z * [new tag] trunk/70d797a5fc109b20a517646fcaa819477cd0d485 -> trunk/70d797a5fc109b20a517646fcaa819477cd0d485 2025-12-04T09:33:41.0318609Z * [new tag] trunk/7348cb355ff0a6f79cd4871215aea72185748734 -> trunk/7348cb355ff0a6f79cd4871215aea72185748734 2025-12-04T09:33:41.0319684Z * [new tag] trunk/74fe26a1ebe32931783569f2e762e3c2c974901f -> trunk/74fe26a1ebe32931783569f2e762e3c2c974901f 2025-12-04T09:33:41.0320762Z * [new tag] trunk/76aeb8c7e0f795b3fddca134cbea9a69da3ee696 -> trunk/76aeb8c7e0f795b3fddca134cbea9a69da3ee696 2025-12-04T09:33:41.0321597Z * [new tag] trunk/7716da9fb23f27a65b41f9f016a2afadf281c18f -> trunk/7716da9fb23f27a65b41f9f016a2afadf281c18f 2025-12-04T09:33:41.0322705Z * [new tag] trunk/7741edd4ed665f3988052e260863efb508d61a03 -> trunk/7741edd4ed665f3988052e260863efb508d61a03 2025-12-04T09:33:41.0323804Z * [new tag] trunk/78adb3b3df41b45d2368b67226d2f864b78939a6 -> trunk/78adb3b3df41b45d2368b67226d2f864b78939a6 2025-12-04T09:33:41.0324876Z * [new tag] trunk/79d7b178225e5ed24d4e1db74e5abbff848f5fb7 -> trunk/79d7b178225e5ed24d4e1db74e5abbff848f5fb7 2025-12-04T09:33:41.0325733Z * [new tag] trunk/7a1e316115fc6996b3f2336822ba5d5f6179f0c3 -> trunk/7a1e316115fc6996b3f2336822ba5d5f6179f0c3 2025-12-04T09:33:41.0326796Z * [new tag] trunk/7a41b66367c38d0af3e8a90f7be48d6b281e7bca -> trunk/7a41b66367c38d0af3e8a90f7be48d6b281e7bca 2025-12-04T09:33:41.0327772Z * [new tag] trunk/7b7af390ea8541c611d1ce2018a6934188fc197b -> trunk/7b7af390ea8541c611d1ce2018a6934188fc197b 2025-12-04T09:33:41.0328826Z * [new tag] trunk/7ba4680f3755a560af81aa0f688791e367aa3609 -> trunk/7ba4680f3755a560af81aa0f688791e367aa3609 2025-12-04T09:33:41.0330171Z * [new tag] trunk/7bc2a66ded06a0b2549aa51d807edc5dc3e73d1b -> trunk/7bc2a66ded06a0b2549aa51d807edc5dc3e73d1b 2025-12-04T09:33:41.0330924Z * [new tag] trunk/7c648509a7470ace9fb2bae960dd4790f7e943e9 -> trunk/7c648509a7470ace9fb2bae960dd4790f7e943e9 2025-12-04T09:33:41.0331872Z * [new tag] trunk/7cbc2d034cecd21ab5c9707d0a9c525c17143fb8 -> trunk/7cbc2d034cecd21ab5c9707d0a9c525c17143fb8 2025-12-04T09:33:41.0332893Z * [new tag] trunk/7d1bbaf4ba301ea3fba6f3c7bc02d58f6417aaed -> trunk/7d1bbaf4ba301ea3fba6f3c7bc02d58f6417aaed 2025-12-04T09:33:41.0333973Z * [new tag] trunk/7d2a33e4ebf60b217a3cd77feae19231eb996fc8 -> trunk/7d2a33e4ebf60b217a3cd77feae19231eb996fc8 2025-12-04T09:33:41.0334958Z * [new tag] trunk/7eb625920054b1126a7d2d99818aaa188c6ba95e -> trunk/7eb625920054b1126a7d2d99818aaa188c6ba95e 2025-12-04T09:33:41.0335914Z * [new tag] trunk/7f55ba19c456a3d6cc443dd9edb6bb7cca677ead -> trunk/7f55ba19c456a3d6cc443dd9edb6bb7cca677ead 2025-12-04T09:33:41.0337006Z * [new tag] trunk/81af382128efa094d8702e18f2c133760904c718 -> trunk/81af382128efa094d8702e18f2c133760904c718 2025-12-04T09:33:41.0338595Z * [new tag] trunk/84149583d483e9c973c9a0feda70e4f3964947b0 -> trunk/84149583d483e9c973c9a0feda70e4f3964947b0 2025-12-04T09:33:41.0339958Z * [new tag] trunk/85a315917efe82c24306be805c584ec044951c75 -> trunk/85a315917efe82c24306be805c584ec044951c75 2025-12-04T09:33:41.0340954Z * [new tag] trunk/87329491c82a5f8c1cc4ec11d8f55a5de2551ece -> trunk/87329491c82a5f8c1cc4ec11d8f55a5de2551ece 2025-12-04T09:33:41.0341882Z * [new tag] trunk/892640e25aeefa8007c5af837214b4502b6b62a6 -> trunk/892640e25aeefa8007c5af837214b4502b6b62a6 2025-12-04T09:33:41.0343264Z * [new tag] trunk/89e3bbcb5b5321dc8b9520b4d5a8ee60cea1d0b4 -> trunk/89e3bbcb5b5321dc8b9520b4d5a8ee60cea1d0b4 2025-12-04T09:33:41.0344151Z * [new tag] trunk/8c73bbbb02159223c0c97d268a0a74cb78158a1c -> trunk/8c73bbbb02159223c0c97d268a0a74cb78158a1c 2025-12-04T09:33:41.0345251Z * [new tag] trunk/8d56e98c8db988a22cb2dfaeefb30bc7d2a3cc43 -> trunk/8d56e98c8db988a22cb2dfaeefb30bc7d2a3cc43 2025-12-04T09:33:41.0346544Z * [new tag] trunk/8d9dd9603e5ee26c01007f0cd4f018e584840922 -> trunk/8d9dd9603e5ee26c01007f0cd4f018e584840922 2025-12-04T09:33:41.0347467Z * [new tag] trunk/8ef0c0b02b062d75e7c9be2594914a3e784d23ca -> trunk/8ef0c0b02b062d75e7c9be2594914a3e784d23ca 2025-12-04T09:33:41.0348541Z * [new tag] trunk/90b27e7e8352cde97d32ddad24740ef819633f38 -> trunk/90b27e7e8352cde97d32ddad24740ef819633f38 2025-12-04T09:33:41.0349520Z * [new tag] trunk/90f0139e64b2951815d524b6a373bed20c4fbf90 -> trunk/90f0139e64b2951815d524b6a373bed20c4fbf90 2025-12-04T09:33:41.0350441Z * [new tag] trunk/93d0d6838c56af59b0dba794e6aa08f0c1c7799c -> trunk/93d0d6838c56af59b0dba794e6aa08f0c1c7799c 2025-12-04T09:33:41.0351589Z * [new tag] trunk/94ca8d5f1e81fea3ae488650a0fb6795049a9f87 -> trunk/94ca8d5f1e81fea3ae488650a0fb6795049a9f87 2025-12-04T09:33:41.0352532Z * [new tag] trunk/9844fbeadd5cebdf1281d6fbf79164139c352693 -> trunk/9844fbeadd5cebdf1281d6fbf79164139c352693 2025-12-04T09:33:41.0353602Z * [new tag] trunk/99024dec888ec1e50b546822a32b6fb2f35e5eaa -> trunk/99024dec888ec1e50b546822a32b6fb2f35e5eaa 2025-12-04T09:33:41.0354625Z * [new tag] trunk/9a296e640fc88aa44d275b48cd9cc30c573b169d -> trunk/9a296e640fc88aa44d275b48cd9cc30c573b169d 2025-12-04T09:33:41.0355704Z * [new tag] trunk/9b3e34d8589b29f7b4e7fab6f78711b7ca6e4639 -> trunk/9b3e34d8589b29f7b4e7fab6f78711b7ca6e4639 2025-12-04T09:33:41.0356711Z * [new tag] trunk/9cd055e547e9b67a5f9827f8999c38d7eda1bcb8 -> trunk/9cd055e547e9b67a5f9827f8999c38d7eda1bcb8 2025-12-04T09:33:41.0357788Z * [new tag] trunk/9f0df5686cb4ada94f94620acba2e3c3f363b11d -> trunk/9f0df5686cb4ada94f94620acba2e3c3f363b11d 2025-12-04T09:33:41.0359365Z * [new tag] trunk/9f7fceb887d0cfa0326a59b887821c63ff11340a -> trunk/9f7fceb887d0cfa0326a59b887821c63ff11340a 2025-12-04T09:33:41.0360367Z * [new tag] trunk/9f8ef8855d3078d70f7b782540ff2aaf158d6742 -> trunk/9f8ef8855d3078d70f7b782540ff2aaf158d6742 2025-12-04T09:33:41.0361648Z * [new tag] trunk/9fb52efc797b47a1f425a03aa5e47b866d8b1098 -> trunk/9fb52efc797b47a1f425a03aa5e47b866d8b1098 2025-12-04T09:33:41.0362563Z * [new tag] trunk/9ff4a2ebc5762d46c73e46b1b523d7ff349fedfa -> trunk/9ff4a2ebc5762d46c73e46b1b523d7ff349fedfa 2025-12-04T09:33:41.0363968Z * [new tag] trunk/a0f3937b94422354538ebbd47202d5b0e8a3fd0d -> trunk/a0f3937b94422354538ebbd47202d5b0e8a3fd0d 2025-12-04T09:33:41.0364767Z * [new tag] trunk/a15066c28b3145e6edbfc88359d0411d14cfc70c -> trunk/a15066c28b3145e6edbfc88359d0411d14cfc70c 2025-12-04T09:33:41.0365818Z * [new tag] trunk/a20f775e82564d2a9979221ed7f3b8d7cf54ce90 -> trunk/a20f775e82564d2a9979221ed7f3b8d7cf54ce90 2025-12-04T09:33:41.0366844Z * [new tag] trunk/a2973fb00ec002dd4b6bbf07385f066efb259b8c -> trunk/a2973fb00ec002dd4b6bbf07385f066efb259b8c 2025-12-04T09:33:41.0367788Z * [new tag] trunk/a7dc6dab9ad911259d4801c502907e531594db45 -> trunk/a7dc6dab9ad911259d4801c502907e531594db45 2025-12-04T09:33:41.0368866Z * [new tag] trunk/a951a9cee65c01660bbc6e6fded90ecb10fa6109 -> trunk/a951a9cee65c01660bbc6e6fded90ecb10fa6109 2025-12-04T09:33:41.0369949Z * [new tag] trunk/abfa1a6d65c7c159e35c72c25979b9da4971689e -> trunk/abfa1a6d65c7c159e35c72c25979b9da4971689e 2025-12-04T09:33:41.0371203Z * [new tag] trunk/ae3a2395bf66151078e2d201716f7d63ce1c6f3e -> trunk/ae3a2395bf66151078e2d201716f7d63ce1c6f3e 2025-12-04T09:33:41.0372024Z * [new tag] trunk/afdff7f0325080dedac44d080cb5a3b0e65e6c5e -> trunk/afdff7f0325080dedac44d080cb5a3b0e65e6c5e 2025-12-04T09:33:41.0372982Z * [new tag] trunk/b1aed4e7a72c03a38f44543aaea0dae2e9b76d48 -> trunk/b1aed4e7a72c03a38f44543aaea0dae2e9b76d48 2025-12-04T09:33:41.0374102Z * [new tag] trunk/b1decff555cd50e2123c8c6e25cc0d447c411f62 -> trunk/b1decff555cd50e2123c8c6e25cc0d447c411f62 2025-12-04T09:33:41.0375353Z * [new tag] trunk/b2b6b034c9fd08672c40e63ef243556ad4c49bd2 -> trunk/b2b6b034c9fd08672c40e63ef243556ad4c49bd2 2025-12-04T09:33:41.0376298Z * [new tag] trunk/b39813b4a04931682b0491adba2138d01d716d99 -> trunk/b39813b4a04931682b0491adba2138d01d716d99 2025-12-04T09:33:41.0377447Z * [new tag] trunk/b3a7edb2311367974cc7cd764cfb11a5d6758b24 -> trunk/b3a7edb2311367974cc7cd764cfb11a5d6758b24 2025-12-04T09:33:41.0396229Z * [new tag] trunk/b4cc1329c86acaef6d42c1fac7169b8d870ab0d7 -> trunk/b4cc1329c86acaef6d42c1fac7169b8d870ab0d7 2025-12-04T09:33:41.0396897Z * [new tag] trunk/b555c39217f765759954a4f9f9bd1e9b87bed11a -> trunk/b555c39217f765759954a4f9f9bd1e9b87bed11a 2025-12-04T09:33:41.0397366Z * [new tag] trunk/b6b6c80379388b7f9932c3e6a0f9907bf430e417 -> trunk/b6b6c80379388b7f9932c3e6a0f9907bf430e417 2025-12-04T09:33:41.0397826Z * [new tag] trunk/b6b6d912df0b6f4082f8e50b18bd1de1dd7325f4 -> trunk/b6b6d912df0b6f4082f8e50b18bd1de1dd7325f4 2025-12-04T09:33:41.0398290Z * [new tag] trunk/b7d60685f8cbc939b68a20871e90db67e729329b -> trunk/b7d60685f8cbc939b68a20871e90db67e729329b 2025-12-04T09:33:41.0398747Z * [new tag] trunk/b7f6b9a4fc6259f7af068f31868b3119bb1bac3e -> trunk/b7f6b9a4fc6259f7af068f31868b3119bb1bac3e 2025-12-04T09:33:41.0399203Z * [new tag] trunk/b8c4ba3593761e7b2a3ebd86f040fb07b47c02cf -> trunk/b8c4ba3593761e7b2a3ebd86f040fb07b47c02cf 2025-12-04T09:33:41.0399671Z * [new tag] trunk/b9c8f3a4884befb965ff42620ce44a71b04887f5 -> trunk/b9c8f3a4884befb965ff42620ce44a71b04887f5 2025-12-04T09:33:41.0400115Z * [new tag] trunk/ba1412546f3082c0958c077acc2025e4dbc33f1f -> trunk/ba1412546f3082c0958c077acc2025e4dbc33f1f 2025-12-04T09:33:41.0400586Z * [new tag] trunk/bac403c0b38c63bdbcc0c31f1c2b0bc0260f610f -> trunk/bac403c0b38c63bdbcc0c31f1c2b0bc0260f610f 2025-12-04T09:33:41.0401105Z * [new tag] trunk/bb3034198b459401fabeab254e1b99f0115046e2 -> trunk/bb3034198b459401fabeab254e1b99f0115046e2 2025-12-04T09:33:41.0401656Z * [new tag] trunk/bc39b2b3bc7a6e19a42e62bd576974035086fe55 -> trunk/bc39b2b3bc7a6e19a42e62bd576974035086fe55 2025-12-04T09:33:41.0402309Z * [new tag] trunk/bc43d5b297f207a11d83d77ddf0152bdaabe15a8 -> trunk/bc43d5b297f207a11d83d77ddf0152bdaabe15a8 2025-12-04T09:33:41.0402764Z * [new tag] trunk/bc6a4863c7246a6493d16d4ea6eee71ec07c6a09 -> trunk/bc6a4863c7246a6493d16d4ea6eee71ec07c6a09 2025-12-04T09:33:41.0403244Z * [new tag] trunk/bea4912944defdbcb8b061800caab6cbbbd01df5 -> trunk/bea4912944defdbcb8b061800caab6cbbbd01df5 2025-12-04T09:33:41.0403694Z * [new tag] trunk/c04e2c656f48d82d1521b867bbbf03967b9b7564 -> trunk/c04e2c656f48d82d1521b867bbbf03967b9b7564 2025-12-04T09:33:41.0404135Z * [new tag] trunk/c0660bcee27e7d7731634e274576a7081882bede -> trunk/c0660bcee27e7d7731634e274576a7081882bede 2025-12-04T09:33:41.0404614Z * [new tag] trunk/c178ed43d3d99cbefe84fbfb21d6f282b20d62ac -> trunk/c178ed43d3d99cbefe84fbfb21d6f282b20d62ac 2025-12-04T09:33:41.0405059Z * [new tag] trunk/c55b1e8f61d041ee436d697449eb028931d574fb -> trunk/c55b1e8f61d041ee436d697449eb028931d574fb 2025-12-04T09:33:41.0405526Z * [new tag] trunk/c6ae7579fe12fe75f1a8f7043a494c90567273f1 -> trunk/c6ae7579fe12fe75f1a8f7043a494c90567273f1 2025-12-04T09:33:41.0405983Z * [new tag] trunk/c8210e7d94bad5ae21ac389fa4ba8a463c76c4d0 -> trunk/c8210e7d94bad5ae21ac389fa4ba8a463c76c4d0 2025-12-04T09:33:41.0406535Z * [new tag] trunk/cc0853af42122f8185321f542616f4474e717f09 -> trunk/cc0853af42122f8185321f542616f4474e717f09 2025-12-04T09:33:41.0407060Z * [new tag] trunk/cddec6562eabfa390d014fa3741a5659cf9c94c9 -> trunk/cddec6562eabfa390d014fa3741a5659cf9c94c9 2025-12-04T09:33:41.0408152Z * [new tag] trunk/ce5e7e3bf1f4b69a4f4f93d288ba75b906df492a -> trunk/ce5e7e3bf1f4b69a4f4f93d288ba75b906df492a 2025-12-04T09:33:41.0409440Z * [new tag] trunk/d038b0130ec7c20ebcac219301292fd8e98a1ace -> trunk/d038b0130ec7c20ebcac219301292fd8e98a1ace 2025-12-04T09:33:41.0410243Z * [new tag] trunk/d16447dacaf2420ea175f0c275c75da951f57d39 -> trunk/d16447dacaf2420ea175f0c275c75da951f57d39 2025-12-04T09:33:41.0411325Z * [new tag] trunk/d19f1e8cab6810bb2e99141f9976665954c67a50 -> trunk/d19f1e8cab6810bb2e99141f9976665954c67a50 2025-12-04T09:33:41.0412386Z * [new tag] trunk/d1c9f03b2a5af4104721712f8cdffe9b4f340c01 -> trunk/d1c9f03b2a5af4104721712f8cdffe9b4f340c01 2025-12-04T09:33:41.0413690Z * [new tag] trunk/d40f4950f2b7f7aa380a22fe0f6166e71680fbcf -> trunk/d40f4950f2b7f7aa380a22fe0f6166e71680fbcf 2025-12-04T09:33:41.0414597Z * [new tag] trunk/d5038950bacfe36bbf24a47a455fe76901deb8e8 -> trunk/d5038950bacfe36bbf24a47a455fe76901deb8e8 2025-12-04T09:33:41.0415576Z * [new tag] trunk/d54ff42903c2ae0533931ff11d23b35f875bdb3d -> trunk/d54ff42903c2ae0533931ff11d23b35f875bdb3d 2025-12-04T09:33:41.0416651Z * [new tag] trunk/d76697633a2d2b9cced1ae21161849b33bfe7e47 -> trunk/d76697633a2d2b9cced1ae21161849b33bfe7e47 2025-12-04T09:33:41.0417807Z * [new tag] trunk/d78f52b199c547106d4cd9d2856dd0805c118bf1 -> trunk/d78f52b199c547106d4cd9d2856dd0805c118bf1 2025-12-04T09:33:41.0418928Z * [new tag] trunk/d8fd5c6eed28e5004150691d048a3f6785e19a8e -> trunk/d8fd5c6eed28e5004150691d048a3f6785e19a8e 2025-12-04T09:33:41.0419995Z * [new tag] trunk/d900f5e86745dec76713f4b0ef07005ef36b2f5a -> trunk/d900f5e86745dec76713f4b0ef07005ef36b2f5a 2025-12-04T09:33:41.0421091Z * [new tag] trunk/d973dc6b87d763859fe1c5bd1287e3b6b1c49d1b -> trunk/d973dc6b87d763859fe1c5bd1287e3b6b1c49d1b 2025-12-04T09:33:41.0422164Z * [new tag] trunk/d998c03304cb6ede76e1ed535b4ddeb6c2bf40ec -> trunk/d998c03304cb6ede76e1ed535b4ddeb6c2bf40ec 2025-12-04T09:33:41.0423494Z * [new tag] trunk/d9cb8a70833101dbbe16b99520cfbdd70d0a87bf -> trunk/d9cb8a70833101dbbe16b99520cfbdd70d0a87bf 2025-12-04T09:33:41.0424468Z * [new tag] trunk/d9d5e91b43f70eb8637af55db6856d49be391ffd -> trunk/d9d5e91b43f70eb8637af55db6856d49be391ffd 2025-12-04T09:33:41.0425533Z * [new tag] trunk/dd18a75336a4fbd7497955cc5665904724fce889 -> trunk/dd18a75336a4fbd7497955cc5665904724fce889 2025-12-04T09:33:41.0426565Z * [new tag] trunk/ded9bcd61a059bf723e6e84689552962b480ea77 -> trunk/ded9bcd61a059bf723e6e84689552962b480ea77 2025-12-04T09:33:41.0427658Z * [new tag] trunk/dfbd3714d15c37a7b83b322a6b60f997fc00f50c -> trunk/dfbd3714d15c37a7b83b322a6b60f997fc00f50c 2025-12-04T09:33:41.0428944Z * [new tag] trunk/e115f9f4e4b039f8e9a642aaa2bd8254a920541b -> trunk/e115f9f4e4b039f8e9a642aaa2bd8254a920541b 2025-12-04T09:33:41.0429808Z * [new tag] trunk/e3f24fd73ad74c6e7176687986436956c7c18235 -> trunk/e3f24fd73ad74c6e7176687986436956c7c18235 2025-12-04T09:33:41.0431037Z * [new tag] trunk/e7d24d3ff93d1503ba63860b7057438ad93f918e -> trunk/e7d24d3ff93d1503ba63860b7057438ad93f918e 2025-12-04T09:33:41.0432082Z * [new tag] trunk/ea7035f462a0d2830865ee86c832bd101e1427fc -> trunk/ea7035f462a0d2830865ee86c832bd101e1427fc 2025-12-04T09:33:41.0433746Z * [new tag] trunk/eabb7ad2128580ef674446027b95bcf4e21e8df3 -> trunk/eabb7ad2128580ef674446027b95bcf4e21e8df3 2025-12-04T09:33:41.0434778Z * [new tag] trunk/eb5c63652a33da42e7018c23df5f20a3eb4c6ccf -> trunk/eb5c63652a33da42e7018c23df5f20a3eb4c6ccf 2025-12-04T09:33:41.0435923Z * [new tag] trunk/ec2c71f5c85021b8938cdafadce24c15a36fd93e -> trunk/ec2c71f5c85021b8938cdafadce24c15a36fd93e 2025-12-04T09:33:41.0436951Z * [new tag] trunk/ecbcc3f6bf327856b435b259ac63cc2f328c4b4e -> trunk/ecbcc3f6bf327856b435b259ac63cc2f328c4b4e 2025-12-04T09:33:41.0438556Z * [new tag] trunk/ee87bbe876c42575e961b32a0827d76bc9782ca2 -> trunk/ee87bbe876c42575e961b32a0827d76bc9782ca2 2025-12-04T09:33:41.0439558Z * [new tag] trunk/ef019d1d431c4c5a95b594cb90d40a50cd00f5e4 -> trunk/ef019d1d431c4c5a95b594cb90d40a50cd00f5e4 2025-12-04T09:33:41.0440616Z * [new tag] trunk/ef8ecc13830a86c4b231f1aad9aba7851db61b53 -> trunk/ef8ecc13830a86c4b231f1aad9aba7851db61b53 2025-12-04T09:33:41.0441679Z * [new tag] trunk/f1076f5510920044912247b1abb8760cb820f598 -> trunk/f1076f5510920044912247b1abb8760cb820f598 2025-12-04T09:33:41.0442897Z * [new tag] trunk/f2d6a75a00a1d648ca9a0abc6a33e14c3dea6c40 -> trunk/f2d6a75a00a1d648ca9a0abc6a33e14c3dea6c40 2025-12-04T09:33:41.0443882Z * [new tag] trunk/f47dd0ddef1359e5b43e4b962412f67b30ecde56 -> trunk/f47dd0ddef1359e5b43e4b962412f67b30ecde56 2025-12-04T09:33:41.0445087Z * [new tag] trunk/f49d32dfa4730dcfb1b60eeeb369b5889da983c8 -> trunk/f49d32dfa4730dcfb1b60eeeb369b5889da983c8 2025-12-04T09:33:41.0446021Z * [new tag] trunk/f4dedf78fc30fd4b93975787ca6074ee89db9467 -> trunk/f4dedf78fc30fd4b93975787ca6074ee89db9467 2025-12-04T09:33:41.0447096Z * [new tag] trunk/f7c0d03819ebed05c4038f095d66d1b8c54aca17 -> trunk/f7c0d03819ebed05c4038f095d66d1b8c54aca17 2025-12-04T09:33:41.0448145Z * [new tag] trunk/f7e1bd80a063e17453c361837ba6ea2570920a73 -> trunk/f7e1bd80a063e17453c361837ba6ea2570920a73 2025-12-04T09:33:41.0449082Z * [new tag] trunk/f9bd6c53624c7c0ea3772de78498326e84c2f0e7 -> trunk/f9bd6c53624c7c0ea3772de78498326e84c2f0e7 2025-12-04T09:33:41.0450298Z * [new tag] trunk/fb5be221a46b51bfc9509013b0d85bc5a9d4f15b -> trunk/fb5be221a46b51bfc9509013b0d85bc5a9d4f15b 2025-12-04T09:33:41.0451385Z * [new tag] trunk/fdf863d5e1de3b2688c9511e96876e34581dbfd7 -> trunk/fdf863d5e1de3b2688c9511e96876e34581dbfd7 2025-12-04T09:33:41.0453023Z * [new tag] trunk/fe0e65adfc0e7ca6e5f57e6ea8b16bd5cc967307 -> trunk/fe0e65adfc0e7ca6e5f57e6ea8b16bd5cc967307 2025-12-04T09:33:41.0454031Z * [new tag] trunk/fec710bf89173f5355468a7ce1afe9157c3d9009 -> trunk/fec710bf89173f5355468a7ce1afe9157c3d9009 2025-12-04T09:33:41.0455175Z * [new tag] trunk/ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 -> trunk/ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:33:41.0455900Z * [new tag] v0.1.1 -> v0.1.1 2025-12-04T09:33:41.0457052Z * [new tag] v0.1.10 -> v0.1.10 2025-12-04T09:33:41.0458112Z * [new tag] v0.1.11 -> v0.1.11 2025-12-04T09:33:41.0459259Z * [new tag] v0.1.12 -> v0.1.12 2025-12-04T09:33:41.0460221Z * [new tag] v0.1.2 -> v0.1.2 2025-12-04T09:33:41.0461160Z * [new tag] v0.1.3 -> v0.1.3 2025-12-04T09:33:41.0462069Z * [new tag] v0.1.4 -> v0.1.4 2025-12-04T09:33:41.0463020Z * [new tag] v0.1.5 -> v0.1.5 2025-12-04T09:33:41.0463999Z * [new tag] v0.1.6 -> v0.1.6 2025-12-04T09:33:41.0464923Z * [new tag] v0.1.7 -> v0.1.7 2025-12-04T09:33:41.0465843Z * [new tag] v0.1.8 -> v0.1.8 2025-12-04T09:33:41.0467288Z * [new tag] v0.1.9 -> v0.1.9 2025-12-04T09:33:41.0468316Z * [new tag] v0.2.0 -> v0.2.0 2025-12-04T09:33:41.0469363Z * [new tag] v0.3.0 -> v0.3.0 2025-12-04T09:33:41.0470462Z * [new tag] v0.3.1 -> v0.3.1 2025-12-04T09:33:41.0471452Z * [new tag] v0.4.0 -> v0.4.0 2025-12-04T09:33:41.0472398Z * [new tag] v0.4.1 -> v0.4.1 2025-12-04T09:33:41.0473354Z * [new tag] v1.0.0 -> v1.0.0 2025-12-04T09:33:41.0474356Z * [new tag] v1.0.0a0 -> v1.0.0a0 2025-12-04T09:33:41.0475337Z * [new tag] v1.0.1 -> v1.0.1 2025-12-04T09:33:41.0476352Z * [new tag] v1.0rc0 -> v1.0rc0 2025-12-04T09:33:41.0477078Z * [new tag] v1.0rc1 -> v1.0rc1 2025-12-04T09:33:41.0478198Z * [new tag] v1.1.0 -> v1.1.0 2025-12-04T09:33:41.0479218Z * [new tag] v1.1.0a0 -> v1.1.0a0 2025-12-04T09:33:41.0480425Z * [new tag] v1.10.0 -> v1.10.0 2025-12-04T09:33:41.0481484Z * [new tag] v1.10.0-rc1 -> v1.10.0-rc1 2025-12-04T09:33:41.0482633Z * [new tag] v1.10.0-rc2 -> v1.10.0-rc2 2025-12-04T09:33:41.0483375Z * [new tag] v1.10.0-rc3 -> v1.10.0-rc3 2025-12-04T09:33:41.0484467Z * [new tag] v1.10.1 -> v1.10.1 2025-12-04T09:33:41.0485220Z * [new tag] v1.10.1-rc1 -> v1.10.1-rc1 2025-12-04T09:33:41.0486010Z * [new tag] v1.10.2 -> v1.10.2 2025-12-04T09:33:41.0486807Z * [new tag] v1.10.2-rc1 -> v1.10.2-rc1 2025-12-04T09:33:41.0487901Z * [new tag] v1.11.0 -> v1.11.0 2025-12-04T09:33:41.0489005Z * [new tag] v1.11.0-rc1 -> v1.11.0-rc1 2025-12-04T09:33:41.0490172Z * [new tag] v1.11.0-rc2 -> v1.11.0-rc2 2025-12-04T09:33:41.0491367Z * [new tag] v1.11.0-rc3 -> v1.11.0-rc3 2025-12-04T09:33:41.0492465Z * [new tag] v1.11.0-rc4 -> v1.11.0-rc4 2025-12-04T09:33:41.0493545Z * [new tag] v1.11.0-rc5 -> v1.11.0-rc5 2025-12-04T09:33:41.0494330Z * [new tag] v1.11.0-rc6 -> v1.11.0-rc6 2025-12-04T09:33:41.0495157Z * [new tag] v1.11.0-rc7 -> v1.11.0-rc7 2025-12-04T09:33:41.0496552Z * [new tag] v1.12.0 -> v1.12.0 2025-12-04T09:33:41.0497645Z * [new tag] v1.12.0-rc1 -> v1.12.0-rc1 2025-12-04T09:33:41.0498694Z * [new tag] v1.12.0-rc2 -> v1.12.0-rc2 2025-12-04T09:33:41.0499743Z * [new tag] v1.12.0-rc3 -> v1.12.0-rc3 2025-12-04T09:33:41.0500791Z * [new tag] v1.12.0-rc4 -> v1.12.0-rc4 2025-12-04T09:33:41.0501802Z * [new tag] v1.12.0-rc5 -> v1.12.0-rc5 2025-12-04T09:33:41.0502985Z * [new tag] v1.12.0-rc6 -> v1.12.0-rc6 2025-12-04T09:33:41.0503712Z * [new tag] v1.12.0-rc7 -> v1.12.0-rc7 2025-12-04T09:33:41.0504529Z * [new tag] v1.12.0-rc8 -> v1.12.0-rc8 2025-12-04T09:33:41.0505419Z * [new tag] v1.12.1 -> v1.12.1 2025-12-04T09:33:41.0506572Z * [new tag] v1.12.1-rc1 -> v1.12.1-rc1 2025-12-04T09:33:41.0507621Z * [new tag] v1.12.1-rc2 -> v1.12.1-rc2 2025-12-04T09:33:41.0508740Z * [new tag] v1.12.1-rc3 -> v1.12.1-rc3 2025-12-04T09:33:41.0509815Z * [new tag] v1.12.1-rc4 -> v1.12.1-rc4 2025-12-04T09:33:41.0510541Z * [new tag] v1.12.1-rc5 -> v1.12.1-rc5 2025-12-04T09:33:41.0511689Z * [new tag] v1.13.0 -> v1.13.0 2025-12-04T09:33:41.0512684Z * [new tag] v1.13.0-rc1 -> v1.13.0-rc1 2025-12-04T09:33:41.0513711Z * [new tag] v1.13.0-rc2 -> v1.13.0-rc2 2025-12-04T09:33:41.0514732Z * [new tag] v1.13.0-rc3 -> v1.13.0-rc3 2025-12-04T09:33:41.0515916Z * [new tag] v1.13.0-rc4 -> v1.13.0-rc4 2025-12-04T09:33:41.0516641Z * [new tag] v1.13.0-rc5 -> v1.13.0-rc5 2025-12-04T09:33:41.0517442Z * [new tag] v1.13.0-rc6 -> v1.13.0-rc6 2025-12-04T09:33:41.0518638Z * [new tag] v1.13.1 -> v1.13.1 2025-12-04T09:33:41.0519400Z * [new tag] v1.13.1-rc1 -> v1.13.1-rc1 2025-12-04T09:33:41.0520428Z * [new tag] v1.2.0 -> v1.2.0 2025-12-04T09:33:41.0521471Z * [new tag] v1.2.0a0 -> v1.2.0a0 2025-12-04T09:33:41.0522449Z * [new tag] v1.3.0 -> v1.3.0 2025-12-04T09:33:41.0523542Z * [new tag] v1.3.0a0 -> v1.3.0a0 2025-12-04T09:33:41.0524258Z * [new tag] v1.3.1 -> v1.3.1 2025-12-04T09:33:41.0525343Z * [new tag] v1.4.0 -> v1.4.0 2025-12-04T09:33:41.0526352Z * [new tag] v1.4.0a0 -> v1.4.0a0 2025-12-04T09:33:41.0527091Z * [new tag] v1.4.1 -> v1.4.1 2025-12-04T09:33:41.0528229Z * [new tag] v1.5.0 -> v1.5.0 2025-12-04T09:33:41.0529339Z * [new tag] v1.5.0-rc1 -> v1.5.0-rc1 2025-12-04T09:33:41.0530438Z * [new tag] v1.5.0-rc2 -> v1.5.0-rc2 2025-12-04T09:33:41.0532017Z * [new tag] v1.5.0-rc3 -> v1.5.0-rc3 2025-12-04T09:33:41.0532986Z * [new tag] v1.5.0-rc4 -> v1.5.0-rc4 2025-12-04T09:33:41.0533762Z * [new tag] v1.5.0-rc5 -> v1.5.0-rc5 2025-12-04T09:33:41.0534873Z * [new tag] v1.5.1 -> v1.5.1 2025-12-04T09:33:41.0535612Z * [new tag] v1.5.1-rc1 -> v1.5.1-rc1 2025-12-04T09:33:41.0536420Z * [new tag] v1.6.0 -> v1.6.0 2025-12-04T09:33:41.0537639Z * [new tag] v1.6.0-rc1 -> v1.6.0-rc1 2025-12-04T09:33:41.0538894Z * [new tag] v1.6.0-rc2 -> v1.6.0-rc2 2025-12-04T09:33:41.0539841Z * [new tag] v1.6.0-rc3 -> v1.6.0-rc3 2025-12-04T09:33:41.0540840Z * [new tag] v1.6.0-rc4 -> v1.6.0-rc4 2025-12-04T09:33:41.0541908Z * [new tag] v1.6.0-rc5 -> v1.6.0-rc5 2025-12-04T09:33:41.0542916Z * [new tag] v1.6.0-rc6 -> v1.6.0-rc6 2025-12-04T09:33:41.0543656Z * [new tag] v1.6.0-rc7 -> v1.6.0-rc7 2025-12-04T09:33:41.0544784Z * [new tag] v1.7.0 -> v1.7.0 2025-12-04T09:33:41.0545940Z * [new tag] v1.7.0-rc1 -> v1.7.0-rc1 2025-12-04T09:33:41.0547093Z * [new tag] v1.7.0-rc2 -> v1.7.0-rc2 2025-12-04T09:33:41.0548139Z * [new tag] v1.7.0-rc3 -> v1.7.0-rc3 2025-12-04T09:33:41.0548850Z * [new tag] v1.7.0-rc4 -> v1.7.0-rc4 2025-12-04T09:33:41.0549938Z * [new tag] v1.7.1 -> v1.7.1 2025-12-04T09:33:41.0551163Z * [new tag] v1.7.1-rc1 -> v1.7.1-rc1 2025-12-04T09:33:41.0552231Z * [new tag] v1.7.1-rc2 -> v1.7.1-rc2 2025-12-04T09:33:41.0553002Z * [new tag] v1.7.1-rc3 -> v1.7.1-rc3 2025-12-04T09:33:41.0554086Z * [new tag] v1.8.0 -> v1.8.0 2025-12-04T09:33:41.0554856Z * [new tag] v1.8.0-rc1 -> v1.8.0-rc1 2025-12-04T09:33:41.0556099Z * [new tag] v1.8.0-rc2 -> v1.8.0-rc2 2025-12-04T09:33:41.0557168Z * [new tag] v1.8.0-rc3 -> v1.8.0-rc3 2025-12-04T09:33:41.0558141Z * [new tag] v1.8.0-rc4 -> v1.8.0-rc4 2025-12-04T09:33:41.0558881Z * [new tag] v1.8.0-rc5 -> v1.8.0-rc5 2025-12-04T09:33:41.0559683Z * [new tag] v1.8.1 -> v1.8.1 2025-12-04T09:33:41.0560812Z * [new tag] v1.8.1-rc1 -> v1.8.1-rc1 2025-12-04T09:33:41.0562049Z * [new tag] v1.8.1-rc2 -> v1.8.1-rc2 2025-12-04T09:33:41.0562850Z * [new tag] v1.8.1-rc3 -> v1.8.1-rc3 2025-12-04T09:33:41.0564514Z * [new tag] v1.8.2 -> v1.8.2 2025-12-04T09:33:41.0565275Z * [new tag] v1.8.2-rc1 -> v1.8.2-rc1 2025-12-04T09:33:41.0566398Z * [new tag] v1.9.0 -> v1.9.0 2025-12-04T09:33:41.0567459Z * [new tag] v1.9.0-rc1 -> v1.9.0-rc1 2025-12-04T09:33:41.0568643Z * [new tag] v1.9.0-rc2 -> v1.9.0-rc2 2025-12-04T09:33:41.0569655Z * [new tag] v1.9.0-rc3 -> v1.9.0-rc3 2025-12-04T09:33:41.0570419Z * [new tag] v1.9.0-rc4 -> v1.9.0-rc4 2025-12-04T09:33:41.0571525Z * [new tag] v1.9.1 -> v1.9.1 2025-12-04T09:33:41.0572806Z * [new tag] v1.9.1-rc1 -> v1.9.1-rc1 2025-12-04T09:33:41.0573518Z * [new tag] v1.9.1-rc2 -> v1.9.1-rc2 2025-12-04T09:33:41.0574665Z * [new tag] v2.0.0 -> v2.0.0 2025-12-04T09:33:41.0575701Z * [new tag] v2.0.0-rc1 -> v2.0.0-rc1 2025-12-04T09:33:41.0576779Z * [new tag] v2.0.0-rc2 -> v2.0.0-rc2 2025-12-04T09:33:41.0577977Z * [new tag] v2.0.0-rc3 -> v2.0.0-rc3 2025-12-04T09:33:41.0578974Z * [new tag] v2.0.0-rc4 -> v2.0.0-rc4 2025-12-04T09:33:41.0580056Z * [new tag] v2.0.0-rc5 -> v2.0.0-rc5 2025-12-04T09:33:41.0580864Z * [new tag] v2.0.0-rc6 -> v2.0.0-rc6 2025-12-04T09:33:41.0581978Z * [new tag] v2.0.1 -> v2.0.1 2025-12-04T09:33:41.0583087Z * [new tag] v2.0.1-rc1 -> v2.0.1-rc1 2025-12-04T09:33:41.0583791Z * [new tag] v2.0.1-rc2 -> v2.0.1-rc2 2025-12-04T09:33:41.0584854Z * [new tag] v2.0.1-rc3 -> v2.0.1-rc3 2025-12-04T09:33:41.0585582Z * [new tag] v2.0.1-rc4 -> v2.0.1-rc4 2025-12-04T09:33:41.0587281Z * [new tag] v2.1.0 -> v2.1.0 2025-12-04T09:33:41.0588295Z * [new tag] v2.1.0-rc1 -> v2.1.0-rc1 2025-12-04T09:33:41.0589496Z * [new tag] v2.1.0-rc2 -> v2.1.0-rc2 2025-12-04T09:33:41.0590656Z * [new tag] v2.1.0-rc3 -> v2.1.0-rc3 2025-12-04T09:33:41.0591750Z * [new tag] v2.1.0-rc4 -> v2.1.0-rc4 2025-12-04T09:33:41.0592812Z * [new tag] v2.1.0-rc5 -> v2.1.0-rc5 2025-12-04T09:33:41.0593561Z * [new tag] v2.1.0-rc6 -> v2.1.0-rc6 2025-12-04T09:33:41.0594781Z * [new tag] v2.1.1 -> v2.1.1 2025-12-04T09:33:41.0596107Z * [new tag] v2.1.1-rc1 -> v2.1.1-rc1 2025-12-04T09:33:41.0597203Z * [new tag] v2.1.1-rc2 -> v2.1.1-rc2 2025-12-04T09:33:41.0598857Z * [new tag] v2.1.1-rc3 -> v2.1.1-rc3 2025-12-04T09:33:41.0599953Z * [new tag] v2.1.1-rc4 -> v2.1.1-rc4 2025-12-04T09:33:41.0600967Z * [new tag] v2.1.1-rc5 -> v2.1.1-rc5 2025-12-04T09:33:41.0601715Z * [new tag] v2.1.1-rc6 -> v2.1.1-rc6 2025-12-04T09:33:41.0602779Z * [new tag] v2.1.2 -> v2.1.2 2025-12-04T09:33:41.0603907Z * [new tag] v2.1.2-rc1 -> v2.1.2-rc1 2025-12-04T09:33:41.0605055Z * [new tag] v2.1.2-rc2 -> v2.1.2-rc2 2025-12-04T09:33:41.0605770Z * [new tag] v2.1.2-rc3 -> v2.1.2-rc3 2025-12-04T09:33:41.0606916Z * [new tag] v2.2.0 -> v2.2.0 2025-12-04T09:33:41.0607928Z * [new tag] v2.2.0-rc1 -> v2.2.0-rc1 2025-12-04T09:33:41.0608979Z * [new tag] v2.2.0-rc2 -> v2.2.0-rc2 2025-12-04T09:33:41.0609936Z * [new tag] v2.2.0-rc3 -> v2.2.0-rc3 2025-12-04T09:33:41.0610977Z * [new tag] v2.2.0-rc4 -> v2.2.0-rc4 2025-12-04T09:33:41.0611941Z * [new tag] v2.2.0-rc5 -> v2.2.0-rc5 2025-12-04T09:33:41.0612969Z * [new tag] v2.2.0-rc6 -> v2.2.0-rc6 2025-12-04T09:33:41.0613739Z * [new tag] v2.2.0-rc7 -> v2.2.0-rc7 2025-12-04T09:33:41.0614523Z * [new tag] v2.2.0-rc8 -> v2.2.0-rc8 2025-12-04T09:33:41.0615741Z * [new tag] v2.2.1 -> v2.2.1 2025-12-04T09:33:41.0616843Z * [new tag] v2.2.1-rc1 -> v2.2.1-rc1 2025-12-04T09:33:41.0617806Z * [new tag] v2.2.1-rc2 -> v2.2.1-rc2 2025-12-04T09:33:41.0618525Z * [new tag] v2.2.1-rc3 -> v2.2.1-rc3 2025-12-04T09:33:41.0619332Z * [new tag] v2.2.2 -> v2.2.2 2025-12-04T09:33:41.0620661Z * [new tag] v2.2.2-rc1 -> v2.2.2-rc1 2025-12-04T09:33:41.0621385Z * [new tag] v2.2.2-rc2 -> v2.2.2-rc2 2025-12-04T09:33:41.0622377Z * [new tag] v2.2.2-rc3 -> v2.2.2-rc3 2025-12-04T09:33:41.0623569Z * [new tag] v2.3.0 -> v2.3.0 2025-12-04T09:33:41.0624538Z * [new tag] v2.3.0-rc1 -> v2.3.0-rc1 2025-12-04T09:33:41.0625699Z * [new tag] v2.3.0-rc10 -> v2.3.0-rc10 2025-12-04T09:33:41.0626796Z * [new tag] v2.3.0-rc11 -> v2.3.0-rc11 2025-12-04T09:33:41.0627545Z * [new tag] v2.3.0-rc12 -> v2.3.0-rc12 2025-12-04T09:33:41.0628648Z * [new tag] v2.3.0-rc2 -> v2.3.0-rc2 2025-12-04T09:33:41.0629752Z * [new tag] v2.3.0-rc3 -> v2.3.0-rc3 2025-12-04T09:33:41.0630807Z * [new tag] v2.3.0-rc4 -> v2.3.0-rc4 2025-12-04T09:33:41.0631811Z * [new tag] v2.3.0-rc5 -> v2.3.0-rc5 2025-12-04T09:33:41.0632615Z * [new tag] v2.3.0-rc6 -> v2.3.0-rc6 2025-12-04T09:33:41.0633764Z * [new tag] v2.3.0-rc7 -> v2.3.0-rc7 2025-12-04T09:33:41.0634816Z * [new tag] v2.3.0-rc8 -> v2.3.0-rc8 2025-12-04T09:33:41.0635523Z * [new tag] v2.3.0-rc9 -> v2.3.0-rc9 2025-12-04T09:33:41.0636333Z * [new tag] v2.3.1 -> v2.3.1 2025-12-04T09:33:41.0637523Z * [new tag] v2.3.1-rc1 -> v2.3.1-rc1 2025-12-04T09:33:41.0638659Z * [new tag] v2.3.1-rc2 -> v2.3.1-rc2 2025-12-04T09:33:41.0639744Z * [new tag] v2.3.1-rc3 -> v2.3.1-rc3 2025-12-04T09:33:41.0640815Z * [new tag] v2.4.0 -> v2.4.0 2025-12-04T09:33:41.0641902Z * [new tag] v2.4.0-rc1 -> v2.4.0-rc1 2025-12-04T09:33:41.0642902Z * [new tag] v2.4.0-rc2 -> v2.4.0-rc2 2025-12-04T09:33:41.0643923Z * [new tag] v2.4.0-rc3 -> v2.4.0-rc3 2025-12-04T09:33:41.0644979Z * [new tag] v2.4.0-rc4 -> v2.4.0-rc4 2025-12-04T09:33:41.0646081Z * [new tag] v2.4.0-rc5 -> v2.4.0-rc5 2025-12-04T09:33:41.0647179Z * [new tag] v2.4.0-rc6 -> v2.4.0-rc6 2025-12-04T09:33:41.0648265Z * [new tag] v2.4.0-rc7 -> v2.4.0-rc7 2025-12-04T09:33:41.0649296Z * [new tag] v2.4.0-rc8 -> v2.4.0-rc8 2025-12-04T09:33:41.0650432Z * [new tag] v2.4.0-rc9 -> v2.4.0-rc9 2025-12-04T09:33:41.0651177Z * [new tag] v2.4.1 -> v2.4.1 2025-12-04T09:33:41.0652389Z * [new tag] v2.4.1-rc1 -> v2.4.1-rc1 2025-12-04T09:33:41.0653486Z * [new tag] v2.4.1-rc2 -> v2.4.1-rc2 2025-12-04T09:33:41.0654571Z * [new tag] v2.4.1-rc3 -> v2.4.1-rc3 2025-12-04T09:33:41.0655715Z * [new tag] v2.5.0 -> v2.5.0 2025-12-04T09:33:41.0656771Z * [new tag] v2.5.0-rc1 -> v2.5.0-rc1 2025-12-04T09:33:41.0657660Z * [new tag] v2.5.0-rc10 -> v2.5.0-rc10 2025-12-04T09:33:41.0658745Z * [new tag] v2.5.0-rc2 -> v2.5.0-rc2 2025-12-04T09:33:41.0659774Z * [new tag] v2.5.0-rc3 -> v2.5.0-rc3 2025-12-04T09:33:41.0660829Z * [new tag] v2.5.0-rc4 -> v2.5.0-rc4 2025-12-04T09:33:41.0661884Z * [new tag] v2.5.0-rc5 -> v2.5.0-rc5 2025-12-04T09:33:41.0663078Z * [new tag] v2.5.0-rc6 -> v2.5.0-rc6 2025-12-04T09:33:41.0664525Z * [new tag] v2.5.0-rc7 -> v2.5.0-rc7 2025-12-04T09:33:41.0665615Z * [new tag] v2.5.0-rc8 -> v2.5.0-rc8 2025-12-04T09:33:41.0666788Z * [new tag] v2.5.0-rc9 -> v2.5.0-rc9 2025-12-04T09:33:41.0667482Z * [new tag] v2.5.1 -> v2.5.1 2025-12-04T09:33:41.0668290Z * [new tag] v2.5.1-rc1 -> v2.5.1-rc1 2025-12-04T09:33:41.0669085Z * [new tag] v2.6.0 -> v2.6.0 2025-12-04T09:33:41.0670318Z * [new tag] v2.6.0-rc1 -> v2.6.0-rc1 2025-12-04T09:33:41.0671470Z * [new tag] v2.6.0-rc2 -> v2.6.0-rc2 2025-12-04T09:33:41.0672552Z * [new tag] v2.6.0-rc3 -> v2.6.0-rc3 2025-12-04T09:33:41.0673573Z * [new tag] v2.6.0-rc4 -> v2.6.0-rc4 2025-12-04T09:33:41.0674902Z * [new tag] v2.6.0-rc5 -> v2.6.0-rc5 2025-12-04T09:33:41.0676075Z * [new tag] v2.6.0-rc6 -> v2.6.0-rc6 2025-12-04T09:33:41.0677194Z * [new tag] v2.6.0-rc7 -> v2.6.0-rc7 2025-12-04T09:33:41.0678453Z * [new tag] v2.6.0-rc8 -> v2.6.0-rc8 2025-12-04T09:33:41.0679518Z * [new tag] v2.6.0-rc9 -> v2.6.0-rc9 2025-12-04T09:33:41.0680788Z * [new tag] v2.7.0 -> v2.7.0 2025-12-04T09:33:41.0681834Z * [new tag] v2.7.0-rc1 -> v2.7.0-rc1 2025-12-04T09:33:41.0682611Z * [new tag] v2.7.0-rc10 -> v2.7.0-rc10 2025-12-04T09:33:41.0683900Z * [new tag] v2.7.0-rc2 -> v2.7.0-rc2 2025-12-04T09:33:41.0685032Z * [new tag] v2.7.0-rc3 -> v2.7.0-rc3 2025-12-04T09:33:41.0686156Z * [new tag] v2.7.0-rc4 -> v2.7.0-rc4 2025-12-04T09:33:41.0687157Z * [new tag] v2.7.0-rc5 -> v2.7.0-rc5 2025-12-04T09:33:41.0688178Z * [new tag] v2.7.0-rc6 -> v2.7.0-rc6 2025-12-04T09:33:41.0689306Z * [new tag] v2.7.0-rc7 -> v2.7.0-rc7 2025-12-04T09:33:41.0690484Z * [new tag] v2.7.0-rc8 -> v2.7.0-rc8 2025-12-04T09:33:41.0691629Z * [new tag] v2.7.0-rc9 -> v2.7.0-rc9 2025-12-04T09:33:41.0692381Z * [new tag] v2.7.1 -> v2.7.1 2025-12-04T09:33:41.0693680Z * [new tag] v2.7.1-rc1 -> v2.7.1-rc1 2025-12-04T09:33:41.0694811Z * [new tag] v2.7.1-rc2 -> v2.7.1-rc2 2025-12-04T09:33:41.0695940Z * [new tag] v2.7.1-rc3 -> v2.7.1-rc3 2025-12-04T09:33:41.0697359Z * [new tag] v2.7.1-rc4 -> v2.7.1-rc4 2025-12-04T09:33:41.0698390Z * [new tag] v2.7.1-rc5 -> v2.7.1-rc5 2025-12-04T09:33:41.0699201Z * [new tag] v2.8.0 -> v2.8.0 2025-12-04T09:33:41.0700397Z * [new tag] v2.8.0-rc1 -> v2.8.0-rc1 2025-12-04T09:33:41.0701477Z * [new tag] v2.8.0-rc2 -> v2.8.0-rc2 2025-12-04T09:33:41.0702645Z * [new tag] v2.8.0-rc3 -> v2.8.0-rc3 2025-12-04T09:33:41.0703862Z * [new tag] v2.8.0-rc4 -> v2.8.0-rc4 2025-12-04T09:33:41.0704976Z * [new tag] v2.8.0-rc5 -> v2.8.0-rc5 2025-12-04T09:33:41.0706101Z * [new tag] v2.8.0-rc6 -> v2.8.0-rc6 2025-12-04T09:33:41.0707182Z * [new tag] v2.8.0-rc7 -> v2.8.0-rc7 2025-12-04T09:33:41.0708238Z * [new tag] v2.8.0-rc8 -> v2.8.0-rc8 2025-12-04T09:33:41.0709375Z * [new tag] v2.9.0 -> v2.9.0 2025-12-04T09:33:41.0710479Z * [new tag] v2.9.0-rc1 -> v2.9.0-rc1 2025-12-04T09:33:41.0711669Z * [new tag] v2.9.0-rc10 -> v2.9.0-rc10 2025-12-04T09:33:41.0712742Z * [new tag] v2.9.0-rc11 -> v2.9.0-rc11 2025-12-04T09:33:41.0714074Z * [new tag] v2.9.0-rc2 -> v2.9.0-rc2 2025-12-04T09:33:41.0715239Z * [new tag] v2.9.0-rc3 -> v2.9.0-rc3 2025-12-04T09:33:41.0716335Z * [new tag] v2.9.0-rc4 -> v2.9.0-rc4 2025-12-04T09:33:41.0717446Z * [new tag] v2.9.0-rc5 -> v2.9.0-rc5 2025-12-04T09:33:41.0718818Z * [new tag] v2.9.0-rc6 -> v2.9.0-rc6 2025-12-04T09:33:41.0719909Z * [new tag] v2.9.0-rc7 -> v2.9.0-rc7 2025-12-04T09:33:41.0721258Z * [new tag] v2.9.0-rc8 -> v2.9.0-rc8 2025-12-04T09:33:41.0722013Z * [new tag] v2.9.0-rc9 -> v2.9.0-rc9 2025-12-04T09:33:41.0722950Z * [new tag] v2.9.1 -> v2.9.1 2025-12-04T09:33:41.0724074Z * [new tag] v2.9.1-rc1 -> v2.9.1-rc1 2025-12-04T09:33:41.0725257Z * [new tag] v2.9.1-rc2 -> v2.9.1-rc2 2025-12-04T09:33:41.0726902Z * [new tag] viable/strict/1759343184 -> viable/strict/1759343184 2025-12-04T09:33:41.0727870Z * [new tag] viable/strict/1759346540 -> viable/strict/1759346540 2025-12-04T09:33:41.0728969Z * [new tag] viable/strict/1759348181 -> viable/strict/1759348181 2025-12-04T09:33:41.0729969Z * [new tag] viable/strict/1759350324 -> viable/strict/1759350324 2025-12-04T09:33:41.0730989Z * [new tag] viable/strict/1759351793 -> viable/strict/1759351793 2025-12-04T09:33:41.0731868Z * [new tag] viable/strict/1759353844 -> viable/strict/1759353844 2025-12-04T09:33:41.0732917Z * [new tag] viable/strict/1759355374 -> viable/strict/1759355374 2025-12-04T09:33:41.0733832Z * [new tag] viable/strict/1759357472 -> viable/strict/1759357472 2025-12-04T09:33:41.0734867Z * [new tag] viable/strict/1759361002 -> viable/strict/1759361002 2025-12-04T09:33:41.0736216Z * [new tag] viable/strict/1759362585 -> viable/strict/1759362585 2025-12-04T09:33:41.0737544Z * [new tag] viable/strict/1759365359 -> viable/strict/1759365359 2025-12-04T09:33:41.0738643Z * [new tag] viable/strict/1759370089 -> viable/strict/1759370089 2025-12-04T09:33:41.0739697Z * [new tag] viable/strict/1759377554 -> viable/strict/1759377554 2025-12-04T09:33:41.0740755Z * [new tag] viable/strict/1759379133 -> viable/strict/1759379133 2025-12-04T09:33:41.0741796Z * [new tag] viable/strict/1759389871 -> viable/strict/1759389871 2025-12-04T09:33:41.0742839Z * [new tag] viable/strict/1759393562 -> viable/strict/1759393562 2025-12-04T09:33:41.0743981Z * [new tag] viable/strict/1759395076 -> viable/strict/1759395076 2025-12-04T09:33:41.0745063Z * [new tag] viable/strict/1759398579 -> viable/strict/1759398579 2025-12-04T09:33:41.0746135Z * [new tag] viable/strict/1759404142 -> viable/strict/1759404142 2025-12-04T09:33:41.0747136Z * [new tag] viable/strict/1759405773 -> viable/strict/1759405773 2025-12-04T09:33:41.0748197Z * [new tag] viable/strict/1759408041 -> viable/strict/1759408041 2025-12-04T09:33:41.0749075Z * [new tag] viable/strict/1759411593 -> viable/strict/1759411593 2025-12-04T09:33:41.0750199Z * [new tag] viable/strict/1759427395 -> viable/strict/1759427395 2025-12-04T09:33:41.0751256Z * [new tag] viable/strict/1759434582 -> viable/strict/1759434582 2025-12-04T09:33:41.0752334Z * [new tag] viable/strict/1759436720 -> viable/strict/1759436720 2025-12-04T09:33:41.0753431Z * [new tag] viable/strict/1759440219 -> viable/strict/1759440219 2025-12-04T09:33:41.0754300Z * [new tag] viable/strict/1759441948 -> viable/strict/1759441948 2025-12-04T09:33:41.0755402Z * [new tag] viable/strict/1759443860 -> viable/strict/1759443860 2025-12-04T09:33:41.0756614Z * [new tag] viable/strict/1759445377 -> viable/strict/1759445377 2025-12-04T09:33:41.0757626Z * [new tag] viable/strict/1759447415 -> viable/strict/1759447415 2025-12-04T09:33:41.0758795Z * [new tag] viable/strict/1759451750 -> viable/strict/1759451750 2025-12-04T09:33:41.0759858Z * [new tag] viable/strict/1759453910 -> viable/strict/1759453910 2025-12-04T09:33:41.0761512Z * [new tag] viable/strict/1759456483 -> viable/strict/1759456483 2025-12-04T09:33:41.0762586Z * [new tag] viable/strict/1759459279 -> viable/strict/1759459279 2025-12-04T09:33:41.0763640Z * [new tag] viable/strict/1759460742 -> viable/strict/1759460742 2025-12-04T09:33:41.0764738Z * [new tag] viable/strict/1759462025 -> viable/strict/1759462025 2025-12-04T09:33:41.0765831Z * [new tag] viable/strict/1759469086 -> viable/strict/1759469086 2025-12-04T09:33:41.0766781Z * [new tag] viable/strict/1759470581 -> viable/strict/1759470581 2025-12-04T09:33:41.0767867Z * [new tag] viable/strict/1759472786 -> viable/strict/1759472786 2025-12-04T09:33:41.0768977Z * [new tag] viable/strict/1759476294 -> viable/strict/1759476294 2025-12-04T09:33:41.0770022Z * [new tag] viable/strict/1759479963 -> viable/strict/1759479963 2025-12-04T09:33:41.0771027Z * [new tag] viable/strict/1759492177 -> viable/strict/1759492177 2025-12-04T09:33:41.0772068Z * [new tag] viable/strict/1759519278 -> viable/strict/1759519278 2025-12-04T09:33:41.0773059Z * [new tag] viable/strict/1759524580 -> viable/strict/1759524580 2025-12-04T09:33:41.0774108Z * [new tag] viable/strict/1759528193 -> viable/strict/1759528193 2025-12-04T09:33:41.0775383Z * [new tag] viable/strict/1759533797 -> viable/strict/1759533797 2025-12-04T09:33:41.0776447Z * [new tag] viable/strict/1759542780 -> viable/strict/1759542780 2025-12-04T09:33:41.0777607Z * [new tag] viable/strict/1759549779 -> viable/strict/1759549779 2025-12-04T09:33:41.0778662Z * [new tag] viable/strict/1759555455 -> viable/strict/1759555455 2025-12-04T09:33:41.0779755Z * [new tag] viable/strict/1759559176 -> viable/strict/1759559176 2025-12-04T09:33:41.0780788Z * [new tag] viable/strict/1759560629 -> viable/strict/1759560629 2025-12-04T09:33:41.0781852Z * [new tag] viable/strict/1759569848 -> viable/strict/1759569848 2025-12-04T09:33:41.0783105Z * [new tag] viable/strict/1759571382 -> viable/strict/1759571382 2025-12-04T09:33:41.0784108Z * [new tag] viable/strict/1759573474 -> viable/strict/1759573474 2025-12-04T09:33:41.0785149Z * [new tag] viable/strict/1759618187 -> viable/strict/1759618187 2025-12-04T09:33:41.0786187Z * [new tag] viable/strict/1759626742 -> viable/strict/1759626742 2025-12-04T09:33:41.0787299Z * [new tag] viable/strict/1759632427 -> viable/strict/1759632427 2025-12-04T09:33:41.0788307Z * [new tag] viable/strict/1759634971 -> viable/strict/1759634971 2025-12-04T09:33:41.0789458Z * [new tag] viable/strict/1759661382 -> viable/strict/1759661382 2025-12-04T09:33:41.0790546Z * [new tag] viable/strict/1759663294 -> viable/strict/1759663294 2025-12-04T09:33:41.0791340Z * [new tag] viable/strict/1759708178 -> viable/strict/1759708178 2025-12-04T09:33:41.0792640Z * [new tag] viable/strict/1759715695 -> viable/strict/1759715695 2025-12-04T09:33:41.0793628Z * [new tag] viable/strict/1759728293 -> viable/strict/1759728293 2025-12-04T09:33:41.0794701Z * [new tag] viable/strict/1759735513 -> viable/strict/1759735513 2025-12-04T09:33:41.0795773Z * [new tag] viable/strict/1759739177 -> viable/strict/1759739177 2025-12-04T09:33:41.0800792Z * [new tag] viable/strict/1759758635 -> viable/strict/1759758635 2025-12-04T09:33:41.0801997Z * [new tag] viable/strict/1759765784 -> viable/strict/1759765784 2025-12-04T09:33:41.0803145Z * [new tag] viable/strict/1759767948 -> viable/strict/1759767948 2025-12-04T09:33:41.0804271Z * [new tag] viable/strict/1759771461 -> viable/strict/1759771461 2025-12-04T09:33:41.0805067Z * [new tag] viable/strict/1759776706 -> viable/strict/1759776706 2025-12-04T09:33:41.0806221Z * [new tag] viable/strict/1759782317 -> viable/strict/1759782317 2025-12-04T09:33:41.0807386Z * [new tag] viable/strict/1759783777 -> viable/strict/1759783777 2025-12-04T09:33:41.0808455Z * [new tag] viable/strict/1759785815 -> viable/strict/1759785815 2025-12-04T09:33:41.0809575Z * [new tag] viable/strict/1759789459 -> viable/strict/1759789459 2025-12-04T09:33:41.0810576Z * [new tag] viable/strict/1759790974 -> viable/strict/1759790974 2025-12-04T09:33:41.0811419Z * [new tag] viable/strict/1759794583 -> viable/strict/1759794583 2025-12-04T09:33:41.0812538Z * [new tag] viable/strict/1759797408 -> viable/strict/1759797408 2025-12-04T09:33:41.0813646Z * [new tag] viable/strict/1759799518 -> viable/strict/1759799518 2025-12-04T09:33:41.0814673Z * [new tag] viable/strict/1759804909 -> viable/strict/1759804909 2025-12-04T09:33:41.0815712Z * [new tag] viable/strict/1759807643 -> viable/strict/1759807643 2025-12-04T09:33:41.0816758Z * [new tag] viable/strict/1759809089 -> viable/strict/1759809089 2025-12-04T09:33:41.0817889Z * [new tag] viable/strict/1759811145 -> viable/strict/1759811145 2025-12-04T09:33:41.0818985Z * [new tag] viable/strict/1759812581 -> viable/strict/1759812581 2025-12-04T09:33:41.0819977Z * [new tag] viable/strict/1759814683 -> viable/strict/1759814683 2025-12-04T09:33:41.0821008Z * [new tag] viable/strict/1759821889 -> viable/strict/1759821889 2025-12-04T09:33:41.0822139Z * [new tag] viable/strict/1759823376 -> viable/strict/1759823376 2025-12-04T09:33:41.0823184Z * [new tag] viable/strict/1759827107 -> viable/strict/1759827107 2025-12-04T09:33:41.0824200Z * [new tag] viable/strict/1759830577 -> viable/strict/1759830577 2025-12-04T09:33:41.0825340Z * [new tag] viable/strict/1759832720 -> viable/strict/1759832720 2025-12-04T09:33:41.0826409Z * [new tag] viable/strict/1759842063 -> viable/strict/1759842063 2025-12-04T09:33:41.0827398Z * [new tag] viable/strict/1759847121 -> viable/strict/1759847121 2025-12-04T09:33:41.0828800Z * [new tag] viable/strict/1759850721 -> viable/strict/1759850721 2025-12-04T09:33:41.0829793Z * [new tag] viable/strict/1759857870 -> viable/strict/1759857870 2025-12-04T09:33:41.0830884Z * [new tag] viable/strict/1759863143 -> viable/strict/1759863143 2025-12-04T09:33:41.0832005Z * [new tag] viable/strict/1759875874 -> viable/strict/1759875874 2025-12-04T09:33:41.0832784Z * [new tag] viable/strict/1759877385 -> viable/strict/1759877385 2025-12-04T09:33:41.0834372Z * [new tag] viable/strict/1759883801 -> viable/strict/1759883801 2025-12-04T09:33:41.0835535Z * [new tag] viable/strict/1759885922 -> viable/strict/1759885922 2025-12-04T09:33:41.0836389Z * [new tag] viable/strict/1759888488 -> viable/strict/1759888488 2025-12-04T09:33:41.0837483Z * [new tag] viable/strict/1759895471 -> viable/strict/1759895471 2025-12-04T09:33:41.0838522Z * [new tag] viable/strict/1759904803 -> viable/strict/1759904803 2025-12-04T09:33:41.0839788Z * [new tag] viable/strict/1759908300 -> viable/strict/1759908300 2025-12-04T09:33:41.0840872Z * [new tag] viable/strict/1759915520 -> viable/strict/1759915520 2025-12-04T09:33:41.0841913Z * [new tag] viable/strict/1759916978 -> viable/strict/1759916978 2025-12-04T09:33:41.0842692Z * [new tag] viable/strict/1759930024 -> viable/strict/1759930024 2025-12-04T09:33:41.0843889Z * [new tag] viable/strict/1759948122 -> viable/strict/1759948122 2025-12-04T09:33:41.0844987Z * [new tag] viable/strict/1759952983 -> viable/strict/1759952983 2025-12-04T09:33:41.0846037Z * [new tag] viable/strict/1759955121 -> viable/strict/1759955121 2025-12-04T09:33:41.0847098Z * [new tag] viable/strict/1759962298 -> viable/strict/1759962298 2025-12-04T09:33:41.0848174Z * [new tag] viable/strict/1759965837 -> viable/strict/1759965837 2025-12-04T09:33:41.0849330Z * [new tag] viable/strict/1759970213 -> viable/strict/1759970213 2025-12-04T09:33:41.0850330Z * [new tag] viable/strict/1759974894 -> viable/strict/1759974894 2025-12-04T09:33:41.0851356Z * [new tag] viable/strict/1759977763 -> viable/strict/1759977763 2025-12-04T09:33:41.0852453Z * [new tag] viable/strict/1759979241 -> viable/strict/1759979241 2025-12-04T09:33:41.0853514Z * [new tag] viable/strict/1759985417 -> viable/strict/1759985417 2025-12-04T09:33:41.0854566Z * [new tag] viable/strict/1759987490 -> viable/strict/1759987490 2025-12-04T09:33:41.0855650Z * [new tag] viable/strict/1759996180 -> viable/strict/1759996180 2025-12-04T09:33:41.0856716Z * [new tag] viable/strict/1760065682 -> viable/strict/1760065682 2025-12-04T09:33:41.0857916Z * [new tag] viable/strict/1760066894 -> viable/strict/1760066894 2025-12-04T09:33:41.0859083Z * [new tag] viable/strict/1760070345 -> viable/strict/1760070345 2025-12-04T09:33:41.0860165Z * [new tag] viable/strict/1760089782 -> viable/strict/1760089782 2025-12-04T09:33:41.0861171Z * [new tag] viable/strict/1760091921 -> viable/strict/1760091921 2025-12-04T09:33:41.0862232Z * [new tag] viable/strict/1760127924 -> viable/strict/1760127924 2025-12-04T09:33:41.0863223Z * [new tag] viable/strict/1760129489 -> viable/strict/1760129489 2025-12-04T09:33:41.0864415Z * [new tag] viable/strict/1760132980 -> viable/strict/1760132980 2025-12-04T09:33:41.0865489Z * [new tag] viable/strict/1760135060 -> viable/strict/1760135060 2025-12-04T09:33:41.0866679Z * [new tag] viable/strict/1760215782 -> viable/strict/1760215782 2025-12-04T09:33:41.0867788Z * [new tag] viable/strict/1760273849 -> viable/strict/1760273849 2025-12-04T09:33:41.0868794Z * [new tag] viable/strict/1760275517 -> viable/strict/1760275517 2025-12-04T09:33:41.0869853Z * [new tag] viable/strict/1760276979 -> viable/strict/1760276979 2025-12-04T09:33:41.0870934Z * [new tag] viable/strict/1760279007 -> viable/strict/1760279007 2025-12-04T09:33:41.0871739Z * [new tag] viable/strict/1760286328 -> viable/strict/1760286328 2025-12-04T09:33:41.0872723Z * [new tag] viable/strict/1760493304 -> viable/strict/1760493304 2025-12-04T09:33:41.0873841Z * [new tag] viable/strict/1760496298 -> viable/strict/1760496298 2025-12-04T09:33:41.0874849Z * [new tag] viable/strict/1760518396 -> viable/strict/1760518396 2025-12-04T09:33:41.0875906Z * [new tag] viable/strict/1760534864 -> viable/strict/1760534864 2025-12-04T09:33:41.0876993Z * [new tag] viable/strict/1760549062 -> viable/strict/1760549062 2025-12-04T09:33:41.0878109Z * [new tag] viable/strict/1760552799 -> viable/strict/1760552799 2025-12-04T09:33:41.0879198Z * [new tag] viable/strict/1760554355 -> viable/strict/1760554355 2025-12-04T09:33:41.0880303Z * [new tag] viable/strict/1760556275 -> viable/strict/1760556275 2025-12-04T09:33:41.0881366Z * [new tag] viable/strict/1760564979 -> viable/strict/1760564979 2025-12-04T09:33:41.0882473Z * [new tag] viable/strict/1760567049 -> viable/strict/1760567049 2025-12-04T09:33:41.0883957Z * [new tag] viable/strict/1760568585 -> viable/strict/1760568585 2025-12-04T09:33:41.0884983Z * [new tag] viable/strict/1760570630 -> viable/strict/1760570630 2025-12-04T09:33:41.0886128Z * [new tag] viable/strict/1760572180 -> viable/strict/1760572180 2025-12-04T09:33:41.0887056Z * [new tag] viable/strict/1760575094 -> viable/strict/1760575094 2025-12-04T09:33:41.0888303Z * [new tag] viable/strict/1760579709 -> viable/strict/1760579709 2025-12-04T09:33:41.0889884Z * [new tag] viable/strict/1760582614 -> viable/strict/1760582614 2025-12-04T09:33:41.0891171Z * [new tag] viable/strict/1760586815 -> viable/strict/1760586815 2025-12-04T09:33:41.0892113Z * [new tag] viable/strict/1760588829 -> viable/strict/1760588829 2025-12-04T09:33:41.0893118Z * [new tag] viable/strict/1760590200 -> viable/strict/1760590200 2025-12-04T09:33:41.0894228Z * [new tag] viable/strict/1760592311 -> viable/strict/1760592311 2025-12-04T09:33:41.0895254Z * [new tag] viable/strict/1760619733 -> viable/strict/1760619733 2025-12-04T09:33:41.0896200Z * [new tag] viable/strict/1760628335 -> viable/strict/1760628335 2025-12-04T09:33:41.0897448Z * [new tag] viable/strict/1760635490 -> viable/strict/1760635490 2025-12-04T09:33:41.0898560Z * [new tag] viable/strict/1760640743 -> viable/strict/1760640743 2025-12-04T09:33:41.0899576Z * [new tag] viable/strict/1760642528 -> viable/strict/1760642528 2025-12-04T09:33:41.0900635Z * [new tag] viable/strict/1760646330 -> viable/strict/1760646330 2025-12-04T09:33:41.0901810Z * [new tag] viable/strict/1760666101 -> viable/strict/1760666101 2025-12-04T09:33:41.0902999Z * [new tag] viable/strict/1760668990 -> viable/strict/1760668990 2025-12-04T09:33:41.0904403Z * [new tag] viable/strict/1760670600 -> viable/strict/1760670600 2025-12-04T09:33:41.0905432Z * [new tag] viable/strict/1760671704 -> viable/strict/1760671704 2025-12-04T09:33:41.0906533Z * [new tag] viable/strict/1760673121 -> viable/strict/1760673121 2025-12-04T09:33:41.0907539Z * [new tag] viable/strict/1760675352 -> viable/strict/1760675352 2025-12-04T09:33:41.0908603Z * [new tag] viable/strict/1760696731 -> viable/strict/1760696731 2025-12-04T09:33:41.0911320Z * [new tag] viable/strict/1760723515 -> viable/strict/1760723515 2025-12-04T09:33:41.0912364Z * [new tag] viable/strict/1760727234 -> viable/strict/1760727234 2025-12-04T09:33:41.0913507Z * [new tag] viable/strict/1760730578 -> viable/strict/1760730578 2025-12-04T09:33:41.0914545Z * [new tag] viable/strict/1760732726 -> viable/strict/1760732726 2025-12-04T09:33:41.0915751Z * [new tag] viable/strict/1760734180 -> viable/strict/1760734180 2025-12-04T09:33:41.0916577Z * [new tag] viable/strict/1760736251 -> viable/strict/1760736251 2025-12-04T09:33:41.0917755Z * [new tag] viable/strict/1760737772 -> viable/strict/1760737772 2025-12-04T09:33:41.0918871Z * [new tag] viable/strict/1760758005 -> viable/strict/1760758005 2025-12-04T09:33:41.0919883Z * [new tag] viable/strict/1760761532 -> viable/strict/1760761532 2025-12-04T09:33:41.0920979Z * [new tag] viable/strict/1760802581 -> viable/strict/1760802581 2025-12-04T09:33:41.0922025Z * [new tag] viable/strict/1760827772 -> viable/strict/1760827772 2025-12-04T09:33:41.0923070Z * [new tag] viable/strict/1760834524 -> viable/strict/1760834524 2025-12-04T09:33:41.0924208Z * [new tag] viable/strict/1760845009 -> viable/strict/1760845009 2025-12-04T09:33:41.0925328Z * [new tag] viable/strict/1760876836 -> viable/strict/1760876836 2025-12-04T09:33:41.0926397Z * [new tag] viable/strict/1760880329 -> viable/strict/1760880329 2025-12-04T09:33:41.0927438Z * [new tag] viable/strict/1760888987 -> viable/strict/1760888987 2025-12-04T09:33:41.0928531Z * [new tag] viable/strict/1760912664 -> viable/strict/1760912664 2025-12-04T09:33:41.0929542Z * [new tag] viable/strict/1760925321 -> viable/strict/1760925321 2025-12-04T09:33:41.0930531Z * [new tag] viable/strict/1760931488 -> viable/strict/1760931488 2025-12-04T09:33:41.0931620Z * [new tag] viable/strict/1760932693 -> viable/strict/1760932693 2025-12-04T09:33:41.0932685Z * [new tag] viable/strict/1761004184 -> viable/strict/1761004184 2025-12-04T09:33:41.0933742Z * [new tag] viable/strict/1761014748 -> viable/strict/1761014748 2025-12-04T09:33:41.0934814Z * [new tag] viable/strict/1761017491 -> viable/strict/1761017491 2025-12-04T09:33:41.0935876Z * [new tag] viable/strict/1761018806 -> viable/strict/1761018806 2025-12-04T09:33:41.0937084Z * [new tag] viable/strict/1761020754 -> viable/strict/1761020754 2025-12-04T09:33:41.0938255Z * [new tag] viable/strict/1761024303 -> viable/strict/1761024303 2025-12-04T09:33:41.0939296Z * [new tag] viable/strict/1761029582 -> viable/strict/1761029582 2025-12-04T09:33:41.0940321Z * [new tag] viable/strict/1761031535 -> viable/strict/1761031535 2025-12-04T09:33:41.0941420Z * [new tag] viable/strict/1761035196 -> viable/strict/1761035196 2025-12-04T09:33:41.0942542Z * [new tag] viable/strict/1761045825 -> viable/strict/1761045825 2025-12-04T09:33:41.0943705Z * [new tag] viable/strict/1761054796 -> viable/strict/1761054796 2025-12-04T09:33:41.0944761Z * [new tag] viable/strict/1761060314 -> viable/strict/1761060314 2025-12-04T09:33:41.0945874Z * [new tag] viable/strict/1761071198 -> viable/strict/1761071198 2025-12-04T09:33:41.0946966Z * [new tag] viable/strict/1761074628 -> viable/strict/1761074628 2025-12-04T09:33:41.0948024Z * [new tag] viable/strict/1761078351 -> viable/strict/1761078351 2025-12-04T09:33:41.0949143Z * [new tag] viable/strict/1761079822 -> viable/strict/1761079822 2025-12-04T09:33:41.0950214Z * [new tag] viable/strict/1761081873 -> viable/strict/1761081873 2025-12-04T09:33:41.0951246Z * [new tag] viable/strict/1761083392 -> viable/strict/1761083392 2025-12-04T09:33:41.0952328Z * [new tag] viable/strict/1761085465 -> viable/strict/1761085465 2025-12-04T09:33:41.0953496Z * [new tag] viable/strict/1761089099 -> viable/strict/1761089099 2025-12-04T09:33:41.0954619Z * [new tag] viable/strict/1761095535 -> viable/strict/1761095535 2025-12-04T09:33:41.0955589Z * [new tag] viable/strict/1761098119 -> viable/strict/1761098119 2025-12-04T09:33:41.0957152Z * [new tag] viable/strict/1761101330 -> viable/strict/1761101330 2025-12-04T09:33:41.0958204Z * [new tag] viable/strict/1761114425 -> viable/strict/1761114425 2025-12-04T09:33:41.0959319Z * [new tag] viable/strict/1761116036 -> viable/strict/1761116036 2025-12-04T09:33:41.0960360Z * [new tag] viable/strict/1761119379 -> viable/strict/1761119379 2025-12-04T09:33:41.0961447Z * [new tag] viable/strict/1761121601 -> viable/strict/1761121601 2025-12-04T09:33:41.0962490Z * [new tag] viable/strict/1761123234 -> viable/strict/1761123234 2025-12-04T09:33:41.0963483Z * [new tag] viable/strict/1761126621 -> viable/strict/1761126621 2025-12-04T09:33:41.0964582Z * [new tag] viable/strict/1761132259 -> viable/strict/1761132259 2025-12-04T09:33:41.0965660Z * [new tag] viable/strict/1761146746 -> viable/strict/1761146746 2025-12-04T09:33:41.0966713Z * [new tag] viable/strict/1761164752 -> viable/strict/1761164752 2025-12-04T09:33:41.0967664Z * [new tag] viable/strict/1761166198 -> viable/strict/1761166198 2025-12-04T09:33:41.0968822Z * [new tag] viable/strict/1761175424 -> viable/strict/1761175424 2025-12-04T09:33:41.0969899Z * [new tag] viable/strict/1761176983 -> viable/strict/1761176983 2025-12-04T09:33:41.0971075Z * [new tag] viable/strict/1761179891 -> viable/strict/1761179891 2025-12-04T09:33:41.0972172Z * [new tag] viable/strict/1761181930 -> viable/strict/1761181930 2025-12-04T09:33:41.0973296Z * [new tag] viable/strict/1761184516 -> viable/strict/1761184516 2025-12-04T09:33:41.0974474Z * [new tag] viable/strict/1761190179 -> viable/strict/1761190179 2025-12-04T09:33:41.0975935Z * [new tag] viable/strict/1761193558 -> viable/strict/1761193558 2025-12-04T09:33:41.0977021Z * [new tag] viable/strict/1761207990 -> viable/strict/1761207990 2025-12-04T09:33:41.0978136Z * [new tag] viable/strict/1761229539 -> viable/strict/1761229539 2025-12-04T09:33:41.0979444Z * [new tag] viable/strict/1761244031 -> viable/strict/1761244031 2025-12-04T09:33:41.0980535Z * [new tag] viable/strict/1761248986 -> viable/strict/1761248986 2025-12-04T09:33:41.0981555Z * [new tag] viable/strict/1761259791 -> viable/strict/1761259791 2025-12-04T09:33:41.0982614Z * [new tag] viable/strict/1761266139 -> viable/strict/1761266139 2025-12-04T09:33:41.0983793Z * [new tag] viable/strict/1761268316 -> viable/strict/1761268316 2025-12-04T09:33:41.0984812Z * [new tag] viable/strict/1761273805 -> viable/strict/1761273805 2025-12-04T09:33:41.0985843Z * [new tag] viable/strict/1761275261 -> viable/strict/1761275261 2025-12-04T09:33:41.0986911Z * [new tag] viable/strict/1761277913 -> viable/strict/1761277913 2025-12-04T09:33:41.0988070Z * [new tag] viable/strict/1761290701 -> viable/strict/1761290701 2025-12-04T09:33:41.0989183Z * [new tag] viable/strict/1761294396 -> viable/strict/1761294396 2025-12-04T09:33:41.0990247Z * [new tag] viable/strict/1761303047 -> viable/strict/1761303047 2025-12-04T09:33:41.0991287Z * [new tag] viable/strict/1761335388 -> viable/strict/1761335388 2025-12-04T09:33:41.0992341Z * [new tag] viable/strict/1761337551 -> viable/strict/1761337551 2025-12-04T09:33:41.0993605Z * [new tag] viable/strict/1761339007 -> viable/strict/1761339007 2025-12-04T09:33:41.0994473Z * [new tag] viable/strict/1761341050 -> viable/strict/1761341050 2025-12-04T09:33:41.0995605Z * [new tag] viable/strict/1761346188 -> viable/strict/1761346188 2025-12-04T09:33:41.0996985Z * [new tag] viable/strict/1761349792 -> viable/strict/1761349792 2025-12-04T09:33:41.0998055Z * [new tag] viable/strict/1761352620 -> viable/strict/1761352620 2025-12-04T09:33:41.0999113Z * [new tag] viable/strict/1761354730 -> viable/strict/1761354730 2025-12-04T09:33:41.1000223Z * [new tag] viable/strict/1761357298 -> viable/strict/1761357298 2025-12-04T09:33:41.1001282Z * [new tag] viable/strict/1761360201 -> viable/strict/1761360201 2025-12-04T09:33:41.1002394Z * [new tag] viable/strict/1761361753 -> viable/strict/1761361753 2025-12-04T09:33:41.1003425Z * [new tag] viable/strict/1761364351 -> viable/strict/1761364351 2025-12-04T09:33:41.1004492Z * [new tag] viable/strict/1761366338 -> viable/strict/1761366338 2025-12-04T09:33:41.1005714Z * [new tag] viable/strict/1761367802 -> viable/strict/1761367802 2025-12-04T09:33:41.1006808Z * [new tag] viable/strict/1761369889 -> viable/strict/1761369889 2025-12-04T09:33:41.1007787Z * [new tag] viable/strict/1761371385 -> viable/strict/1761371385 2025-12-04T09:33:41.1009082Z * [new tag] viable/strict/1761373581 -> viable/strict/1761373581 2025-12-04T09:33:41.1010273Z * [new tag] viable/strict/1761375054 -> viable/strict/1761375054 2025-12-04T09:33:41.1011438Z * [new tag] viable/strict/1761421785 -> viable/strict/1761421785 2025-12-04T09:33:41.1012571Z * [new tag] viable/strict/1761434614 -> viable/strict/1761434614 2025-12-04T09:33:41.1014098Z * [new tag] viable/strict/1761439254 -> viable/strict/1761439254 2025-12-04T09:33:41.1015180Z * [new tag] viable/strict/1761454187 -> viable/strict/1761454187 2025-12-04T09:33:41.1016344Z * [new tag] viable/strict/1761459991 -> viable/strict/1761459991 2025-12-04T09:33:41.1017708Z * [new tag] viable/strict/1761470668 -> viable/strict/1761470668 2025-12-04T09:33:41.1019215Z * [new tag] viable/strict/1761472188 -> viable/strict/1761472188 2025-12-04T09:33:41.1020283Z * [new tag] viable/strict/1761503178 -> viable/strict/1761503178 2025-12-04T09:33:41.1021403Z * [new tag] viable/strict/1761517492 -> viable/strict/1761517492 2025-12-04T09:33:41.1022427Z * [new tag] viable/strict/1761518981 -> viable/strict/1761518981 2025-12-04T09:33:41.1023563Z * [new tag] viable/strict/1761533609 -> viable/strict/1761533609 2025-12-04T09:33:41.1024411Z * [new tag] viable/strict/1761546438 -> viable/strict/1761546438 2025-12-04T09:33:41.1025802Z * [new tag] viable/strict/1761548133 -> viable/strict/1761548133 2025-12-04T09:33:41.1027118Z * [new tag] viable/strict/1761555186 -> viable/strict/1761555186 2025-12-04T09:33:41.1028289Z * [new tag] viable/strict/1761557178 -> viable/strict/1761557178 2025-12-04T09:33:41.1029376Z * [new tag] viable/strict/1761560772 -> viable/strict/1761560772 2025-12-04T09:33:41.1030435Z * [new tag] viable/strict/1761562266 -> viable/strict/1761562266 2025-12-04T09:33:41.1031593Z * [new tag] viable/strict/1761564260 -> viable/strict/1761564260 2025-12-04T09:33:41.1032630Z * [new tag] viable/strict/1761568072 -> viable/strict/1761568072 2025-12-04T09:33:41.1033689Z * [new tag] viable/strict/1761571683 -> viable/strict/1761571683 2025-12-04T09:33:41.1034640Z * [new tag] viable/strict/1761580199 -> viable/strict/1761580199 2025-12-04T09:33:41.1035672Z * [new tag] viable/strict/1761587383 -> viable/strict/1761587383 2025-12-04T09:33:41.1036850Z * [new tag] viable/strict/1761591165 -> viable/strict/1761591165 2025-12-04T09:33:41.1037906Z * [new tag] viable/strict/1761594575 -> viable/strict/1761594575 2025-12-04T09:33:41.1038967Z * [new tag] viable/strict/1761596710 -> viable/strict/1761596710 2025-12-04T09:33:41.1040026Z * [new tag] viable/strict/1761598189 -> viable/strict/1761598189 2025-12-04T09:33:41.1041098Z * [new tag] viable/strict/1761600254 -> viable/strict/1761600254 2025-12-04T09:33:41.1042192Z * [new tag] viable/strict/1761603879 -> viable/strict/1761603879 2025-12-04T09:33:41.1043298Z * [new tag] viable/strict/1761605429 -> viable/strict/1761605429 2025-12-04T09:33:41.1044496Z * [new tag] viable/strict/1761607468 -> viable/strict/1761607468 2025-12-04T09:33:41.1045648Z * [new tag] viable/strict/1761608983 -> viable/strict/1761608983 2025-12-04T09:33:41.1046756Z * [new tag] viable/strict/1761611846 -> viable/strict/1761611846 2025-12-04T09:33:41.1048343Z * [new tag] viable/strict/1761613922 -> viable/strict/1761613922 2025-12-04T09:33:41.1049185Z * [new tag] viable/strict/1761616504 -> viable/strict/1761616504 2025-12-04T09:33:41.1050197Z * [new tag] viable/strict/1761619599 -> viable/strict/1761619599 2025-12-04T09:33:41.1051268Z * [new tag] viable/strict/1761686693 -> viable/strict/1761686693 2025-12-04T09:33:41.1052343Z * [new tag] viable/strict/1761688179 -> viable/strict/1761688179 2025-12-04T09:33:41.1053366Z * [new tag] viable/strict/1761691973 -> viable/strict/1761691973 2025-12-04T09:33:41.1054592Z * [new tag] viable/strict/1761693884 -> viable/strict/1761693884 2025-12-04T09:33:41.1055703Z * [new tag] viable/strict/1761695389 -> viable/strict/1761695389 2025-12-04T09:33:41.1056793Z * [new tag] viable/strict/1761698408 -> viable/strict/1761698408 2025-12-04T09:33:41.1057995Z * [new tag] viable/strict/1761702931 -> viable/strict/1761702931 2025-12-04T09:33:41.1059059Z * [new tag] viable/strict/1761706307 -> viable/strict/1761706307 2025-12-04T09:33:41.1060197Z * [new tag] viable/strict/1761709065 -> viable/strict/1761709065 2025-12-04T09:33:41.1061294Z * [new tag] viable/strict/1761710285 -> viable/strict/1761710285 2025-12-04T09:33:41.1062411Z * [new tag] viable/strict/1761711983 -> viable/strict/1761711983 2025-12-04T09:33:41.1063590Z * [new tag] viable/strict/1761713514 -> viable/strict/1761713514 2025-12-04T09:33:41.1064778Z * [new tag] viable/strict/1761715523 -> viable/strict/1761715523 2025-12-04T09:33:41.1065929Z * [new tag] viable/strict/1761727973 -> viable/strict/1761727973 2025-12-04T09:33:41.1067043Z * [new tag] viable/strict/1761751558 -> viable/strict/1761751558 2025-12-04T09:33:41.1068212Z * [new tag] viable/strict/1761755187 -> viable/strict/1761755187 2025-12-04T09:33:41.1069406Z * [new tag] viable/strict/1761756826 -> viable/strict/1761756826 2025-12-04T09:33:41.1070538Z * [new tag] viable/strict/1761769551 -> viable/strict/1761769551 2025-12-04T09:33:41.1071719Z * [new tag] viable/strict/1761771032 -> viable/strict/1761771032 2025-12-04T09:33:41.1072722Z * [new tag] viable/strict/1761773101 -> viable/strict/1761773101 2025-12-04T09:33:41.1073943Z * [new tag] viable/strict/1761781792 -> viable/strict/1761781792 2025-12-04T09:33:41.1075147Z * [new tag] viable/strict/1761784788 -> viable/strict/1761784788 2025-12-04T09:33:41.1076206Z * [new tag] viable/strict/1761786740 -> viable/strict/1761786740 2025-12-04T09:33:41.1077311Z * [new tag] viable/strict/1761789332 -> viable/strict/1761789332 2025-12-04T09:33:41.1078970Z * [new tag] viable/strict/1761792569 -> viable/strict/1761792569 2025-12-04T09:33:41.1080077Z * [new tag] viable/strict/1761795289 -> viable/strict/1761795289 2025-12-04T09:33:41.1081177Z * [new tag] viable/strict/1761798345 -> viable/strict/1761798345 2025-12-04T09:33:41.1082375Z * [new tag] viable/strict/1761799827 -> viable/strict/1761799827 2025-12-04T09:33:41.1083582Z * [new tag] viable/strict/1761805604 -> viable/strict/1761805604 2025-12-04T09:33:41.1084683Z * [new tag] viable/strict/1761807202 -> viable/strict/1761807202 2025-12-04T09:33:41.1085808Z * [new tag] viable/strict/1761809094 -> viable/strict/1761809094 2025-12-04T09:33:41.1086867Z * [new tag] viable/strict/1761810576 -> viable/strict/1761810576 2025-12-04T09:33:41.1087991Z * [new tag] viable/strict/1761812771 -> viable/strict/1761812771 2025-12-04T09:33:41.1089101Z * [new tag] viable/strict/1761814363 -> viable/strict/1761814363 2025-12-04T09:33:41.1090217Z * [new tag] viable/strict/1761857410 -> viable/strict/1761857410 2025-12-04T09:33:41.1091399Z * [new tag] viable/strict/1761860985 -> viable/strict/1761860985 2025-12-04T09:33:41.1092502Z * [new tag] viable/strict/1761863094 -> viable/strict/1761863094 2025-12-04T09:33:41.1093596Z * [new tag] viable/strict/1761864590 -> viable/strict/1761864590 2025-12-04T09:33:41.1094688Z * [new tag] viable/strict/1761866675 -> viable/strict/1761866675 2025-12-04T09:33:41.1096216Z * [new tag] viable/strict/1761868178 -> viable/strict/1761868178 2025-12-04T09:33:41.1097457Z * [new tag] viable/strict/1761871111 -> viable/strict/1761871111 2025-12-04T09:33:41.1098595Z * [new tag] viable/strict/1761873126 -> viable/strict/1761873126 2025-12-04T09:33:41.1099721Z * [new tag] viable/strict/1761875714 -> viable/strict/1761875714 2025-12-04T09:33:41.1100844Z * [new tag] viable/strict/1761878924 -> viable/strict/1761878924 2025-12-04T09:33:41.1101986Z * [new tag] viable/strict/1761881727 -> viable/strict/1761881727 2025-12-04T09:33:41.1103147Z * [new tag] viable/strict/1761882959 -> viable/strict/1761882959 2025-12-04T09:33:41.1104230Z * [new tag] viable/strict/1761886268 -> viable/strict/1761886268 2025-12-04T09:33:41.1105325Z * [new tag] viable/strict/1761893641 -> viable/strict/1761893641 2025-12-04T09:33:41.1106485Z * [new tag] viable/strict/1761931517 -> viable/strict/1761931517 2025-12-04T09:33:41.1107608Z * [new tag] viable/strict/1761933080 -> viable/strict/1761933080 2025-12-04T09:33:41.1108690Z * [new tag] viable/strict/1761935217 -> viable/strict/1761935217 2025-12-04T09:33:41.1109866Z * [new tag] viable/strict/1761938533 -> viable/strict/1761938533 2025-12-04T09:33:41.1111024Z * [new tag] viable/strict/1761940184 -> viable/strict/1761940184 2025-12-04T09:33:41.1112165Z * [new tag] viable/strict/1761942338 -> viable/strict/1761942338 2025-12-04T09:33:41.1113306Z * [new tag] viable/strict/1761946100 -> viable/strict/1761946100 2025-12-04T09:33:41.1114405Z * [new tag] viable/strict/1761947374 -> viable/strict/1761947374 2025-12-04T09:33:41.1115534Z * [new tag] viable/strict/1761950978 -> viable/strict/1761950978 2025-12-04T09:33:41.1116778Z * [new tag] viable/strict/1761957727 -> viable/strict/1761957727 2025-12-04T09:33:41.1117894Z * [new tag] viable/strict/1761959532 -> viable/strict/1761959532 2025-12-04T09:33:41.1119166Z * [new tag] viable/strict/1761965366 -> viable/strict/1761965366 2025-12-04T09:33:41.1120388Z * [new tag] viable/strict/1761968066 -> viable/strict/1761968066 2025-12-04T09:33:41.1121983Z * [new tag] viable/strict/1761969322 -> viable/strict/1761969322 2025-12-04T09:33:41.1123083Z * [new tag] viable/strict/1761974723 -> viable/strict/1761974723 2025-12-04T09:33:41.1124307Z * [new tag] viable/strict/1761981837 -> viable/strict/1761981837 2025-12-04T09:33:41.1125383Z * [new tag] viable/strict/1761985546 -> viable/strict/1761985546 2025-12-04T09:33:41.1126624Z * [new tag] viable/strict/1761987030 -> viable/strict/1761987030 2025-12-04T09:33:41.1127793Z * [new tag] viable/strict/1762003554 -> viable/strict/1762003554 2025-12-04T09:33:41.1128927Z * [new tag] viable/strict/1762021560 -> viable/strict/1762021560 2025-12-04T09:33:41.1129985Z * [new tag] viable/strict/1762032190 -> viable/strict/1762032190 2025-12-04T09:33:41.1131109Z * [new tag] viable/strict/1762040981 -> viable/strict/1762040981 2025-12-04T09:33:41.1132293Z * [new tag] viable/strict/1762048525 -> viable/strict/1762048525 2025-12-04T09:33:41.1133469Z * [new tag] viable/strict/1762104223 -> viable/strict/1762104223 2025-12-04T09:33:41.1134666Z * [new tag] viable/strict/1762105778 -> viable/strict/1762105778 2025-12-04T09:33:41.1135564Z * [new tag] viable/strict/1762115109 -> viable/strict/1762115109 2025-12-04T09:33:41.1136733Z * [new tag] viable/strict/1762125840 -> viable/strict/1762125840 2025-12-04T09:33:41.1137853Z * [new tag] viable/strict/1762127377 -> viable/strict/1762127377 2025-12-04T09:33:41.1139385Z * [new tag] viable/strict/1762134925 -> viable/strict/1762134925 2025-12-04T09:33:41.1140438Z * [new tag] viable/strict/1762138338 -> viable/strict/1762138338 2025-12-04T09:33:41.1141553Z * [new tag] viable/strict/1762148993 -> viable/strict/1762148993 2025-12-04T09:33:41.1142684Z * [new tag] viable/strict/1762152871 -> viable/strict/1762152871 2025-12-04T09:33:41.1143764Z * [new tag] viable/strict/1762156183 -> viable/strict/1762156183 2025-12-04T09:33:41.1144900Z * [new tag] viable/strict/1762163457 -> viable/strict/1762163457 2025-12-04T09:33:41.1145988Z * [new tag] viable/strict/1762165569 -> viable/strict/1762165569 2025-12-04T09:33:41.1147086Z * [new tag] viable/strict/1762169035 -> viable/strict/1762169035 2025-12-04T09:33:41.1149967Z * [new tag] viable/strict/1762174936 -> viable/strict/1762174936 2025-12-04T09:33:41.1150256Z * [new tag] viable/strict/1762194412 -> viable/strict/1762194412 2025-12-04T09:33:41.1150744Z * [new tag] viable/strict/1762195876 -> viable/strict/1762195876 2025-12-04T09:33:41.1151613Z * [new tag] viable/strict/1762197788 -> viable/strict/1762197788 2025-12-04T09:33:41.1152720Z * [new tag] viable/strict/1762199389 -> viable/strict/1762199389 2025-12-04T09:33:41.1154069Z * [new tag] viable/strict/1762206585 -> viable/strict/1762206585 2025-12-04T09:33:41.1155293Z * [new tag] viable/strict/1762210184 -> viable/strict/1762210184 2025-12-04T09:33:41.1156384Z * [new tag] viable/strict/1762218736 -> viable/strict/1762218736 2025-12-04T09:33:41.1157554Z * [new tag] viable/strict/1762224529 -> viable/strict/1762224529 2025-12-04T09:33:41.1158851Z * [new tag] viable/strict/1762227253 -> viable/strict/1762227253 2025-12-04T09:33:41.1159638Z * [new tag] viable/strict/1762228515 -> viable/strict/1762228515 2025-12-04T09:33:41.1160948Z * [new tag] viable/strict/1762230349 -> viable/strict/1762230349 2025-12-04T09:33:41.1162099Z * [new tag] viable/strict/1762231859 -> viable/strict/1762231859 2025-12-04T09:33:41.1163217Z * [new tag] viable/strict/1762233925 -> viable/strict/1762233925 2025-12-04T09:33:41.1164614Z * [new tag] viable/strict/1762237630 -> viable/strict/1762237630 2025-12-04T09:33:41.1165512Z * [new tag] viable/strict/1762253522 -> viable/strict/1762253522 2025-12-04T09:33:41.1166829Z * [new tag] viable/strict/1762278588 -> viable/strict/1762278588 2025-12-04T09:33:41.1167962Z * [new tag] viable/strict/1762284203 -> viable/strict/1762284203 2025-12-04T09:33:41.1169076Z * [new tag] viable/strict/1762289446 -> viable/strict/1762289446 2025-12-04T09:33:41.1170271Z * [new tag] viable/strict/1762291515 -> viable/strict/1762291515 2025-12-04T09:33:41.1171384Z * [new tag] viable/strict/1762295100 -> viable/strict/1762295100 2025-12-04T09:33:41.1172267Z * [new tag] viable/strict/1762296590 -> viable/strict/1762296590 2025-12-04T09:33:41.1173331Z * [new tag] viable/strict/1762300179 -> viable/strict/1762300179 2025-12-04T09:33:41.1174192Z * [new tag] viable/strict/1762303207 -> viable/strict/1762303207 2025-12-04T09:33:41.1175436Z * [new tag] viable/strict/1762386584 -> viable/strict/1762386584 2025-12-04T09:33:41.1176549Z * [new tag] viable/strict/1762391537 -> viable/strict/1762391537 2025-12-04T09:33:41.1177636Z * [new tag] viable/strict/1762394119 -> viable/strict/1762394119 2025-12-04T09:33:41.1179166Z * [new tag] viable/strict/1762397437 -> viable/strict/1762397437 2025-12-04T09:33:41.1180283Z * [new tag] viable/strict/1762400256 -> viable/strict/1762400256 2025-12-04T09:33:41.1181432Z * [new tag] viable/strict/1762401469 -> viable/strict/1762401469 2025-12-04T09:33:41.1182650Z * [new tag] viable/strict/1762408195 -> viable/strict/1762408195 2025-12-04T09:33:41.1183877Z * [new tag] viable/strict/1762410411 -> viable/strict/1762410411 2025-12-04T09:33:41.1185000Z * [new tag] viable/strict/1762417613 -> viable/strict/1762417613 2025-12-04T09:33:41.1186115Z * [new tag] viable/strict/1762419198 -> viable/strict/1762419198 2025-12-04T09:33:41.1187335Z * [new tag] viable/strict/1762422656 -> viable/strict/1762422656 2025-12-04T09:33:41.1188872Z * [new tag] viable/strict/1762424746 -> viable/strict/1762424746 2025-12-04T09:33:41.1190074Z * [new tag] viable/strict/1762446386 -> viable/strict/1762446386 2025-12-04T09:33:41.1191170Z * [new tag] viable/strict/1762449912 -> viable/strict/1762449912 2025-12-04T09:33:41.1192303Z * [new tag] viable/strict/1762457031 -> viable/strict/1762457031 2025-12-04T09:33:41.1193560Z * [new tag] viable/strict/1762462441 -> viable/strict/1762462441 2025-12-04T09:33:41.1194702Z * [new tag] viable/strict/1762467909 -> viable/strict/1762467909 2025-12-04T09:33:41.1196515Z * [new tag] viable/strict/1762471493 -> viable/strict/1762471493 2025-12-04T09:33:41.1202191Z * [new tag] viable/strict/1762475990 -> viable/strict/1762475990 2025-12-04T09:33:41.1203909Z * [new tag] viable/strict/1762477933 -> viable/strict/1762477933 2025-12-04T09:33:41.1204845Z * [new tag] viable/strict/1762491053 -> viable/strict/1762491053 2025-12-04T09:33:41.1206308Z * [new tag] viable/strict/1762493118 -> viable/strict/1762493118 2025-12-04T09:33:41.1207312Z * [new tag] viable/strict/1762498442 -> viable/strict/1762498442 2025-12-04T09:33:41.1208265Z * [new tag] viable/strict/1762501778 -> viable/strict/1762501778 2025-12-04T09:33:41.1209345Z * [new tag] viable/strict/1762504001 -> viable/strict/1762504001 2025-12-04T09:33:41.1210775Z * [new tag] viable/strict/1762505583 -> viable/strict/1762505583 2025-12-04T09:33:41.1211854Z * [new tag] viable/strict/1762507523 -> viable/strict/1762507523 2025-12-04T09:33:41.1213057Z * [new tag] viable/strict/1762511140 -> viable/strict/1762511140 2025-12-04T09:33:41.1214504Z * [new tag] viable/strict/1762512632 -> viable/strict/1762512632 2025-12-04T09:33:41.1215567Z * [new tag] viable/strict/1762520467 -> viable/strict/1762520467 2025-12-04T09:33:41.1216716Z * [new tag] viable/strict/1762522016 -> viable/strict/1762522016 2025-12-04T09:33:41.1217962Z * [new tag] viable/strict/1762530591 -> viable/strict/1762530591 2025-12-04T09:33:41.1219150Z * [new tag] viable/strict/1762543405 -> viable/strict/1762543405 2025-12-04T09:33:41.1220086Z * [new tag] viable/strict/1762544998 -> viable/strict/1762544998 2025-12-04T09:33:41.1221215Z * [new tag] viable/strict/1762552182 -> viable/strict/1762552182 2025-12-04T09:33:41.1222391Z * [new tag] viable/strict/1762554297 -> viable/strict/1762554297 2025-12-04T09:33:41.1223321Z * [new tag] viable/strict/1762559381 -> viable/strict/1762559381 2025-12-04T09:33:41.1224492Z * [new tag] viable/strict/1762562222 -> viable/strict/1762562222 2025-12-04T09:33:41.1225606Z * [new tag] viable/strict/1762564319 -> viable/strict/1762564319 2025-12-04T09:33:41.1226617Z * [new tag] viable/strict/1762566904 -> viable/strict/1762566904 2025-12-04T09:33:41.1227721Z * [new tag] viable/strict/1762569781 -> viable/strict/1762569781 2025-12-04T09:33:41.1228760Z * [new tag] viable/strict/1762575940 -> viable/strict/1762575940 2025-12-04T09:33:41.1229944Z * [new tag] viable/strict/1762580974 -> viable/strict/1762580974 2025-12-04T09:33:41.1231132Z * [new tag] viable/strict/1762583185 -> viable/strict/1762583185 2025-12-04T09:33:41.1232246Z * [new tag] viable/strict/1762586647 -> viable/strict/1762586647 2025-12-04T09:33:41.1233336Z * [new tag] viable/strict/1762588183 -> viable/strict/1762588183 2025-12-04T09:33:41.1234449Z * [new tag] viable/strict/1762593886 -> viable/strict/1762593886 2025-12-04T09:33:41.1235818Z * [new tag] viable/strict/1762650743 -> viable/strict/1762650743 2025-12-04T09:33:41.1236939Z * [new tag] viable/strict/1762653328 -> viable/strict/1762653328 2025-12-04T09:33:41.1238071Z * [new tag] viable/strict/1762659342 -> viable/strict/1762659342 2025-12-04T09:33:41.1239203Z * [new tag] viable/strict/1762662360 -> viable/strict/1762662360 2025-12-04T09:33:41.1240936Z * [new tag] viable/strict/1762667377 -> viable/strict/1762667377 2025-12-04T09:33:41.1241900Z * [new tag] viable/strict/1762671090 -> viable/strict/1762671090 2025-12-04T09:33:41.1243062Z * [new tag] viable/strict/1762680284 -> viable/strict/1762680284 2025-12-04T09:33:41.1244266Z * [new tag] viable/strict/1762683900 -> viable/strict/1762683900 2025-12-04T09:33:41.1245390Z * [new tag] viable/strict/1762705541 -> viable/strict/1762705541 2025-12-04T09:33:41.1246447Z * [new tag] viable/strict/1762709004 -> viable/strict/1762709004 2025-12-04T09:33:41.1247703Z * [new tag] viable/strict/1762746004 -> viable/strict/1762746004 2025-12-04T09:33:41.1249144Z * [new tag] viable/strict/1762748799 -> viable/strict/1762748799 2025-12-04T09:33:41.1250236Z * [new tag] viable/strict/1762759504 -> viable/strict/1762759504 2025-12-04T09:33:41.1251607Z * [new tag] viable/strict/1762760973 -> viable/strict/1762760973 2025-12-04T09:33:41.1252668Z * [new tag] viable/strict/1762775374 -> viable/strict/1762775374 2025-12-04T09:33:41.1253874Z * [new tag] viable/strict/1762777661 -> viable/strict/1762777661 2025-12-04T09:33:41.1254984Z * [new tag] viable/strict/1762779774 -> viable/strict/1762779774 2025-12-04T09:33:41.1256414Z * [new tag] viable/strict/1762781259 -> viable/strict/1762781259 2025-12-04T09:33:41.1257646Z * [new tag] viable/strict/1762793628 -> viable/strict/1762793628 2025-12-04T09:33:41.1258859Z * [new tag] viable/strict/1762800711 -> viable/strict/1762800711 2025-12-04T09:33:41.1259979Z * [new tag] viable/strict/1762809894 -> viable/strict/1762809894 2025-12-04T09:33:41.1261089Z * [new tag] viable/strict/1762811384 -> viable/strict/1762811384 2025-12-04T09:33:41.1262355Z * [new tag] viable/strict/1762813841 -> viable/strict/1762813841 2025-12-04T09:33:41.1263546Z * [new tag] viable/strict/1762815047 -> viable/strict/1762815047 2025-12-04T09:33:41.1265001Z * [new tag] viable/strict/1762817094 -> viable/strict/1762817094 2025-12-04T09:33:41.1266015Z * [new tag] viable/strict/1762818582 -> viable/strict/1762818582 2025-12-04T09:33:41.1267140Z * [new tag] viable/strict/1762821623 -> viable/strict/1762821623 2025-12-04T09:33:41.1268130Z * [new tag] viable/strict/1762823531 -> viable/strict/1762823531 2025-12-04T09:33:41.1269380Z * [new tag] viable/strict/1762849583 -> viable/strict/1762849583 2025-12-04T09:33:41.1270496Z * [new tag] viable/strict/1762851200 -> viable/strict/1762851200 2025-12-04T09:33:41.1271616Z * [new tag] viable/strict/1762854603 -> viable/strict/1762854603 2025-12-04T09:33:41.1272780Z * [new tag] viable/strict/1762858276 -> viable/strict/1762858276 2025-12-04T09:33:41.1274192Z * [new tag] viable/strict/1762860891 -> viable/strict/1762860891 2025-12-04T09:33:41.1276007Z * [new tag] viable/strict/1762866174 -> viable/strict/1762866174 2025-12-04T09:33:41.1277812Z * [new tag] viable/strict/1762867653 -> viable/strict/1762867653 2025-12-04T09:33:41.1278916Z * [new tag] viable/strict/1762872669 -> viable/strict/1762872669 2025-12-04T09:33:41.1279901Z * [new tag] viable/strict/1762878380 -> viable/strict/1762878380 2025-12-04T09:33:41.1281074Z * [new tag] viable/strict/1762889003 -> viable/strict/1762889003 2025-12-04T09:33:41.1282272Z * [new tag] viable/strict/1762890589 -> viable/strict/1762890589 2025-12-04T09:33:41.1283387Z * [new tag] viable/strict/1762892743 -> viable/strict/1762892743 2025-12-04T09:33:41.1284571Z * [new tag] viable/strict/1762894271 -> viable/strict/1762894271 2025-12-04T09:33:41.1285571Z * [new tag] viable/strict/1762896287 -> viable/strict/1762896287 2025-12-04T09:33:41.1286706Z * [new tag] viable/strict/1762915871 -> viable/strict/1762915871 2025-12-04T09:33:41.1287885Z * [new tag] viable/strict/1762918569 -> viable/strict/1762918569 2025-12-04T09:33:41.1289100Z * [new tag] viable/strict/1762919776 -> viable/strict/1762919776 2025-12-04T09:33:41.1290275Z * [new tag] viable/strict/1762923072 -> viable/strict/1762923072 2025-12-04T09:33:41.1291524Z * [new tag] viable/strict/1762928826 -> viable/strict/1762928826 2025-12-04T09:33:41.1292750Z * [new tag] viable/strict/1762930451 -> viable/strict/1762930451 2025-12-04T09:33:41.1293869Z * [new tag] viable/strict/1762933780 -> viable/strict/1762933780 2025-12-04T09:33:41.1295056Z * [new tag] viable/strict/1762937638 -> viable/strict/1762937638 2025-12-04T09:33:41.1296615Z * [new tag] viable/strict/1762939545 -> viable/strict/1762939545 2025-12-04T09:33:41.1298095Z * [new tag] viable/strict/1762962692 -> viable/strict/1762962692 2025-12-04T09:33:41.1299154Z * [new tag] viable/strict/1762979143 -> viable/strict/1762979143 2025-12-04T09:33:41.1300349Z * [new tag] viable/strict/1762984188 -> viable/strict/1762984188 2025-12-04T09:33:41.1301333Z * [new tag] viable/strict/1762986306 -> viable/strict/1762986306 2025-12-04T09:33:41.1302718Z * [new tag] viable/strict/1762989903 -> viable/strict/1762989903 2025-12-04T09:33:41.1303631Z * [new tag] viable/strict/1762991377 -> viable/strict/1762991377 2025-12-04T09:33:41.1304783Z * [new tag] viable/strict/1762998921 -> viable/strict/1762998921 2025-12-04T09:33:41.1306098Z * [new tag] viable/strict/1763002287 -> viable/strict/1763002287 2025-12-04T09:33:41.1307383Z * [new tag] viable/strict/1763016840 -> viable/strict/1763016840 2025-12-04T09:33:41.1308483Z * [new tag] viable/strict/1763020180 -> viable/strict/1763020180 2025-12-04T09:33:41.1309701Z * [new tag] viable/strict/1763027421 -> viable/strict/1763027421 2025-12-04T09:33:41.1310893Z * [new tag] viable/strict/1763031120 -> viable/strict/1763031120 2025-12-04T09:33:41.1312090Z * [new tag] viable/strict/1763036861 -> viable/strict/1763036861 2025-12-04T09:33:41.1313332Z * [new tag] viable/strict/1763038993 -> viable/strict/1763038993 2025-12-04T09:33:41.1314561Z * [new tag] viable/strict/1763054703 -> viable/strict/1763054703 2025-12-04T09:33:41.1315591Z * [new tag] viable/strict/1763067061 -> viable/strict/1763067061 2025-12-04T09:33:41.1316744Z * [new tag] viable/strict/1763070847 -> viable/strict/1763070847 2025-12-04T09:33:41.1317916Z * [new tag] viable/strict/1763072706 -> viable/strict/1763072706 2025-12-04T09:33:41.1319110Z * [new tag] viable/strict/1763076302 -> viable/strict/1763076302 2025-12-04T09:33:41.1320338Z * [new tag] viable/strict/1763080816 -> viable/strict/1763080816 2025-12-04T09:33:41.1321493Z * [new tag] viable/strict/1763082732 -> viable/strict/1763082732 2025-12-04T09:33:41.1322554Z * [new tag] viable/strict/1763085329 -> viable/strict/1763085329 2025-12-04T09:33:41.1323685Z * [new tag] viable/strict/1763088623 -> viable/strict/1763088623 2025-12-04T09:33:41.1325083Z * [new tag] viable/strict/1763091402 -> viable/strict/1763091402 2025-12-04T09:33:41.1326138Z * [new tag] viable/strict/1763092602 -> viable/strict/1763092602 2025-12-04T09:33:41.1327268Z * [new tag] viable/strict/1763094355 -> viable/strict/1763094355 2025-12-04T09:33:41.1328399Z * [new tag] viable/strict/1763099390 -> viable/strict/1763099390 2025-12-04T09:33:41.1329569Z * [new tag] viable/strict/1763101608 -> viable/strict/1763101608 2025-12-04T09:33:41.1330783Z * [new tag] viable/strict/1763105102 -> viable/strict/1763105102 2025-12-04T09:33:41.1332017Z * [new tag] viable/strict/1763112347 -> viable/strict/1763112347 2025-12-04T09:33:41.1333163Z * [new tag] viable/strict/1763119471 -> viable/strict/1763119471 2025-12-04T09:33:41.1334262Z * [new tag] viable/strict/1763126835 -> viable/strict/1763126835 2025-12-04T09:33:41.1335127Z * [new tag] viable/strict/1763149779 -> viable/strict/1763149779 2025-12-04T09:33:41.1336273Z * [new tag] viable/strict/1763164178 -> viable/strict/1763164178 2025-12-04T09:33:41.1337569Z * [new tag] viable/strict/1763167104 -> viable/strict/1763167104 2025-12-04T09:33:41.1338679Z * [new tag] viable/strict/1763169132 -> viable/strict/1763169132 2025-12-04T09:33:41.1339833Z * [new tag] viable/strict/1763171708 -> viable/strict/1763171708 2025-12-04T09:33:41.1340960Z * [new tag] viable/strict/1763174759 -> viable/strict/1763174759 2025-12-04T09:33:41.1342172Z * [new tag] viable/strict/1763180744 -> viable/strict/1763180744 2025-12-04T09:33:41.1343329Z * [new tag] viable/strict/1763182227 -> viable/strict/1763182227 2025-12-04T09:33:41.1344436Z * [new tag] viable/strict/1763184309 -> viable/strict/1763184309 2025-12-04T09:33:41.1346386Z * [new tag] viable/strict/1763187991 -> viable/strict/1763187991 2025-12-04T09:33:41.1347409Z * [new tag] viable/strict/1763191445 -> viable/strict/1763191445 2025-12-04T09:33:41.1348695Z * [new tag] viable/strict/1763195152 -> viable/strict/1763195152 2025-12-04T09:33:41.1349655Z * [new tag] viable/strict/1763205769 -> viable/strict/1763205769 2025-12-04T09:33:41.1351117Z * [new tag] viable/strict/1763246990 -> viable/strict/1763246990 2025-12-04T09:33:41.1352237Z * [new tag] viable/strict/1763261578 -> viable/strict/1763261578 2025-12-04T09:33:41.1353826Z * [new tag] viable/strict/1763286573 -> viable/strict/1763286573 2025-12-04T09:33:41.1354751Z * [new tag] viable/strict/1763292167 -> viable/strict/1763292167 2025-12-04T09:33:41.1355910Z * [new tag] viable/strict/1763333386 -> viable/strict/1763333386 2025-12-04T09:33:41.1357025Z * [new tag] viable/strict/1763340082 -> viable/strict/1763340082 2025-12-04T09:33:41.1359167Z * [new tag] viable/strict/1763364324 -> viable/strict/1763364324 2025-12-04T09:33:41.1360206Z * [new tag] viable/strict/1763371569 -> viable/strict/1763371569 2025-12-04T09:33:41.1361361Z * [new tag] viable/strict/1763373067 -> viable/strict/1763373067 2025-12-04T09:33:41.1362476Z * [new tag] viable/strict/1763375157 -> viable/strict/1763375157 2025-12-04T09:33:41.1363635Z * [new tag] viable/strict/1763382462 -> viable/strict/1763382462 2025-12-04T09:33:41.1365000Z * [new tag] viable/strict/1763394661 -> viable/strict/1763394661 2025-12-04T09:33:41.1366373Z * [new tag] viable/strict/1763396797 -> viable/strict/1763396797 2025-12-04T09:33:41.1367460Z * [new tag] viable/strict/1763398542 -> viable/strict/1763398542 2025-12-04T09:33:41.1368635Z * [new tag] viable/strict/1763401807 -> viable/strict/1763401807 2025-12-04T09:33:41.1369635Z * [new tag] viable/strict/1763414698 -> viable/strict/1763414698 2025-12-04T09:33:41.1370860Z * [new tag] viable/strict/1763419807 -> viable/strict/1763419807 2025-12-04T09:33:41.1371999Z * [new tag] viable/strict/1763426369 -> viable/strict/1763426369 2025-12-04T09:33:41.1373229Z * [new tag] viable/strict/1763428331 -> viable/strict/1763428331 2025-12-04T09:33:41.1374394Z * [new tag] viable/strict/1763430922 -> viable/strict/1763430922 2025-12-04T09:33:41.1375434Z * [new tag] viable/strict/1763434184 -> viable/strict/1763434184 2025-12-04T09:33:41.1376552Z * [new tag] viable/strict/1763439973 -> viable/strict/1763439973 2025-12-04T09:33:41.1378174Z * [new tag] viable/strict/1763444995 -> viable/strict/1763444995 2025-12-04T09:33:41.1379072Z * [new tag] viable/strict/1763447206 -> viable/strict/1763447206 2025-12-04T09:33:41.1380244Z * [new tag] viable/strict/1763448826 -> viable/strict/1763448826 2025-12-04T09:33:41.1381425Z * [new tag] viable/strict/1763450717 -> viable/strict/1763450717 2025-12-04T09:33:41.1382592Z * [new tag] viable/strict/1763452183 -> viable/strict/1763452183 2025-12-04T09:33:41.1383799Z * [new tag] viable/strict/1763457945 -> viable/strict/1763457945 2025-12-04T09:33:41.1384936Z * [new tag] viable/strict/1763459439 -> viable/strict/1763459439 2025-12-04T09:33:41.1385930Z * [new tag] viable/strict/1763461556 -> viable/strict/1763461556 2025-12-04T09:33:41.1387091Z * [new tag] viable/strict/1763463103 -> viable/strict/1763463103 2025-12-04T09:33:41.1388332Z * [new tag] viable/strict/1763465100 -> viable/strict/1763465100 2025-12-04T09:33:41.1389406Z * [new tag] viable/strict/1763468866 -> viable/strict/1763468866 2025-12-04T09:33:41.1390378Z * [new tag] viable/strict/1763493823 -> viable/strict/1763493823 2025-12-04T09:33:41.1391408Z * [new tag] viable/strict/1763496249 -> viable/strict/1763496249 2025-12-04T09:33:41.1392510Z * [new tag] viable/strict/1763502620 -> viable/strict/1763502620 2025-12-04T09:33:41.1393728Z * [new tag] viable/strict/1763504715 -> viable/strict/1763504715 2025-12-04T09:33:41.1394860Z * [new tag] viable/strict/1763506208 -> viable/strict/1763506208 2025-12-04T09:33:41.1396153Z * [new tag] viable/strict/1763520590 -> viable/strict/1763520590 2025-12-04T09:33:41.1397555Z * [new tag] viable/strict/1763523357 -> viable/strict/1763523357 2025-12-04T09:33:41.1398717Z * [new tag] viable/strict/1763529922 -> viable/strict/1763529922 2025-12-04T09:33:41.1399914Z * [new tag] viable/strict/1763531408 -> viable/strict/1763531408 2025-12-04T09:33:41.1401069Z * [new tag] viable/strict/1763533622 -> viable/strict/1763533622 2025-12-04T09:33:41.1402225Z * [new tag] viable/strict/1763538576 -> viable/strict/1763538576 2025-12-04T09:33:41.1403580Z * [new tag] viable/strict/1763545823 -> viable/strict/1763545823 2025-12-04T09:33:41.1404468Z * [new tag] viable/strict/1763547951 -> viable/strict/1763547951 2025-12-04T09:33:41.1405688Z * [new tag] viable/strict/1763551477 -> viable/strict/1763551477 2025-12-04T09:33:41.1406865Z * [new tag] viable/strict/1763552982 -> viable/strict/1763552982 2025-12-04T09:33:41.1408131Z * [new tag] viable/strict/1763594698 -> viable/strict/1763594698 2025-12-04T09:33:41.1409223Z * [new tag] viable/strict/1763596178 -> viable/strict/1763596178 2025-12-04T09:33:41.1410384Z * [new tag] viable/strict/1763599155 -> viable/strict/1763599155 2025-12-04T09:33:41.1411518Z * [new tag] viable/strict/1763603717 -> viable/strict/1763603717 2025-12-04T09:33:41.1412691Z * [new tag] viable/strict/1763606923 -> viable/strict/1763606923 2025-12-04T09:33:41.1413876Z * [new tag] viable/strict/1763609715 -> viable/strict/1763609715 2025-12-04T09:33:41.1415004Z * [new tag] viable/strict/1763612757 -> viable/strict/1763612757 2025-12-04T09:33:41.1416151Z * [new tag] viable/strict/1763616325 -> viable/strict/1763616325 2025-12-04T09:33:41.1417344Z * [new tag] viable/strict/1763623509 -> viable/strict/1763623509 2025-12-04T09:33:41.1418810Z * [new tag] viable/strict/1763624984 -> viable/strict/1763624984 2025-12-04T09:33:41.1419962Z * [new tag] viable/strict/1763628796 -> viable/strict/1763628796 2025-12-04T09:33:41.1420948Z * [new tag] viable/strict/1763634343 -> viable/strict/1763634343 2025-12-04T09:33:41.1422092Z * [new tag] viable/strict/1763635867 -> viable/strict/1763635867 2025-12-04T09:33:41.1423553Z * [new tag] viable/strict/1763639382 -> viable/strict/1763639382 2025-12-04T09:33:41.1424587Z * [new tag] viable/strict/1763646626 -> viable/strict/1763646626 2025-12-04T09:33:41.1425942Z * [new tag] viable/strict/1763655997 -> viable/strict/1763655997 2025-12-04T09:33:41.1427158Z * [new tag] viable/strict/1763659444 -> viable/strict/1763659444 2025-12-04T09:33:41.1428323Z * [new tag] viable/strict/1763660992 -> viable/strict/1763660992 2025-12-04T09:33:41.1429925Z * [new tag] viable/strict/1763663201 -> viable/strict/1763663201 2025-12-04T09:33:41.1431061Z * [new tag] viable/strict/1763670362 -> viable/strict/1763670362 2025-12-04T09:33:41.1432041Z * [new tag] viable/strict/1763675378 -> viable/strict/1763675378 2025-12-04T09:33:41.1433176Z * [new tag] viable/strict/1763693343 -> viable/strict/1763693343 2025-12-04T09:33:41.1434285Z * [new tag] viable/strict/1763696088 -> viable/strict/1763696088 2025-12-04T09:33:41.1435683Z * [new tag] viable/strict/1763697343 -> viable/strict/1763697343 2025-12-04T09:33:41.1436739Z * [new tag] viable/strict/1763699165 -> viable/strict/1763699165 2025-12-04T09:33:41.1437873Z * [new tag] viable/strict/1763700660 -> viable/strict/1763700660 2025-12-04T09:33:41.1438973Z * [new tag] viable/strict/1763704209 -> viable/strict/1763704209 2025-12-04T09:33:41.1440145Z * [new tag] viable/strict/1763706411 -> viable/strict/1763706411 2025-12-04T09:33:41.1441239Z * [new tag] viable/strict/1763708082 -> viable/strict/1763708082 2025-12-04T09:33:41.1442297Z * [new tag] viable/strict/1763711381 -> viable/strict/1763711381 2025-12-04T09:33:41.1443370Z * [new tag] viable/strict/1763713593 -> viable/strict/1763713593 2025-12-04T09:33:41.1444561Z * [new tag] viable/strict/1763715201 -> viable/strict/1763715201 2025-12-04T09:33:41.1445665Z * [new tag] viable/strict/1763733017 -> viable/strict/1763733017 2025-12-04T09:33:41.1446857Z * [new tag] viable/strict/1763735108 -> viable/strict/1763735108 2025-12-04T09:33:41.1447960Z * [new tag] viable/strict/1763749579 -> viable/strict/1763749579 2025-12-04T09:33:41.1449072Z * [new tag] viable/strict/1763751113 -> viable/strict/1763751113 2025-12-04T09:33:41.1450198Z * [new tag] viable/strict/1763753035 -> viable/strict/1763753035 2025-12-04T09:33:41.1451414Z * [new tag] viable/strict/1763754578 -> viable/strict/1763754578 2025-12-04T09:33:41.1452555Z * [new tag] viable/strict/1763756748 -> viable/strict/1763756748 2025-12-04T09:33:41.1453660Z * [new tag] viable/strict/1763758205 -> viable/strict/1763758205 2025-12-04T09:33:41.1454611Z * [new tag] viable/strict/1763764050 -> viable/strict/1763764050 2025-12-04T09:33:41.1455740Z * [new tag] viable/strict/1763771887 -> viable/strict/1763771887 2025-12-04T09:33:41.1457225Z * [new tag] viable/strict/1763773920 -> viable/strict/1763773920 2025-12-04T09:33:41.1458309Z * [new tag] viable/strict/1763776501 -> viable/strict/1763776501 2025-12-04T09:33:41.1459395Z * [new tag] viable/strict/1763779437 -> viable/strict/1763779437 2025-12-04T09:33:41.1460848Z * [new tag] viable/strict/1763781038 -> viable/strict/1763781038 2025-12-04T09:33:41.1461822Z * [new tag] viable/strict/1763782245 -> viable/strict/1763782245 2025-12-04T09:33:41.1462900Z * [new tag] viable/strict/1763785568 -> viable/strict/1763785568 2025-12-04T09:33:41.1464307Z * [new tag] viable/strict/1763787006 -> viable/strict/1763787006 2025-12-04T09:33:41.1465382Z * [new tag] viable/strict/1763789103 -> viable/strict/1763789103 2025-12-04T09:33:41.1466528Z * [new tag] viable/strict/1763790578 -> viable/strict/1763790578 2025-12-04T09:33:41.1467677Z * [new tag] viable/strict/1763796275 -> viable/strict/1763796275 2025-12-04T09:33:41.1469177Z * [new tag] viable/strict/1763801465 -> viable/strict/1763801465 2025-12-04T09:33:41.1470198Z * [new tag] viable/strict/1763803522 -> viable/strict/1763803522 2025-12-04T09:33:41.1471275Z * [new tag] viable/strict/1763808581 -> viable/strict/1763808581 2025-12-04T09:33:41.1472405Z * [new tag] viable/strict/1763840977 -> viable/strict/1763840977 2025-12-04T09:33:41.1473515Z * [new tag] viable/strict/1763846659 -> viable/strict/1763846659 2025-12-04T09:33:41.1474587Z * [new tag] viable/strict/1763872065 -> viable/strict/1763872065 2025-12-04T09:33:41.1475818Z * [new tag] viable/strict/1763873648 -> viable/strict/1763873648 2025-12-04T09:33:41.1476972Z * [new tag] viable/strict/1763875506 -> viable/strict/1763875506 2025-12-04T09:33:41.1477974Z * [new tag] viable/strict/1763889904 -> viable/strict/1763889904 2025-12-04T09:33:41.1479091Z * [new tag] viable/strict/1763930999 -> viable/strict/1763930999 2025-12-04T09:33:41.1480527Z * [new tag] viable/strict/1763944964 -> viable/strict/1763944964 2025-12-04T09:33:41.1481278Z * [new tag] viable/strict/1763958474 -> viable/strict/1763958474 2025-12-04T09:33:41.1482392Z * [new tag] viable/strict/1763967263 -> viable/strict/1763967263 2025-12-04T09:33:41.1483560Z * [new tag] viable/strict/1763972803 -> viable/strict/1763972803 2025-12-04T09:33:41.1484653Z * [new tag] viable/strict/1763976376 -> viable/strict/1763976376 2025-12-04T09:33:41.1485784Z * [new tag] viable/strict/1763989404 -> viable/strict/1763989404 2025-12-04T09:33:41.1486878Z * [new tag] viable/strict/1763990887 -> viable/strict/1763990887 2025-12-04T09:33:41.1487980Z * [new tag] viable/strict/1764019919 -> viable/strict/1764019919 2025-12-04T09:33:41.1489163Z * [new tag] viable/strict/1764023134 -> viable/strict/1764023134 2025-12-04T09:33:41.1490130Z * [new tag] viable/strict/1764024593 -> viable/strict/1764024593 2025-12-04T09:33:41.1491290Z * [new tag] viable/strict/1764026706 -> viable/strict/1764026706 2025-12-04T09:33:41.1492776Z * [new tag] viable/strict/1764031139 -> viable/strict/1764031139 2025-12-04T09:33:41.1493832Z * [new tag] viable/strict/1764033131 -> viable/strict/1764033131 2025-12-04T09:33:41.1494788Z * [new tag] viable/strict/1764035725 -> viable/strict/1764035725 2025-12-04T09:33:41.1495776Z * [new tag] viable/strict/1764624265 -> viable/strict/1764624265 2025-12-04T09:33:41.1496941Z * [new tag] viable/strict/1764631514 -> viable/strict/1764631514 2025-12-04T09:33:41.1497968Z * [new tag] viable/strict/1764632987 -> viable/strict/1764632987 2025-12-04T09:33:41.1498940Z * [new tag] viable/strict/1764636063 -> viable/strict/1764636063 2025-12-04T09:33:41.1499984Z * [new tag] viable/strict/1764643975 -> viable/strict/1764643975 2025-12-04T09:33:41.1500913Z * [new tag] viable/strict/1764646859 -> viable/strict/1764646859 2025-12-04T09:33:41.1502486Z * [new tag] viable/strict/1764653120 -> viable/strict/1764653120 2025-12-04T09:33:41.1503254Z * [new tag] viable/strict/1764654632 -> viable/strict/1764654632 2025-12-04T09:33:41.1504255Z * [new tag] viable/strict/1764656821 -> viable/strict/1764656821 2025-12-04T09:33:41.1505141Z * [new tag] viable/strict/1764658557 -> viable/strict/1764658557 2025-12-04T09:33:41.1506100Z * [new tag] viable/strict/1764660333 -> viable/strict/1764660333 2025-12-04T09:33:41.1507058Z * [new tag] viable/strict/1764661812 -> viable/strict/1764661812 2025-12-04T09:33:41.1508016Z * [new tag] viable/strict/1764664023 -> viable/strict/1764664023 2025-12-04T09:33:41.1508928Z * [new tag] viable/strict/1764669150 -> viable/strict/1764669150 2025-12-04T09:33:41.1509925Z * [new tag] viable/strict/1764680709 -> viable/strict/1764680709 2025-12-04T09:33:41.1510884Z * [new tag] viable/strict/1764687619 -> viable/strict/1764687619 2025-12-04T09:33:41.1511879Z * [new tag] viable/strict/1764696355 -> viable/strict/1764696355 2025-12-04T09:33:41.1512796Z * [new tag] viable/strict/1764701767 -> viable/strict/1764701767 2025-12-04T09:33:41.1513800Z * [new tag] viable/strict/1764710768 -> viable/strict/1764710768 2025-12-04T09:33:41.1514724Z * [new tag] viable/strict/1764716202 -> viable/strict/1764716202 2025-12-04T09:33:41.1515668Z * [new tag] viable/strict/1764793566 -> viable/strict/1764793566 2025-12-04T09:33:41.1516629Z * [new tag] viable/strict/1764797093 -> viable/strict/1764797093 2025-12-04T09:33:41.1517597Z * [new tag] viable/strict/1764800729 -> viable/strict/1764800729 2025-12-04T09:33:41.1518903Z * [new tag] whc_flight_1 -> whc_flight_1 2025-12-04T09:33:41.1519931Z * [new tag] whc_flight_2 -> whc_flight_2 2025-12-04T09:33:41.1521356Z * [new tag] whc_flight_4 -> whc_flight_4 2025-12-04T09:33:41.2372349Z [command]/usr/bin/git rev-parse --verify --quiet ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32^{object} 2025-12-04T09:33:41.2401844Z ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:33:41.2405445Z ##[endgroup] 2025-12-04T09:33:41.2405914Z ##[group]Determining the checkout info 2025-12-04T09:33:41.2406875Z ##[endgroup] 2025-12-04T09:33:41.2411120Z [command]/usr/bin/git sparse-checkout disable 2025-12-04T09:33:41.2445316Z [command]/usr/bin/git config --local --unset-all extensions.worktreeConfig 2025-12-04T09:33:41.2471992Z ##[group]Checking out the ref 2025-12-04T09:33:41.2475808Z [command]/usr/bin/git checkout --progress --force ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:33:42.2931341Z Updating files: 75% (15218/20121) 2025-12-04T09:33:42.3089794Z Updating files: 76% (15292/20121) 2025-12-04T09:33:42.3234172Z Updating files: 77% (15494/20121) 2025-12-04T09:33:42.3460815Z Updating files: 78% (15695/20121) 2025-12-04T09:33:42.3755551Z Updating files: 79% (15896/20121) 2025-12-04T09:33:42.4115347Z Updating files: 80% (16097/20121) 2025-12-04T09:33:42.4434465Z Updating files: 81% (16299/20121) 2025-12-04T09:33:42.4669224Z Updating files: 82% (16500/20121) 2025-12-04T09:33:42.4836381Z Updating files: 83% (16701/20121) 2025-12-04T09:33:42.4989410Z Updating files: 84% (16902/20121) 2025-12-04T09:33:42.5166547Z Updating files: 85% (17103/20121) 2025-12-04T09:33:42.5336202Z Updating files: 86% (17305/20121) 2025-12-04T09:33:42.5485760Z Updating files: 87% (17506/20121) 2025-12-04T09:33:42.5612479Z Updating files: 88% (17707/20121) 2025-12-04T09:33:42.5762372Z Updating files: 89% (17908/20121) 2025-12-04T09:33:42.5953936Z Updating files: 90% (18109/20121) 2025-12-04T09:33:42.6081941Z Updating files: 91% (18311/20121) 2025-12-04T09:33:42.6253098Z Updating files: 92% (18512/20121) 2025-12-04T09:33:42.6458244Z Updating files: 93% (18713/20121) 2025-12-04T09:33:42.6683685Z Updating files: 94% (18914/20121) 2025-12-04T09:33:42.6878051Z Updating files: 95% (19115/20121) 2025-12-04T09:33:42.7050465Z Updating files: 96% (19317/20121) 2025-12-04T09:33:42.7235239Z Updating files: 97% (19518/20121) 2025-12-04T09:33:42.7551288Z Updating files: 98% (19719/20121) 2025-12-04T09:33:42.7747950Z Updating files: 99% (19920/20121) 2025-12-04T09:33:42.7748335Z Updating files: 100% (20121/20121) 2025-12-04T09:33:42.7748678Z Updating files: 100% (20121/20121), done. 2025-12-04T09:33:42.8058684Z Note: switching to 'ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32'. 2025-12-04T09:33:42.8059248Z 2025-12-04T09:33:42.8059614Z You are in 'detached HEAD' state. You can look around, make experimental 2025-12-04T09:33:42.8060594Z changes and commit them, and you can discard any commits you make in this 2025-12-04T09:33:42.8061599Z state without impacting any branches by switching back to a branch. 2025-12-04T09:33:42.8062245Z 2025-12-04T09:33:42.8062618Z If you want to create a new branch to retain commits you create, you may 2025-12-04T09:33:42.8063499Z do so (now or later) by using -c with the switch command. Example: 2025-12-04T09:33:42.8063864Z 2025-12-04T09:33:42.8064013Z git switch -c 2025-12-04T09:33:42.8064532Z 2025-12-04T09:33:42.8064658Z Or undo this operation with: 2025-12-04T09:33:42.8064887Z 2025-12-04T09:33:42.8064990Z git switch - 2025-12-04T09:33:42.8065157Z 2025-12-04T09:33:42.8065436Z Turn off this advice by setting config variable advice.detachedHead to false 2025-12-04T09:33:42.8065848Z 2025-12-04T09:33:42.8066188Z HEAD is now at ffd9b0fb435 Resolve collective autotuning test failure on arm (#168919) 2025-12-04T09:33:42.8149245Z ##[endgroup] 2025-12-04T09:33:42.8149929Z ##[group]Setting up auth for fetching submodules 2025-12-04T09:33:42.8156602Z [command]/usr/bin/git config --global http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-12-04T09:33:42.8209806Z [command]/usr/bin/git config --global --unset-all url.https://github.com/.insteadOf 2025-12-04T09:33:42.8238877Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf git@github.com: 2025-12-04T09:33:42.8267384Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf org-21003710@github.com: 2025-12-04T09:33:42.8292220Z ##[endgroup] 2025-12-04T09:33:42.8292856Z ##[group]Fetching submodules 2025-12-04T09:33:42.8296479Z [command]/usr/bin/git submodule sync --recursive 2025-12-04T09:33:42.8637343Z [command]/usr/bin/git -c protocol.version=2 submodule update --init --force --recursive 2025-12-04T09:33:42.8973908Z Submodule 'android/libs/fbjni' (https://github.com/facebookincubator/fbjni.git) registered for path 'android/libs/fbjni' 2025-12-04T09:33:42.8975338Z Submodule 'third_party/NNPACK_deps/FP16' (https://github.com/Maratyszcza/FP16.git) registered for path 'third_party/FP16' 2025-12-04T09:33:42.8978599Z Submodule 'third_party/NNPACK_deps/FXdiv' (https://github.com/Maratyszcza/FXdiv.git) registered for path 'third_party/FXdiv' 2025-12-04T09:33:42.8981629Z Submodule 'third_party/NNPACK' (https://github.com/Maratyszcza/NNPACK.git) registered for path 'third_party/NNPACK' 2025-12-04T09:33:42.8984663Z Submodule 'third_party/NVTX' (https://github.com/NVIDIA/NVTX.git) registered for path 'third_party/NVTX' 2025-12-04T09:33:42.8988370Z Submodule 'third_party/VulkanMemoryAllocator' (https://github.com/GPUOpen-LibrariesAndSDKs/VulkanMemoryAllocator.git) registered for path 'third_party/VulkanMemoryAllocator' 2025-12-04T09:33:42.8991310Z Submodule 'third_party/XNNPACK' (https://github.com/google/XNNPACK.git) registered for path 'third_party/XNNPACK' 2025-12-04T09:33:42.8994808Z Submodule 'third_party/aiter' (https://github.com/ROCm/aiter.git) registered for path 'third_party/aiter' 2025-12-04T09:33:42.8998833Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/benchmark' 2025-12-04T09:33:42.9002936Z Submodule 'third_party/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/composable_kernel' 2025-12-04T09:33:42.9006504Z Submodule 'third_party/cpp-httplib' (https://github.com/yhirose/cpp-httplib.git) registered for path 'third_party/cpp-httplib' 2025-12-04T09:33:42.9010464Z Submodule 'third_party/cpuinfo' (https://github.com/pytorch/cpuinfo.git) registered for path 'third_party/cpuinfo' 2025-12-04T09:33:42.9015441Z Submodule 'third_party/cudnn_frontend' (https://github.com/NVIDIA/cudnn-frontend.git) registered for path 'third_party/cudnn_frontend' 2025-12-04T09:33:42.9019547Z Submodule 'third_party/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/cutlass' 2025-12-04T09:33:42.9024024Z Submodule 'third_party/fbgemm' (https://github.com/pytorch/fbgemm) registered for path 'third_party/fbgemm' 2025-12-04T09:33:42.9029036Z Submodule 'third_party/flash-attention' (https://github.com/Dao-AILab/flash-attention.git) registered for path 'third_party/flash-attention' 2025-12-04T09:33:42.9035484Z Submodule 'third_party/flatbuffers' (https://github.com/google/flatbuffers.git) registered for path 'third_party/flatbuffers' 2025-12-04T09:33:42.9040571Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/fmt' 2025-12-04T09:33:42.9045941Z Submodule 'third_party/gemmlowp/gemmlowp' (https://github.com/google/gemmlowp.git) registered for path 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:33:42.9051083Z Submodule 'third_party/gloo' (https://github.com/pytorch/gloo) registered for path 'third_party/gloo' 2025-12-04T09:33:42.9056660Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/googletest' 2025-12-04T09:33:42.9062214Z Submodule 'third_party/ideep' (https://github.com/intel/ideep) registered for path 'third_party/ideep' 2025-12-04T09:33:42.9067880Z Submodule 'third_party/ittapi' (https://github.com/intel/ittapi.git) registered for path 'third_party/ittapi' 2025-12-04T09:33:42.9073519Z Submodule 'third_party/kineto' (https://github.com/pytorch/kineto) registered for path 'third_party/kineto' 2025-12-04T09:33:42.9079439Z Submodule 'third_party/kleidiai' (https://github.com/ARM-software/kleidiai.git) registered for path 'third_party/kleidiai' 2025-12-04T09:33:42.9085354Z Submodule 'third_party/mimalloc' (https://github.com/microsoft/mimalloc.git) registered for path 'third_party/mimalloc' 2025-12-04T09:33:42.9091427Z Submodule 'third_party/nlohmann' (https://github.com/nlohmann/json.git) registered for path 'third_party/nlohmann' 2025-12-04T09:33:42.9097841Z Submodule 'third_party/onnx' (https://github.com/onnx/onnx.git) registered for path 'third_party/onnx' 2025-12-04T09:33:42.9104411Z Submodule 'third_party/opentelemetry-cpp' (https://github.com/open-telemetry/opentelemetry-cpp.git) registered for path 'third_party/opentelemetry-cpp' 2025-12-04T09:33:42.9110588Z Submodule 'third_party/pocketfft' (https://github.com/mreineck/pocketfft) registered for path 'third_party/pocketfft' 2025-12-04T09:33:42.9117464Z Submodule 'third_party/protobuf' (https://github.com/protocolbuffers/protobuf.git) registered for path 'third_party/protobuf' 2025-12-04T09:33:42.9124054Z Submodule 'third_party/NNPACK_deps/psimd' (https://github.com/Maratyszcza/psimd.git) registered for path 'third_party/psimd' 2025-12-04T09:33:42.9130852Z Submodule 'third_party/NNPACK_deps/pthreadpool' (https://github.com/Maratyszcza/pthreadpool.git) registered for path 'third_party/pthreadpool' 2025-12-04T09:33:42.9139328Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/pybind11' 2025-12-04T09:33:42.9146425Z Submodule 'third_party/python-peachpy' (https://github.com/malfet/PeachPy.git) registered for path 'third_party/python-peachpy' 2025-12-04T09:33:42.9153352Z Submodule 'third_party/sleef' (https://github.com/shibatch/sleef) registered for path 'third_party/sleef' 2025-12-04T09:33:42.9176015Z Submodule 'third_party/tensorpipe' (https://github.com/pytorch/tensorpipe.git) registered for path 'third_party/tensorpipe' 2025-12-04T09:33:42.9194970Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/android/libs/fbjni'... 2025-12-04T09:33:43.1528001Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FXdiv'... 2025-12-04T09:33:43.1529649Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/FP16'... 2025-12-04T09:33:43.1561832Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fmt'... 2025-12-04T09:33:46.9467943Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NNPACK'... 2025-12-04T09:33:46.9469699Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/benchmark'... 2025-12-04T09:33:46.9471366Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/NVTX'... 2025-12-04T09:33:46.9473381Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gloo'... 2025-12-04T09:33:46.9475295Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/gemmlowp/gemmlowp'... 2025-12-04T09:33:46.9477270Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention'... 2025-12-04T09:33:46.9479149Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpuinfo'... 2025-12-04T09:33:46.9481387Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cpp-httplib'... 2025-12-04T09:33:46.9483153Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep'... 2025-12-04T09:33:46.9484892Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ittapi'... 2025-12-04T09:33:46.9486692Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kleidiai'... 2025-12-04T09:33:46.9488437Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pocketfft'... 2025-12-04T09:33:46.9490315Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cudnn_frontend'... 2025-12-04T09:33:46.9492062Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/mimalloc'... 2025-12-04T09:33:46.9493765Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/psimd'... 2025-12-04T09:33:46.9495576Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/googletest'... 2025-12-04T09:33:46.9497685Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pthreadpool'... 2025-12-04T09:33:46.9778882Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flatbuffers'... 2025-12-04T09:33:47.1098365Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/python-peachpy'... 2025-12-04T09:33:47.1842442Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp'... 2025-12-04T09:34:08.3963639Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/VulkanMemoryAllocator'... 2025-12-04T09:34:08.3965544Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe'... 2025-12-04T09:34:08.3966973Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto'... 2025-12-04T09:34:08.3968470Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/sleef'... 2025-12-04T09:34:08.3970051Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/pybind11'... 2025-12-04T09:34:08.3971685Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/cutlass'... 2025-12-04T09:34:08.3973287Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm'... 2025-12-04T09:34:08.3974300Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx'... 2025-12-04T09:34:08.3975966Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/composable_kernel'... 2025-12-04T09:34:08.3978109Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/nlohmann'... 2025-12-04T09:34:08.3979192Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/XNNPACK'... 2025-12-04T09:34:11.8632058Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter'... 2025-12-04T09:34:11.8632980Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf'... 2025-12-04T09:34:11.8814568Z Submodule path 'android/libs/fbjni': checked out '7e1e1fe3858c63c251c637ae41a20de425dde96f' 2025-12-04T09:34:11.8958768Z Submodule path 'third_party/FP16': checked out '4dfe081cf6bcd15db339cf2680b9281b8451eeb3' 2025-12-04T09:34:11.9066628Z Submodule path 'third_party/FXdiv': checked out 'b408327ac2a15ec3e43352421954f5b1967701d1' 2025-12-04T09:34:11.9355677Z Submodule path 'third_party/NNPACK': checked out 'c07e3a0400713d546e0dea2d5466dd22ea389c73' 2025-12-04T09:34:12.0316435Z Submodule path 'third_party/NVTX': checked out '3ebbc93ded7285963bff932c678fa367eb393ba6' 2025-12-04T09:34:12.0897245Z Submodule path 'third_party/VulkanMemoryAllocator': checked out '1d8f600fd424278486eade7ed3e877c99f0846b1' 2025-12-04T09:34:12.9454402Z Submodule path 'third_party/XNNPACK': checked out '51a0103656eff6fc9bfd39a4597923c4b542c883' 2025-12-04T09:34:13.1640023Z Submodule path 'third_party/aiter': checked out '01aae101b9e5e94d6c16a9514c9fb8df99c93150' 2025-12-04T09:34:13.1662616Z Submodule '3rdparty/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:34:13.1692284Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/aiter/3rdparty/composable_kernel'... 2025-12-04T09:34:18.5402928Z Submodule path 'third_party/aiter/3rdparty/composable_kernel': checked out 'cffe8fa2a442ac8e80dd236a1a5d24fe3d7e0cbf' 2025-12-04T09:34:18.5684661Z Submodule path 'third_party/benchmark': checked out '299e5928955cc62af9968370293b916f5130916f' 2025-12-04T09:34:18.9848926Z Submodule path 'third_party/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-12-04T09:34:19.0414726Z Submodule path 'third_party/cpp-httplib': checked out '89c932f313c6437c38f2982869beacc89c2f2246' 2025-12-04T09:34:19.1550907Z Submodule path 'third_party/cpuinfo': checked out 'f858c30bcb16f8effd5ff46996f0514539e17abc' 2025-12-04T09:34:19.2116413Z Submodule path 'third_party/cudnn_frontend': checked out '0b1577c8c83401237d601d0d0db5210506705396' 2025-12-04T09:34:19.9602599Z Submodule path 'third_party/cutlass': checked out 'f88806b1e31dfa579842638740216dd41fc6c588' 2025-12-04T09:34:20.1411930Z Submodule path 'third_party/fbgemm': checked out 'c0b988d39a9e47c794d699f29930ed4d7c7e13a4' 2025-12-04T09:34:20.1437444Z Submodule 'external/asmjit' (https://github.com/asmjit/asmjit.git) registered for path 'third_party/fbgemm/external/asmjit' 2025-12-04T09:34:20.1440674Z Submodule 'external/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:34:20.1443832Z Submodule 'external/cpuinfo' (https://github.com/pytorch/cpuinfo) registered for path 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:34:20.1447168Z Submodule 'external/cutlass' (https://github.com/jwfromm/cutlass) registered for path 'third_party/fbgemm/external/cutlass' 2025-12-04T09:34:20.1450711Z Submodule 'external/googletest' (https://github.com/google/googletest) registered for path 'third_party/fbgemm/external/googletest' 2025-12-04T09:34:20.1454379Z Submodule 'external/hipify_torch' (https://github.com/ROCmSoftwarePlatform/hipify_torch.git) registered for path 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:34:20.1457999Z Submodule 'external/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/fbgemm/external/json' 2025-12-04T09:34:20.1490460Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/asmjit'... 2025-12-04T09:34:21.4923478Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/hipify_torch'... 2025-12-04T09:34:21.4924623Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cpuinfo'... 2025-12-04T09:34:21.4925661Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/googletest'... 2025-12-04T09:34:21.5925590Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/composable_kernel'... 2025-12-04T09:34:25.1504040Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/cutlass'... 2025-12-04T09:34:25.2504661Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/fbgemm/external/json'... 2025-12-04T09:34:28.4240839Z Submodule path 'third_party/fbgemm/external/asmjit': checked out 'a3199e8857792cd10b7589ff5d58343d2c9008ea' 2025-12-04T09:34:28.8337942Z Submodule path 'third_party/fbgemm/external/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-12-04T09:34:28.9499401Z Submodule path 'third_party/fbgemm/external/cpuinfo': checked out '6543fec09b2f04ac4a666882998b534afc9c1349' 2025-12-04T09:34:29.6813261Z Submodule path 'third_party/fbgemm/external/cutlass': checked out '98125ce499b0fdf7ffbe0e3052f5b8709f4840f8' 2025-12-04T09:34:29.7348325Z Submodule path 'third_party/fbgemm/external/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T09:34:29.7484261Z Submodule path 'third_party/fbgemm/external/hipify_torch': checked out '63b6a7b541fa7f08f8475ca7d74054db36ff2691' 2025-12-04T09:34:29.8722856Z Submodule path 'third_party/fbgemm/external/json': checked out '9cca280a4d0ccf0c08f47a99aa71d1b0e52f8d03' 2025-12-04T09:34:29.9533135Z Submodule path 'third_party/flash-attention': checked out '979702c87a8713a8e0a5e9fee122b90d2ef13be5' 2025-12-04T09:34:29.9556070Z Submodule 'csrc/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:34:29.9558688Z Submodule 'csrc/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:34:29.9589175Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/composable_kernel'... 2025-12-04T09:34:34.7898306Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/cutlass'... 2025-12-04T09:34:35.0750506Z Submodule path 'third_party/flash-attention/csrc/composable_kernel': checked out '888317e698e9803c62bd38568abc9e05d7709f33' 2025-12-04T09:34:35.7247276Z Submodule path 'third_party/flash-attention/csrc/cutlass': checked out 'c506e16788cb08416a4a57e11a9067beeee29420' 2025-12-04T09:34:35.8848914Z Submodule path 'third_party/flatbuffers': checked out 'a2cd1ea3b6d3fee220106b5fed3f7ce8da9eb757' 2025-12-04T09:34:35.9187297Z Submodule path 'third_party/fmt': checked out '407c905e45ad75fc29bf0f9bb7c5c2fd3475976f' 2025-12-04T09:34:35.9648620Z Submodule path 'third_party/gemmlowp/gemmlowp': checked out '3fb5c176c17c765a3492cd2f0321b0dab712f350' 2025-12-04T09:34:35.9941224Z Submodule path 'third_party/gloo': checked out '54cbae0d3a67fa890b4c3d9ee162b7860315e341' 2025-12-04T09:34:36.0461085Z Submodule path 'third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T09:34:36.0613993Z Submodule path 'third_party/ideep': checked out '719d8e6cd7f7a0e01b155657526d693acf97c2b3' 2025-12-04T09:34:36.0633452Z Submodule 'mkl-dnn' (https://github.com/intel/mkl-dnn.git) registered for path 'third_party/ideep/mkl-dnn' 2025-12-04T09:34:36.0660755Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/ideep/mkl-dnn'... 2025-12-04T09:34:53.2467722Z Submodule path 'third_party/ideep/mkl-dnn': checked out '8d263e693366ef8db40acc569cc7d8edf644556d' 2025-12-04T09:34:53.2702349Z Submodule path 'third_party/ittapi': checked out 'dec1d23ca65ab069d225dfe40dea14f455170959' 2025-12-04T09:34:53.3681404Z Submodule path 'third_party/kineto': checked out '31f85df8fbd89c188f14ef10f1ec65379786b943' 2025-12-04T09:34:53.3702937Z Submodule 'libkineto/third_party/dynolog' (https://github.com/facebookincubator/dynolog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:34:53.3705718Z Submodule 'libkineto/third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:34:53.3708937Z Submodule 'libkineto/third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:34:53.3739818Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog'... 2025-12-04T09:34:54.1699200Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/fmt'... 2025-12-04T09:34:55.0183446Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/googletest'... 2025-12-04T09:34:55.1224592Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog': checked out 'd2ffe0a4e3acace628db49974246b66fc3e85fb1' 2025-12-04T09:34:55.1244332Z Submodule 'third_party/DCGM' (https://github.com/NVIDIA/DCGM.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:34:55.1247314Z Submodule 'third_party/cpr' (https://github.com/libcpr/cpr.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:34:55.1250413Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:34:55.1253811Z Submodule 'third_party/gflags' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:34:55.1257209Z Submodule 'third_party/glog' (https://github.com/google/glog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:34:55.1260851Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:34:55.1264392Z Submodule 'third_party/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:34:55.1268045Z Submodule 'third_party/pfs' (https://github.com/dtrugman/pfs.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:34:55.1271976Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:34:55.1305052Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM'... 2025-12-04T09:34:57.1528881Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/pfs'... 2025-12-04T09:34:57.1530350Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp'... 2025-12-04T09:34:57.1531822Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags'... 2025-12-04T09:34:57.1533188Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/cpr'... 2025-12-04T09:34:57.1534535Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/glog'... 2025-12-04T09:34:57.1536672Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/googletest'... 2025-12-04T09:34:57.1538386Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/fmt'... 2025-12-04T09:34:57.2529405Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/json'... 2025-12-04T09:35:03.7829048Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM': checked out 'ffde4e54bc7249a6039a5e6b45b395141e1217f9' 2025-12-04T09:35:03.8038619Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr': checked out '871ed52d350214a034f6ef8a3b8f51c5ce1bd400' 2025-12-04T09:35:03.8464986Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt': checked out 'cd4af11efc9c622896a3e4cb599fa28668ca3d05' 2025-12-04T09:35:03.8625351Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags': checked out 'e171aa2d15ed9eb17054558e0b3a6a413bb01067' 2025-12-04T09:35:03.8645469Z Submodule 'doc' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:35:03.8674632Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc'... 2025-12-04T09:35:04.1496744Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc': checked out '8411df715cf522606e3b1aca386ddfc0b63d34b4' 2025-12-04T09:35:04.1718965Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog': checked out 'b33e3bad4c46c8a6345525fd822af355e5ef9446' 2025-12-04T09:35:04.2250569Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T09:35:04.3407459Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json': checked out '4f8fba14066156b73f1189a2b8bd568bde5284c5' 2025-12-04T09:35:04.3607783Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs': checked out 'f68a2fa8ea36c783bdd760371411fcb495aa3150' 2025-12-04T09:35:04.3807731Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp': checked out 'b1234816facfdda29845c46696a02998a4af115a' 2025-12-04T09:35:04.3826822Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:35:04.3829057Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:35:04.3859972Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-12-04T09:35:06.9166693Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest'... 2025-12-04T09:35:07.2098734Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'd7ba35bbb649209c66e582d5a0244ba988a15159' 2025-12-04T09:35:07.2650093Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-12-04T09:35:07.3030700Z Submodule path 'third_party/kineto/libkineto/third_party/fmt': checked out '40626af88bd7df9a5fb80be7b25ac85b122d6c21' 2025-12-04T09:35:07.3573588Z Submodule path 'third_party/kineto/libkineto/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T09:35:07.4351326Z Submodule path 'third_party/kleidiai': checked out 'd7770c89632329a9914ef1a90289917597639cbe' 2025-12-04T09:35:07.4807946Z Submodule path 'third_party/mimalloc': checked out 'fbd8b99c2b828428947d70fdc046bb55609be93e' 2025-12-04T09:35:07.5983405Z Submodule path 'third_party/nlohmann': checked out '55f93686c01528224f448c19128836e7df245f72' 2025-12-04T09:35:08.0776744Z Submodule path 'third_party/onnx': checked out 'e709452ef2bbc1d113faf678c24e6d3467696e83' 2025-12-04T09:35:08.0820112Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/onnx/third_party/pybind11' 2025-12-04T09:35:08.0849780Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/onnx/third_party/pybind11'... 2025-12-04T09:35:08.9777926Z Submodule path 'third_party/onnx/third_party/pybind11': checked out 'a2e59f0e7065404b44dfe92a28aca47ba1378dc4' 2025-12-04T09:35:09.0591148Z Submodule path 'third_party/opentelemetry-cpp': checked out 'a799f4aed9c94b765dcdaabaeab7d5e7e2310878' 2025-12-04T09:35:09.0614634Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark) registered for path 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:35:09.0617189Z Submodule 'third_party/googletest' (https://github.com/google/googletest) registered for path 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:35:09.0620284Z Submodule 'third_party/ms-gsl' (https://github.com/microsoft/GSL) registered for path 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:35:09.0623516Z Submodule 'third_party/nlohmann-json' (https://github.com/nlohmann/json) registered for path 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:35:09.0626937Z Submodule 'third_party/opentelemetry-proto' (https://github.com/open-telemetry/opentelemetry-proto) registered for path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:35:09.0630207Z Submodule 'third_party/opentracing-cpp' (https://github.com/opentracing/opentracing-cpp.git) registered for path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:35:09.0633641Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:35:09.0637035Z Submodule 'tools/vcpkg' (https://github.com/Microsoft/vcpkg) registered for path 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:35:09.0668779Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/benchmark'... 2025-12-04T09:35:09.5164069Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentracing-cpp'... 2025-12-04T09:35:09.5165448Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentelemetry-proto'... 2025-12-04T09:35:09.5166704Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/ms-gsl'... 2025-12-04T09:35:09.5167929Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp'... 2025-12-04T09:35:09.6165682Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/googletest'... 2025-12-04T09:35:10.3666698Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/nlohmann-json'... 2025-12-04T09:35:17.9710984Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/tools/vcpkg'... 2025-12-04T09:35:18.7088842Z Submodule path 'third_party/opentelemetry-cpp/third_party/benchmark': checked out 'd572f4777349d43653b21d6c2fc63020ab326db2' 2025-12-04T09:35:18.7566162Z Submodule path 'third_party/opentelemetry-cpp/third_party/googletest': checked out 'b796f7d44681514f58a683a3a71ff17c94edb0c1' 2025-12-04T09:35:18.7763845Z Submodule path 'third_party/opentelemetry-cpp/third_party/ms-gsl': checked out '6f4529395c5b7c2d661812257cd6780c67e54afa' 2025-12-04T09:35:18.8969840Z Submodule path 'third_party/opentelemetry-cpp/third_party/nlohmann-json': checked out 'bc889afb4c5bf1c0d8ee29ef35eaaf4c8bef8a5d' 2025-12-04T09:35:18.9130963Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto': checked out '4ca4f0335c63cda7ab31ea7ed70d6553aee14dce' 2025-12-04T09:35:18.9305994Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp': checked out '06b57f48ded1fa3bdd3d4346f6ef29e40e08eaf5' 2025-12-04T09:35:18.9495794Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp': checked out 'c9ffcdda9086ffd9e1283ea7a0276d831f3c8a8d' 2025-12-04T09:35:18.9515269Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:35:18.9518243Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:35:18.9548865Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-12-04T09:35:21.2060820Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest'... 2025-12-04T09:35:21.4926806Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'eefb26f82b233268fc98577d265352720d477ba4' 2025-12-04T09:35:21.5458021Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-12-04T09:35:22.1016170Z Submodule path 'third_party/opentelemetry-cpp/tools/vcpkg': checked out '8eb57355a4ffb410a2e94c07b4dca2dffbee8e50' 2025-12-04T09:35:22.1154121Z Submodule path 'third_party/pocketfft': checked out '0fa0ef591e38c2758e3184c6c23e497b9f732ffa' 2025-12-04T09:35:22.4292958Z Submodule path 'third_party/protobuf': checked out 'd1eca4e4b421cd2997495c4b4e65cea6be4e9b8a' 2025-12-04T09:35:22.4317415Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:35:22.4320142Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/protobuf/third_party/googletest' 2025-12-04T09:35:22.4352021Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/benchmark'... 2025-12-04T09:35:22.9550295Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/protobuf/third_party/googletest'... 2025-12-04T09:35:23.4550349Z Submodule path 'third_party/protobuf/third_party/benchmark': checked out '5b7683f49e1e9223cf9927b24f6fd3d6bd82e3f8' 2025-12-04T09:35:23.5385843Z Submodule path 'third_party/protobuf/third_party/googletest': checked out '5ec7f0c4a113e2f18ac2c6cc7df51ad6afc24081' 2025-12-04T09:35:23.5500875Z Submodule path 'third_party/psimd': checked out '072586a71b55b7f8c584153d223e95687148a900' 2025-12-04T09:35:23.5646222Z Submodule path 'third_party/pthreadpool': checked out '4fe0e1e183925bf8cfa6aae24237e724a96479b8' 2025-12-04T09:35:23.6143190Z Submodule path 'third_party/pybind11': checked out 'f5fbe867d2d26e4a0a9177a51f6e568868ad3dc8' 2025-12-04T09:35:23.6485569Z Submodule path 'third_party/python-peachpy': checked out 'f45429b087dd7d5bc78bb40dc7cf06425c252d67' 2025-12-04T09:35:23.6996300Z Submodule path 'third_party/sleef': checked out '5a1d179df9cf652951b59010a2d2075372d67f68' 2025-12-04T09:35:23.7324851Z Submodule path 'third_party/tensorpipe': checked out '2b4cd91092d335a697416b2a3cb398283246849d' 2025-12-04T09:35:23.7346646Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:35:23.7349590Z Submodule 'third_party/libnop' (https://github.com/google/libnop.git) registered for path 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:35:23.7352883Z Submodule 'third_party/libuv' (https://github.com/libuv/libuv.git) registered for path 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:35:23.7355834Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:35:23.7388532Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/googletest'... 2025-12-04T09:35:24.9817495Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libnop'... 2025-12-04T09:35:24.9819734Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11'... 2025-12-04T09:35:25.0379749Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libuv'... 2025-12-04T09:35:25.1072402Z Submodule path 'third_party/tensorpipe/third_party/googletest': checked out 'aee0f9d9b5b87796ee8a0ab26b7587ec30e8858e' 2025-12-04T09:35:25.1265691Z Submodule path 'third_party/tensorpipe/third_party/libnop': checked out '910b55815be16109f04f4180e9adee14fb4ce281' 2025-12-04T09:35:25.2154403Z Submodule path 'third_party/tensorpipe/third_party/libuv': checked out '5152db2cbfeb5582e9c27c5ea1dba2cd9e10759b' 2025-12-04T09:35:25.2501713Z Submodule path 'third_party/tensorpipe/third_party/pybind11': checked out 'a23996fce38ff6ccfbcdc09f1e63f2c4be5ea2ef' 2025-12-04T09:35:25.2521846Z Submodule 'tools/clang' (https://github.com/wjakob/clang-cindex-python3) registered for path 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:35:25.2552399Z Cloning into '/home/ec2-user/actions-runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11/tools/clang'... 2025-12-04T09:35:25.4441016Z Submodule path 'third_party/tensorpipe/third_party/pybind11/tools/clang': checked out '6a00cbc4a9b8e68b71caf7f774b3f9c753ae84d5' 2025-12-04T09:35:25.4482400Z [command]/usr/bin/git submodule foreach --recursive git config --local gc.auto 0 2025-12-04T09:35:25.4823600Z Entering 'android/libs/fbjni' 2025-12-04T09:35:25.4872203Z Entering 'third_party/FP16' 2025-12-04T09:35:25.4920439Z Entering 'third_party/FXdiv' 2025-12-04T09:35:25.4968631Z Entering 'third_party/NNPACK' 2025-12-04T09:35:25.5017388Z Entering 'third_party/NVTX' 2025-12-04T09:35:25.5068750Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:35:25.5117519Z Entering 'third_party/XNNPACK' 2025-12-04T09:35:25.5183163Z Entering 'third_party/aiter' 2025-12-04T09:35:25.5231557Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:35:25.5287972Z Entering 'third_party/benchmark' 2025-12-04T09:35:25.5337987Z Entering 'third_party/composable_kernel' 2025-12-04T09:35:25.5395647Z Entering 'third_party/cpp-httplib' 2025-12-04T09:35:25.5442759Z Entering 'third_party/cpuinfo' 2025-12-04T09:35:25.5491488Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:35:25.5538521Z Entering 'third_party/cutlass' 2025-12-04T09:35:25.5598565Z Entering 'third_party/fbgemm' 2025-12-04T09:35:25.5648120Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:35:25.5693509Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:35:25.5749017Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:35:25.5796494Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:35:25.5853280Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:35:25.5899315Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:35:25.5943511Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:35:25.5992500Z Entering 'third_party/flash-attention' 2025-12-04T09:35:25.6041724Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:35:25.6094376Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:35:25.6153833Z Entering 'third_party/flatbuffers' 2025-12-04T09:35:25.6204279Z Entering 'third_party/fmt' 2025-12-04T09:35:25.6252249Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:35:25.6300372Z Entering 'third_party/gloo' 2025-12-04T09:35:25.6348577Z Entering 'third_party/googletest' 2025-12-04T09:35:25.6395804Z Entering 'third_party/ideep' 2025-12-04T09:35:25.6442813Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:35:25.6499195Z Entering 'third_party/ittapi' 2025-12-04T09:35:25.6546920Z Entering 'third_party/kineto' 2025-12-04T09:35:25.6594583Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:35:25.6640397Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:35:25.6687868Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:35:25.6733395Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:35:25.6780628Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:35:25.6825781Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:35:25.6873826Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:35:25.6920771Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:35:25.6966793Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:35:25.7015638Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:35:25.7062409Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:35:25.7108269Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:35:25.7157155Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:35:25.7208186Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:35:25.7253554Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:35:25.7301872Z Entering 'third_party/kleidiai' 2025-12-04T09:35:25.7350552Z Entering 'third_party/mimalloc' 2025-12-04T09:35:25.7397785Z Entering 'third_party/nlohmann' 2025-12-04T09:35:25.7446863Z Entering 'third_party/onnx' 2025-12-04T09:35:25.7515000Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:35:25.7565601Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:35:25.7614741Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:35:25.7659516Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:35:25.7705258Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:35:25.7750736Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:35:25.7799180Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:35:25.7845966Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:35:25.7892791Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:35:25.7940725Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:35:25.7988934Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:35:25.8038056Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:35:25.8109006Z Entering 'third_party/pocketfft' 2025-12-04T09:35:25.8157584Z Entering 'third_party/protobuf' 2025-12-04T09:35:25.8210233Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:35:25.8256651Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:35:25.8307614Z Entering 'third_party/psimd' 2025-12-04T09:35:25.8355644Z Entering 'third_party/pthreadpool' 2025-12-04T09:35:25.8404818Z Entering 'third_party/pybind11' 2025-12-04T09:35:25.8452446Z Entering 'third_party/python-peachpy' 2025-12-04T09:35:25.8502081Z Entering 'third_party/sleef' 2025-12-04T09:35:25.8549893Z Entering 'third_party/tensorpipe' 2025-12-04T09:35:25.8598875Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:35:25.8644343Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:35:25.8691032Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:35:25.8736922Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:35:25.8781931Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:35:25.8845754Z ##[endgroup] 2025-12-04T09:35:25.8846309Z ##[group]Persisting credentials for submodules 2025-12-04T09:35:25.8852726Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'url\.https\:\/\/github\.com\/\.insteadOf' && git config --local --unset-all 'url.https://github.com/.insteadOf' || :" 2025-12-04T09:35:25.9189007Z Entering 'android/libs/fbjni' 2025-12-04T09:35:25.9251787Z Entering 'third_party/FP16' 2025-12-04T09:35:25.9315487Z Entering 'third_party/FXdiv' 2025-12-04T09:35:25.9377875Z Entering 'third_party/NNPACK' 2025-12-04T09:35:25.9439753Z Entering 'third_party/NVTX' 2025-12-04T09:35:25.9503461Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:35:25.9566953Z Entering 'third_party/XNNPACK' 2025-12-04T09:35:25.9646711Z Entering 'third_party/aiter' 2025-12-04T09:35:25.9711045Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:35:25.9786661Z Entering 'third_party/benchmark' 2025-12-04T09:35:25.9849144Z Entering 'third_party/composable_kernel' 2025-12-04T09:35:25.9922173Z Entering 'third_party/cpp-httplib' 2025-12-04T09:35:25.9984333Z Entering 'third_party/cpuinfo' 2025-12-04T09:35:26.0048596Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:35:26.0111639Z Entering 'third_party/cutlass' 2025-12-04T09:35:26.0184937Z Entering 'third_party/fbgemm' 2025-12-04T09:35:26.0252974Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:35:26.0316385Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:35:26.0389166Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:35:26.0451751Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:35:26.0525502Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:35:26.0587394Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:35:26.0648811Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:35:26.0715579Z Entering 'third_party/flash-attention' 2025-12-04T09:35:26.0780906Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:35:26.0850102Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:35:26.0925143Z Entering 'third_party/flatbuffers' 2025-12-04T09:35:26.0991296Z Entering 'third_party/fmt' 2025-12-04T09:35:26.1056488Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:35:26.1119720Z Entering 'third_party/gloo' 2025-12-04T09:35:26.1182694Z Entering 'third_party/googletest' 2025-12-04T09:35:26.1245617Z Entering 'third_party/ideep' 2025-12-04T09:35:26.1309927Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:35:26.1380620Z Entering 'third_party/ittapi' 2025-12-04T09:35:26.1442963Z Entering 'third_party/kineto' 2025-12-04T09:35:26.1506634Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:35:26.1568889Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:35:26.1633566Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:35:26.1695632Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:35:26.1758323Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:35:26.1821093Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:35:26.1885499Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:35:26.1948166Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:35:26.2011830Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:35:26.2075597Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:35:26.2139155Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:35:26.2202397Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:35:26.2266990Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:35:26.2334964Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:35:26.2398561Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:35:26.2461790Z Entering 'third_party/kleidiai' 2025-12-04T09:35:26.2527174Z Entering 'third_party/mimalloc' 2025-12-04T09:35:26.2590225Z Entering 'third_party/nlohmann' 2025-12-04T09:35:26.2658530Z Entering 'third_party/onnx' 2025-12-04T09:35:26.2742842Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:35:26.2808896Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:35:26.2873227Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:35:26.2935249Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:35:26.2997655Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:35:26.3059833Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:35:26.3124039Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:35:26.3187559Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:35:26.3248720Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:35:26.3310921Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:35:26.3373771Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:35:26.3438994Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:35:26.3526550Z Entering 'third_party/pocketfft' 2025-12-04T09:35:26.3589706Z Entering 'third_party/protobuf' 2025-12-04T09:35:26.3656285Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:35:26.3720764Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:35:26.3785785Z Entering 'third_party/psimd' 2025-12-04T09:35:26.3849093Z Entering 'third_party/pthreadpool' 2025-12-04T09:35:26.3912712Z Entering 'third_party/pybind11' 2025-12-04T09:35:26.3977247Z Entering 'third_party/python-peachpy' 2025-12-04T09:35:26.4039116Z Entering 'third_party/sleef' 2025-12-04T09:35:26.4103489Z Entering 'third_party/tensorpipe' 2025-12-04T09:35:26.4166540Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:35:26.4234028Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:35:26.4296253Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:35:26.4359476Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:35:26.4421518Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:35:26.4505624Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local 'http.https://github.com/.extraheader' 'AUTHORIZATION: basic ***' && git config --local --show-origin --name-only --get-regexp remote.origin.url" 2025-12-04T09:35:26.4855617Z Entering 'android/libs/fbjni' 2025-12-04T09:35:26.4916331Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-12-04T09:35:26.4937098Z Entering 'third_party/FP16' 2025-12-04T09:35:26.4997137Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-12-04T09:35:26.5016967Z Entering 'third_party/FXdiv' 2025-12-04T09:35:26.5077624Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-12-04T09:35:26.5098610Z Entering 'third_party/NNPACK' 2025-12-04T09:35:26.5156784Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-12-04T09:35:26.5177123Z Entering 'third_party/NVTX' 2025-12-04T09:35:26.5236537Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-12-04T09:35:26.5256961Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:35:26.5318817Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-12-04T09:35:26.5338911Z Entering 'third_party/XNNPACK' 2025-12-04T09:35:26.5399385Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-12-04T09:35:26.5436393Z Entering 'third_party/aiter' 2025-12-04T09:35:26.5495653Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-12-04T09:35:26.5516814Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:35:26.5575833Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-12-04T09:35:26.5606005Z Entering 'third_party/benchmark' 2025-12-04T09:35:26.5664655Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-12-04T09:35:26.5685093Z Entering 'third_party/composable_kernel' 2025-12-04T09:35:26.5744003Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-12-04T09:35:26.5773320Z Entering 'third_party/cpp-httplib' 2025-12-04T09:35:26.5833638Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-12-04T09:35:26.5853762Z Entering 'third_party/cpuinfo' 2025-12-04T09:35:26.5914489Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-12-04T09:35:26.5935271Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:35:26.5995265Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-12-04T09:35:26.6015208Z Entering 'third_party/cutlass' 2025-12-04T09:35:26.6075159Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-12-04T09:35:26.6106663Z Entering 'third_party/fbgemm' 2025-12-04T09:35:26.6166025Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-12-04T09:35:26.6188337Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:35:26.6246644Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-12-04T09:35:26.6265199Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:35:26.6324937Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-12-04T09:35:26.6354617Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:35:26.6418477Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-12-04T09:35:26.6438327Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:35:26.6496925Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-12-04T09:35:26.6526099Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:35:26.6585531Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-12-04T09:35:26.6605364Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:35:26.6664208Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-12-04T09:35:26.6683744Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:35:26.6742036Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-12-04T09:35:26.6764232Z Entering 'third_party/flash-attention' 2025-12-04T09:35:26.6824805Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-12-04T09:35:26.6844331Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:35:26.6904086Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-12-04T09:35:26.6929536Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:35:26.6987487Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-12-04T09:35:26.7017162Z Entering 'third_party/flatbuffers' 2025-12-04T09:35:26.7076338Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-12-04T09:35:26.7099171Z Entering 'third_party/fmt' 2025-12-04T09:35:26.7157809Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-12-04T09:35:26.7177508Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:35:26.7235661Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-12-04T09:35:26.7255174Z Entering 'third_party/gloo' 2025-12-04T09:35:26.7314345Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-12-04T09:35:26.7334216Z Entering 'third_party/googletest' 2025-12-04T09:35:26.7393624Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:35:26.7413568Z Entering 'third_party/ideep' 2025-12-04T09:35:26.7472093Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-12-04T09:35:26.7490162Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:35:26.7547228Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-12-04T09:35:26.7574528Z Entering 'third_party/ittapi' 2025-12-04T09:35:26.7633416Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-12-04T09:35:26.7652932Z Entering 'third_party/kineto' 2025-12-04T09:35:26.7714971Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-12-04T09:35:26.7733734Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:35:26.7793473Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-12-04T09:35:26.7811157Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:35:26.7870159Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-12-04T09:35:26.7889314Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:35:26.7950989Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-12-04T09:35:26.7969387Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:35:26.8028872Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-12-04T09:35:26.8046920Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:35:26.8106903Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-12-04T09:35:26.8123347Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:35:26.8183672Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-12-04T09:35:26.8205126Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:35:26.8264197Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-12-04T09:35:26.8282441Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:35:26.8341254Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:35:26.8359496Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:35:26.8418047Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-12-04T09:35:26.8437290Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:35:26.8496488Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-12-04T09:35:26.8514261Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:35:26.8573903Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T09:35:26.8592175Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:35:26.8654811Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T09:35:26.8675851Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:35:26.8736638Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T09:35:26.8759600Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:35:26.8817881Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-12-04T09:35:26.8836470Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:35:26.8896859Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-12-04T09:35:26.8916374Z Entering 'third_party/kleidiai' 2025-12-04T09:35:26.8976357Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-12-04T09:35:26.8996266Z Entering 'third_party/mimalloc' 2025-12-04T09:35:26.9054362Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-12-04T09:35:26.9073202Z Entering 'third_party/nlohmann' 2025-12-04T09:35:26.9130833Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-12-04T09:35:26.9151535Z Entering 'third_party/onnx' 2025-12-04T09:35:26.9210791Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-12-04T09:35:26.9249210Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:35:26.9307742Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-12-04T09:35:26.9328166Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:35:26.9388315Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-12-04T09:35:26.9408732Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:35:26.9465756Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-12-04T09:35:26.9483115Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:35:26.9541029Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:35:26.9558647Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:35:26.9616919Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-12-04T09:35:26.9634158Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:35:26.9695750Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-12-04T09:35:26.9714884Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:35:26.9773489Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-12-04T09:35:26.9791489Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:35:26.9849213Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-12-04T09:35:26.9866431Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:35:26.9927589Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T09:35:26.9944691Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:35:27.0003008Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T09:35:27.0022411Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:35:27.0082681Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T09:35:27.0103027Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:35:27.0159824Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-12-04T09:35:27.0201942Z Entering 'third_party/pocketfft' 2025-12-04T09:35:27.0260298Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-12-04T09:35:27.0278241Z Entering 'third_party/protobuf' 2025-12-04T09:35:27.0339004Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-12-04T09:35:27.0360762Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:35:27.0418651Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-12-04T09:35:27.0436321Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:35:27.0499625Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:35:27.0519363Z Entering 'third_party/psimd' 2025-12-04T09:35:27.0577017Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-12-04T09:35:27.0595615Z Entering 'third_party/pthreadpool' 2025-12-04T09:35:27.0653002Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-12-04T09:35:27.0671162Z Entering 'third_party/pybind11' 2025-12-04T09:35:27.0731138Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-12-04T09:35:27.0750205Z Entering 'third_party/python-peachpy' 2025-12-04T09:35:27.0808153Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-12-04T09:35:27.0826643Z Entering 'third_party/sleef' 2025-12-04T09:35:27.0884538Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-12-04T09:35:27.0903384Z Entering 'third_party/tensorpipe' 2025-12-04T09:35:27.0961313Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-12-04T09:35:27.0980254Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:35:27.1037774Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-12-04T09:35:27.1055168Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:35:27.1114093Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-12-04T09:35:27.1131732Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:35:27.1189421Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-12-04T09:35:27.1207329Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:35:27.1265293Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-12-04T09:35:27.1281866Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:35:27.1339819Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-12-04T09:35:27.1979184Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'git@github.com:' 2025-12-04T09:35:27.2316233Z Entering 'android/libs/fbjni' 2025-12-04T09:35:27.2364003Z Entering 'third_party/FP16' 2025-12-04T09:35:27.2410671Z Entering 'third_party/FXdiv' 2025-12-04T09:35:27.2457556Z Entering 'third_party/NNPACK' 2025-12-04T09:35:27.2506626Z Entering 'third_party/NVTX' 2025-12-04T09:35:27.2555987Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:35:27.2602499Z Entering 'third_party/XNNPACK' 2025-12-04T09:35:27.2665394Z Entering 'third_party/aiter' 2025-12-04T09:35:27.2712608Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:35:27.2770008Z Entering 'third_party/benchmark' 2025-12-04T09:35:27.2817062Z Entering 'third_party/composable_kernel' 2025-12-04T09:35:27.2874378Z Entering 'third_party/cpp-httplib' 2025-12-04T09:35:27.2924149Z Entering 'third_party/cpuinfo' 2025-12-04T09:35:27.2971028Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:35:27.3019158Z Entering 'third_party/cutlass' 2025-12-04T09:35:27.3075293Z Entering 'third_party/fbgemm' 2025-12-04T09:35:27.3124275Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:35:27.3169340Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:35:27.3225456Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:35:27.3272109Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:35:27.3328553Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:35:27.3374045Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:35:27.3419672Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:35:27.3468557Z Entering 'third_party/flash-attention' 2025-12-04T09:35:27.3516421Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:35:27.3568856Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:35:27.3625937Z Entering 'third_party/flatbuffers' 2025-12-04T09:35:27.3675663Z Entering 'third_party/fmt' 2025-12-04T09:35:27.3723145Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:35:27.3770339Z Entering 'third_party/gloo' 2025-12-04T09:35:27.3817058Z Entering 'third_party/googletest' 2025-12-04T09:35:27.3864413Z Entering 'third_party/ideep' 2025-12-04T09:35:27.3911176Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:35:27.3967683Z Entering 'third_party/ittapi' 2025-12-04T09:35:27.4016711Z Entering 'third_party/kineto' 2025-12-04T09:35:27.4063392Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:35:27.4110116Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:35:27.4157943Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:35:27.4205022Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:35:27.4252187Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:35:27.4297107Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:35:27.4345380Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:35:27.4391283Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:35:27.4440653Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:35:27.4490213Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:35:27.4538002Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:35:27.4586986Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:35:27.4638227Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:35:27.4691797Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:35:27.4739471Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:35:27.4793999Z Entering 'third_party/kleidiai' 2025-12-04T09:35:27.4843190Z Entering 'third_party/mimalloc' 2025-12-04T09:35:27.4893128Z Entering 'third_party/nlohmann' 2025-12-04T09:35:27.4942837Z Entering 'third_party/onnx' 2025-12-04T09:35:27.5011554Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:35:27.5061379Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:35:27.5117244Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:35:27.5163402Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:35:27.5210965Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:35:27.5256498Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:35:27.5305862Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:35:27.5352783Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:35:27.5400813Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:35:27.5445239Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:35:27.5493167Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:35:27.5545951Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:35:27.5620809Z Entering 'third_party/pocketfft' 2025-12-04T09:35:27.5669082Z Entering 'third_party/protobuf' 2025-12-04T09:35:27.5720823Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:35:27.5770796Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:35:27.5820365Z Entering 'third_party/psimd' 2025-12-04T09:35:27.5869339Z Entering 'third_party/pthreadpool' 2025-12-04T09:35:27.5926272Z Entering 'third_party/pybind11' 2025-12-04T09:35:27.5974761Z Entering 'third_party/python-peachpy' 2025-12-04T09:35:27.6022553Z Entering 'third_party/sleef' 2025-12-04T09:35:27.6070160Z Entering 'third_party/tensorpipe' 2025-12-04T09:35:27.6117073Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:35:27.6163598Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:35:27.6211333Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:35:27.6258112Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:35:27.6305528Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:35:27.6370331Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'org-21003710@github.com:' 2025-12-04T09:35:27.6715969Z Entering 'android/libs/fbjni' 2025-12-04T09:35:27.6763719Z Entering 'third_party/FP16' 2025-12-04T09:35:27.6814230Z Entering 'third_party/FXdiv' 2025-12-04T09:35:27.6862615Z Entering 'third_party/NNPACK' 2025-12-04T09:35:27.6910112Z Entering 'third_party/NVTX' 2025-12-04T09:35:27.6959719Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:35:27.7007879Z Entering 'third_party/XNNPACK' 2025-12-04T09:35:27.7072999Z Entering 'third_party/aiter' 2025-12-04T09:35:27.7124613Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:35:27.7181652Z Entering 'third_party/benchmark' 2025-12-04T09:35:27.7229886Z Entering 'third_party/composable_kernel' 2025-12-04T09:35:27.7285084Z Entering 'third_party/cpp-httplib' 2025-12-04T09:35:27.7331838Z Entering 'third_party/cpuinfo' 2025-12-04T09:35:27.7379502Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:35:27.7427182Z Entering 'third_party/cutlass' 2025-12-04T09:35:27.7484046Z Entering 'third_party/fbgemm' 2025-12-04T09:35:27.7534591Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:35:27.7580645Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:35:27.7635829Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:35:27.7682660Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:35:27.7739410Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:35:27.7786118Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:35:27.7833070Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:35:27.7884639Z Entering 'third_party/flash-attention' 2025-12-04T09:35:27.7934233Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:35:27.7988907Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:35:27.8046784Z Entering 'third_party/flatbuffers' 2025-12-04T09:35:27.8100620Z Entering 'third_party/fmt' 2025-12-04T09:35:27.8147893Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:35:27.8195953Z Entering 'third_party/gloo' 2025-12-04T09:35:27.8244883Z Entering 'third_party/googletest' 2025-12-04T09:35:27.8294304Z Entering 'third_party/ideep' 2025-12-04T09:35:27.8343499Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:35:27.8400680Z Entering 'third_party/ittapi' 2025-12-04T09:35:27.8449513Z Entering 'third_party/kineto' 2025-12-04T09:35:27.8498422Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:35:27.8544723Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:35:27.8593852Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:35:27.8644262Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:35:27.8689802Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:35:27.8737402Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:35:27.8787325Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:35:27.8834345Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:35:27.8881976Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:35:27.8931836Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:35:27.8978656Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:35:27.9025874Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:35:27.9074578Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:35:27.9126567Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:35:27.9173952Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:35:27.9224626Z Entering 'third_party/kleidiai' 2025-12-04T09:35:27.9274161Z Entering 'third_party/mimalloc' 2025-12-04T09:35:27.9324866Z Entering 'third_party/nlohmann' 2025-12-04T09:35:27.9374549Z Entering 'third_party/onnx' 2025-12-04T09:35:27.9442772Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:35:27.9493146Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:35:27.9546320Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:35:27.9592900Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:35:27.9640231Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:35:27.9686290Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:35:27.9733601Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:35:27.9781403Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:35:27.9827286Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:35:27.9873837Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:35:27.9925093Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:35:27.9977145Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:35:28.0051834Z Entering 'third_party/pocketfft' 2025-12-04T09:35:28.0100960Z Entering 'third_party/protobuf' 2025-12-04T09:35:28.0152141Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:35:28.0199133Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:35:28.0247915Z Entering 'third_party/psimd' 2025-12-04T09:35:28.0297026Z Entering 'third_party/pthreadpool' 2025-12-04T09:35:28.0343752Z Entering 'third_party/pybind11' 2025-12-04T09:35:28.0391122Z Entering 'third_party/python-peachpy' 2025-12-04T09:35:28.0438325Z Entering 'third_party/sleef' 2025-12-04T09:35:28.0485695Z Entering 'third_party/tensorpipe' 2025-12-04T09:35:28.0532995Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:35:28.0578216Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:35:28.0624016Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:35:28.0669583Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:35:28.0713480Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:35:28.0773189Z ##[endgroup] 2025-12-04T09:35:28.0812111Z [command]/usr/bin/git log -1 --format=%H 2025-12-04T09:35:28.0836838Z ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:35:28.0948379Z ##[group]Run cd "${GITHUB_WORKSPACE}" 2025-12-04T09:35:28.0948803Z cd "${GITHUB_WORKSPACE}" 2025-12-04T09:35:28.0949329Z # Clean stale submodule dirs 2025-12-04T09:35:28.0949695Z if [ -z "${NO_SUDO}" ]; then 2025-12-04T09:35:28.0950163Z  sudo git submodule foreach --recursive git clean -ffdx 2025-12-04T09:35:28.0950626Z else 2025-12-04T09:35:28.0950989Z  git submodule foreach --recursive git clean -ffdx 2025-12-04T09:35:28.0951426Z fi 2025-12-04T09:35:28.0962029Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:35:28.0962491Z env: 2025-12-04T09:35:28.0962736Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:35:28.0963050Z NO_SUDO: true 2025-12-04T09:35:28.0963316Z ##[endgroup] 2025-12-04T09:35:28.1319449Z Entering 'android/libs/fbjni' 2025-12-04T09:35:28.1357129Z Entering 'third_party/FP16' 2025-12-04T09:35:28.1391823Z Entering 'third_party/FXdiv' 2025-12-04T09:35:28.1426469Z Entering 'third_party/NNPACK' 2025-12-04T09:35:28.1465180Z Entering 'third_party/NVTX' 2025-12-04T09:35:28.1508642Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T09:35:28.1546247Z Entering 'third_party/XNNPACK' 2025-12-04T09:35:28.1682728Z Entering 'third_party/aiter' 2025-12-04T09:35:28.1729870Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T09:35:28.1849590Z Entering 'third_party/benchmark' 2025-12-04T09:35:28.1885510Z Entering 'third_party/composable_kernel' 2025-12-04T09:35:28.2015669Z Entering 'third_party/cpp-httplib' 2025-12-04T09:35:28.2053895Z Entering 'third_party/cpuinfo' 2025-12-04T09:35:28.2094545Z Entering 'third_party/cudnn_frontend' 2025-12-04T09:35:28.2134243Z Entering 'third_party/cutlass' 2025-12-04T09:35:28.2249398Z Entering 'third_party/fbgemm' 2025-12-04T09:35:28.2317340Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T09:35:28.2351705Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T09:35:28.2483656Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T09:35:28.2527791Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T09:35:28.2641550Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T09:35:28.2678077Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T09:35:28.2710890Z Entering 'third_party/fbgemm/external/json' 2025-12-04T09:35:28.2759436Z Entering 'third_party/flash-attention' 2025-12-04T09:35:28.2805372Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T09:35:28.2913757Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T09:35:28.3017685Z Entering 'third_party/flatbuffers' 2025-12-04T09:35:28.3094012Z Entering 'third_party/fmt' 2025-12-04T09:35:28.3130182Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T09:35:28.3166846Z Entering 'third_party/gloo' 2025-12-04T09:35:28.3205382Z Entering 'third_party/googletest' 2025-12-04T09:35:28.3242740Z Entering 'third_party/ideep' 2025-12-04T09:35:28.3275729Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T09:35:28.3369956Z Entering 'third_party/ittapi' 2025-12-04T09:35:28.3408325Z Entering 'third_party/kineto' 2025-12-04T09:35:28.3447130Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T09:35:28.3488381Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T09:35:28.3538628Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T09:35:28.3573812Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T09:35:28.3611129Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T09:35:28.3643617Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T09:35:28.3679382Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T09:35:28.3716852Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T09:35:28.3754224Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T09:35:28.3800173Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T09:35:28.3835356Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T09:35:28.3871305Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:35:28.3925684Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:35:28.3967424Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T09:35:28.4003421Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T09:35:28.4042395Z Entering 'third_party/kleidiai' 2025-12-04T09:35:28.4087011Z Entering 'third_party/mimalloc' 2025-12-04T09:35:28.4124696Z Entering 'third_party/nlohmann' 2025-12-04T09:35:28.4174460Z Entering 'third_party/onnx' 2025-12-04T09:35:28.4538952Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T09:35:28.4578818Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T09:35:28.4639706Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T09:35:28.4674134Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T09:35:28.4711239Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T09:35:28.4744275Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T09:35:28.4790499Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T09:35:28.4825009Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T09:35:28.4861617Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T09:35:28.4898202Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T09:35:28.4950903Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T09:35:28.4990729Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T09:35:28.5283087Z Entering 'third_party/pocketfft' 2025-12-04T09:35:28.5320799Z Entering 'third_party/protobuf' 2025-12-04T09:35:28.5408538Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T09:35:28.5443140Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T09:35:28.5484068Z Entering 'third_party/psimd' 2025-12-04T09:35:28.5518355Z Entering 'third_party/pthreadpool' 2025-12-04T09:35:28.5552753Z Entering 'third_party/pybind11' 2025-12-04T09:35:28.5591624Z Entering 'third_party/python-peachpy' 2025-12-04T09:35:28.5627654Z Entering 'third_party/sleef' 2025-12-04T09:35:28.5665763Z Entering 'third_party/tensorpipe' 2025-12-04T09:35:28.5704198Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T09:35:28.5741242Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T09:35:28.5775560Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T09:35:28.5816131Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T09:35:28.5849644Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T09:35:28.6055231Z Prepare all required actions 2025-12-04T09:35:28.6055891Z Getting action download info 2025-12-04T09:35:28.7717167Z ##[group]Run ./.github/actions/setup-linux 2025-12-04T09:35:28.7717546Z env: 2025-12-04T09:35:28.7717803Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:35:28.7718110Z ##[endgroup] 2025-12-04T09:35:28.7769050Z ##[group]Run set -euo pipefail 2025-12-04T09:35:28.7769482Z set -euo pipefail 2025-12-04T09:35:28.7769823Z function get_ec2_metadata() { 2025-12-04T09:35:28.7770270Z  # Pulled from instance metadata endpoint for EC2 2025-12-04T09:35:28.7771014Z  # see https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/instancedata-data-retrieval.html 2025-12-04T09:35:28.7771686Z  category=$1 2025-12-04T09:35:28.7772111Z  # If it is GCP runner (runner name contains gcp), do not run this 2025-12-04T09:35:28.7772611Z  runner_name_str=i-092818c7270e5db43 2025-12-04T09:35:28.7773068Z  if [[ -f /.inarc ]]; then 2025-12-04T09:35:28.7773461Z  echo "ARC Runner, no info on ec2 metadata" 2025-12-04T09:35:28.7773915Z  elif [[ $runner_name_str == *"gcp"* ]]; then 2025-12-04T09:35:28.7774464Z  echo "Runner is from Google Cloud Platform, No info on ec2 metadata" 2025-12-04T09:35:28.7774964Z  else 2025-12-04T09:35:28.7775972Z  curl -H "X-aws-ec2-metadata-token: $(curl -s -X PUT "http://169.254.169.254/latest/api/token" -H "X-aws-ec2-metadata-token-ttl-seconds: 30")" -fsSL "http://169.254.169.254/latest/meta-data/${category}" 2025-12-04T09:35:28.7777180Z  fi 2025-12-04T09:35:28.7777429Z } 2025-12-04T09:35:28.7777731Z echo "ami-id: $(get_ec2_metadata ami-id)" 2025-12-04T09:35:28.7778218Z echo "instance-id: $(get_ec2_metadata instance-id)" 2025-12-04T09:35:28.7778784Z echo "instance-type: $(get_ec2_metadata instance-type)" 2025-12-04T09:35:28.7779271Z echo "system info $(uname -a)" 2025-12-04T09:35:28.7786160Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:35:28.7786603Z env: 2025-12-04T09:35:28.7786852Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:35:28.7787140Z ##[endgroup] 2025-12-04T09:35:28.7945436Z ami-id: ami-08982f1c5bf93d976 2025-12-04T09:35:28.8058702Z instance-id: i-092818c7270e5db43 2025-12-04T09:35:28.8172352Z instance-type: g4dn.4xlarge 2025-12-04T09:35:28.8184658Z system info Linux ip-10-0-50-234.ec2.internal 6.1.150-174.273.amzn2023.x86_64 #1 SMP PREEMPT_DYNAMIC Tue Sep 9 12:21:26 UTC 2025 x86_64 x86_64 x86_64 GNU/Linux 2025-12-04T09:35:28.8208396Z ##[group]Run if [ -f /usr/bin/nvidia-smi ]; then nvidia-smi; fi 2025-12-04T09:35:28.8208965Z if [ -f /usr/bin/nvidia-smi ]; then nvidia-smi; fi 2025-12-04T09:35:28.8215870Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:35:28.8216323Z env: 2025-12-04T09:35:28.8216564Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:35:28.8216961Z ##[endgroup] 2025-12-04T09:35:30.1834730Z Thu Dec 4 09:35:30 2025 2025-12-04T09:35:30.1835444Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:35:30.1836388Z | NVIDIA-SMI 580.82.07 Driver Version: 580.82.07 CUDA Version: 13.0 | 2025-12-04T09:35:30.1837274Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:35:30.1838252Z | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T09:35:30.1839199Z | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | 2025-12-04T09:35:30.1839948Z | | | MIG M. | 2025-12-04T09:35:30.1840573Z |=========================================+========================+======================| 2025-12-04T09:35:30.1938036Z | 0 Tesla T4 Off | 00000000:00:1E.0 Off | 0 | 2025-12-04T09:35:30.1939347Z | N/A 31C P0 26W / 70W | 0MiB / 15360MiB | 9% Default | 2025-12-04T09:35:30.1940035Z | | | N/A | 2025-12-04T09:35:30.1940779Z +-----------------------------------------+------------------------+----------------------+ 2025-12-04T09:35:30.1941408Z 2025-12-04T09:35:30.1941787Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:35:30.1942702Z | Processes: | 2025-12-04T09:35:30.1943502Z | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T09:35:30.1944250Z | ID ID Usage | 2025-12-04T09:35:30.1944893Z |=========================================================================================| 2025-12-04T09:35:30.1945668Z | No running processes found | 2025-12-04T09:35:30.1946505Z +-----------------------------------------------------------------------------------------+ 2025-12-04T09:35:30.6144926Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:35:30.6146072Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:35:30.6155119Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:35:30.6155568Z env: 2025-12-04T09:35:30.6155808Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:35:30.6156117Z ##[endgroup] 2025-12-04T09:35:30.6218722Z ##[group]Run if systemctl is-active --quiet docker; then 2025-12-04T09:35:30.6219238Z if systemctl is-active --quiet docker; then 2025-12-04T09:35:30.6219693Z  echo "Docker daemon is running..."; 2025-12-04T09:35:30.6220094Z else 2025-12-04T09:35:30.6220500Z  echo "Starting docker daemon..." && sudo systemctl start docker; 2025-12-04T09:35:30.6221003Z fi 2025-12-04T09:35:30.6227326Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:35:30.6227774Z env: 2025-12-04T09:35:30.6228035Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:35:30.6228334Z ##[endgroup] 2025-12-04T09:35:30.6318346Z Docker daemon is running... 2025-12-04T09:35:30.6362005Z ##[group]Run nick-fields/retry@v3.0.0 2025-12-04T09:35:30.6362355Z with: 2025-12-04T09:35:30.6362593Z shell: bash 2025-12-04T09:35:30.6362847Z timeout_minutes: 5 2025-12-04T09:35:30.6363128Z max_attempts: 3 2025-12-04T09:35:30.6363389Z retry_wait_seconds: 30 2025-12-04T09:35:30.6366135Z command: AWS_ACCOUNT_ID=$(aws sts get-caller-identity|grep Account|cut -f4 -d\") aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" # For LF Runners we need to make sure we also login to Meta's ECR docker registry too. META_AWS_ACCOUNT_ID=308535385114 if [ "$AWS_ACCOUNT_ID" != "$META_AWS_ACCOUNT_ID" ] ; then aws ecr get-login-password --region "$AWS_DEFAULT_REGION" | docker login --username AWS \ --password-stdin "$META_AWS_ACCOUNT_ID.dkr.ecr.$AWS_DEFAULT_REGION.amazonaws.com" fi 2025-12-04T09:35:30.6368886Z polling_interval_seconds: 1 2025-12-04T09:35:30.6369217Z warning_on_retry: true 2025-12-04T09:35:30.6369528Z continue_on_error: false 2025-12-04T09:35:30.6369812Z env: 2025-12-04T09:35:30.6370059Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:35:30.6370370Z AWS_RETRY_MODE: standard 2025-12-04T09:35:30.6370662Z AWS_MAX_ATTEMPTS: 5 2025-12-04T09:35:30.6370959Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:35:30.6371276Z ##[endgroup] 2025-12-04T09:35:31.9227860Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T09:35:31.9228878Z Configure a credential helper to remove this warning. See 2025-12-04T09:35:31.9229562Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T09:35:31.9230033Z 2025-12-04T09:35:31.9230140Z Login Succeeded 2025-12-04T09:35:32.7330541Z Command completed after 1 attempt(s). 2025-12-04T09:35:32.7387149Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:35:32.7387800Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:35:32.7388337Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:35:32.7395882Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:35:32.7396582Z env: 2025-12-04T09:35:32.7396833Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:35:32.7397150Z ##[endgroup] 2025-12-04T09:35:32.7485571Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T09:35:32.7486240Z # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T09:35:32.7486773Z # shellcheck disable=SC2046 2025-12-04T09:35:32.7487174Z docker stop $(docker ps -q) || true 2025-12-04T09:35:32.7487588Z # Prune all of the docker images 2025-12-04T09:35:32.7487969Z docker system prune -af 2025-12-04T09:35:32.7494916Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:35:32.7495372Z env: 2025-12-04T09:35:32.7495633Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:35:32.7495941Z ##[endgroup] 2025-12-04T09:35:32.7782812Z "docker stop" requires at least 1 argument. 2025-12-04T09:35:32.7783272Z See 'docker stop --help'. 2025-12-04T09:35:32.7783475Z 2025-12-04T09:35:32.7783687Z Usage: docker stop [OPTIONS] CONTAINER [CONTAINER...] 2025-12-04T09:35:32.7784007Z 2025-12-04T09:35:32.7784145Z Stop one or more running containers 2025-12-04T09:35:32.8136583Z Total reclaimed space: 0B 2025-12-04T09:35:32.8347990Z ##[group]Run pytorch/test-infra/.github/actions/calculate-docker-image@main 2025-12-04T09:35:32.8348556Z with: 2025-12-04T09:35:32.8349504Z docker-image-name: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:35:32.8350579Z use-custom-docker-registry: true 2025-12-04T09:35:32.8350949Z docker-build-dir: .ci/docker 2025-12-04T09:35:32.8351277Z docker-build-script: ./build.sh 2025-12-04T09:35:32.8351799Z working-directory: . 2025-12-04T09:35:32.8352209Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:35:32.8352667Z force-push: false 2025-12-04T09:35:32.8352932Z env: 2025-12-04T09:35:32.8353177Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:35:32.8353461Z ##[endgroup] 2025-12-04T09:35:32.8375005Z ##[group]Run set -ex 2025-12-04T09:35:32.8375356Z set -ex 2025-12-04T09:35:32.8375617Z  2025-12-04T09:35:32.8376134Z # If the docker build directory or the build script doesn't exist, the action will 2025-12-04T09:35:32.8377029Z # gracefully return the docker image name as it is. Pulling docker image in Linux 2025-12-04T09:35:32.8377703Z # job could then download the pre-built image as usual 2025-12-04T09:35:32.8378524Z if [[ -d "${DOCKER_BUILD_DIR}" ]] && [[ -f "${DOCKER_BUILD_DIR}/${DOCKER_BUILD_SCRIPT}" ]] && [[ "${USE_CUSTOM_DOCKER_REGISTRY}" == "true" ]]; then 2025-12-04T09:35:32.8379274Z  echo "skip=false" >> "${GITHUB_OUTPUT}" 2025-12-04T09:35:32.8379677Z else 2025-12-04T09:35:32.8379976Z  echo "skip=true" >> "${GITHUB_OUTPUT}" 2025-12-04T09:35:32.8380503Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:35:32.8380971Z  2025-12-04T09:35:32.8381623Z  echo "Not using custom ECR registry. Either it was not requested or there is no Docker build script in the ${REPO_NAME} repo..." 2025-12-04T09:35:32.8382383Z  exit 0 2025-12-04T09:35:32.8382629Z fi 2025-12-04T09:35:32.8382866Z  2025-12-04T09:35:32.8383259Z if [[ "${DOCKER_IMAGE_NAME}" == *"${DOCKER_REGISTRY}/${REPO_NAME}"* ]]; then 2025-12-04T09:35:32.8383970Z  # The docker image name already includes the ECR prefix and tag, so we can just 2025-12-04T09:35:32.8384589Z  # use it as it is, but first let's extract the tag 2025-12-04T09:35:32.8385153Z  DOCKER_TAG=$(echo "${DOCKER_IMAGE_NAME}" | awk -F '[:,]' '{print $2}') 2025-12-04T09:35:32.8385750Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:35:32.8386311Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:35:32.8386785Z else 2025-12-04T09:35:32.8387099Z  if [[ "${DOCKER_IMAGE_NAME}" == *:* ]]; then 2025-12-04T09:35:32.8387545Z  CUSTOM_TAG_PREFIX=${DOCKER_IMAGE_NAME#*:} 2025-12-04T09:35:32.8387994Z  DOCKER_IMAGE_NAME=${DOCKER_IMAGE_NAME%%:*} 2025-12-04T09:35:32.8388385Z  fi 2025-12-04T09:35:32.8388918Z  DOCKER_TAG=${CUSTOM_TAG_PREFIX:+${CUSTOM_TAG_PREFIX}-}$(git rev-parse HEAD:"${DOCKER_BUILD_DIR}") 2025-12-04T09:35:32.8389628Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:35:32.8390383Z  echo "docker-image=${DOCKER_REGISTRY}/${REPO_NAME}/${DOCKER_IMAGE_NAME}:${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:35:32.8391217Z  echo "custom-tag-prefix=${CUSTOM_TAG_PREFIX}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:35:32.8391725Z fi 2025-12-04T09:35:32.8398931Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:35:32.8399375Z env: 2025-12-04T09:35:32.8399626Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:35:32.8399922Z REPO_NAME: pytorch 2025-12-04T09:35:32.8401054Z DOCKER_IMAGE_NAME: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:35:32.8402109Z DOCKER_BUILD_DIR: .ci/docker 2025-12-04T09:35:32.8402445Z DOCKER_BUILD_SCRIPT: ./build.sh 2025-12-04T09:35:32.8402876Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:35:32.8403350Z USE_CUSTOM_DOCKER_REGISTRY: true 2025-12-04T09:35:32.8403691Z CUSTOM_TAG_PREFIX: 2025-12-04T09:35:32.8403954Z ##[endgroup] 2025-12-04T09:35:32.8434720Z + [[ -d .ci/docker ]] 2025-12-04T09:35:32.8435031Z + [[ -f .ci/docker/./build.sh ]] 2025-12-04T09:35:32.8435545Z + [[ true == \t\r\u\e ]] 2025-12-04T09:35:32.8435836Z + echo skip=false 2025-12-04T09:35:32.8437109Z + [[ 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a == *\3\0\8\5\3\5\3\8\5\1\1\4\.\d\k\r\.\e\c\r\.\u\s\-\e\a\s\t\-\1\.\a\m\a\z\o\n\a\w\s\.\c\o\m\/\p\y\t\o\r\c\h* ]] 2025-12-04T09:35:32.8442826Z ++ echo 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:35:32.8443800Z ++ awk -F '[:,]' '{print $2}' 2025-12-04T09:35:32.8467281Z + DOCKER_TAG=pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:35:32.8468354Z + echo docker-tag=pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:35:32.8469907Z + echo docker-image=308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:35:32.8496374Z ##[group]Run set +e 2025-12-04T09:35:32.8496731Z set +e 2025-12-04T09:35:32.8497046Z set -x 2025-12-04T09:35:32.8497305Z  2025-12-04T09:35:32.8497546Z login() { 2025-12-04T09:35:32.8498087Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-12-04T09:35:32.8498699Z } 2025-12-04T09:35:32.8498950Z  2025-12-04T09:35:32.8499175Z retry () { 2025-12-04T09:35:32.8499489Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-12-04T09:35:32.8499859Z } 2025-12-04T09:35:32.8500083Z  2025-12-04T09:35:32.8500348Z retry login "${DOCKER_REGISTRY}" 2025-12-04T09:35:32.8500704Z  2025-12-04T09:35:32.8500954Z START_TIME=$(date +%s) 2025-12-04T09:35:32.8501282Z # Wait up to 120 minutes 2025-12-04T09:35:32.8501708Z while [[ $(( $(date +%s) - 7200 )) -lt $START_TIME ]]; do 2025-12-04T09:35:32.8502297Z  # Check if image already exists, if it does then skip building it 2025-12-04T09:35:32.8502867Z  if docker manifest inspect "${DOCKER_IMAGE}"; then 2025-12-04T09:35:32.8503292Z  exit 0 2025-12-04T09:35:32.8503560Z  fi 2025-12-04T09:35:32.8503791Z  2025-12-04T09:35:32.8504244Z  # NB: This flag is used by Docker build workflow to push the image to ECR, so we can 2025-12-04T09:35:32.8505025Z  # use this to differentiate between the Docker build and regular build jobs. For the 2025-12-04T09:35:32.8505810Z  # latter, it will wait for the Docker images to become available before continuing 2025-12-04T09:35:32.8506410Z  if [ "${DOCKER_PUSH:-false}" == "true" ]; then 2025-12-04T09:35:32.8506877Z  # It's a Docker build job, let's build the image 2025-12-04T09:35:32.8507282Z  break 2025-12-04T09:35:32.8507556Z  else 2025-12-04T09:35:32.8507962Z  # It's a regular build job, wait for the image to become available 2025-12-04T09:35:32.8508455Z  sleep 300 2025-12-04T09:35:32.8508728Z  fi 2025-12-04T09:35:32.8508978Z done 2025-12-04T09:35:32.8509227Z  2025-12-04T09:35:32.8509630Z # NB: This part requires a full checkout. Otherwise, the merge base will 2025-12-04T09:35:32.8510499Z # be empty. The default action would be to continue rebuild the image 2025-12-04T09:35:32.8511113Z if [[ "$BASE_REVISION" = "$(git rev-parse HEAD)" ]]; then 2025-12-04T09:35:32.8511650Z  # if we're on the base branch then use the parent commit 2025-12-04T09:35:32.8512118Z  MERGE_BASE=$(git rev-parse HEAD~) 2025-12-04T09:35:32.8512496Z else 2025-12-04T09:35:32.8512882Z  # otherwise we're on a PR, so use the most recent base commit 2025-12-04T09:35:32.8513435Z  MERGE_BASE=$(git merge-base HEAD "$BASE_REVISION") 2025-12-04T09:35:32.8513976Z fi 2025-12-04T09:35:32.8514224Z  2025-12-04T09:35:32.8514491Z if [[ -z "${MERGE_BASE}" ]]; then 2025-12-04T09:35:32.8514894Z  echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-12-04T09:35:32.8515286Z  2025-12-04T09:35:32.8515838Z  echo "Finding merge base only works with full checkout, please set fetch-depth to 0, continuing ..." 2025-12-04T09:35:32.8516486Z  exit 0 2025-12-04T09:35:32.8516740Z fi 2025-12-04T09:35:32.8516974Z  2025-12-04T09:35:32.8517311Z if ! git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}"; then 2025-12-04T09:35:32.8518120Z  echo "Directory '${DOCKER_BUILD_DIR}' not found in commit $MERGE_BASE, you should rebase onto a more recent commit" 2025-12-04T09:35:32.8518812Z  exit 1 2025-12-04T09:35:32.8519068Z fi 2025-12-04T09:35:32.8519292Z  2025-12-04T09:35:32.8519723Z PREVIOUS_DOCKER_TAG=$(git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}") 2025-12-04T09:35:32.8520505Z # If no image exists but the hash is the same as the previous hash then we should error out here 2025-12-04T09:35:32.8521209Z if [[ "${PREVIOUS_DOCKER_TAG}" == "${DOCKER_TAG}" ]]; then 2025-12-04T09:35:32.8521999Z  echo "WARNING: Something has gone wrong and the previous image isn't available for the merge-base of your branch" 2025-12-04T09:35:32.8522916Z  echo " Will re-build docker image to store in local cache, TTS may be longer" 2025-12-04T09:35:32.8523460Z fi 2025-12-04T09:35:32.8523688Z  2025-12-04T09:35:32.8523985Z echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-12-04T09:35:32.8530818Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:35:32.8531244Z env: 2025-12-04T09:35:32.8531490Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:35:32.8531808Z DOCKER_BUILD_DIR: .ci/docker 2025-12-04T09:35:32.8532212Z BASE_REVISION: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:35:32.8533306Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:35:32.8534654Z DOCKER_TAG: pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:35:32.8535457Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:35:32.8535918Z DOCKER_PUSH: 2025-12-04T09:35:32.8536165Z ##[endgroup] 2025-12-04T09:35:32.8563355Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:35:32.8563861Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:35:32.8566400Z + aws ecr get-login-password --region us-east-1 2025-12-04T09:35:32.8567953Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:35:33.4387506Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T09:35:33.4388267Z Configure a credential helper to remove this warning. See 2025-12-04T09:35:33.4389049Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T09:35:33.4389647Z 2025-12-04T09:35:33.4389754Z Login Succeeded 2025-12-04T09:35:33.4405460Z ++ date +%s 2025-12-04T09:35:33.4416086Z + START_TIME=1764840933 2025-12-04T09:35:33.4419925Z ++ date +%s 2025-12-04T09:35:33.4430539Z + [[ 1764833733 -lt 1764840933 ]] 2025-12-04T09:35:33.4431636Z + docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:35:33.7257271Z { 2025-12-04T09:35:33.7257744Z "schemaVersion": 2, 2025-12-04T09:35:33.7258422Z "mediaType": "application/vnd.docker.distribution.manifest.v2+json", 2025-12-04T09:35:33.7259146Z "config": { 2025-12-04T09:35:33.7259553Z "mediaType": "application/vnd.docker.container.image.v1+json", 2025-12-04T09:35:33.7260712Z "size": 34787, 2025-12-04T09:35:33.7261440Z "digest": "sha256:5465aa79632b68f6240c23f0d0b021df4d0fd595333b61a40d36a0cf73656024" 2025-12-04T09:35:33.7262073Z }, 2025-12-04T09:35:33.7262466Z "layers": [ 2025-12-04T09:35:33.7262845Z { 2025-12-04T09:35:33.7263241Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7263737Z "size": 30447951, 2025-12-04T09:35:33.7264270Z "digest": "sha256:63e5bc7682b85ae57a1221210f64d62e7a90b0a30f19af4ca734b8242ae49d63" 2025-12-04T09:35:33.7265019Z }, 2025-12-04T09:35:33.7265262Z { 2025-12-04T09:35:33.7265949Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7266722Z "size": 1554, 2025-12-04T09:35:33.7267283Z "digest": "sha256:835841cca3b7e1464290cdb78e48773e03583413fbed852c3cc5165a392ea44d" 2025-12-04T09:35:33.7267834Z }, 2025-12-04T09:35:33.7268051Z { 2025-12-04T09:35:33.7268417Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7268917Z "size": 313276213, 2025-12-04T09:35:33.7269425Z "digest": "sha256:1bf1bb125deaa5b8a3adf121671e87ba2fa7e229f9eb1dff7ade581cb737175a" 2025-12-04T09:35:33.7269976Z }, 2025-12-04T09:35:33.7270192Z { 2025-12-04T09:35:33.7270568Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7271037Z "size": 787, 2025-12-04T09:35:33.7271514Z "digest": "sha256:b21856d1bf420da6fa8ec7331b82ab355d4f4178644e7d3a3d3d0fbc3610109a" 2025-12-04T09:35:33.7272076Z }, 2025-12-04T09:35:33.7272279Z { 2025-12-04T09:35:33.7272654Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7273137Z "size": 106, 2025-12-04T09:35:33.7273621Z "digest": "sha256:848ba2c095e2b9e6acfb0ecf077adb526fb2fa82ed44cf6648ebde97f296f8ec" 2025-12-04T09:35:33.7274170Z }, 2025-12-04T09:35:33.7274384Z { 2025-12-04T09:35:33.7274757Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7275224Z "size": 704, 2025-12-04T09:35:33.7275702Z "digest": "sha256:029495b23122c840ca0e52d487afa8d2c4dbf1991cd7f204ec3e434dcf947bf4" 2025-12-04T09:35:33.7276556Z }, 2025-12-04T09:35:33.7276915Z { 2025-12-04T09:35:33.7277621Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7278311Z "size": 1216, 2025-12-04T09:35:33.7278789Z "digest": "sha256:073bb82063cfba4639b11fea43753dbb128f9238353189fc02d2e2aa0b2ad359" 2025-12-04T09:35:33.7279440Z + exit 0 2025-12-04T09:35:33.7279676Z }, 2025-12-04T09:35:33.7279896Z { 2025-12-04T09:35:33.7280257Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7280742Z "size": 484, 2025-12-04T09:35:33.7281212Z "digest": "sha256:59b63930883363c7d2aaab27cc61555d9f3e119dc18247a8624c98ebdaa354a5" 2025-12-04T09:35:33.7281747Z }, 2025-12-04T09:35:33.7281964Z { 2025-12-04T09:35:33.7282335Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7282818Z "size": 110362071, 2025-12-04T09:35:33.7283304Z "digest": "sha256:1c6177b2970db2d7743b4337c420a35f2ec79f338c30d97d534a1f0987c00913" 2025-12-04T09:35:33.7283846Z }, 2025-12-04T09:35:33.7284059Z { 2025-12-04T09:35:33.7284423Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7284904Z "size": 4961, 2025-12-04T09:35:33.7285387Z "digest": "sha256:fabe466dd5f33c3209a56abf5cb46b9b07fe21c57fb43b98e13308c8665c0864" 2025-12-04T09:35:33.7285932Z }, 2025-12-04T09:35:33.7286345Z { 2025-12-04T09:35:33.7286722Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7287191Z "size": 1755, 2025-12-04T09:35:33.7287662Z "digest": "sha256:2b5a11b41761d8ea3b829e4772e4064cb6c4e4989126af324d0057661e4493a1" 2025-12-04T09:35:33.7288209Z }, 2025-12-04T09:35:33.7288412Z { 2025-12-04T09:35:33.7288790Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7289271Z "size": 724, 2025-12-04T09:35:33.7289738Z "digest": "sha256:9681563a88ff9e62494a2740e537440d3df978d466c9478d6a941fae8b57b084" 2025-12-04T09:35:33.7290356Z }, 2025-12-04T09:35:33.7290574Z { 2025-12-04T09:35:33.7290954Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7291426Z "size": 544, 2025-12-04T09:35:33.7291902Z "digest": "sha256:dc0780902fca810498f16efa71f8e5990385f141a0cfcc552616a4acc434f79a" 2025-12-04T09:35:33.7292455Z }, 2025-12-04T09:35:33.7292657Z { 2025-12-04T09:35:33.7293034Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7293521Z "size": 3185191720, 2025-12-04T09:35:33.7294006Z "digest": "sha256:5b09a2b135c8e540e2b9374b68991afdd63a5dfaba75fb44efe054a591f400c2" 2025-12-04T09:35:33.7294562Z }, 2025-12-04T09:35:33.7294769Z { 2025-12-04T09:35:33.7295163Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7295648Z "size": 32, 2025-12-04T09:35:33.7296354Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:35:33.7296989Z }, 2025-12-04T09:35:33.7297211Z { 2025-12-04T09:35:33.7297592Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7298064Z "size": 396, 2025-12-04T09:35:33.7298560Z "digest": "sha256:5bfdaeb5578d6ffcd7db29c48303cbceb13c591210feaa216a8daa7a6d445b4b" 2025-12-04T09:35:33.7299136Z }, 2025-12-04T09:35:33.7299344Z { 2025-12-04T09:35:33.7299727Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7300223Z "size": 236865, 2025-12-04T09:35:33.7300683Z "digest": "sha256:0ef42867f370b8a14b8c301388793b78a0bd2533bb2a317b129b03c8667dc767" 2025-12-04T09:35:33.7301226Z }, 2025-12-04T09:35:33.7301447Z { 2025-12-04T09:35:33.7301810Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7302298Z "size": 230, 2025-12-04T09:35:33.7302763Z "digest": "sha256:446083e497f322789c2d87933a77fb2dfd94e18d2e85f6d4362e6e9521b82c4e" 2025-12-04T09:35:33.7303305Z }, 2025-12-04T09:35:33.7303514Z { 2025-12-04T09:35:33.7303886Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7304362Z "size": 3043500, 2025-12-04T09:35:33.7304831Z "digest": "sha256:d8a170bef0f4e0e28f5ba0952320dd465552adf74f0864b4f47cc11f4c4f82f7" 2025-12-04T09:35:33.7305383Z }, 2025-12-04T09:35:33.7305595Z { 2025-12-04T09:35:33.7305951Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7306438Z "size": 1472, 2025-12-04T09:35:33.7306914Z "digest": "sha256:e2b6cd6a5bd0418a1e4aca3f37942324d4d9f9b0177597e37fc8d1a5626048e1" 2025-12-04T09:35:33.7307451Z }, 2025-12-04T09:35:33.7307667Z { 2025-12-04T09:35:33.7308043Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7308509Z "size": 481, 2025-12-04T09:35:33.7308981Z "digest": "sha256:93efc0181a22218a544413f1d57e9e0e7a0f492e41bef598084c5b9177e3987a" 2025-12-04T09:35:33.7309528Z }, 2025-12-04T09:35:33.7309741Z { 2025-12-04T09:35:33.7310111Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7310589Z "size": 202, 2025-12-04T09:35:33.7311059Z "digest": "sha256:7454c938f17425bcf167ad28a62b42b95f638a7d2cf0840885cfe5ffe8480a12" 2025-12-04T09:35:33.7311599Z }, 2025-12-04T09:35:33.7311812Z { 2025-12-04T09:35:33.7312182Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7312651Z "size": 607, 2025-12-04T09:35:33.7313270Z "digest": "sha256:4d57ff55f6d4161cb6c29e2c0b08d47e65898427db3938479158684899f0023d" 2025-12-04T09:35:33.7313822Z }, 2025-12-04T09:35:33.7314026Z { 2025-12-04T09:35:33.7314401Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7314891Z "size": 6243016141, 2025-12-04T09:35:33.7315375Z "digest": "sha256:b0301534b4a58072d5b140b08a7608bbead41d126fa29fdc78c1e8a43ebb865d" 2025-12-04T09:35:33.7315928Z }, 2025-12-04T09:35:33.7316145Z { 2025-12-04T09:35:33.7316522Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7317089Z "size": 829, 2025-12-04T09:35:33.7317564Z "digest": "sha256:1969e15d0c13874ea5883ed829235a19ef6dc21c8aa6172032b78a8ffa6ff262" 2025-12-04T09:35:33.7318109Z }, 2025-12-04T09:35:33.7318311Z { 2025-12-04T09:35:33.7318685Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7319170Z "size": 33450177, 2025-12-04T09:35:33.7319656Z "digest": "sha256:73180a0f2d5a961a0cc0ba2c3cf375fdcfb43ae5e4e5c63a000c4b4366d52a64" 2025-12-04T09:35:33.7320213Z }, 2025-12-04T09:35:33.7320428Z { 2025-12-04T09:35:33.7320789Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7321269Z "size": 104, 2025-12-04T09:35:33.7321742Z "digest": "sha256:ad81b25cb69f8cf42a4a96678a64b7d0598a8f95236a3e63d1fec4e53edff613" 2025-12-04T09:35:33.7322292Z }, 2025-12-04T09:35:33.7322493Z { 2025-12-04T09:35:33.7322866Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7323351Z "size": 1496, 2025-12-04T09:35:33.7323812Z "digest": "sha256:8165374f8dccf88a7791a5d31afbe29e4d4542b4f1cf1904945e07f9af6bf8ba" 2025-12-04T09:35:33.7324364Z }, 2025-12-04T09:35:33.7324577Z { 2025-12-04T09:35:33.7324933Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7325417Z "size": 458786969, 2025-12-04T09:35:33.7325915Z "digest": "sha256:7779c0bb9be2030df9060b526b98d0afeed1ce5b61ee0530321ef04a4e145e8c" 2025-12-04T09:35:33.7326456Z }, 2025-12-04T09:35:33.7326668Z { 2025-12-04T09:35:33.7327039Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7327509Z "size": 164, 2025-12-04T09:35:33.7327980Z "digest": "sha256:4d0a1c027262ed8c83181b931b64afa1c41c3cac97580231c4cae3a524ebd7d5" 2025-12-04T09:35:33.7328526Z }, 2025-12-04T09:35:33.7328739Z { 2025-12-04T09:35:33.7329101Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7329589Z "size": 346, 2025-12-04T09:35:33.7330059Z "digest": "sha256:a51e0dab2d596e6563483f27c12660007160847d177ba4c31812a8f44ada5754" 2025-12-04T09:35:33.7330588Z }, 2025-12-04T09:35:33.7330802Z { 2025-12-04T09:35:33.7331176Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7331645Z "size": 32, 2025-12-04T09:35:33.7332121Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:35:33.7332685Z }, 2025-12-04T09:35:33.7332886Z { 2025-12-04T09:35:33.7333259Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7333741Z "size": 106, 2025-12-04T09:35:33.7334210Z "digest": "sha256:3eb6d4ff040b8761b1e3e1da768bdb884ce0e5324e3d0f6471b0a8b2ddf4736f" 2025-12-04T09:35:33.7334763Z }, 2025-12-04T09:35:33.7334979Z { 2025-12-04T09:35:33.7335353Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7335823Z "size": 424, 2025-12-04T09:35:33.7336297Z "digest": "sha256:b168858b85373f8ddca549d79267a06de4fa945d04bf791c55c9ddc93957fa3c" 2025-12-04T09:35:33.7336934Z }, 2025-12-04T09:35:33.7337146Z { 2025-12-04T09:35:33.7337525Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7338010Z "size": 19309367, 2025-12-04T09:35:33.7338478Z "digest": "sha256:d77a39278026a8899e2f97643918bdcf96e711ca26951880b4841b319dc71321" 2025-12-04T09:35:33.7339025Z }, 2025-12-04T09:35:33.7339345Z { 2025-12-04T09:35:33.7339711Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7340199Z "size": 108, 2025-12-04T09:35:33.7340683Z "digest": "sha256:36fbd357280b6b40e90f36ac3d19da3da10e5dbf0027a5cfe8e2f29d1870d347" 2025-12-04T09:35:33.7341245Z }, 2025-12-04T09:35:33.7341450Z { 2025-12-04T09:35:33.7341823Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7342312Z "size": 826, 2025-12-04T09:35:33.7342776Z "digest": "sha256:4e3b10a5dd6aed29f238d604925e2a4f873141c1087c8dd4fdde5c61e7560893" 2025-12-04T09:35:33.7343438Z }, 2025-12-04T09:35:33.7343652Z { 2025-12-04T09:35:33.7344016Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7344501Z "size": 724, 2025-12-04T09:35:33.7344961Z "digest": "sha256:9681563a88ff9e62494a2740e537440d3df978d466c9478d6a941fae8b57b084" 2025-12-04T09:35:33.7345497Z }, 2025-12-04T09:35:33.7345706Z { 2025-12-04T09:35:33.7346082Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7386769Z "size": 149, 2025-12-04T09:35:33.7387648Z "digest": "sha256:3092fab73b59190b9facfc49bf18f58612172bc2fd68dfa339a1118632616939" 2025-12-04T09:35:33.7388488Z }, 2025-12-04T09:35:33.7388750Z { 2025-12-04T09:35:33.7389221Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7389856Z "size": 136, 2025-12-04T09:35:33.7390526Z "digest": "sha256:20020dd28a15ba092fcbfe906ee39cdddfcc9d0b7eb42fdd6f4c08a984fa9c00" 2025-12-04T09:35:33.7391432Z }, 2025-12-04T09:35:33.7391731Z { 2025-12-04T09:35:33.7392294Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7393066Z "size": 140, 2025-12-04T09:35:33.7393848Z "digest": "sha256:ae5280ce969dcff08c091e9a5f7641f13561b2b0ee44d78b7c3f81d8fe8e6d32" 2025-12-04T09:35:33.7394815Z }, 2025-12-04T09:35:33.7395169Z { 2025-12-04T09:35:33.7395820Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7396852Z "size": 32, 2025-12-04T09:35:33.7397643Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:35:33.7398340Z }, 2025-12-04T09:35:33.7398560Z { 2025-12-04T09:35:33.7398928Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7399408Z "size": 223, 2025-12-04T09:35:33.7399870Z "digest": "sha256:026e4484b749dfc556dcf7c8f45c1759518a89072e4dbc974d9405ada1582d03" 2025-12-04T09:35:33.7400395Z }, 2025-12-04T09:35:33.7400608Z { 2025-12-04T09:35:33.7400972Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7401436Z "size": 256, 2025-12-04T09:35:33.7401912Z "digest": "sha256:1be9da2ce53d20d8befad5c024ee0eb41ee35984307cbd5621d8effae0353073" 2025-12-04T09:35:33.7402469Z }, 2025-12-04T09:35:33.7402678Z { 2025-12-04T09:35:33.7403029Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7403513Z "size": 32, 2025-12-04T09:35:33.7403983Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:35:33.7404523Z }, 2025-12-04T09:35:33.7404737Z { 2025-12-04T09:35:33.7405107Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7405576Z "size": 106, 2025-12-04T09:35:33.7406044Z "digest": "sha256:6481b7a1d9fb4001fd6f9e2a8d1600192529ddb957128e41671ca4630fa06ad4" 2025-12-04T09:35:33.7406587Z }, 2025-12-04T09:35:33.7406791Z { 2025-12-04T09:35:33.7407165Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7407649Z "size": 312293471, 2025-12-04T09:35:33.7408133Z "digest": "sha256:fa519d18c39d8f297109c056017ebce7efc322d058afd27fdac5880d6c8d35b0" 2025-12-04T09:35:33.7408689Z }, 2025-12-04T09:35:33.7408905Z { 2025-12-04T09:35:33.7409279Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7409755Z "size": 3058012325, 2025-12-04T09:35:33.7410459Z "digest": "sha256:d172f25b97f78fce0f6c6701f0db794b1c994a9cdf8cff9ddc6bdd1a1bea835c" 2025-12-04T09:35:33.7411032Z }, 2025-12-04T09:35:33.7411234Z { 2025-12-04T09:35:33.7411609Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7412094Z "size": 129, 2025-12-04T09:35:33.7412559Z "digest": "sha256:fd60ab6b1c2c85a932e9894b5d0cf5c9e75fa21782e3028ea40d76017ecfbf85" 2025-12-04T09:35:33.7413121Z }, 2025-12-04T09:35:33.7413339Z { 2025-12-04T09:35:33.7413697Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7414294Z "size": 880, 2025-12-04T09:35:33.7414775Z "digest": "sha256:0afe45579c2c87002db8c1abf7b32a748e6cb3b9b57e9b391f91cad9f84df476" 2025-12-04T09:35:33.7415332Z }, 2025-12-04T09:35:33.7415536Z { 2025-12-04T09:35:33.7415914Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7416395Z "size": 724, 2025-12-04T09:35:33.7416944Z "digest": "sha256:9681563a88ff9e62494a2740e537440d3df978d466c9478d6a941fae8b57b084" 2025-12-04T09:35:33.7417493Z }, 2025-12-04T09:35:33.7417708Z { 2025-12-04T09:35:33.7418065Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7418545Z "size": 139, 2025-12-04T09:35:33.7419009Z "digest": "sha256:5884ffd6720b47274f651262d5f9224f55960f9ea717faafe332aa20afb0ffa4" 2025-12-04T09:35:33.7419536Z }, 2025-12-04T09:35:33.7419751Z { 2025-12-04T09:35:33.7420120Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7420594Z "size": 32, 2025-12-04T09:35:33.7421069Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:35:33.7421623Z }, 2025-12-04T09:35:33.7421837Z { 2025-12-04T09:35:33.7422196Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7422681Z "size": 160, 2025-12-04T09:35:33.7423174Z "digest": "sha256:ab7a7c316fa7a9b7a96304ce96fafdffbc5cc6b960a4bb2def9131b36d9225c5" 2025-12-04T09:35:33.7423726Z }, 2025-12-04T09:35:33.7423940Z { 2025-12-04T09:35:33.7424312Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7424778Z "size": 1012, 2025-12-04T09:35:33.7425254Z "digest": "sha256:c7775ce5574bdde75b4c09a1db19f7d0dc027f1f4c1f961022fc55833133e616" 2025-12-04T09:35:33.7425807Z }, 2025-12-04T09:35:33.7426007Z { 2025-12-04T09:35:33.7426379Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7426859Z "size": 724, 2025-12-04T09:35:33.7427313Z "digest": "sha256:9681563a88ff9e62494a2740e537440d3df978d466c9478d6a941fae8b57b084" 2025-12-04T09:35:33.7427852Z }, 2025-12-04T09:35:33.7428068Z { 2025-12-04T09:35:33.7428440Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7428909Z "size": 134, 2025-12-04T09:35:33.7429383Z "digest": "sha256:81945c4fb228ca73f4bac38b6d8a1eca7139585d4a078219dfaa16ea13945949" 2025-12-04T09:35:33.7429938Z }, 2025-12-04T09:35:33.7430145Z { 2025-12-04T09:35:33.7430522Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7431002Z "size": 32, 2025-12-04T09:35:33.7431463Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:35:33.7432024Z }, 2025-12-04T09:35:33.7432239Z { 2025-12-04T09:35:33.7432599Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7433082Z "size": 158, 2025-12-04T09:35:33.7433553Z "digest": "sha256:663cbe24d60bf42bc7a440cb4867e4287cacf54194dd3152406668e61d7e92e5" 2025-12-04T09:35:33.7434109Z }, 2025-12-04T09:35:33.7434312Z { 2025-12-04T09:35:33.7434688Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7435174Z "size": 603, 2025-12-04T09:35:33.7435618Z "digest": "sha256:43f216b027865c8ca16f855703465445f3a548614a4d7e29387337b9651ac25c" 2025-12-04T09:35:33.7436153Z }, 2025-12-04T09:35:33.7436368Z { 2025-12-04T09:35:33.7436828Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7437322Z "size": 724, 2025-12-04T09:35:33.7437790Z "digest": "sha256:9681563a88ff9e62494a2740e537440d3df978d466c9478d6a941fae8b57b084" 2025-12-04T09:35:33.7438312Z }, 2025-12-04T09:35:33.7438525Z { 2025-12-04T09:35:33.7438894Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7439361Z "size": 155, 2025-12-04T09:35:33.7439839Z "digest": "sha256:c47c3cfeb68763aa19727693ad52fe0c80561a98139adaa2ab5eccea35c2d1b4" 2025-12-04T09:35:33.7440460Z }, 2025-12-04T09:35:33.7440676Z { 2025-12-04T09:35:33.7441030Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7441513Z "size": 32, 2025-12-04T09:35:33.7441988Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:35:33.7442529Z }, 2025-12-04T09:35:33.7442742Z { 2025-12-04T09:35:33.7443119Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7443595Z "size": 188, 2025-12-04T09:35:33.7444064Z "digest": "sha256:7d326b9e267322de9337ac2a71ddeac4cb61f28a018a6155863f83a164ad9437" 2025-12-04T09:35:33.7444615Z }, 2025-12-04T09:35:33.7444815Z { 2025-12-04T09:35:33.7445186Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7445665Z "size": 1370, 2025-12-04T09:35:33.7446127Z "digest": "sha256:7ec8f17141c8335192fa21b660dfe1fe0ad16b202bc234e7d4ef063b35124158" 2025-12-04T09:35:33.7446675Z }, 2025-12-04T09:35:33.7446896Z { 2025-12-04T09:35:33.7447263Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7447729Z "size": 32, 2025-12-04T09:35:33.7448204Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:35:33.7448761Z }, 2025-12-04T09:35:33.7448962Z { 2025-12-04T09:35:33.7449334Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7449821Z "size": 136, 2025-12-04T09:35:33.7450278Z "digest": "sha256:26249ea175bf816b87c4c83e5efb78fd386a800fa10e819ba85b06858bcf877e" 2025-12-04T09:35:33.7450826Z }, 2025-12-04T09:35:33.7451038Z { 2025-12-04T09:35:33.7451395Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7451878Z "size": 529, 2025-12-04T09:35:33.7452345Z "digest": "sha256:5e8e9ccb36f30a8c3a7e6a5011ee5001152f36c9c749397f3e234b1822326dd0" 2025-12-04T09:35:33.7452896Z }, 2025-12-04T09:35:33.7453097Z { 2025-12-04T09:35:33.7453470Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7453950Z "size": 32, 2025-12-04T09:35:33.7454411Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:35:33.7454962Z }, 2025-12-04T09:35:33.7455176Z { 2025-12-04T09:35:33.7455532Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7456014Z "size": 104, 2025-12-04T09:35:33.7456489Z "digest": "sha256:5bc72d4e1de83a1a254e8808f727118dd54cf048c14ff298a5299e015a116bfd" 2025-12-04T09:35:33.7457114Z }, 2025-12-04T09:35:33.7457331Z { 2025-12-04T09:35:33.7457701Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7458171Z "size": 436, 2025-12-04T09:35:33.7458639Z "digest": "sha256:83cddbd497794c27254e11c4c00105d1f61399e7fef9d208a0be250724efd2c0" 2025-12-04T09:35:33.7459179Z }, 2025-12-04T09:35:33.7459391Z { 2025-12-04T09:35:33.7459752Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7460238Z "size": 32, 2025-12-04T09:35:33.7460707Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:35:33.7461249Z }, 2025-12-04T09:35:33.7461474Z { 2025-12-04T09:35:33.7461848Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7462316Z "size": 109, 2025-12-04T09:35:33.7462867Z "digest": "sha256:60c25d8c3dd2d78785f659204d0b1e64954ca581f89874b68ffe8fee23c6b661" 2025-12-04T09:35:33.7463418Z }, 2025-12-04T09:35:33.7463619Z { 2025-12-04T09:35:33.7463994Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7464476Z "size": 1896, 2025-12-04T09:35:33.7464951Z "digest": "sha256:a534dcf4b9a9e5fabed742c8a8fc43c9cfe7346ea88ab3c177c3b14fd3afe00a" 2025-12-04T09:35:33.7465514Z }, 2025-12-04T09:35:33.7465725Z { 2025-12-04T09:35:33.7466096Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7466638Z "size": 245582017, 2025-12-04T09:35:33.7467128Z "digest": "sha256:10138310c65c78d7de8375225ce37f5f7bfae7898e4e8bbcb90bd56a1bd05db4" 2025-12-04T09:35:33.7467680Z }, 2025-12-04T09:35:33.7467886Z { 2025-12-04T09:35:33.7468267Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7468753Z "size": 106, 2025-12-04T09:35:33.7469220Z "digest": "sha256:8487679f252b6fb703dc9398d73aaeec68df724bfc961579ec5bdae62ebe3a37" 2025-12-04T09:35:33.7469771Z }, 2025-12-04T09:35:33.7469986Z { 2025-12-04T09:35:33.7470341Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7470819Z "size": 162, 2025-12-04T09:35:33.7471280Z "digest": "sha256:52580ee2caa9ab69b0ac640315ee350e847cd0955c0a1eafa933a076669e87ad" 2025-12-04T09:35:33.7471820Z }, 2025-12-04T09:35:33.7472011Z { 2025-12-04T09:35:33.7472371Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7472839Z "size": 7944, 2025-12-04T09:35:33.7473311Z "digest": "sha256:741c215cb2ffb295ab6a07fab3f0dfdde029463779ff9c0bbff4add26a340cfb" 2025-12-04T09:35:33.7473862Z }, 2025-12-04T09:35:33.7474069Z { 2025-12-04T09:35:33.7474424Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7474896Z "size": 8070, 2025-12-04T09:35:33.7475354Z "digest": "sha256:d17f5aba17a608d1c7851cb3940a25d43f063385813051127074f693d0ede19b" 2025-12-04T09:35:33.7475882Z }, 2025-12-04T09:35:33.7476088Z { 2025-12-04T09:35:33.7476455Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7476926Z "size": 304, 2025-12-04T09:35:33.7477393Z "digest": "sha256:bc08246bb4ba18c3ec5bc69e16b6b4e929c5bd0f3fae10eeb0b1a622a63d6fa2" 2025-12-04T09:35:33.7477955Z }, 2025-12-04T09:35:33.7478158Z { 2025-12-04T09:35:33.7478516Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7478988Z "size": 23755574, 2025-12-04T09:35:33.7479462Z "digest": "sha256:7323bf084bf98f915db061b178c56525a0f95bd34d211b381c7527ad242c5a58" 2025-12-04T09:35:33.7479990Z }, 2025-12-04T09:35:33.7480199Z { 2025-12-04T09:35:33.7480570Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7481040Z "size": 108, 2025-12-04T09:35:33.7481520Z "digest": "sha256:d344ecc97fd77c7d12fd68ddb67aeb6cc3dd2e723de5ad1ca2c80b45c8d6bd77" 2025-12-04T09:35:33.7482081Z }, 2025-12-04T09:35:33.7482277Z { 2025-12-04T09:35:33.7482641Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7483115Z "size": 54145663, 2025-12-04T09:35:33.7483591Z "digest": "sha256:fb60b2d2147ff57c218f449f5b680132af8f7f8032ed69f422b48a3c3c1424f4" 2025-12-04T09:35:33.7484120Z }, 2025-12-04T09:35:33.7484319Z { 2025-12-04T09:35:33.7484680Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T09:35:33.7485145Z "size": 32, 2025-12-04T09:35:33.7485607Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T09:35:33.7486157Z } 2025-12-04T09:35:33.7486352Z ] 2025-12-04T09:35:33.7486555Z } 2025-12-04T09:35:33.7517173Z ##[group]Run set -eux 2025-12-04T09:35:33.7517484Z set -eux 2025-12-04T09:35:33.7517945Z # It's ok if this steps fails, it would then be an anonymous user like what we used to have 2025-12-04T09:35:33.7519447Z aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token | jq --raw-output '.SecretString' | jq -r .docker_hub_readonly_token | docker login --username pytorchbot --password-stdin || true 2025-12-04T09:35:33.7527010Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:35:33.7527433Z env: 2025-12-04T09:35:33.7527658Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:35:33.7527945Z ##[endgroup] 2025-12-04T09:35:33.7558199Z + aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token 2025-12-04T09:35:33.7559031Z + jq --raw-output .SecretString 2025-12-04T09:35:33.7560452Z + jq -r .docker_hub_readonly_token 2025-12-04T09:35:33.7561594Z + docker login --username pytorchbot --password-stdin 2025-12-04T09:35:34.4207739Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T09:35:34.4208479Z Configure a credential helper to remove this warning. See 2025-12-04T09:35:34.4209161Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T09:35:34.4209620Z 2025-12-04T09:35:34.4211251Z Login Succeeded 2025-12-04T09:35:34.4303713Z ##[group]Run tag=${ECR_DOCKER_IMAGE##*:} 2025-12-04T09:35:34.4304153Z tag=${ECR_DOCKER_IMAGE##*:} 2025-12-04T09:35:34.4304622Z echo "docker pull ghcr.io/pytorch/ci-image:${tag/:/-}" 2025-12-04T09:35:34.4311510Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:35:34.4311954Z env: 2025-12-04T09:35:34.4312207Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:35:34.4313198Z ECR_DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:35:34.4314232Z ##[endgroup] 2025-12-04T09:35:34.4344158Z docker pull ghcr.io/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:35:34.4395605Z ##[group]Run pytorch/test-infra/.github/actions/pull-docker-image@main 2025-12-04T09:35:34.4396462Z with: 2025-12-04T09:35:34.4397384Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:35:34.4398520Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:35:34.4398972Z env: 2025-12-04T09:35:34.4399199Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:35:34.4399505Z ##[endgroup] 2025-12-04T09:35:34.4416722Z ##[group]Run set -x 2025-12-04T09:35:34.4417137Z set -x 2025-12-04T09:35:34.4417386Z set +e 2025-12-04T09:35:34.4417644Z  2025-12-04T09:35:34.4417916Z login() { 2025-12-04T09:35:34.4418470Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-12-04T09:35:34.4419085Z } 2025-12-04T09:35:34.4419313Z  2025-12-04T09:35:34.4419612Z retry () { 2025-12-04T09:35:34.4419916Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-12-04T09:35:34.4420287Z } 2025-12-04T09:35:34.4420529Z  2025-12-04T09:35:34.4420797Z retry login "${DOCKER_REGISTRY}" 2025-12-04T09:35:34.4421144Z  2025-12-04T09:35:34.4421720Z IMAGE_SIZE=$(docker manifest inspect "${DOCKER_IMAGE}" | jq '[.layers[].size, .config.size] | add / 1024 / 1024') 2025-12-04T09:35:34.4422511Z echo "Compressed size of image in MB: ${IMAGE_SIZE}" 2025-12-04T09:35:34.4422940Z  2025-12-04T09:35:34.4423181Z set -e 2025-12-04T09:35:34.4423580Z # ignore output since only exit code is used for conditional 2025-12-04T09:35:34.4424174Z # only pull docker image if it's not available locally 2025-12-04T09:35:34.4424803Z if ! docker inspect --type=image "${DOCKER_IMAGE}" >/dev/null 2>/dev/null; then 2025-12-04T09:35:34.4425403Z  retry docker pull "${DOCKER_IMAGE}" 2025-12-04T09:35:34.4425778Z fi 2025-12-04T09:35:34.4432487Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:35:34.4432925Z env: 2025-12-04T09:35:34.4433172Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:35:34.4434149Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:35:34.4435274Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:35:34.4435731Z ##[endgroup] 2025-12-04T09:35:34.4462253Z + set +e 2025-12-04T09:35:34.4462859Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:35:34.4463640Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:35:34.4466103Z + aws ecr get-login-password --region us-east-1 2025-12-04T09:35:34.4467823Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T09:35:35.0577503Z WARNING! Your password will be stored unencrypted in /home/ec2-user/.docker/config.json. 2025-12-04T09:35:35.0578221Z Configure a credential helper to remove this warning. See 2025-12-04T09:35:35.0586466Z https://docs.docker.com/engine/reference/commandline/login/#credentials-store 2025-12-04T09:35:35.0586936Z 2025-12-04T09:35:35.0587043Z Login Succeeded 2025-12-04T09:35:35.0606028Z ++ docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:35:35.0607186Z ++ jq '[.layers[].size, .config.size] | add / 1024 / 1024' 2025-12-04T09:35:35.3148728Z + IMAGE_SIZE=13438.219573020935 2025-12-04T09:35:35.3149523Z + echo 'Compressed size of image in MB: 13438.219573020935' 2025-12-04T09:35:35.3150236Z + set -e 2025-12-04T09:35:35.3151808Z + docker inspect --type=image 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:35:35.3153280Z Compressed size of image in MB: 13438.219573020935 2025-12-04T09:35:35.3282994Z + retry docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:35:35.3284764Z + docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:35:35.5648110Z pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a: Pulling from pytorch/ci-image 2025-12-04T09:35:35.5650827Z 63e5bc7682b8: Pulling fs layer 2025-12-04T09:35:35.5651192Z 835841cca3b7: Pulling fs layer 2025-12-04T09:35:35.5651549Z 1bf1bb125dea: Pulling fs layer 2025-12-04T09:35:35.5651904Z b21856d1bf42: Pulling fs layer 2025-12-04T09:35:35.5652227Z 848ba2c095e2: Pulling fs layer 2025-12-04T09:35:35.5652535Z 029495b23122: Pulling fs layer 2025-12-04T09:35:35.5652870Z 073bb82063cf: Pulling fs layer 2025-12-04T09:35:35.5653191Z 59b639308833: Pulling fs layer 2025-12-04T09:35:35.5653562Z 1c6177b2970d: Pulling fs layer 2025-12-04T09:35:35.5653886Z fabe466dd5f3: Pulling fs layer 2025-12-04T09:35:35.5654212Z 2b5a11b41761: Pulling fs layer 2025-12-04T09:35:35.5654544Z 9681563a88ff: Pulling fs layer 2025-12-04T09:35:35.5654855Z dc0780902fca: Pulling fs layer 2025-12-04T09:35:35.5655179Z 5b09a2b135c8: Pulling fs layer 2025-12-04T09:35:35.5655507Z 4f4fb700ef54: Pulling fs layer 2025-12-04T09:35:35.5655822Z 5bfdaeb5578d: Pulling fs layer 2025-12-04T09:35:35.5656147Z 0ef42867f370: Pulling fs layer 2025-12-04T09:35:35.5656588Z 446083e497f3: Pulling fs layer 2025-12-04T09:35:35.5657145Z d8a170bef0f4: Pulling fs layer 2025-12-04T09:35:35.5657489Z e2b6cd6a5bd0: Pulling fs layer 2025-12-04T09:35:35.5657822Z 93efc0181a22: Pulling fs layer 2025-12-04T09:35:35.5658132Z 7454c938f174: Pulling fs layer 2025-12-04T09:35:35.5658462Z 4d57ff55f6d4: Pulling fs layer 2025-12-04T09:35:35.5658798Z b0301534b4a5: Pulling fs layer 2025-12-04T09:35:35.5659103Z 1969e15d0c13: Pulling fs layer 2025-12-04T09:35:35.5659423Z 73180a0f2d5a: Pulling fs layer 2025-12-04T09:35:35.5659746Z ad81b25cb69f: Pulling fs layer 2025-12-04T09:35:35.5660077Z 8165374f8dcc: Pulling fs layer 2025-12-04T09:35:35.5660384Z 7779c0bb9be2: Pulling fs layer 2025-12-04T09:35:35.5660832Z 4d0a1c027262: Pulling fs layer 2025-12-04T09:35:35.5661155Z a51e0dab2d59: Pulling fs layer 2025-12-04T09:35:35.5661560Z 3eb6d4ff040b: Pulling fs layer 2025-12-04T09:35:35.5661879Z b168858b8537: Pulling fs layer 2025-12-04T09:35:35.5662195Z d77a39278026: Pulling fs layer 2025-12-04T09:35:35.5662502Z 36fbd357280b: Pulling fs layer 2025-12-04T09:35:35.5662825Z 4e3b10a5dd6a: Pulling fs layer 2025-12-04T09:35:35.5663378Z 3092fab73b59: Pulling fs layer 2025-12-04T09:35:35.5663687Z 20020dd28a15: Pulling fs layer 2025-12-04T09:35:35.5664156Z ae5280ce969d: Pulling fs layer 2025-12-04T09:35:35.5664574Z 026e4484b749: Pulling fs layer 2025-12-04T09:35:35.5664904Z 1be9da2ce53d: Pulling fs layer 2025-12-04T09:35:35.5665233Z 6481b7a1d9fb: Pulling fs layer 2025-12-04T09:35:35.5665562Z fa519d18c39d: Pulling fs layer 2025-12-04T09:35:35.5665890Z d172f25b97f7: Pulling fs layer 2025-12-04T09:35:35.5666200Z fd60ab6b1c2c: Pulling fs layer 2025-12-04T09:35:35.5666524Z 0afe45579c2c: Pulling fs layer 2025-12-04T09:35:35.5666846Z 5884ffd6720b: Pulling fs layer 2025-12-04T09:35:35.5667153Z ab7a7c316fa7: Pulling fs layer 2025-12-04T09:35:35.5667514Z c7775ce5574b: Pulling fs layer 2025-12-04T09:35:35.5667833Z 81945c4fb228: Pulling fs layer 2025-12-04T09:35:35.5668141Z 663cbe24d60b: Pulling fs layer 2025-12-04T09:35:35.5668511Z 43f216b02786: Pulling fs layer 2025-12-04T09:35:35.5668833Z c47c3cfeb687: Pulling fs layer 2025-12-04T09:35:35.5669164Z 7d326b9e2673: Pulling fs layer 2025-12-04T09:35:35.5669472Z 7ec8f17141c8: Pulling fs layer 2025-12-04T09:35:35.5669784Z 4d57ff55f6d4: Waiting 2025-12-04T09:35:35.5670058Z 59b639308833: Waiting 2025-12-04T09:35:35.5670312Z 029495b23122: Waiting 2025-12-04T09:35:35.5670763Z b0301534b4a5: Waiting 2025-12-04T09:35:35.5671038Z 848ba2c095e2: Waiting 2025-12-04T09:35:35.5671313Z ab7a7c316fa7: Waiting 2025-12-04T09:35:35.5671568Z 073bb82063cf: Waiting 2025-12-04T09:35:35.5671836Z 663cbe24d60b: Waiting 2025-12-04T09:35:35.5672108Z 5bfdaeb5578d: Waiting 2025-12-04T09:35:35.5672363Z c7775ce5574b: Waiting 2025-12-04T09:35:35.5672627Z 4f4fb700ef54: Waiting 2025-12-04T09:35:35.5672898Z 026e4484b749: Waiting 2025-12-04T09:35:35.5673201Z 2b5a11b41761: Waiting 2025-12-04T09:35:35.5673493Z 26249ea175bf: Pulling fs layer 2025-12-04T09:35:35.5673807Z c47c3cfeb687: Waiting 2025-12-04T09:35:35.5674070Z d8a170bef0f4: Waiting 2025-12-04T09:35:35.5674343Z e2b6cd6a5bd0: Waiting 2025-12-04T09:35:35.5674618Z fabe466dd5f3: Waiting 2025-12-04T09:35:35.5674903Z 5e8e9ccb36f3: Pulling fs layer 2025-12-04T09:35:35.5675214Z 1969e15d0c13: Waiting 2025-12-04T09:35:35.5675480Z 1c6177b2970d: Waiting 2025-12-04T09:35:35.5675896Z 446083e497f3: Waiting 2025-12-04T09:35:35.5676168Z fd60ab6b1c2c: Waiting 2025-12-04T09:35:35.5676449Z b21856d1bf42: Waiting 2025-12-04T09:35:35.5676740Z 73180a0f2d5a: Waiting 2025-12-04T09:35:35.5677006Z d172f25b97f7: Waiting 2025-12-04T09:35:35.5677312Z 0ef42867f370: Waiting 2025-12-04T09:35:35.5677565Z 81945c4fb228: Waiting 2025-12-04T09:35:35.5677832Z 93efc0181a22: Waiting 2025-12-04T09:35:35.5678101Z 0afe45579c2c: Waiting 2025-12-04T09:35:35.5678369Z 9681563a88ff: Waiting 2025-12-04T09:35:35.5678623Z 7454c938f174: Waiting 2025-12-04T09:35:35.5678888Z 26249ea175bf: Waiting 2025-12-04T09:35:35.5679158Z ad81b25cb69f: Waiting 2025-12-04T09:35:35.5679413Z 5884ffd6720b: Waiting 2025-12-04T09:35:35.5679692Z dc0780902fca: Waiting 2025-12-04T09:35:35.5679960Z 5b09a2b135c8: Waiting 2025-12-04T09:35:35.5680216Z 20020dd28a15: Waiting 2025-12-04T09:35:35.5680514Z 5bc72d4e1de8: Pulling fs layer 2025-12-04T09:35:35.5680830Z d77a39278026: Waiting 2025-12-04T09:35:35.5681099Z 83cddbd49779: Pulling fs layer 2025-12-04T09:35:35.5681426Z 60c25d8c3dd2: Pulling fs layer 2025-12-04T09:35:35.5681746Z 3eb6d4ff040b: Waiting 2025-12-04T09:35:35.5682005Z b168858b8537: Waiting 2025-12-04T09:35:35.5682332Z 4d0a1c027262: Waiting 2025-12-04T09:35:35.5682605Z 4e3b10a5dd6a: Waiting 2025-12-04T09:35:35.5682886Z a534dcf4b9a9: Pulling fs layer 2025-12-04T09:35:35.5683203Z 7779c0bb9be2: Waiting 2025-12-04T09:35:35.5683478Z 5bc72d4e1de8: Waiting 2025-12-04T09:35:35.5683752Z 8165374f8dcc: Waiting 2025-12-04T09:35:35.5684020Z 10138310c65c: Pulling fs layer 2025-12-04T09:35:35.5684341Z 8487679f252b: Pulling fs layer 2025-12-04T09:35:35.5684718Z 52580ee2caa9: Pulling fs layer 2025-12-04T09:35:35.5685034Z 83cddbd49779: Waiting 2025-12-04T09:35:35.5685316Z 741c215cb2ff: Pulling fs layer 2025-12-04T09:35:35.5685624Z 60c25d8c3dd2: Waiting 2025-12-04T09:35:35.5685898Z d17f5aba17a6: Pulling fs layer 2025-12-04T09:35:35.5686347Z bc08246bb4ba: Pulling fs layer 2025-12-04T09:35:35.5686672Z 7323bf084bf9: Pulling fs layer 2025-12-04T09:35:35.5686965Z 36fbd357280b: Waiting 2025-12-04T09:35:35.5687293Z 8487679f252b: Waiting 2025-12-04T09:35:35.5687592Z d344ecc97fd7: Pulling fs layer 2025-12-04T09:35:35.5687891Z 3092fab73b59: Waiting 2025-12-04T09:35:35.5688173Z fb60b2d2147f: Pulling fs layer 2025-12-04T09:35:35.5688481Z 10138310c65c: Waiting 2025-12-04T09:35:35.5688803Z a534dcf4b9a9: Waiting 2025-12-04T09:35:35.5689078Z bc08246bb4ba: Waiting 2025-12-04T09:35:35.5689346Z 7323bf084bf9: Waiting 2025-12-04T09:35:35.5689603Z 741c215cb2ff: Waiting 2025-12-04T09:35:35.5689871Z 52580ee2caa9: Waiting 2025-12-04T09:35:35.5690144Z d17f5aba17a6: Waiting 2025-12-04T09:35:35.5690419Z fb60b2d2147f: Waiting 2025-12-04T09:35:35.5690677Z d344ecc97fd7: Waiting 2025-12-04T09:35:35.5690947Z ae5280ce969d: Waiting 2025-12-04T09:35:35.5691220Z 43f216b02786: Waiting 2025-12-04T09:35:35.5691472Z 6481b7a1d9fb: Waiting 2025-12-04T09:35:35.5691747Z a51e0dab2d59: Waiting 2025-12-04T09:35:35.5692020Z fa519d18c39d: Waiting 2025-12-04T09:35:35.5692274Z 7d326b9e2673: Waiting 2025-12-04T09:35:35.6581758Z 835841cca3b7: Verifying Checksum 2025-12-04T09:35:35.6582360Z 835841cca3b7: Download complete 2025-12-04T09:35:35.7377050Z b21856d1bf42: Download complete 2025-12-04T09:35:35.8133426Z 848ba2c095e2: Verifying Checksum 2025-12-04T09:35:35.8133860Z 848ba2c095e2: Download complete 2025-12-04T09:35:35.9043341Z 029495b23122: Download complete 2025-12-04T09:35:35.9422842Z 63e5bc7682b8: Verifying Checksum 2025-12-04T09:35:35.9423206Z 63e5bc7682b8: Download complete 2025-12-04T09:35:35.9848580Z 073bb82063cf: Download complete 2025-12-04T09:35:36.0203711Z 59b639308833: Download complete 2025-12-04T09:35:36.1463293Z fabe466dd5f3: Verifying Checksum 2025-12-04T09:35:36.1463700Z fabe466dd5f3: Download complete 2025-12-04T09:35:36.2215631Z 2b5a11b41761: Download complete 2025-12-04T09:35:36.3007604Z 9681563a88ff: Verifying Checksum 2025-12-04T09:35:36.3008173Z 9681563a88ff: Download complete 2025-12-04T09:35:36.3666295Z dc0780902fca: Verifying Checksum 2025-12-04T09:35:36.3666870Z dc0780902fca: Download complete 2025-12-04T09:35:36.9183877Z 63e5bc7682b8: Pull complete 2025-12-04T09:35:36.9403900Z 835841cca3b7: Pull complete 2025-12-04T09:35:37.7894115Z 1c6177b2970d: Verifying Checksum 2025-12-04T09:35:37.7894656Z 1c6177b2970d: Download complete 2025-12-04T09:35:37.7959307Z 4f4fb700ef54: Download complete 2025-12-04T09:35:37.8816385Z 5bfdaeb5578d: Verifying Checksum 2025-12-04T09:35:37.8817057Z 5bfdaeb5578d: Download complete 2025-12-04T09:35:37.9934212Z 0ef42867f370: Verifying Checksum 2025-12-04T09:35:37.9934840Z 0ef42867f370: Download complete 2025-12-04T09:35:38.0639139Z 446083e497f3: Verifying Checksum 2025-12-04T09:35:38.0639530Z 446083e497f3: Download complete 2025-12-04T09:35:38.4629934Z d8a170bef0f4: Download complete 2025-12-04T09:35:38.5779283Z e2b6cd6a5bd0: Download complete 2025-12-04T09:35:38.6543915Z 93efc0181a22: Verifying Checksum 2025-12-04T09:35:38.6544563Z 93efc0181a22: Download complete 2025-12-04T09:35:38.7401632Z 7454c938f174: Download complete 2025-12-04T09:35:38.8275858Z 4d57ff55f6d4: Download complete 2025-12-04T09:35:40.9344899Z 1bf1bb125dea: Verifying Checksum 2025-12-04T09:35:40.9360799Z 1bf1bb125dea: Download complete 2025-12-04T09:35:41.0114913Z 1969e15d0c13: Verifying Checksum 2025-12-04T09:35:41.0115600Z 1969e15d0c13: Download complete 2025-12-04T09:35:41.4242814Z 73180a0f2d5a: Verifying Checksum 2025-12-04T09:35:41.4243384Z 73180a0f2d5a: Download complete 2025-12-04T09:35:41.5268871Z ad81b25cb69f: Verifying Checksum 2025-12-04T09:35:41.5269442Z ad81b25cb69f: Download complete 2025-12-04T09:35:41.6003144Z 8165374f8dcc: Download complete 2025-12-04T09:35:49.3076589Z 7779c0bb9be2: Verifying Checksum 2025-12-04T09:35:49.3077024Z 7779c0bb9be2: Download complete 2025-12-04T09:35:49.4226776Z 4d0a1c027262: Verifying Checksum 2025-12-04T09:35:49.4227195Z 4d0a1c027262: Download complete 2025-12-04T09:35:49.4267055Z 1bf1bb125dea: Pull complete 2025-12-04T09:35:49.5119379Z a51e0dab2d59: Download complete 2025-12-04T09:35:49.5813964Z b21856d1bf42: Pull complete 2025-12-04T09:35:49.6302471Z 3eb6d4ff040b: Verifying Checksum 2025-12-04T09:35:49.6303075Z 3eb6d4ff040b: Download complete 2025-12-04T09:35:49.7283440Z b168858b8537: Verifying Checksum 2025-12-04T09:35:49.7283834Z b168858b8537: Download complete 2025-12-04T09:35:49.7979663Z 848ba2c095e2: Pull complete 2025-12-04T09:35:49.9644481Z 029495b23122: Pull complete 2025-12-04T09:35:50.1362732Z d77a39278026: Verifying Checksum 2025-12-04T09:35:50.1363155Z d77a39278026: Download complete 2025-12-04T09:35:50.1466831Z 073bb82063cf: Pull complete 2025-12-04T09:35:50.2425974Z 36fbd357280b: Verifying Checksum 2025-12-04T09:35:50.2426398Z 36fbd357280b: Download complete 2025-12-04T09:35:50.2900978Z 59b639308833: Pull complete 2025-12-04T09:35:50.3084413Z 4e3b10a5dd6a: Verifying Checksum 2025-12-04T09:35:50.3084814Z 4e3b10a5dd6a: Download complete 2025-12-04T09:35:50.3938352Z 3092fab73b59: Download complete 2025-12-04T09:35:50.4813378Z 20020dd28a15: Download complete 2025-12-04T09:35:50.5555078Z ae5280ce969d: Verifying Checksum 2025-12-04T09:35:50.5555737Z ae5280ce969d: Download complete 2025-12-04T09:35:50.6330764Z 026e4484b749: Verifying Checksum 2025-12-04T09:35:50.6331847Z 026e4484b749: Download complete 2025-12-04T09:35:50.7156252Z 1be9da2ce53d: Verifying Checksum 2025-12-04T09:35:50.7156893Z 1be9da2ce53d: Download complete 2025-12-04T09:35:50.8057839Z 6481b7a1d9fb: Verifying Checksum 2025-12-04T09:35:50.8058447Z 6481b7a1d9fb: Download complete 2025-12-04T09:35:52.9750025Z 1c6177b2970d: Pull complete 2025-12-04T09:35:53.1927236Z fabe466dd5f3: Pull complete 2025-12-04T09:35:53.4135700Z 2b5a11b41761: Pull complete 2025-12-04T09:35:53.6411698Z 9681563a88ff: Pull complete 2025-12-04T09:35:53.8491865Z dc0780902fca: Pull complete 2025-12-04T09:35:55.7692407Z fa519d18c39d: Verifying Checksum 2025-12-04T09:35:55.7692835Z fa519d18c39d: Download complete 2025-12-04T09:36:25.4928117Z 5b09a2b135c8: Verifying Checksum 2025-12-04T09:36:25.4928580Z 5b09a2b135c8: Download complete 2025-12-04T09:36:25.5957543Z fd60ab6b1c2c: Download complete 2025-12-04T09:36:25.6805196Z 0afe45579c2c: Download complete 2025-12-04T09:36:25.7800833Z 5884ffd6720b: Verifying Checksum 2025-12-04T09:36:25.7801282Z 5884ffd6720b: Download complete 2025-12-04T09:36:25.8663139Z ab7a7c316fa7: Verifying Checksum 2025-12-04T09:36:25.8663581Z ab7a7c316fa7: Download complete 2025-12-04T09:36:25.9643297Z c7775ce5574b: Verifying Checksum 2025-12-04T09:36:25.9643897Z c7775ce5574b: Download complete 2025-12-04T09:36:26.0256944Z 81945c4fb228: Verifying Checksum 2025-12-04T09:36:26.0257644Z 81945c4fb228: Download complete 2025-12-04T09:36:26.0960614Z 663cbe24d60b: Verifying Checksum 2025-12-04T09:36:26.0961231Z 663cbe24d60b: Download complete 2025-12-04T09:36:26.1866081Z 43f216b02786: Verifying Checksum 2025-12-04T09:36:26.1866763Z 43f216b02786: Download complete 2025-12-04T09:36:26.3015681Z c47c3cfeb687: Verifying Checksum 2025-12-04T09:36:26.3016132Z c47c3cfeb687: Download complete 2025-12-04T09:36:26.3990292Z 7d326b9e2673: Download complete 2025-12-04T09:36:26.4926552Z 7ec8f17141c8: Verifying Checksum 2025-12-04T09:36:26.4927145Z 7ec8f17141c8: Download complete 2025-12-04T09:36:26.5600513Z 26249ea175bf: Verifying Checksum 2025-12-04T09:36:26.5600970Z 26249ea175bf: Download complete 2025-12-04T09:36:26.6808642Z 5e8e9ccb36f3: Download complete 2025-12-04T09:36:26.7659972Z 5bc72d4e1de8: Verifying Checksum 2025-12-04T09:36:26.7660394Z 5bc72d4e1de8: Download complete 2025-12-04T09:36:26.8721333Z 83cddbd49779: Verifying Checksum 2025-12-04T09:36:26.8721948Z 83cddbd49779: Download complete 2025-12-04T09:36:26.9650075Z 60c25d8c3dd2: Download complete 2025-12-04T09:36:27.0584945Z a534dcf4b9a9: Download complete 2025-12-04T09:36:30.6447693Z 10138310c65c: Verifying Checksum 2025-12-04T09:36:30.6448110Z 10138310c65c: Download complete 2025-12-04T09:36:30.7223024Z 8487679f252b: Download complete 2025-12-04T09:36:30.8007488Z 52580ee2caa9: Verifying Checksum 2025-12-04T09:36:30.8008220Z 52580ee2caa9: Download complete 2025-12-04T09:36:30.9015005Z 741c215cb2ff: Verifying Checksum 2025-12-04T09:36:31.0234064Z d17f5aba17a6: Download complete 2025-12-04T09:36:31.1104167Z bc08246bb4ba: Download complete 2025-12-04T09:36:31.5412421Z 7323bf084bf9: Verifying Checksum 2025-12-04T09:36:31.5412863Z 7323bf084bf9: Download complete 2025-12-04T09:36:31.6401254Z d344ecc97fd7: Download complete 2025-12-04T09:36:32.6197483Z fb60b2d2147f: Verifying Checksum 2025-12-04T09:36:32.6198190Z fb60b2d2147f: Download complete 2025-12-04T09:36:46.7144944Z d172f25b97f7: Verifying Checksum 2025-12-04T09:36:46.7146405Z d172f25b97f7: Download complete 2025-12-04T09:37:17.5637200Z 5b09a2b135c8: Pull complete 2025-12-04T09:37:17.7786443Z 4f4fb700ef54: Pull complete 2025-12-04T09:37:17.9231092Z 5bfdaeb5578d: Pull complete 2025-12-04T09:37:18.1088195Z 0ef42867f370: Pull complete 2025-12-04T09:37:18.2934070Z 446083e497f3: Pull complete 2025-12-04T09:37:18.5103194Z d8a170bef0f4: Pull complete 2025-12-04T09:37:18.6938990Z e2b6cd6a5bd0: Pull complete 2025-12-04T09:37:18.8720181Z 93efc0181a22: Pull complete 2025-12-04T09:37:19.0363431Z 7454c938f174: Pull complete 2025-12-04T09:37:19.2040991Z 4d57ff55f6d4: Pull complete 2025-12-04T09:37:20.5523975Z b0301534b4a5: Verifying Checksum 2025-12-04T09:37:20.5524672Z b0301534b4a5: Download complete 2025-12-04T09:38:38.3581996Z b0301534b4a5: Pull complete 2025-12-04T09:38:38.5806086Z 1969e15d0c13: Pull complete 2025-12-04T09:38:39.3527569Z 73180a0f2d5a: Pull complete 2025-12-04T09:38:39.5772643Z ad81b25cb69f: Pull complete 2025-12-04T09:38:39.7971792Z 8165374f8dcc: Pull complete 2025-12-04T09:38:48.0473008Z 7779c0bb9be2: Pull complete 2025-12-04T09:38:48.2674956Z 4d0a1c027262: Pull complete 2025-12-04T09:38:48.4862883Z a51e0dab2d59: Pull complete 2025-12-04T09:38:48.8804798Z 3eb6d4ff040b: Pull complete 2025-12-04T09:38:49.0978424Z b168858b8537: Pull complete 2025-12-04T09:38:49.5099622Z d77a39278026: Pull complete 2025-12-04T09:38:49.6467378Z 36fbd357280b: Pull complete 2025-12-04T09:38:49.7672619Z 4e3b10a5dd6a: Pull complete 2025-12-04T09:38:50.0597710Z 3092fab73b59: Pull complete 2025-12-04T09:38:50.2766891Z 20020dd28a15: Pull complete 2025-12-04T09:38:50.4981896Z ae5280ce969d: Pull complete 2025-12-04T09:38:50.9192812Z 026e4484b749: Pull complete 2025-12-04T09:38:51.1270637Z 1be9da2ce53d: Pull complete 2025-12-04T09:38:51.5402800Z 6481b7a1d9fb: Pull complete 2025-12-04T09:38:53.3747735Z fa519d18c39d: Pull complete 2025-12-04T09:39:52.8398514Z d172f25b97f7: Pull complete 2025-12-04T09:39:53.0603335Z fd60ab6b1c2c: Pull complete 2025-12-04T09:39:53.2869681Z 0afe45579c2c: Pull complete 2025-12-04T09:39:53.6815228Z 5884ffd6720b: Pull complete 2025-12-04T09:39:54.0496245Z ab7a7c316fa7: Pull complete 2025-12-04T09:39:54.2566079Z c7775ce5574b: Pull complete 2025-12-04T09:39:54.6672278Z 81945c4fb228: Pull complete 2025-12-04T09:39:55.0753699Z 663cbe24d60b: Pull complete 2025-12-04T09:39:55.3066043Z 43f216b02786: Pull complete 2025-12-04T09:39:55.7184754Z c47c3cfeb687: Pull complete 2025-12-04T09:39:56.1152693Z 7d326b9e2673: Pull complete 2025-12-04T09:39:56.3324367Z 7ec8f17141c8: Pull complete 2025-12-04T09:39:56.7484153Z 26249ea175bf: Pull complete 2025-12-04T09:39:56.9739463Z 5e8e9ccb36f3: Pull complete 2025-12-04T09:39:57.3657019Z 5bc72d4e1de8: Pull complete 2025-12-04T09:39:57.5923793Z 83cddbd49779: Pull complete 2025-12-04T09:39:58.0202785Z 60c25d8c3dd2: Pull complete 2025-12-04T09:39:58.2322493Z a534dcf4b9a9: Pull complete 2025-12-04T09:40:05.0215497Z 10138310c65c: Pull complete 2025-12-04T09:40:05.2236696Z 8487679f252b: Pull complete 2025-12-04T09:40:05.2889933Z 52580ee2caa9: Pull complete 2025-12-04T09:40:05.4667360Z 741c215cb2ff: Pull complete 2025-12-04T09:40:05.6836455Z d17f5aba17a6: Pull complete 2025-12-04T09:40:05.8325622Z bc08246bb4ba: Pull complete 2025-12-04T09:40:07.2755852Z 7323bf084bf9: Pull complete 2025-12-04T09:40:07.4913782Z d344ecc97fd7: Pull complete 2025-12-04T09:40:09.4097830Z fb60b2d2147f: Pull complete 2025-12-04T09:40:09.7382724Z Digest: sha256:ae30f11a5b50741bd652aa0c94ad89ef791c4e50157eff642748620825cf7940 2025-12-04T09:40:09.7799339Z Status: Downloaded newer image for 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:40:09.7983059Z 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:40:09.8067415Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:40:09.8068567Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:40:09.8079023Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:40:09.8079470Z env: 2025-12-04T09:40:09.8079705Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:40:09.8080009Z ##[endgroup] 2025-12-04T09:40:09.8278367Z ##[group]Run pytorch/test-infra/.github/actions/setup-nvidia@main 2025-12-04T09:40:09.8278865Z with: 2025-12-04T09:40:09.8279108Z driver-version: 525.105.17 2025-12-04T09:40:09.8279410Z env: 2025-12-04T09:40:09.8279651Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:40:09.8279936Z ##[endgroup] 2025-12-04T09:40:09.8366010Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:40:09.8367116Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T09:40:09.8374135Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:40:09.8374575Z env: 2025-12-04T09:40:09.8374818Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:40:09.8375105Z ##[endgroup] 2025-12-04T09:40:09.8487810Z ##[group]Run set -euo pipefail 2025-12-04T09:40:09.8488197Z set -euo pipefail 2025-12-04T09:40:09.8488561Z  2025-12-04T09:40:09.8488817Z has_gpu=false 2025-12-04T09:40:09.8489106Z devices="" 2025-12-04T09:40:09.8489375Z  2025-12-04T09:40:09.8489683Z if command -v nvidia-smi >/dev/null 2>&1; then 2025-12-04T09:40:09.8490210Z  if nvidia-smi -L >/tmp/nvidia_devices 2>/dev/null; then 2025-12-04T09:40:09.8490663Z  has_gpu=true 2025-12-04T09:40:09.8491004Z  devices=$(cat /tmp/nvidia_devices) 2025-12-04T09:40:09.8491364Z  fi 2025-12-04T09:40:09.8491606Z fi 2025-12-04T09:40:09.8491839Z  2025-12-04T09:40:09.8492084Z if [ "$has_gpu" = false ]; then 2025-12-04T09:40:09.8492542Z  if ls /dev/nvidia* >/tmp/nvidia_devices 2>/dev/null; then 2025-12-04T09:40:09.8492991Z  has_gpu=true 2025-12-04T09:40:09.8493331Z  devices=$(cat /tmp/nvidia_devices) 2025-12-04T09:40:09.8493690Z  fi 2025-12-04T09:40:09.8493931Z fi 2025-12-04T09:40:09.8494184Z  2025-12-04T09:40:09.8494527Z if [ "$has_gpu" = false ] && command -v lspci >/dev/null 2>&1; then 2025-12-04T09:40:09.8495126Z  if lspci | grep -i 'nvidia' >/tmp/nvidia_devices 2>/dev/null; then 2025-12-04T09:40:09.8495615Z  has_gpu=true 2025-12-04T09:40:09.8495942Z  devices=$(cat /tmp/nvidia_devices) 2025-12-04T09:40:09.8496563Z  fi 2025-12-04T09:40:09.8496868Z fi 2025-12-04T09:40:09.8497092Z  2025-12-04T09:40:09.8497443Z printf 'HAS_NVIDIA=%s\n' "$has_gpu" >> "$GITHUB_OUTPUT" 2025-12-04T09:40:09.8498081Z printf 'DETECTED_DEVICES<> "$GITHUB_OUTPUT" 2025-12-04T09:40:09.8504963Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:40:09.8505389Z env: 2025-12-04T09:40:09.8505638Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:40:09.8505938Z ##[endgroup] 2025-12-04T09:40:11.3998653Z ##[group]Run if [ "${HAS_NVIDIA}" = "true" ]; then 2025-12-04T09:40:11.3999272Z if [ "${HAS_NVIDIA}" = "true" ]; then 2025-12-04T09:40:11.3999718Z  echo "HAS_NVIDIA_GPU=true" >> "${GITHUB_ENV}" 2025-12-04T09:40:11.4000336Z  echo "GPU_FLAG=--gpus all -e NVIDIA_DRIVER_CAPABILITIES=all" >> "${GITHUB_ENV}" 2025-12-04T09:40:11.4000891Z else 2025-12-04T09:40:11.4001205Z  echo "HAS_NVIDIA_GPU=false" >> "${GITHUB_ENV}" 2025-12-04T09:40:11.4001608Z fi 2025-12-04T09:40:11.4008449Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:40:11.4008873Z env: 2025-12-04T09:40:11.4009126Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:40:11.4009434Z HAS_NVIDIA: true 2025-12-04T09:40:11.4009697Z ##[endgroup] 2025-12-04T09:40:11.4096567Z ##[group]Run nick-fields/retry@3e91a01664abd3c5cd539100d10d33b9c5b68482 2025-12-04T09:40:11.4097132Z with: 2025-12-04T09:40:11.4097366Z timeout_minutes: 10 2025-12-04T09:40:11.4097656Z max_attempts: 3 2025-12-04T09:40:11.4130336Z command: # Is it disgusting to have a full shell script here in this github action? Sure # But is it the best way to make it so that this action relies on nothing else? Absolutely set -eou pipefail DISTRIBUTION=$(. /etc/os-release;echo $ID$VERSION_ID) DRIVER_FN="NVIDIA-Linux-x86_64-${DRIVER_VERSION}.run" install_nvidia_docker2_amzn2() { ( set -x # Needed for yum-config-manager sudo yum install -y yum-utils if [[ "${DISTRIBUTION}" == "amzn2023" ]] ; then YUM_REPO_URL="https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo" else # Amazon Linux 2 YUM_REPO_URL="https://nvidia.github.io/nvidia-docker/${DISTRIBUTION}/nvidia-docker.repo" fi sudo yum-config-manager --add-repo "${YUM_REPO_URL}" sudo yum install -y \ nvidia-container-toolkit-1.17.8 \ libnvidia-container-tools-1.17.8 \ libnvidia-container1-1.17.8 \ nvidia-container-toolkit-base-1.17.8 sudo systemctl restart docker ) } install_nvidia_docker2_ubuntu20() { ( set -x # Install nvidia-driver package if not installed status="$(dpkg-query -W --showformat='${db:Status-Status}' nvidia-docker2 2>&1)" if [ ! $? = 0 ] || [ ! "$status" = installed ]; then sudo apt-get install -y nvidia-container-toolkit-1.17.8 sudo systemctl restart docker fi ) } pre_install_nvidia_driver_amzn2() { ( # Purge any nvidia driver installed from RHEL repo sudo yum remove -y nvidia-driver-latest-dkms ) } install_nvidia_driver_common() { ( # Try to gather more information about the runner and its existing NVIDIA driver if any echo "Before installing NVIDIA driver" lspci lsmod modinfo nvidia || true HAS_NVIDIA_DRIVER=0 # Check if NVIDIA driver has already been installed if [ -x "$(command -v nvidia-smi)" ]; then set +e # The driver exists, check its version next. Also check only the first GPU if there are more than one of them # so that the same driver version is not print over multiple lines INSTALLED_DRIVER_VERSION=$(nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0) NVIDIA_SMI_STATUS=$? if [ "$NVIDIA_SMI_STATUS" -ne 0 ] && [ "$NVIDIA_SMI_STATUS" -ne 14 ]; then echo "Failed to get NVIDIA driver version ($INSTALLED_DRIVER_VERSION). Continuing" elif [ "$INSTALLED_DRIVER_VERSION" != "$DRIVER_VERSION" ]; then echo "NVIDIA driver ($INSTALLED_DRIVER_VERSION) has been installed, but we expect to have $DRIVER_VERSION instead. Continuing" # Turn off persistent mode so that the installation script can unload the kernel module sudo killall nvidia-persistenced || true else HAS_NVIDIA_DRIVER=1 echo "NVIDIA driver ($INSTALLED_DRIVER_VERSION) has already been installed. Skipping NVIDIA driver installation" fi set -e fi if [ "$HAS_NVIDIA_DRIVER" -eq 0 ]; then # CAUTION: this may need to be updated in future if [ "${DISTRIBUTION}" != ubuntu20.04 ]; then sudo yum groupinstall -y "Development Tools" # ensure our kernel install is the same as our underlying kernel, # groupinstall "Development Tools" has a habit of mismatching kernel headers sudo yum install -y "kernel-devel-uname-r == $(uname -r)" sudo modprobe backlight fi sudo curl -fsL -o /tmp/nvidia_driver "https://s3.amazonaws.com/ossci-linux/nvidia_driver/$DRIVER_FN" set +e sudo /bin/bash /tmp/nvidia_driver -s --no-drm NVIDIA_INSTALLATION_STATUS=$? RESET_GPU=0 if [ "$NVIDIA_INSTALLATION_STATUS" -ne 0 ]; then sudo cat /var/log/nvidia-installer.log # Fail to install NVIDIA driver, try to reset the GPU RESET_GPU=1 elif [ -x "$(command -v nvidia-smi)" ]; then # Check again if nvidia-smi works even if the driver installation completes successfully INSTALLED_DRIVER_VERSION=$(nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0) NVIDIA_SMI_STATUS=$? if [ "$NVIDIA_SMI_STATUS" -ne 0 ] && [ "$NVIDIA_SMI_STATUS" -ne 14 ]; then RESET_GPU=1 fi fi if [ "$RESET_GPU" -eq 1 ]; then NVIDIA_DEVICES=$(lspci -D | grep -i NVIDIA | cut -d' ' -f1) # The GPU can get stuck in a failure state if somehow the test crashs the GPU microcode. When this # happens, we'll try to reset all NVIDIA devices https://github.com/pytorch/pytorch/issues/88388 for PCI_ID in $NVIDIA_DEVICES; do DEVICE_ENABLED=$(cat /sys/bus/pci/devices/$PCI_ID/enable) echo "Reseting $PCI_ID (enabled state: $DEVICE_ENABLED)" # This requires sudo permission of course echo "1" | sudo tee /sys/bus/pci/devices/$PCI_ID/reset sleep 1 done fi sudo rm -fv /tmp/nvidia_driver set -e fi ) } post_install_nvidia_driver_common() { ( sudo modprobe nvidia || true echo "After installing NVIDIA driver" lspci lsmod modinfo nvidia || true ( set +e nvidia-smi # NB: Annoyingly, nvidia-smi command returns successfully with return code 0 even in # the case where the driver has already crashed as it still can get the driver version # and some basic information like the bus ID. However, the rest of the information # would be missing (ERR!), for example: # # +-----------------------------------------------------------------------------+ # | NVIDIA-SMI 525.89.02 Driver Version: 525.89.02 CUDA Version: 12.0 | # |-------------------------------+----------------------+----------------------+ # | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | # | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | # | | | MIG M. | # |===============================+======================+======================| # | 0 ERR! Off | 00000000:00:1E.0 Off | ERR! | # |ERR! ERR! ERR! ERR! / ERR! | 4184MiB / 23028MiB | ERR! Default | # | | | ERR! | # +-------------------------------+----------------------+----------------------+ # # +-----------------------------------------------------------------------------+ # | Processes: | # | GPU GI CI PID Type Process name GPU Memory | # | ID ID Usage | # |=============================================================================| # +-----------------------------------------------------------------------------+ # # This should be reported as a failure instead as it will guarantee to fail when # Docker tries to run with --gpus all # # So, the correct check here is to query one of the missing piece of info like # GPU name, so that the command can fail accordingly nvidia-smi --query-gpu=gpu_name --format=csv,noheader --id=0 NVIDIA_SMI_STATUS=$? # Allowable exit statuses for nvidia-smi, see: https://github.com/NVIDIA/gpu-operator/issues/285 if [ "$NVIDIA_SMI_STATUS" -eq 0 ] || [ "$NVIDIA_SMI_STATUS" -eq 14 ]; then echo "INFO: Ignoring allowed status ${NVIDIA_SMI_STATUS}" else echo "ERROR: nvidia-smi exited with unresolved status ${NVIDIA_SMI_STATUS}" exit ${NVIDIA_SMI_STATUS} fi set -e ) ) } install_nvidia_driver_amzn2() { ( set -x pre_install_nvidia_driver_amzn2 install_nvidia_driver_common post_install_nvidia_driver_common ) } install_nvidia_driver_ubuntu20() { ( set -x install_nvidia_driver_common post_install_nvidia_driver_common ) } echo "== Installing nvidia driver ${DRIVER_FN} ==" case "${DISTRIBUTION}" in amzn*) install_nvidia_driver_amzn2 ;; ubuntu20.04) install_nvidia_driver_ubuntu20 ;; *) echo "ERROR: Unknown distribution ${DISTRIBUTION}" exit 1 ;; esac # Install container toolkit based on distribution echo "== Installing nvidia container toolkit for ${DISTRIBUTION} ==" case "${DISTRIBUTION}" in amzn*) install_nvidia_docker2_amzn2 ;; ubuntu20.04) install_nvidia_docker2_ubuntu20 ;; *) echo "ERROR: Unknown distribution ${DISTRIBUTION}" exit 1 ;; esac # Fix https://github.com/NVIDIA/nvidia-docker/issues/1648 on runners with # more than one GPUs. This just needs to be run once. The command fails # on subsequent runs and complains that the mode is already on, but that's # ok sudo nvidia-persistenced || true # This should show persistence mode ON nvidia-smi # check if the container-toolkit is correctly installed and CUDA is available inside a container docker run --rm -t --gpus=all public.ecr.aws/docker/library/python:3.13 nvidia-smi 2025-12-04T09:40:11.4164019Z retry_wait_seconds: 10 2025-12-04T09:40:11.4164342Z polling_interval_seconds: 1 2025-12-04T09:40:11.4164661Z warning_on_retry: true 2025-12-04T09:40:11.4164978Z continue_on_error: false 2025-12-04T09:40:11.4165345Z env: 2025-12-04T09:40:11.4165575Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:40:11.4165879Z HAS_NVIDIA_GPU: true 2025-12-04T09:40:11.4166247Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:40:11.4166667Z DRIVER_VERSION: 525.105.17 2025-12-04T09:40:11.4166971Z ##[endgroup] 2025-12-04T09:40:11.5462809Z == Installing nvidia driver NVIDIA-Linux-x86_64-525.105.17.run == 2025-12-04T09:40:11.5463828Z + pre_install_nvidia_driver_amzn2 2025-12-04T09:40:11.5464451Z + sudo yum remove -y nvidia-driver-latest-dkms 2025-12-04T09:40:12.2830822Z No match for argument: nvidia-driver-latest-dkms 2025-12-04T09:40:12.2831325Z No packages marked for removal. 2025-12-04T09:40:12.2904582Z Dependencies resolved. 2025-12-04T09:40:12.2915821Z Nothing to do. 2025-12-04T09:40:12.2916605Z Complete! 2025-12-04T09:40:12.3974595Z + install_nvidia_driver_common 2025-12-04T09:40:12.3977958Z + echo 'Before installing NVIDIA driver' 2025-12-04T09:40:12.3978907Z + lspci 2025-12-04T09:40:12.3980790Z Before installing NVIDIA driver 2025-12-04T09:40:12.5352278Z 00:00.0 Host bridge: Intel Corporation 440FX - 82441FX PMC [Natoma] 2025-12-04T09:40:12.5352951Z 00:01.0 ISA bridge: Intel Corporation 82371SB PIIX3 ISA [Natoma/Triton II] 2025-12-04T09:40:12.5353667Z 00:01.3 Non-VGA unclassified device: Intel Corporation 82371AB/EB/MB PIIX4 ACPI (rev 08) 2025-12-04T09:40:12.5354377Z 00:03.0 VGA compatible controller: Amazon.com, Inc. Device 1111 2025-12-04T09:40:12.5354994Z 00:04.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe EBS Controller 2025-12-04T09:40:12.5355672Z 00:05.0 Ethernet controller: Amazon.com, Inc. Elastic Network Adapter (ENA) 2025-12-04T09:40:12.5356310Z 00:1e.0 3D controller: NVIDIA Corporation TU104GL [Tesla T4] (rev a1) 2025-12-04T09:40:12.5356929Z 00:1f.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-12-04T09:40:12.5357442Z + lsmod 2025-12-04T09:40:12.5397358Z Module Size Used by 2025-12-04T09:40:12.5398010Z nvidia_uvm 1925120 0 2025-12-04T09:40:12.5398353Z nvidia 14286848 1 nvidia_uvm 2025-12-04T09:40:12.5398718Z drm 602112 1 nvidia 2025-12-04T09:40:12.5399107Z drm_panel_orientation_quirks 32768 1 drm 2025-12-04T09:40:12.5399486Z backlight 24576 1 drm 2025-12-04T09:40:12.5399843Z i2c_core 110592 2 nvidia,drm 2025-12-04T09:40:12.5400208Z xt_conntrack 16384 1 2025-12-04T09:40:12.5400515Z nft_chain_nat 16384 3 2025-12-04T09:40:12.5400839Z xt_MASQUERADE 20480 1 2025-12-04T09:40:12.5401206Z nf_nat 57344 2 nft_chain_nat,xt_MASQUERADE 2025-12-04T09:40:12.5401605Z nf_conntrack_netlink 57344 0 2025-12-04T09:40:12.5402135Z nf_conntrack 184320 4 xt_conntrack,nf_nat,nf_conntrack_netlink,xt_MASQUERADE 2025-12-04T09:40:12.5402690Z nf_defrag_ipv6 24576 1 nf_conntrack 2025-12-04T09:40:12.5403173Z nf_defrag_ipv4 16384 1 nf_conntrack 2025-12-04T09:40:12.5403535Z xfrm_user 57344 1 2025-12-04T09:40:12.5403887Z xfrm_algo 16384 1 xfrm_user 2025-12-04T09:40:12.5404247Z xt_addrtype 16384 2 2025-12-04T09:40:12.5404551Z nft_compat 20480 4 2025-12-04T09:40:12.5404928Z nf_tables 311296 57 nft_compat,nft_chain_nat 2025-12-04T09:40:12.5405447Z nfnetlink 20480 4 nft_compat,nf_conntrack_netlink,nf_tables 2025-12-04T09:40:12.5405921Z br_netfilter 36864 0 2025-12-04T09:40:12.5406251Z bridge 323584 1 br_netfilter 2025-12-04T09:40:12.5406622Z stp 16384 1 bridge 2025-12-04T09:40:12.5406976Z llc 16384 2 bridge,stp 2025-12-04T09:40:12.5407316Z overlay 167936 0 2025-12-04T09:40:12.5407622Z tls 139264 0 2025-12-04T09:40:12.5407925Z nls_ascii 16384 1 2025-12-04T09:40:12.5408218Z nls_cp437 20480 1 2025-12-04T09:40:12.5408522Z vfat 24576 1 2025-12-04T09:40:12.5408834Z fat 86016 1 vfat 2025-12-04T09:40:12.5409154Z sunrpc 700416 1 2025-12-04T09:40:12.5409460Z ena 184320 0 2025-12-04T09:40:12.5409757Z i8042 45056 0 2025-12-04T09:40:12.5410134Z skx_edac_common 28672 0 2025-12-04T09:40:12.5410456Z serio 28672 3 i8042 2025-12-04T09:40:12.5410801Z ghash_clmulni_intel 16384 0 2025-12-04T09:40:12.5411166Z button 24576 0 2025-12-04T09:40:12.5411468Z sch_fq_codel 20480 17 2025-12-04T09:40:12.5411827Z dm_mod 188416 0 2025-12-04T09:40:12.5412134Z fuse 184320 1 2025-12-04T09:40:12.5412426Z loop 36864 0 2025-12-04T09:40:12.5412736Z configfs 57344 1 2025-12-04T09:40:12.5413045Z dmi_sysfs 20480 0 2025-12-04T09:40:12.5413345Z crc32_pclmul 16384 0 2025-12-04T09:40:12.5413658Z crc32c_intel 24576 0 2025-12-04T09:40:12.5413971Z efivarfs 24576 1 2025-12-04T09:40:12.5414456Z + modinfo nvidia 2025-12-04T09:40:12.5416052Z filename: /lib/modules/6.1.150-174.273.amzn2023.x86_64/kernel/drivers/video/nvidia.ko 2025-12-04T09:40:12.5416618Z import_ns: DMA_BUF 2025-12-04T09:40:12.5416988Z alias: char-major-195-* 2025-12-04T09:40:12.5417307Z version: 580.82.07 2025-12-04T09:40:12.5417613Z supported: external 2025-12-04T09:40:12.5417925Z license: Dual MIT/GPL 2025-12-04T09:40:12.5418270Z firmware: nvidia/580.82.07/gsp_tu10x.bin 2025-12-04T09:40:12.5418688Z firmware: nvidia/580.82.07/gsp_ga10x.bin 2025-12-04T09:40:12.5419089Z srcversion: BA7240A71DCF7DC6FE88C1D 2025-12-04T09:40:12.5419487Z alias: of:N*T*Cnvidia,tegra264-displayC* 2025-12-04T09:40:12.5419921Z alias: of:N*T*Cnvidia,tegra264-display 2025-12-04T09:40:12.5420354Z alias: of:N*T*Cnvidia,tegra234-displayC* 2025-12-04T09:40:12.5420775Z alias: of:N*T*Cnvidia,tegra234-display 2025-12-04T09:40:12.5421303Z alias: pci:v000010DEd*sv*sd*bc06sc80i00* 2025-12-04T09:40:12.5421732Z alias: pci:v000010DEd*sv*sd*bc03sc02i00* 2025-12-04T09:40:12.5422173Z alias: pci:v000010DEd*sv*sd*bc03sc00i00* 2025-12-04T09:40:12.5422547Z depends: i2c-core,drm 2025-12-04T09:40:12.5422867Z retpoline: Y 2025-12-04T09:40:12.5423132Z name: nvidia 2025-12-04T09:40:12.5423589Z vermagic: 6.1.150-174.273.amzn2023.x86_64 SMP preempt mod_unload modversions 2025-12-04T09:40:12.5424167Z parm: NvSwitchRegDwords:NvSwitch regkey (charp) 2025-12-04T09:40:12.5424728Z parm: NvSwitchBlacklist:NvSwitchBlacklist=uuid[,uuid...] (charp) 2025-12-04T09:40:12.5425247Z parm: NVreg_ResmanDebugLevel:int 2025-12-04T09:40:12.5425619Z parm: NVreg_RmLogonRC:int 2025-12-04T09:40:12.5425986Z parm: NVreg_ModifyDeviceFiles:int 2025-12-04T09:40:12.5426372Z parm: NVreg_DeviceFileUID:int 2025-12-04T09:40:12.5426730Z parm: NVreg_DeviceFileGID:int 2025-12-04T09:40:12.5427107Z parm: NVreg_DeviceFileMode:int 2025-12-04T09:40:12.5427556Z parm: NVreg_InitializeSystemMemoryAllocations:int 2025-12-04T09:40:12.5428032Z parm: NVreg_UsePageAttributeTable:int 2025-12-04T09:40:12.5428428Z parm: NVreg_EnablePCIeGen3:int 2025-12-04T09:40:12.5428802Z parm: NVreg_EnableMSI:int 2025-12-04T09:40:12.5429176Z parm: NVreg_EnableStreamMemOPs:int 2025-12-04T09:40:12.5429609Z parm: NVreg_RestrictProfilingToAdminUsers:int 2025-12-04T09:40:12.5430091Z parm: NVreg_PreserveVideoMemoryAllocations:int 2025-12-04T09:40:12.5430562Z parm: NVreg_EnableS0ixPowerManagement:int 2025-12-04T09:40:12.5431051Z parm: NVreg_S0ixPowerManagementVideoMemoryThreshold:int 2025-12-04T09:40:12.5431585Z parm: NVreg_DynamicPowerManagement:int 2025-12-04T09:40:12.5432099Z parm: NVreg_DynamicPowerManagementVideoMemoryThreshold:int 2025-12-04T09:40:12.5432604Z parm: NVreg_EnableGpuFirmware:int 2025-12-04T09:40:12.5433017Z parm: NVreg_EnableGpuFirmwareLogs:int 2025-12-04T09:40:12.5433474Z parm: NVreg_OpenRmEnableUnsupportedGpus:int 2025-12-04T09:40:12.5433933Z parm: NVreg_EnableUserNUMAManagement:int 2025-12-04T09:40:12.5434339Z parm: NVreg_MemoryPoolSize:int 2025-12-04T09:40:12.5434734Z parm: NVreg_KMallocHeapMaxSize:int 2025-12-04T09:40:12.5435138Z parm: NVreg_VMallocHeapMaxSize:int 2025-12-04T09:40:12.5435522Z parm: NVreg_IgnoreMMIOCheck:int 2025-12-04T09:40:12.5435906Z parm: NVreg_NvLinkDisable:int 2025-12-04T09:40:12.5436330Z parm: NVreg_EnablePCIERelaxedOrderingMode:int 2025-12-04T09:40:12.5436777Z parm: NVreg_RegisterPCIDriver:int 2025-12-04T09:40:12.5437202Z parm: NVreg_RegisterPlatformDeviceDriver:int 2025-12-04T09:40:12.5437645Z parm: NVreg_EnableResizableBar:int 2025-12-04T09:40:12.5438058Z parm: NVreg_EnableDbgBreakpoint:int 2025-12-04T09:40:12.5438479Z parm: NVreg_EnableNonblockingOpen:int 2025-12-04T09:40:12.5439013Z parm: NVreg_CoherentGPUMemoryMode:charp 2025-12-04T09:40:12.5439438Z parm: NVreg_RegistryDwords:charp 2025-12-04T09:40:12.5439846Z parm: NVreg_RegistryDwordsPerDevice:charp 2025-12-04T09:40:12.5440258Z parm: NVreg_RmMsg:charp 2025-12-04T09:40:12.5440619Z parm: NVreg_GpuBlacklist:charp 2025-12-04T09:40:12.5441024Z parm: NVreg_TemporaryFilePath:charp 2025-12-04T09:40:12.5441413Z parm: NVreg_ExcludedGpus:charp 2025-12-04T09:40:12.5441801Z parm: NVreg_DmaRemapPeerMmio:int 2025-12-04T09:40:12.5442205Z parm: NVreg_RmNvlinkBandwidth:charp 2025-12-04T09:40:12.5442631Z parm: NVreg_RmNvlinkBandwidthLinkCount:int 2025-12-04T09:40:12.5443060Z parm: NVreg_ImexChannelCount:int 2025-12-04T09:40:12.5443484Z parm: NVreg_CreateImexChannel0:int 2025-12-04T09:40:12.5443934Z parm: NVreg_GrdmaPciTopoCheckOverride:int 2025-12-04T09:40:12.5444438Z parm: rm_firmware_active:charp 2025-12-04T09:40:12.5444808Z + HAS_NVIDIA_DRIVER=0 2025-12-04T09:40:12.5445089Z ++ command -v nvidia-smi 2025-12-04T09:40:12.5445412Z + '[' -x /usr/bin/nvidia-smi ']' 2025-12-04T09:40:12.5445723Z + set +e 2025-12-04T09:40:12.5446100Z ++ nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0 2025-12-04T09:40:14.0741508Z + INSTALLED_DRIVER_VERSION=580.82.07 2025-12-04T09:40:14.0741938Z + NVIDIA_SMI_STATUS=0 2025-12-04T09:40:14.0742445Z + '[' 0 -ne 0 ']' 2025-12-04T09:40:14.0742697Z + '[' 580.82.07 '!=' 525.105.17 ']' 2025-12-04T09:40:14.0743312Z + echo 'NVIDIA driver (580.82.07) has been installed, but we expect to have 525.105.17 instead. Continuing' 2025-12-04T09:40:14.0743969Z + sudo killall nvidia-persistenced 2025-12-04T09:40:14.0744541Z NVIDIA driver (580.82.07) has been installed, but we expect to have 525.105.17 instead. Continuing 2025-12-04T09:40:14.2027938Z nvidia-persistenced: no process found 2025-12-04T09:40:14.2047437Z + true 2025-12-04T09:40:14.2047783Z + set -e 2025-12-04T09:40:14.2048019Z + '[' 0 -eq 0 ']' 2025-12-04T09:40:14.2048289Z + '[' amzn2023 '!=' ubuntu20.04 ']' 2025-12-04T09:40:14.2048696Z + sudo yum groupinstall -y 'Development Tools' 2025-12-04T09:40:14.7964604Z Last metadata expiration check: 0:22:52 ago on Thu Dec 4 09:17:22 2025. 2025-12-04T09:40:14.8409583Z No match for group package "system-rpm-config" 2025-12-04T09:40:14.8429771Z No match for group package "rcs" 2025-12-04T09:40:14.8456493Z No match for group package "pkgconfig" 2025-12-04T09:40:14.9042260Z Dependencies resolved. 2025-12-04T09:40:14.9376662Z ================================================================================ 2025-12-04T09:40:14.9377267Z Package Architecture Version Repository Size 2025-12-04T09:40:14.9377794Z ================================================================================ 2025-12-04T09:40:14.9378179Z Installing Groups: 2025-12-04T09:40:14.9378577Z Development Tools 2025-12-04T09:40:14.9378959Z 2025-12-04T09:40:14.9379064Z Transaction Summary 2025-12-04T09:40:14.9379375Z ================================================================================ 2025-12-04T09:40:14.9379644Z 2025-12-04T09:40:15.8410711Z ================================================================================ 2025-12-04T09:40:15.8411497Z WARNING: 2025-12-04T09:40:15.8411989Z A newer release of "Amazon Linux" is available. 2025-12-04T09:40:15.8412503Z 2025-12-04T09:40:15.8412712Z Available Versions: 2025-12-04T09:40:15.8413012Z 2025-12-04T09:40:15.8413194Z Version 2023.9.20250929: 2025-12-04T09:40:15.8413844Z Run the following command to upgrade to 2023.9.20250929: 2025-12-04T09:40:15.8414430Z 2025-12-04T09:40:15.8414682Z dnf upgrade --releasever=2023.9.20250929 2025-12-04T09:40:15.8415167Z 2025-12-04T09:40:15.8415345Z Release notes: 2025-12-04T09:40:15.8416289Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20250929.html 2025-12-04T09:40:15.8417587Z 2025-12-04T09:40:15.8417776Z Version 2023.9.20251014: 2025-12-04T09:40:15.8418440Z Run the following command to upgrade to 2023.9.20251014: 2025-12-04T09:40:15.8419012Z 2025-12-04T09:40:15.8419249Z dnf upgrade --releasever=2023.9.20251014 2025-12-04T09:40:15.8419733Z 2025-12-04T09:40:15.8419902Z Release notes: 2025-12-04T09:40:15.8420655Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251014.html 2025-12-04T09:40:15.8421359Z 2025-12-04T09:40:15.8421525Z Version 2023.9.20251020: 2025-12-04T09:40:15.8422092Z Run the following command to upgrade to 2023.9.20251020: 2025-12-04T09:40:15.8422610Z 2025-12-04T09:40:15.8422801Z dnf upgrade --releasever=2023.9.20251020 2025-12-04T09:40:15.8423188Z 2025-12-04T09:40:15.8423336Z Release notes: 2025-12-04T09:40:15.8424152Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251020.html 2025-12-04T09:40:15.8424912Z 2025-12-04T09:40:15.8425353Z Version 2023.9.20251027: 2025-12-04T09:40:15.8425939Z Run the following command to upgrade to 2023.9.20251027: 2025-12-04T09:40:15.8426431Z 2025-12-04T09:40:15.8426648Z dnf upgrade --releasever=2023.9.20251027 2025-12-04T09:40:15.8427038Z 2025-12-04T09:40:15.8427189Z Release notes: 2025-12-04T09:40:15.8428024Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251027.html 2025-12-04T09:40:15.8428847Z 2025-12-04T09:40:15.8429036Z Version 2023.9.20251105: 2025-12-04T09:40:15.8429612Z Run the following command to upgrade to 2023.9.20251105: 2025-12-04T09:40:15.8430192Z 2025-12-04T09:40:15.8430407Z dnf upgrade --releasever=2023.9.20251105 2025-12-04T09:40:15.8430903Z 2025-12-04T09:40:15.8431077Z Release notes: 2025-12-04T09:40:15.8431950Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251105.html 2025-12-04T09:40:15.8432821Z 2025-12-04T09:40:15.8433003Z Version 2023.9.20251110: 2025-12-04T09:40:15.8433689Z Run the following command to upgrade to 2023.9.20251110: 2025-12-04T09:40:15.8434296Z 2025-12-04T09:40:15.8434536Z dnf upgrade --releasever=2023.9.20251110 2025-12-04T09:40:15.8435005Z 2025-12-04T09:40:15.8435194Z Release notes: 2025-12-04T09:40:15.8436074Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251110.html 2025-12-04T09:40:15.8436963Z 2025-12-04T09:40:15.8437146Z Version 2023.9.20251117: 2025-12-04T09:40:15.8437820Z Run the following command to upgrade to 2023.9.20251117: 2025-12-04T09:40:15.8438407Z 2025-12-04T09:40:15.8438647Z dnf upgrade --releasever=2023.9.20251117 2025-12-04T09:40:15.8439135Z 2025-12-04T09:40:15.8439299Z Release notes: 2025-12-04T09:40:15.8440198Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251117.html 2025-12-04T09:40:15.8441068Z 2025-12-04T09:40:15.8441314Z ================================================================================ 2025-12-04T09:40:15.8441999Z Complete! 2025-12-04T09:40:15.9345406Z ++ uname -r 2025-12-04T09:40:15.9356324Z + sudo yum install -y 'kernel-devel-uname-r == 6.1.150-174.273.amzn2023.x86_64' 2025-12-04T09:40:16.4663855Z Last metadata expiration check: 0:22:54 ago on Thu Dec 4 09:17:22 2025. 2025-12-04T09:40:16.4962840Z Using '==' operator in reldeps can result in an undefined behavior. It is deprecated and the support will be dropped in future versions. Use '=' operator instead. 2025-12-04T09:40:16.5086477Z Package kernel-devel-1:6.1.150-174.273.amzn2023.x86_64 is already installed. 2025-12-04T09:40:16.5700619Z Dependencies resolved. 2025-12-04T09:40:16.6030604Z Nothing to do. 2025-12-04T09:40:16.6031244Z Complete! 2025-12-04T09:40:16.6984518Z + sudo modprobe backlight 2025-12-04T09:40:16.8682513Z + sudo curl -fsL -o /tmp/nvidia_driver https://s3.amazonaws.com/ossci-linux/nvidia_driver/NVIDIA-Linux-x86_64-525.105.17.run 2025-12-04T09:40:21.2261679Z + set +e 2025-12-04T09:40:21.2262043Z + sudo /bin/bash /tmp/nvidia_driver -s --no-drm 2025-12-04T09:40:22.7457954Z Verifying archive integrity... OK 2025-12-04T09:40:50.1374074Z Uncompressing NVIDIA Accelerated Graphics Driver for Linux-x86_64 525.105.17................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................... 2025-12-04T09:40:50.6924621Z 2025-12-04T09:40:50.6925544Z WARNING: The nvidia-drm module will not be installed. As a result, DRM-KMS will not function with this installation of the NVIDIA driver. 2025-12-04T09:40:50.6926241Z 2025-12-04T09:41:16.4054686Z 2025-12-04T09:41:16.4056587Z WARNING: nvidia-installer was forced to guess the X library path '/usr/lib64' and X module path '/usr/lib64/xorg/modules'; these paths were not queryable from the system. If X fails to find the NVIDIA X driver module, please install the `pkg-config` utility and the X.Org SDK/development package for your distribution and reinstall the driver. 2025-12-04T09:41:16.4058335Z 2025-12-04T09:41:16.4072802Z 2025-12-04T09:41:16.4074207Z WARNING: This NVIDIA driver package includes Vulkan components, but no Vulkan ICD loader was detected on this system. The NVIDIA Vulkan ICD will not function without the loader. Most distributions package the Vulkan loader; try installing the "vulkan-loader", "vulkan-icd-loader", or "libvulkan1" package. 2025-12-04T09:41:16.4075707Z 2025-12-04T09:41:27.9099514Z + NVIDIA_INSTALLATION_STATUS=0 2025-12-04T09:41:27.9100049Z + RESET_GPU=0 2025-12-04T09:41:27.9100414Z + '[' 0 -ne 0 ']' 2025-12-04T09:41:27.9101097Z ++ command -v nvidia-smi 2025-12-04T09:41:27.9105439Z + '[' -x /usr/bin/nvidia-smi ']' 2025-12-04T09:41:27.9110042Z ++ nvidia-smi --query-gpu=driver_version --format=csv,noheader --id=0 2025-12-04T09:41:30.5100134Z + INSTALLED_DRIVER_VERSION=525.105.17 2025-12-04T09:41:30.5100565Z + NVIDIA_SMI_STATUS=0 2025-12-04T09:41:30.5100855Z + '[' 0 -ne 0 ']' 2025-12-04T09:41:30.5101100Z + '[' 0 -eq 1 ']' 2025-12-04T09:41:30.5101383Z + sudo rm -fv /tmp/nvidia_driver 2025-12-04T09:41:30.6700484Z removed '/tmp/nvidia_driver' 2025-12-04T09:41:30.6718488Z + set -e 2025-12-04T09:41:30.6722270Z + post_install_nvidia_driver_common 2025-12-04T09:41:30.6725743Z + sudo modprobe nvidia 2025-12-04T09:41:30.8580349Z + echo 'After installing NVIDIA driver' 2025-12-04T09:41:30.8580782Z + lspci 2025-12-04T09:41:30.8581051Z After installing NVIDIA driver 2025-12-04T09:41:30.8716603Z 00:00.0 Host bridge: Intel Corporation 440FX - 82441FX PMC [Natoma] 2025-12-04T09:41:30.8717264Z 00:01.0 ISA bridge: Intel Corporation 82371SB PIIX3 ISA [Natoma/Triton II] 2025-12-04T09:41:30.8717973Z 00:01.3 Non-VGA unclassified device: Intel Corporation 82371AB/EB/MB PIIX4 ACPI (rev 08) 2025-12-04T09:41:30.8718734Z 00:03.0 VGA compatible controller: Amazon.com, Inc. Device 1111 2025-12-04T09:41:30.8719328Z 00:04.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe EBS Controller 2025-12-04T09:41:30.8720009Z 00:05.0 Ethernet controller: Amazon.com, Inc. Elastic Network Adapter (ENA) 2025-12-04T09:41:30.8720640Z 00:1e.0 3D controller: NVIDIA Corporation TU104GL [Tesla T4] (rev a1) 2025-12-04T09:41:30.8721252Z 00:1f.0 Non-Volatile memory controller: Amazon.com, Inc. NVMe SSD Controller 2025-12-04T09:41:30.8722132Z + lsmod 2025-12-04T09:41:30.8748765Z Module Size Used by 2025-12-04T09:41:30.8749155Z nvidia 56537088 0 2025-12-04T09:41:30.8749468Z drm 602112 1 nvidia 2025-12-04T09:41:30.8749847Z drm_panel_orientation_quirks 32768 1 drm 2025-12-04T09:41:30.8750259Z backlight 24576 1 drm 2025-12-04T09:41:30.8750613Z i2c_core 110592 2 nvidia,drm 2025-12-04T09:41:30.8750962Z xt_conntrack 16384 1 2025-12-04T09:41:30.8751283Z nft_chain_nat 16384 3 2025-12-04T09:41:30.8751606Z xt_MASQUERADE 20480 1 2025-12-04T09:41:30.8751958Z nf_nat 57344 2 nft_chain_nat,xt_MASQUERADE 2025-12-04T09:41:30.8752372Z nf_conntrack_netlink 57344 0 2025-12-04T09:41:30.8752869Z nf_conntrack 184320 4 xt_conntrack,nf_nat,nf_conntrack_netlink,xt_MASQUERADE 2025-12-04T09:41:30.8753409Z nf_defrag_ipv6 24576 1 nf_conntrack 2025-12-04T09:41:30.8754067Z nf_defrag_ipv4 16384 1 nf_conntrack 2025-12-04T09:41:30.8754459Z xfrm_user 57344 1 2025-12-04T09:41:30.8754773Z xfrm_algo 16384 1 xfrm_user 2025-12-04T09:41:30.8755136Z xt_addrtype 16384 2 2025-12-04T09:41:30.8755455Z nft_compat 20480 4 2025-12-04T09:41:30.8755833Z nf_tables 311296 57 nft_compat,nft_chain_nat 2025-12-04T09:41:30.8756335Z nfnetlink 20480 4 nft_compat,nf_conntrack_netlink,nf_tables 2025-12-04T09:41:30.8756808Z br_netfilter 36864 0 2025-12-04T09:41:30.8757148Z bridge 323584 1 br_netfilter 2025-12-04T09:41:30.8757504Z stp 16384 1 bridge 2025-12-04T09:41:30.8757852Z llc 16384 2 bridge,stp 2025-12-04T09:41:30.8758200Z overlay 167936 0 2025-12-04T09:41:30.8758494Z tls 139264 0 2025-12-04T09:41:30.8758797Z nls_ascii 16384 1 2025-12-04T09:41:30.8759100Z nls_cp437 20480 1 2025-12-04T09:41:30.8759396Z vfat 24576 1 2025-12-04T09:41:30.8759704Z fat 86016 1 vfat 2025-12-04T09:41:30.8760032Z sunrpc 700416 1 2025-12-04T09:41:30.8760340Z ena 184320 0 2025-12-04T09:41:30.8760623Z i8042 45056 0 2025-12-04T09:41:30.8760929Z skx_edac_common 28672 0 2025-12-04T09:41:30.8761241Z serio 28672 3 i8042 2025-12-04T09:41:30.8761571Z ghash_clmulni_intel 16384 0 2025-12-04T09:41:30.8761888Z button 24576 0 2025-12-04T09:41:30.8762201Z sch_fq_codel 20480 17 2025-12-04T09:41:30.8762511Z dm_mod 188416 0 2025-12-04T09:41:30.8762814Z fuse 184320 1 2025-12-04T09:41:30.8763113Z loop 36864 0 2025-12-04T09:41:30.8763406Z configfs 57344 1 2025-12-04T09:41:30.8763714Z dmi_sysfs 20480 0 2025-12-04T09:41:30.8764020Z crc32_pclmul 16384 0 2025-12-04T09:41:30.8764317Z crc32c_intel 24576 0 2025-12-04T09:41:30.8764633Z efivarfs 24576 1 2025-12-04T09:41:30.8764943Z + modinfo nvidia 2025-12-04T09:41:30.8766910Z filename: /lib/modules/6.1.150-174.273.amzn2023.x86_64/kernel/drivers/video/nvidia.ko 2025-12-04T09:41:30.8767506Z firmware: nvidia/525.105.17/gsp_tu10x.bin 2025-12-04T09:41:30.8767936Z firmware: nvidia/525.105.17/gsp_ad10x.bin 2025-12-04T09:41:30.8768336Z alias: char-major-195-* 2025-12-04T09:41:30.8768657Z version: 525.105.17 2025-12-04T09:41:30.8768961Z supported: external 2025-12-04T09:41:30.8769256Z license: NVIDIA 2025-12-04T09:41:30.8769544Z srcversion: 98F82D76E0EF3952EEE57A7 2025-12-04T09:41:30.8769937Z alias: pci:v000010DEd*sv*sd*bc06sc80i00* 2025-12-04T09:41:30.8770354Z alias: pci:v000010DEd*sv*sd*bc03sc02i00* 2025-12-04T09:41:30.8770776Z alias: pci:v000010DEd*sv*sd*bc03sc00i00* 2025-12-04T09:41:30.8771149Z depends: i2c-core,drm 2025-12-04T09:41:30.8771459Z retpoline: Y 2025-12-04T09:41:30.8771733Z name: nvidia 2025-12-04T09:41:30.8772282Z vermagic: 6.1.150-174.273.amzn2023.x86_64 SMP preempt mod_unload modversions 2025-12-04T09:41:30.8772908Z parm: NvSwitchRegDwords:NvSwitch regkey (charp) 2025-12-04T09:41:30.8773469Z parm: NvSwitchBlacklist:NvSwitchBlacklist=uuid[,uuid...] (charp) 2025-12-04T09:41:30.8773997Z parm: NVreg_ResmanDebugLevel:int 2025-12-04T09:41:30.8774372Z parm: NVreg_RmLogonRC:int 2025-12-04T09:41:30.8774745Z parm: NVreg_ModifyDeviceFiles:int 2025-12-04T09:41:30.8775140Z parm: NVreg_DeviceFileUID:int 2025-12-04T09:41:30.8775501Z parm: NVreg_DeviceFileGID:int 2025-12-04T09:41:30.8775886Z parm: NVreg_DeviceFileMode:int 2025-12-04T09:41:30.8776334Z parm: NVreg_InitializeSystemMemoryAllocations:int 2025-12-04T09:41:30.8776816Z parm: NVreg_UsePageAttributeTable:int 2025-12-04T09:41:30.8777300Z parm: NVreg_EnablePCIeGen3:int 2025-12-04T09:41:30.8777767Z parm: NVreg_EnableMSI:int 2025-12-04T09:41:30.8778133Z parm: NVreg_TCEBypassMode:int 2025-12-04T09:41:30.8778516Z parm: NVreg_EnableStreamMemOPs:int 2025-12-04T09:41:30.8778967Z parm: NVreg_RestrictProfilingToAdminUsers:int 2025-12-04T09:41:30.8779456Z parm: NVreg_PreserveVideoMemoryAllocations:int 2025-12-04T09:41:30.8779914Z parm: NVreg_EnableS0ixPowerManagement:int 2025-12-04T09:41:30.8780424Z parm: NVreg_S0ixPowerManagementVideoMemoryThreshold:int 2025-12-04T09:41:30.8780930Z parm: NVreg_DynamicPowerManagement:int 2025-12-04T09:41:30.8781447Z parm: NVreg_DynamicPowerManagementVideoMemoryThreshold:int 2025-12-04T09:41:30.8781943Z parm: NVreg_EnableGpuFirmware:int 2025-12-04T09:41:30.8782360Z parm: NVreg_EnableGpuFirmwareLogs:int 2025-12-04T09:41:30.8782814Z parm: NVreg_OpenRmEnableUnsupportedGpus:int 2025-12-04T09:41:30.8783265Z parm: NVreg_EnableUserNUMAManagement:int 2025-12-04T09:41:30.8783685Z parm: NVreg_MemoryPoolSize:int 2025-12-04T09:41:30.8784084Z parm: NVreg_KMallocHeapMaxSize:int 2025-12-04T09:41:30.8784482Z parm: NVreg_VMallocHeapMaxSize:int 2025-12-04T09:41:30.8784879Z parm: NVreg_IgnoreMMIOCheck:int 2025-12-04T09:41:30.8785263Z parm: NVreg_NvLinkDisable:int 2025-12-04T09:41:30.8785687Z parm: NVreg_EnablePCIERelaxedOrderingMode:int 2025-12-04T09:41:30.8786118Z parm: NVreg_RegisterPCIDriver:int 2025-12-04T09:41:30.8786530Z parm: NVreg_EnableDbgBreakpoint:int 2025-12-04T09:41:30.8786939Z parm: NVreg_RegistryDwords:charp 2025-12-04T09:41:30.8787348Z parm: NVreg_RegistryDwordsPerDevice:charp 2025-12-04T09:41:30.8787757Z parm: NVreg_RmMsg:charp 2025-12-04T09:41:30.8788115Z parm: NVreg_GpuBlacklist:charp 2025-12-04T09:41:30.8788501Z parm: NVreg_TemporaryFilePath:charp 2025-12-04T09:41:30.8788904Z parm: NVreg_ExcludedGpus:charp 2025-12-04T09:41:30.8789298Z parm: NVreg_DmaRemapPeerMmio:int 2025-12-04T09:41:30.8789676Z parm: rm_firmware_active:charp 2025-12-04T09:41:30.8790020Z + set +e 2025-12-04T09:41:30.8790249Z + nvidia-smi 2025-12-04T09:41:32.8551713Z Thu Dec 4 09:41:32 2025 2025-12-04T09:41:32.8552212Z +-----------------------------------------------------------------------------+ 2025-12-04T09:41:32.8552805Z | NVIDIA-SMI 525.105.17 Driver Version: 525.105.17 CUDA Version: 12.0 | 2025-12-04T09:41:32.8553395Z |-------------------------------+----------------------+----------------------+ 2025-12-04T09:41:32.8553993Z | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T09:41:32.8554643Z | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | 2025-12-04T09:41:32.8555156Z | | | MIG M. | 2025-12-04T09:41:32.8555560Z |===============================+======================+======================| 2025-12-04T09:41:32.8631018Z | 0 Tesla T4 Off | 00000000:00:1E.0 Off | 0 | 2025-12-04T09:41:32.8631551Z | N/A 26C P0 26W / 70W | 2MiB / 15360MiB | 4% Default | 2025-12-04T09:41:32.8632007Z | | | N/A | 2025-12-04T09:41:32.8632469Z +-------------------------------+----------------------+----------------------+ 2025-12-04T09:41:32.8632930Z 2025-12-04T09:41:32.8633387Z +-----------------------------------------------------------------------------+ 2025-12-04T09:41:32.8633881Z | Processes: | 2025-12-04T09:41:32.8634400Z | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T09:41:32.8634887Z | ID ID Usage | 2025-12-04T09:41:32.8635469Z |=============================================================================| 2025-12-04T09:41:32.8635992Z | No running processes found | 2025-12-04T09:41:32.8636554Z +-----------------------------------------------------------------------------+ 2025-12-04T09:41:33.3236362Z + nvidia-smi --query-gpu=gpu_name --format=csv,noheader --id=0 2025-12-04T09:41:35.2742764Z Tesla T4 2025-12-04T09:41:35.6806681Z + NVIDIA_SMI_STATUS=0 2025-12-04T09:41:35.6807045Z + '[' 0 -eq 0 ']' 2025-12-04T09:41:35.6807331Z + echo 'INFO: Ignoring allowed status 0' 2025-12-04T09:41:35.6807697Z + set -e 2025-12-04T09:41:35.6807952Z INFO: Ignoring allowed status 0 2025-12-04T09:41:35.6813695Z == Installing nvidia container toolkit for amzn2023 == 2025-12-04T09:41:35.6817645Z + sudo yum install -y yum-utils 2025-12-04T09:41:36.1662675Z Last metadata expiration check: 0:24:14 ago on Thu Dec 4 09:17:22 2025. 2025-12-04T09:41:36.1990317Z Package dnf-utils-4.3.0-13.amzn2023.0.5.noarch is already installed. 2025-12-04T09:41:36.2607499Z Dependencies resolved. 2025-12-04T09:41:36.2948630Z Nothing to do. 2025-12-04T09:41:36.2949347Z Complete! 2025-12-04T09:41:36.4029861Z + [[ amzn2023 == \a\m\z\n\2\0\2\3 ]] 2025-12-04T09:41:36.4030637Z + YUM_REPO_URL=https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-12-04T09:41:36.4031741Z + sudo yum-config-manager --add-repo https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-12-04T09:41:36.7319571Z Adding repo from: https://nvidia.github.io/libnvidia-container/stable/rpm/nvidia-container-toolkit.repo 2025-12-04T09:41:36.7870927Z + sudo yum install -y nvidia-container-toolkit-1.17.8 libnvidia-container-tools-1.17.8 libnvidia-container1-1.17.8 nvidia-container-toolkit-base-1.17.8 2025-12-04T09:41:37.3916512Z nvidia-container-toolkit 18 kB/s | 833 B 00:00 2025-12-04T09:41:37.4944770Z Dependencies resolved. 2025-12-04T09:41:37.5279633Z ================================================================================ 2025-12-04T09:41:37.5280188Z Package Arch Version Repository Size 2025-12-04T09:41:37.5280652Z ================================================================================ 2025-12-04T09:41:37.5281031Z Downgrading: 2025-12-04T09:41:37.5281493Z libnvidia-container-tools x86_64 1.17.8-1 nvidia-container-toolkit 40 k 2025-12-04T09:41:37.5282198Z libnvidia-container1 x86_64 1.17.8-1 nvidia-container-toolkit 1.0 M 2025-12-04T09:41:37.5282901Z nvidia-container-toolkit x86_64 1.17.8-1 nvidia-container-toolkit 1.2 M 2025-12-04T09:41:37.5283646Z nvidia-container-toolkit-base x86_64 1.17.8-1 nvidia-container-toolkit 5.8 M 2025-12-04T09:41:37.5284095Z 2025-12-04T09:41:37.5284216Z Transaction Summary 2025-12-04T09:41:37.5284507Z ================================================================================ 2025-12-04T09:41:37.5284891Z Downgrade 4 Packages 2025-12-04T09:41:37.5285069Z 2025-12-04T09:41:37.5285382Z Total download size: 8.0 M 2025-12-04T09:41:37.5287280Z Downloading Packages: 2025-12-04T09:41:37.5764590Z (1/4): libnvidia-container-tools-1.17.8-1.x86_6 879 kB/s | 40 kB 00:00 2025-12-04T09:41:37.6349956Z (2/4): libnvidia-container1-1.17.8-1.x86_64.rpm 9.4 MB/s | 1.0 MB 00:00 2025-12-04T09:41:37.6827375Z (3/4): nvidia-container-toolkit-1.17.8-1.x86_64 8.1 MB/s | 1.2 MB 00:00 2025-12-04T09:41:37.8288081Z (4/4): nvidia-container-toolkit-base-1.17.8-1.x 23 MB/s | 5.8 MB 00:00 2025-12-04T09:41:37.8301253Z -------------------------------------------------------------------------------- 2025-12-04T09:41:37.8305705Z Total 27 MB/s | 8.0 MB 00:00 2025-12-04T09:41:37.8308842Z Running transaction check 2025-12-04T09:41:37.8465280Z Transaction check succeeded. 2025-12-04T09:41:37.8465661Z Running transaction test 2025-12-04T09:41:37.9025276Z Transaction test succeeded. 2025-12-04T09:41:37.9027858Z Running transaction 2025-12-04T09:41:38.9297316Z Preparing : 1/1 2025-12-04T09:41:39.0809840Z Downgrading : nvidia-container-toolkit-base-1.17.8-1.x86_64 1/8 2025-12-04T09:41:39.1116670Z Downgrading : libnvidia-container1-1.17.8-1.x86_64 2/8 2025-12-04T09:41:39.1945431Z Running scriptlet: libnvidia-container1-1.17.8-1.x86_64 2/8 2025-12-04T09:41:39.3529744Z Downgrading : libnvidia-container-tools-1.17.8-1.x86_64 3/8 2025-12-04T09:41:39.3830713Z Downgrading : nvidia-container-toolkit-1.17.8-1.x86_64 4/8 2025-12-04T09:41:39.4734010Z Running scriptlet: nvidia-container-toolkit-1.17.8-1.x86_64 4/8 2025-12-04T09:41:39.4800894Z Running scriptlet: nvidia-container-toolkit-1.18.1-1.x86_64 5/8 2025-12-04T09:41:39.4802031Z Cleanup : nvidia-container-toolkit-1.18.1-1.x86_64 5/8 2025-12-04T09:41:39.5164884Z Running scriptlet: nvidia-container-toolkit-1.18.1-1.x86_64 5/8 2025-12-04T09:41:39.5219693Z Running scriptlet: libnvidia-container-tools-1.18.1-1.x86_64 6/8 2025-12-04T09:41:39.5220689Z Cleanup : libnvidia-container-tools-1.18.1-1.x86_64 6/8 2025-12-04T09:41:39.5599030Z Running scriptlet: libnvidia-container-tools-1.18.1-1.x86_64 6/8 2025-12-04T09:41:39.5653022Z Running scriptlet: libnvidia-container1-1.18.1-1.x86_64 7/8 2025-12-04T09:41:39.5654120Z Cleanup : libnvidia-container1-1.18.1-1.x86_64 7/8 2025-12-04T09:41:39.6048098Z Running scriptlet: libnvidia-container1-1.18.1-1.x86_64 7/8 2025-12-04T09:41:39.6104912Z Running scriptlet: nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8 2025-12-04T09:41:39.6106241Z Cleanup : nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8 2025-12-04T09:41:39.6460215Z Running scriptlet: nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8 2025-12-04T09:41:39.7069294Z Running scriptlet: nvidia-container-toolkit-1.17.8-1.x86_64 8/8 2025-12-04T09:41:41.0368308Z Running scriptlet: nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8 2025-12-04T09:41:41.0369093Z Verifying : libnvidia-container-tools-1.17.8-1.x86_64 1/8 2025-12-04T09:41:41.0369771Z Verifying : libnvidia-container-tools-1.18.1-1.x86_64 2/8 2025-12-04T09:41:41.0370438Z Verifying : libnvidia-container1-1.17.8-1.x86_64 3/8 2025-12-04T09:41:41.0371094Z Verifying : libnvidia-container1-1.18.1-1.x86_64 4/8 2025-12-04T09:41:41.0371759Z Verifying : nvidia-container-toolkit-1.17.8-1.x86_64 5/8 2025-12-04T09:41:41.0372415Z Verifying : nvidia-container-toolkit-1.18.1-1.x86_64 6/8 2025-12-04T09:41:41.0373080Z Verifying : nvidia-container-toolkit-base-1.17.8-1.x86_64 7/8 2025-12-04T09:41:41.2039853Z Verifying : nvidia-container-toolkit-base-1.18.1-1.x86_64 8/8================================================================================ 2025-12-04T09:41:41.2041048Z WARNING: 2025-12-04T09:41:41.2041359Z A newer release of "Amazon Linux" is available. 2025-12-04T09:41:41.2041780Z 2025-12-04T09:41:41.2041903Z Available Versions: 2025-12-04T09:41:41.2042084Z 2025-12-04T09:41:41.2042190Z Version 2023.9.20250929: 2025-12-04T09:41:41.2042576Z Run the following command to upgrade to 2023.9.20250929: 2025-12-04T09:41:41.2042895Z 2025-12-04T09:41:41.2043052Z dnf upgrade --releasever=2023.9.20250929 2025-12-04T09:41:41.2043315Z 2025-12-04T09:41:41.2043415Z Release notes: 2025-12-04T09:41:41.2043928Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20250929.html 2025-12-04T09:41:41.2044413Z 2025-12-04T09:41:41.2044516Z Version 2023.9.20251014: 2025-12-04T09:41:41.2045123Z Run the following command to upgrade to 2023.9.20251014: 2025-12-04T09:41:41.2045460Z 2025-12-04T09:41:41.2045599Z dnf upgrade --releasever=2023.9.20251014 2025-12-04T09:41:41.2045876Z 2025-12-04T09:41:41.2045977Z Release notes: 2025-12-04T09:41:41.2046475Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251014.html 2025-12-04T09:41:41.2046949Z 2025-12-04T09:41:41.2047069Z Version 2023.9.20251020: 2025-12-04T09:41:41.2047435Z Run the following command to upgrade to 2023.9.20251020: 2025-12-04T09:41:41.2047762Z 2025-12-04T09:41:41.2047895Z dnf upgrade --releasever=2023.9.20251020 2025-12-04T09:41:41.2048153Z 2025-12-04T09:41:41.2048264Z Release notes: 2025-12-04T09:41:41.2048744Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251020.html 2025-12-04T09:41:41.2049223Z 2025-12-04T09:41:41.2049327Z Version 2023.9.20251027: 2025-12-04T09:41:41.2049704Z Run the following command to upgrade to 2023.9.20251027: 2025-12-04T09:41:41.2050019Z 2025-12-04T09:41:41.2050172Z dnf upgrade --releasever=2023.9.20251027 2025-12-04T09:41:41.2050435Z 2025-12-04T09:41:41.2050536Z Release notes: 2025-12-04T09:41:41.2051024Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251027.html 2025-12-04T09:41:41.2051489Z 2025-12-04T09:41:41.2051602Z Version 2023.9.20251105: 2025-12-04T09:41:41.2051966Z Run the following command to upgrade to 2023.9.20251105: 2025-12-04T09:41:41.2052290Z 2025-12-04T09:41:41.2052423Z dnf upgrade --releasever=2023.9.20251105 2025-12-04T09:41:41.2052694Z 2025-12-04T09:41:41.2052795Z Release notes: 2025-12-04T09:41:41.2053289Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251105.html 2025-12-04T09:41:41.2053755Z 2025-12-04T09:41:41.2053857Z Version 2023.9.20251110: 2025-12-04T09:41:41.2054232Z Run the following command to upgrade to 2023.9.20251110: 2025-12-04T09:41:41.2054544Z 2025-12-04T09:41:41.2054689Z dnf upgrade --releasever=2023.9.20251110 2025-12-04T09:41:41.2054965Z 2025-12-04T09:41:41.2055076Z Release notes: 2025-12-04T09:41:41.2055554Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251110.html 2025-12-04T09:41:41.2056033Z 2025-12-04T09:41:41.2056154Z Version 2023.9.20251117: 2025-12-04T09:41:41.2056517Z Run the following command to upgrade to 2023.9.20251117: 2025-12-04T09:41:41.2056924Z 2025-12-04T09:41:41.2057060Z dnf upgrade --releasever=2023.9.20251117 2025-12-04T09:41:41.2057340Z 2025-12-04T09:41:41.2057439Z Release notes: 2025-12-04T09:41:41.2057938Z https://docs.aws.amazon.com/linux/al2023/release-notes/relnotes-2023.9.20251117.html 2025-12-04T09:41:41.2058405Z 2025-12-04T09:41:41.2058545Z ================================================================================ 2025-12-04T09:41:41.2720399Z 2025-12-04T09:41:41.2720563Z 2025-12-04T09:41:41.2720662Z Downgraded: 2025-12-04T09:41:41.2721117Z libnvidia-container-tools-1.17.8-1.x86_64 2025-12-04T09:41:41.2722052Z libnvidia-container1-1.17.8-1.x86_64 2025-12-04T09:41:41.2722749Z nvidia-container-toolkit-1.17.8-1.x86_64 2025-12-04T09:41:41.2723478Z nvidia-container-toolkit-base-1.17.8-1.x86_64 2025-12-04T09:41:41.2723914Z 2025-12-04T09:41:41.2724025Z Complete! 2025-12-04T09:41:41.3281612Z + sudo systemctl restart docker 2025-12-04T09:41:47.3513285Z Thu Dec 4 09:41:47 2025 2025-12-04T09:41:47.3513756Z +-----------------------------------------------------------------------------+ 2025-12-04T09:41:47.3514366Z | NVIDIA-SMI 525.105.17 Driver Version: 525.105.17 CUDA Version: 12.0 | 2025-12-04T09:41:47.3514950Z |-------------------------------+----------------------+----------------------+ 2025-12-04T09:41:47.3515551Z | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T09:41:47.3516464Z | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | 2025-12-04T09:41:47.3517014Z | | | MIG M. | 2025-12-04T09:41:47.3517432Z |===============================+======================+======================| 2025-12-04T09:41:47.3612773Z | 0 Tesla T4 On | 00000000:00:1E.0 Off | 0 | 2025-12-04T09:41:47.3613286Z | N/A 26C P0 26W / 70W | 2MiB / 15360MiB | 8% Default | 2025-12-04T09:41:47.3613809Z | | | N/A | 2025-12-04T09:41:47.3614679Z +-------------------------------+----------------------+----------------------+ 2025-12-04T09:41:47.3615134Z 2025-12-04T09:41:47.3615593Z +-----------------------------------------------------------------------------+ 2025-12-04T09:41:47.3616128Z | Processes: | 2025-12-04T09:41:47.3617165Z | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T09:41:47.3617639Z | ID ID Usage | 2025-12-04T09:41:47.3618057Z |=============================================================================| 2025-12-04T09:41:47.3618577Z | No running processes found | 2025-12-04T09:41:47.3619134Z +-----------------------------------------------------------------------------+ 2025-12-04T09:41:47.4354478Z Unable to find image 'public.ecr.aws/docker/library/python:3.13' locally 2025-12-04T09:41:47.6437476Z 3.13: Pulling from docker/library/python 2025-12-04T09:41:47.7505297Z 53c88f1dfeb7: Pulling fs layer 2025-12-04T09:41:47.7505752Z eae668646f44: Pulling fs layer 2025-12-04T09:41:47.7506177Z ff2e6e687b6c: Pulling fs layer 2025-12-04T09:41:47.7506510Z 7c40a3faff76: Pulling fs layer 2025-12-04T09:41:47.7506840Z 967a3b1c8fef: Pulling fs layer 2025-12-04T09:41:47.7507189Z a64e1a44f22a: Pulling fs layer 2025-12-04T09:41:47.7507626Z 52655f8a5bcc: Pulling fs layer 2025-12-04T09:41:47.7508049Z 967a3b1c8fef: Waiting 2025-12-04T09:41:47.7508326Z 52655f8a5bcc: Waiting 2025-12-04T09:41:47.7508587Z 7c40a3faff76: Waiting 2025-12-04T09:41:47.9100116Z eae668646f44: Verifying Checksum 2025-12-04T09:41:47.9100552Z eae668646f44: Download complete 2025-12-04T09:41:48.0232883Z 53c88f1dfeb7: Verifying Checksum 2025-12-04T09:41:48.0233493Z 53c88f1dfeb7: Download complete 2025-12-04T09:41:48.1140058Z 967a3b1c8fef: Verifying Checksum 2025-12-04T09:41:48.1140530Z 967a3b1c8fef: Download complete 2025-12-04T09:41:48.1195973Z ff2e6e687b6c: Verifying Checksum 2025-12-04T09:41:48.1196960Z ff2e6e687b6c: Download complete 2025-12-04T09:41:48.1696580Z 52655f8a5bcc: Verifying Checksum 2025-12-04T09:41:48.1697031Z 52655f8a5bcc: Download complete 2025-12-04T09:41:48.2935290Z a64e1a44f22a: Verifying Checksum 2025-12-04T09:41:48.2935719Z a64e1a44f22a: Download complete 2025-12-04T09:41:49.1102757Z 7c40a3faff76: Verifying Checksum 2025-12-04T09:41:49.1103435Z 7c40a3faff76: Download complete 2025-12-04T09:41:49.4929936Z 53c88f1dfeb7: Pull complete 2025-12-04T09:41:50.0921839Z eae668646f44: Pull complete 2025-12-04T09:41:52.0793900Z ff2e6e687b6c: Pull complete 2025-12-04T09:41:57.8373321Z 7c40a3faff76: Pull complete 2025-12-04T09:41:58.0655694Z 967a3b1c8fef: Pull complete 2025-12-04T09:41:58.7095422Z a64e1a44f22a: Pull complete 2025-12-04T09:41:58.7310882Z 52655f8a5bcc: Pull complete 2025-12-04T09:41:58.7443920Z Digest: sha256:3f986299a7b8b44b0d8cf9bda2b22361ce5c3058ef5d7cb17fb7452506680ab0 2025-12-04T09:41:58.7484439Z Status: Downloaded newer image for public.ecr.aws/docker/library/python:3.13 2025-12-04T09:42:06.0874028Z Thu Dec 4 09:42:06 2025 2025-12-04T09:42:06.0874529Z +-----------------------------------------------------------------------------+ 2025-12-04T09:42:06.0875124Z | NVIDIA-SMI 525.105.17 Driver Version: 525.105.17 CUDA Version: 12.0 | 2025-12-04T09:42:06.0875952Z |-------------------------------+----------------------+----------------------+ 2025-12-04T09:42:06.0876573Z | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T09:42:06.0877227Z | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | 2025-12-04T09:42:06.0877738Z | | | MIG M. | 2025-12-04T09:42:06.0878141Z |===============================+======================+======================| 2025-12-04T09:42:06.1032933Z | 0 Tesla T4 On | 00000000:00:1E.0 Off | 0 | 2025-12-04T09:42:06.1033484Z | N/A 25C P8 9W / 70W | 2MiB / 15360MiB | 0% Default | 2025-12-04T09:42:06.1033966Z | | | N/A | 2025-12-04T09:42:06.1034436Z +-------------------------------+----------------------+----------------------+ 2025-12-04T09:42:06.1034912Z 2025-12-04T09:42:06.1035373Z +-----------------------------------------------------------------------------+ 2025-12-04T09:42:06.1035882Z | Processes: | 2025-12-04T09:42:06.1036587Z | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T09:42:06.1037072Z | ID ID Usage | 2025-12-04T09:42:06.1037472Z |=============================================================================| 2025-12-04T09:42:06.1037983Z | No running processes found | 2025-12-04T09:42:06.1038547Z +-----------------------------------------------------------------------------+ 2025-12-04T09:42:07.5676868Z Command completed after 1 attempt(s). 2025-12-04T09:42:07.5778280Z Prepare all required actions 2025-12-04T09:42:07.5812469Z ##[group]Run ./.github/actions/get-workflow-job-id 2025-12-04T09:42:07.5812862Z with: 2025-12-04T09:42:07.5813558Z github-token: *** 2025-12-04T09:42:07.5813829Z env: 2025-12-04T09:42:07.5814058Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:07.5814370Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:07.5814740Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:07.5815159Z ##[endgroup] 2025-12-04T09:42:07.5831399Z ##[group]Run set -eux 2025-12-04T09:42:07.5831696Z set -eux 2025-12-04T09:42:07.5832218Z python3 .github/scripts/get_workflow_job_id.py "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-12-04T09:42:07.5844124Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:42:07.5844552Z env: 2025-12-04T09:42:07.5844800Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:07.5845113Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:07.5845523Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:07.5846140Z GITHUB_TOKEN: *** 2025-12-04T09:42:07.5862629Z ##[endgroup] 2025-12-04T09:42:07.5901330Z + python3 .github/scripts/get_workflow_job_id.py 19922826259 i-092818c7270e5db43 2025-12-04T09:42:09.7455875Z Setting output job-id=57119749282 2025-12-04T09:42:09.7456826Z Setting output job-name=linux-jammy-cuda12.4-py3.10-gcc11 / test (legacy_nvidia_driver, 3, 5, linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check, unstable) 2025-12-04T09:42:09.7588354Z ##[group]Run python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-12-04T09:42:09.7589231Z python3 -m pip install psutil==5.9.8 dataclasses_json==0.6.7 nvidia-ml-py==11.525.84 2025-12-04T09:42:09.7590372Z python3 -m tools.stats.monitor --log-interval "$MONITOR_LOG_INTERVAL" --data-collect-interval "$MONITOR_DATA_COLLECT_INTERVAL" > usage_log.txt 2>&1 & 2025-12-04T09:42:09.7591386Z echo "monitor-script-pid=${!}" >> "${GITHUB_OUTPUT}" 2025-12-04T09:42:09.7599654Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:42:09.7600083Z env: 2025-12-04T09:42:09.7600335Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:09.7600649Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:09.7601025Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:09.7601442Z JOB_ID: 57119749282 2025-12-04T09:42:09.7602212Z JOB_NAME: linux-jammy-cuda12.4-py3.10-gcc11 / test (legacy_nvidia_driver, 3, 5, linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check, unstable) 2025-12-04T09:42:09.7602995Z WORKFLOW_NAME: periodic 2025-12-04T09:42:09.7603309Z WORKFLOW_RUN_ID: 19922826259 2025-12-04T09:42:09.7603631Z MONITOR_LOG_INTERVAL: 5 2025-12-04T09:42:09.7603936Z MONITOR_DATA_COLLECT_INTERVAL: 1 2025-12-04T09:42:09.7604275Z ##[endgroup] 2025-12-04T09:42:10.0770313Z Defaulting to user installation because normal site-packages is not writeable 2025-12-04T09:42:10.4902066Z Collecting psutil==5.9.8 2025-12-04T09:42:10.5065951Z Downloading psutil-5.9.8-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (288 kB) 2025-12-04T09:42:10.5878204Z Collecting dataclasses_json==0.6.7 2025-12-04T09:42:10.5909777Z Downloading dataclasses_json-0.6.7-py3-none-any.whl (28 kB) 2025-12-04T09:42:10.6219718Z Collecting nvidia-ml-py==11.525.84 2025-12-04T09:42:10.6252503Z Downloading nvidia_ml_py-11.525.84-py3-none-any.whl (34 kB) 2025-12-04T09:42:10.7590248Z Collecting marshmallow<4.0.0,>=3.18.0 2025-12-04T09:42:10.7625705Z Downloading marshmallow-3.26.1-py3-none-any.whl (50 kB) 2025-12-04T09:42:10.7876987Z Collecting typing-inspect<1,>=0.4.0 2025-12-04T09:42:10.7909483Z Downloading typing_inspect-0.9.0-py3-none-any.whl (8.8 kB) 2025-12-04T09:42:10.8526477Z Collecting packaging>=17.0 2025-12-04T09:42:10.8562304Z Downloading packaging-25.0-py3-none-any.whl (66 kB) 2025-12-04T09:42:10.8822790Z Collecting mypy-extensions>=0.3.0 2025-12-04T09:42:10.8856101Z Downloading mypy_extensions-1.1.0-py3-none-any.whl (5.0 kB) 2025-12-04T09:42:10.9394124Z Collecting typing-extensions>=3.7.4 2025-12-04T09:42:10.9428952Z Downloading typing_extensions-4.15.0-py3-none-any.whl (44 kB) 2025-12-04T09:42:11.0441737Z Installing collected packages: typing-extensions, packaging, mypy-extensions, typing-inspect, marshmallow, psutil, nvidia-ml-py, dataclasses-json 2025-12-04T09:42:11.3595357Z Successfully installed dataclasses-json-0.6.7 marshmallow-3.26.1 mypy-extensions-1.1.0 nvidia-ml-py-11.525.84 packaging-25.0 psutil-5.9.8 typing-extensions-4.15.0 typing-inspect-0.9.0 2025-12-04T09:42:11.5637607Z Prepare all required actions 2025-12-04T09:42:11.5638098Z Getting action download info 2025-12-04T09:42:11.7207937Z Download action repository 'seemethere/download-artifact-s3@v4' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-12-04T09:42:11.9635051Z Download action repository 'actions/download-artifact@v4' (SHA:d3f86a106a0bac45b974a628896c90dbdf5c8093) 2025-12-04T09:42:12.3787327Z ##[group]Run ./.github/actions/download-build-artifacts 2025-12-04T09:42:12.3787764Z with: 2025-12-04T09:42:12.3788048Z name: linux-jammy-cuda12.4-py3.10-gcc11 2025-12-04T09:42:12.3788427Z s3-bucket: gha-artifacts 2025-12-04T09:42:12.3788730Z env: 2025-12-04T09:42:12.3789147Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:12.3789441Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:12.3789813Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:12.3790234Z ##[endgroup] 2025-12-04T09:42:12.3824346Z ##[group]Run seemethere/download-artifact-s3@v4 2025-12-04T09:42:12.3824751Z with: 2025-12-04T09:42:12.3825088Z name: linux-jammy-cuda12.4-py3.10-gcc11 2025-12-04T09:42:12.3825478Z s3-bucket: gha-artifacts 2025-12-04T09:42:12.3825788Z region: us-east-1 2025-12-04T09:42:12.3826034Z env: 2025-12-04T09:42:12.3826277Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:12.3826589Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:12.3826959Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:12.3827361Z ##[endgroup] 2025-12-04T09:42:12.9119110Z (node:68794) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-12-04T09:42:12.9119719Z 2025-12-04T09:42:12.9119989Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-12-04T09:42:12.9120674Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-12-04T09:42:12.9121357Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-12-04T09:42:13.1799712Z Found 1 objects with prefix pytorch/pytorch/19922826259/linux-jammy-cuda12.4-py3.10-gcc11/ 2025-12-04T09:42:13.1800594Z Starting download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-12-04T09:42:19.9142872Z Finished download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/artifacts.zip 2025-12-04T09:42:19.9149914Z Artifact download has finished successfully 2025-12-04T09:42:19.9357875Z ##[group]Run unzip -o artifacts.zip 2025-12-04T09:42:19.9358261Z unzip -o artifacts.zip 2025-12-04T09:42:19.9366029Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:42:19.9366471Z env: 2025-12-04T09:42:19.9366704Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:19.9367014Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:19.9367396Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:19.9367799Z ##[endgroup] 2025-12-04T09:42:19.9431968Z Archive: artifacts.zip 2025-12-04T09:42:19.9433356Z creating: dist/ 2025-12-04T09:42:21.9830735Z inflating: dist/torch-2.10.0a0+gitffd9b0f-cp310-cp310-linux_x86_64.whl 2025-12-04T09:42:21.9978648Z inflating: dist/.ninja_log 2025-12-04T09:42:21.9979154Z creating: build/custom_test_artifacts/ 2025-12-04T09:42:21.9979953Z creating: build/custom_test_artifacts/custom-op-build/ 2025-12-04T09:42:21.9980587Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/ 2025-12-04T09:42:21.9981298Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/pkgRedirects/ 2025-12-04T09:42:21.9989204Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeConfigureLog.yaml 2025-12-04T09:42:21.9990029Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/ 2025-12-04T09:42:21.9990831Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-12-04T09:42:21.9991711Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-12-04T09:42:21.9992536Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-12-04T09:42:21.9993998Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-12-04T09:42:21.9995295Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-12-04T09:42:21.9996613Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-12-04T09:42:21.9997512Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-12-04T09:42:21.9998372Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-12-04T09:42:21.9999821Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-12-04T09:42:22.0001357Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-12-04T09:42:22.0002440Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-12-04T09:42:22.0004280Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-12-04T09:42:22.0006247Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-12-04T09:42:22.0007227Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-12-04T09:42:22.0008091Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-12-04T09:42:22.0071873Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-12-04T09:42:22.0139095Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-12-04T09:42:22.0140397Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-12-04T09:42:22.0210074Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-12-04T09:42:22.0211331Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-12-04T09:42:22.0212586Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-12-04T09:42:22.0213894Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-12-04T09:42:22.0215148Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-12-04T09:42:22.0216385Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-12-04T09:42:22.0217736Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-12-04T09:42:22.0218966Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-12-04T09:42:22.0220264Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-12-04T09:42:22.0221408Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-12-04T09:42:22.0222504Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-12-04T09:42:22.0223573Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-12-04T09:42:22.0224660Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-12-04T09:42:22.0225708Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-12-04T09:42:22.0226976Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-12-04T09:42:22.0309576Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-12-04T09:42:22.0310548Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-12-04T09:42:22.0395907Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-12-04T09:42:22.0396964Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeScratch/ 2025-12-04T09:42:22.0397688Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeTmp/ 2025-12-04T09:42:22.0398646Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/cmake.check_cache 2025-12-04T09:42:22.0399451Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/ 2025-12-04T09:42:22.0400340Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.ts 2025-12-04T09:42:22.0401345Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.make 2025-12-04T09:42:22.0402317Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/depend.make 2025-12-04T09:42:22.0403217Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/link.txt 2025-12-04T09:42:22.0404148Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/cmake_clean.cmake 2025-12-04T09:42:22.0405080Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/build.make 2025-12-04T09:42:22.0406020Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/DependInfo.cmake 2025-12-04T09:42:22.0406955Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/flags.make 2025-12-04T09:42:22.0407885Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/progress.make 2025-12-04T09:42:22.0426444Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o.d 2025-12-04T09:42:22.0653570Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o 2025-12-04T09:42:22.0654438Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/ 2025-12-04T09:42:22.0655392Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.ts 2025-12-04T09:42:22.0656450Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.make 2025-12-04T09:42:22.0657547Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/depend.make 2025-12-04T09:42:22.0658500Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/link.txt 2025-12-04T09:42:22.0659665Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/cmake_clean.cmake 2025-12-04T09:42:22.0660655Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/build.make 2025-12-04T09:42:22.0661646Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/DependInfo.cmake 2025-12-04T09:42:22.0662636Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/flags.make 2025-12-04T09:42:22.0663611Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/progress.make 2025-12-04T09:42:22.0682427Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o.d 2025-12-04T09:42:22.0775644Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o 2025-12-04T09:42:22.0776931Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-12-04T09:42:22.0777891Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/TargetDirectories.txt 2025-12-04T09:42:22.0778726Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/progress.marks 2025-12-04T09:42:22.0779506Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile2 2025-12-04T09:42:22.0780293Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile.cmake 2025-12-04T09:42:22.0781065Z inflating: build/custom_test_artifacts/custom-op-build/detect_cuda_version.cc 2025-12-04T09:42:22.0783775Z inflating: build/custom_test_artifacts/custom-op-build/CMakeCache.txt 2025-12-04T09:42:22.0784669Z inflating: build/custom_test_artifacts/custom-op-build/Makefile 2025-12-04T09:42:22.0785353Z inflating: build/custom_test_artifacts/custom-op-build/cmake_install.cmake 2025-12-04T09:42:22.0985307Z inflating: build/custom_test_artifacts/custom-op-build/libcustom_ops.so 2025-12-04T09:42:22.1051935Z inflating: build/custom_test_artifacts/custom-op-build/test_custom_ops 2025-12-04T09:42:22.1052528Z creating: build/custom_test_artifacts/jit-hook-build/ 2025-12-04T09:42:22.1053089Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/ 2025-12-04T09:42:22.1053762Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/pkgRedirects/ 2025-12-04T09:42:22.1061588Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeConfigureLog.yaml 2025-12-04T09:42:22.1062375Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/ 2025-12-04T09:42:22.1063151Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-12-04T09:42:22.1063993Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-12-04T09:42:22.1064797Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-12-04T09:42:22.1065888Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-12-04T09:42:22.1067490Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-12-04T09:42:22.1068754Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-12-04T09:42:22.1069660Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-12-04T09:42:22.1070489Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-12-04T09:42:22.1071858Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-12-04T09:42:22.1073529Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-12-04T09:42:22.1074641Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-12-04T09:42:22.1076281Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-12-04T09:42:22.1078277Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-12-04T09:42:22.1079237Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-12-04T09:42:22.1080092Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-12-04T09:42:22.1144473Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-12-04T09:42:22.1211820Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-12-04T09:42:22.1213113Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-12-04T09:42:22.1283025Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-12-04T09:42:22.1284290Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-12-04T09:42:22.1285541Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-12-04T09:42:22.1286825Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-12-04T09:42:22.1288063Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-12-04T09:42:22.1289279Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-12-04T09:42:22.1290601Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-12-04T09:42:22.1291832Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-12-04T09:42:22.1293020Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-12-04T09:42:22.1294151Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-12-04T09:42:22.1295225Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-12-04T09:42:22.1296447Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-12-04T09:42:22.1297581Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-12-04T09:42:22.1298630Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-12-04T09:42:22.1299677Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-12-04T09:42:22.1382002Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-12-04T09:42:22.1382951Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-12-04T09:42:22.1469503Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-12-04T09:42:22.1470455Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeScratch/ 2025-12-04T09:42:22.1471171Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeTmp/ 2025-12-04T09:42:22.1471943Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/cmake.check_cache 2025-12-04T09:42:22.1472754Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/ 2025-12-04T09:42:22.1473774Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.ts 2025-12-04T09:42:22.1474808Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.make 2025-12-04T09:42:22.1475805Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/depend.make 2025-12-04T09:42:22.1476738Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/link.txt 2025-12-04T09:42:22.1477704Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/cmake_clean.cmake 2025-12-04T09:42:22.1478665Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/build.make 2025-12-04T09:42:22.1479635Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/DependInfo.cmake 2025-12-04T09:42:22.1480598Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/flags.make 2025-12-04T09:42:22.1481743Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/progress.make 2025-12-04T09:42:22.1499467Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o.d 2025-12-04T09:42:22.1572148Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o 2025-12-04T09:42:22.1573178Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-12-04T09:42:22.1574113Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/TargetDirectories.txt 2025-12-04T09:42:22.1574934Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/progress.marks 2025-12-04T09:42:22.1575853Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile2 2025-12-04T09:42:22.1577043Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile.cmake 2025-12-04T09:42:22.1577872Z inflating: build/custom_test_artifacts/jit-hook-build/detect_cuda_version.cc 2025-12-04T09:42:22.1580097Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeCache.txt 2025-12-04T09:42:22.1580931Z inflating: build/custom_test_artifacts/jit-hook-build/Makefile 2025-12-04T09:42:22.1581699Z inflating: build/custom_test_artifacts/jit-hook-build/cmake_install.cmake 2025-12-04T09:42:22.1626835Z inflating: build/custom_test_artifacts/jit-hook-build/test_jit_hooks 2025-12-04T09:42:22.1627467Z creating: build/custom_test_artifacts/custom-backend-build/ 2025-12-04T09:42:22.1628102Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/ 2025-12-04T09:42:22.1628844Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/pkgRedirects/ 2025-12-04T09:42:22.1636454Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeConfigureLog.yaml 2025-12-04T09:42:22.1637301Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/ 2025-12-04T09:42:22.1638154Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-12-04T09:42:22.1639059Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-12-04T09:42:22.1639946Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-12-04T09:42:22.1640957Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-12-04T09:42:22.1642349Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-12-04T09:42:22.1643319Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-12-04T09:42:22.1644270Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-12-04T09:42:22.1645169Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-12-04T09:42:22.1646670Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-12-04T09:42:22.1648241Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-12-04T09:42:22.1649364Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-12-04T09:42:22.1651038Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-12-04T09:42:22.1653005Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-12-04T09:42:22.1654034Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/ 2025-12-04T09:42:22.1654958Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/ 2025-12-04T09:42:22.1718850Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp4.ii 2025-12-04T09:42:22.1785688Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.cpp 2025-12-04T09:42:22.1787035Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.module_id 2025-12-04T09:42:22.1856814Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cpp1.ii 2025-12-04T09:42:22.1858186Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.c 2025-12-04T09:42:22.1859525Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.gpu 2025-12-04T09:42:22.1861010Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.cudafe1.stub.c 2025-12-04T09:42:22.1862346Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.ptx 2025-12-04T09:42:22.1863649Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.sm_52.cubin 2025-12-04T09:42:22.1864952Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin 2025-12-04T09:42:22.1866242Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.fatbin.c 2025-12-04T09:42:22.1867507Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/CMakeCUDACompilerId.o 2025-12-04T09:42:22.1868713Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.sm_52.cubin 2025-12-04T09:42:22.1869880Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.reg.c 2025-12-04T09:42:22.1871013Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin 2025-12-04T09:42:22.1872165Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.fatbin.c 2025-12-04T09:42:22.1873282Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/tmp/a_dlink.o 2025-12-04T09:42:22.1874416Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/CMakeCUDACompilerId.cu 2025-12-04T09:42:22.1955983Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCUDA/a.out 2025-12-04T09:42:22.1957094Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCUDACompiler.cmake 2025-12-04T09:42:22.2042472Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CUDA.bin 2025-12-04T09:42:22.2043481Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeScratch/ 2025-12-04T09:42:22.2044272Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeTmp/ 2025-12-04T09:42:22.2045087Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/cmake.check_cache 2025-12-04T09:42:22.2045964Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/ 2025-12-04T09:42:22.2046950Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.ts 2025-12-04T09:42:22.2048096Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.make 2025-12-04T09:42:22.2049201Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/depend.make 2025-12-04T09:42:22.2050212Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/link.txt 2025-12-04T09:42:22.2051383Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/cmake_clean.cmake 2025-12-04T09:42:22.2052448Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/build.make 2025-12-04T09:42:22.2053507Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/DependInfo.cmake 2025-12-04T09:42:22.2054563Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/flags.make 2025-12-04T09:42:22.2055594Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/progress.make 2025-12-04T09:42:22.2056832Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o.d 2025-12-04T09:42:22.2192240Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o 2025-12-04T09:42:22.2193298Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/ 2025-12-04T09:42:22.2194338Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.ts 2025-12-04T09:42:22.2195530Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.make 2025-12-04T09:42:22.2196921Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/depend.make 2025-12-04T09:42:22.2197995Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/link.txt 2025-12-04T09:42:22.2199085Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/cmake_clean.cmake 2025-12-04T09:42:22.2200221Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/build.make 2025-12-04T09:42:22.2201343Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/DependInfo.cmake 2025-12-04T09:42:22.2202468Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/flags.make 2025-12-04T09:42:22.2203551Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/progress.make 2025-12-04T09:42:22.2221590Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o.d 2025-12-04T09:42:22.2284715Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o 2025-12-04T09:42:22.2285879Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-12-04T09:42:22.2286898Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/TargetDirectories.txt 2025-12-04T09:42:22.2287791Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/progress.marks 2025-12-04T09:42:22.2288629Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile2 2025-12-04T09:42:22.2289453Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile.cmake 2025-12-04T09:42:22.2290281Z inflating: build/custom_test_artifacts/custom-backend-build/detect_cuda_version.cc 2025-12-04T09:42:22.2292483Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeCache.txt 2025-12-04T09:42:22.2293398Z inflating: build/custom_test_artifacts/custom-backend-build/Makefile 2025-12-04T09:42:22.2294403Z inflating: build/custom_test_artifacts/custom-backend-build/cmake_install.cmake 2025-12-04T09:42:22.2413312Z inflating: build/custom_test_artifacts/custom-backend-build/libcustom_backend.so 2025-12-04T09:42:22.2459199Z inflating: build/custom_test_artifacts/custom-backend-build/test_custom_backend 2025-12-04T09:42:22.2459802Z creating: build/lib/ 2025-12-04T09:42:22.2555056Z inflating: build/lib/libprotobuf-lite.a 2025-12-04T09:42:22.3070089Z inflating: build/lib/libprotobuf.a 2025-12-04T09:42:22.3649001Z inflating: build/lib/libprotoc.a 2025-12-04T09:42:22.3660453Z inflating: build/lib/libpthreadpool.a 2025-12-04T09:42:22.3670019Z inflating: build/lib/libcpuinfo.a 2025-12-04T09:42:22.3679107Z inflating: build/lib/libcpuinfo_internals.a 2025-12-04T09:42:22.3680156Z inflating: build/lib/libclog.a 2025-12-04T09:42:22.3702748Z inflating: build/lib/libpytorch_qnnpack.a 2025-12-04T09:42:22.3705591Z inflating: build/lib/libnnpack_reference_layers.a 2025-12-04T09:42:22.3726262Z inflating: build/lib/libnnpack.a 2025-12-04T09:42:22.3943995Z inflating: build/lib/libmicrokernels-prod.a 2025-12-04T09:42:22.4971909Z inflating: build/lib/libmicrokernels-all.a 2025-12-04T09:42:22.5053132Z inflating: build/lib/libgtest.a 2025-12-04T09:42:22.5073293Z inflating: build/lib/libgmock.a 2025-12-04T09:42:22.5074193Z inflating: build/lib/libgtest_main.a 2025-12-04T09:42:22.5075102Z inflating: build/lib/libgmock_main.a 2025-12-04T09:42:22.5180627Z inflating: build/lib/libXNNPACK.a 2025-12-04T09:42:22.5269192Z inflating: build/lib/libbenchmark.a 2025-12-04T09:42:22.5270117Z inflating: build/lib/libbenchmark_main.a 2025-12-04T09:42:22.5279365Z inflating: build/lib/libittnotify.a 2025-12-04T09:42:22.5356877Z inflating: build/lib/libasmjit.a 2025-12-04T09:42:22.5357835Z inflating: build/lib/libjitprofiling.a 2025-12-04T09:42:22.6693782Z inflating: build/lib/libfbgemm.a 2025-12-04T09:42:22.6730064Z inflating: build/lib/libtensorpipe_uv.a 2025-12-04T09:42:22.7355936Z inflating: build/lib/libtensorpipe.a 2025-12-04T09:42:22.7638191Z inflating: build/lib/libtensorpipe_cuda.a 2025-12-04T09:42:22.7795913Z inflating: build/lib/libgloo.a 2025-12-04T09:42:22.7850556Z inflating: build/lib/libonnx_proto.a 2025-12-04T09:42:22.8335628Z inflating: build/lib/libgloo_cuda.a 2025-12-04T09:42:22.9161936Z inflating: build/lib/libonnx.a 2025-12-04T09:42:24.0770502Z inflating: build/lib/libdnnl.a 2025-12-04T09:42:24.0793408Z inflating: build/lib/libfmt.a 2025-12-04T09:42:24.1355111Z inflating: build/lib/libkineto.a 2025-12-04T09:42:24.1492569Z inflating: build/lib/libc10.so 2025-12-04T09:42:24.1550949Z inflating: build/lib/libc10_cuda.so 2025-12-04T09:42:24.1552637Z inflating: build/lib/libtorch_global_deps.so 2025-12-04T09:42:24.1554636Z inflating: build/lib/libcaffe2_nvrtc.so 2025-12-04T09:42:27.7665303Z inflating: build/lib/libtorch_cpu.so 2025-12-04T09:42:29.6645804Z inflating: build/lib/libtorch_cuda.so 2025-12-04T09:42:29.6650916Z inflating: build/lib/libshm.so 2025-12-04T09:42:29.6652410Z inflating: build/lib/libtorch.so 2025-12-04T09:42:29.6709281Z inflating: build/lib/libtorch_cuda_linalg.so 2025-12-04T09:42:29.6712052Z inflating: build/lib/libc10d_cuda_test.so 2025-12-04T09:42:29.6794972Z inflating: build/lib/libtorchbind_test.so 2025-12-04T09:42:29.6818028Z inflating: build/lib/libjitbackend_test.so 2025-12-04T09:42:29.6846013Z inflating: build/lib/libbackend_with_compiler.so 2025-12-04T09:42:29.6876676Z inflating: build/lib/libaoti_custom_ops.so 2025-12-04T09:42:29.9667578Z inflating: build/lib/libtorch_python.so 2025-12-04T09:42:29.9711039Z inflating: build/lib/libnnapi_backend.so 2025-12-04T09:42:29.9711472Z creating: build/bin/ 2025-12-04T09:42:30.0252980Z inflating: build/bin/protoc-3.13.0.0 2025-12-04T09:42:30.0795180Z inflating: build/bin/protoc 2025-12-04T09:42:30.0865568Z inflating: build/bin/c10_AllocatorConfig_test 2025-12-04T09:42:30.0931444Z inflating: build/bin/c10_CompileTimeFunctionPointer_test 2025-12-04T09:42:30.0999234Z inflating: build/bin/c10_DeviceGuard_test 2025-12-04T09:42:30.1067042Z inflating: build/bin/c10_Device_test 2025-12-04T09:42:30.1144540Z inflating: build/bin/c10_DispatchKeySet_test 2025-12-04T09:42:30.1208976Z inflating: build/bin/c10_StreamGuard_test 2025-12-04T09:42:30.1279806Z inflating: build/bin/c10_Scalar_test 2025-12-04T09:42:30.1353166Z inflating: build/bin/c10_SymInt_test 2025-12-04T09:42:30.1426258Z inflating: build/bin/c10_InlineStreamGuard_test 2025-12-04T09:42:30.1497402Z inflating: build/bin/c10_InlineDeviceGuard_test 2025-12-04T09:42:30.1570254Z inflating: build/bin/c10_SizesAndStrides_test 2025-12-04T09:42:30.1635467Z inflating: build/bin/c10_ArrayRef_test 2025-12-04T09:42:30.1700087Z inflating: build/bin/c10_ConstexprCrc_test 2025-12-04T09:42:30.1790665Z inflating: build/bin/c10_cow_test 2025-12-04T09:42:30.1860157Z inflating: build/bin/c10_Bitset_test 2025-12-04T09:42:30.1925977Z inflating: build/bin/c10_DeadlockDetection_test 2025-12-04T09:42:30.2000711Z inflating: build/bin/c10_Enumerate_test 2025-12-04T09:42:30.2067544Z inflating: build/bin/c10_Half_test 2025-12-04T09:42:30.2136782Z inflating: build/bin/c10_IntrusiveList_test 2025-12-04T09:42:30.2206684Z inflating: build/bin/c10_NetworkFlow_test 2025-12-04T09:42:30.2279638Z inflating: build/bin/c10_LeftRight_test 2025-12-04T09:42:30.2345291Z inflating: build/bin/c10_Synchronized_test 2025-12-04T09:42:30.2410617Z inflating: build/bin/c10_Semaphore_test 2025-12-04T09:42:30.2482710Z inflating: build/bin/c10_ThreadLocal_test 2025-12-04T09:42:30.2551075Z inflating: build/bin/c10_TypeIndex_test 2025-12-04T09:42:30.2618896Z inflating: build/bin/c10_accumulate_test 2025-12-04T09:42:30.2691555Z inflating: build/bin/c10_bfloat16_test 2025-12-04T09:42:30.2765480Z inflating: build/bin/c10_complex_math_test 2025-12-04T09:42:30.2831698Z inflating: build/bin/c10_bit_cast_test 2025-12-04T09:42:30.2897530Z inflating: build/bin/c10_error_test 2025-12-04T09:42:30.2969093Z inflating: build/bin/c10_complex_test 2025-12-04T09:42:30.3038016Z inflating: build/bin/c10_exception_test 2025-12-04T09:42:30.3104060Z inflating: build/bin/c10_flags_test 2025-12-04T09:42:30.3170718Z inflating: build/bin/c10_generic_math_test 2025-12-04T09:42:30.3364287Z inflating: build/bin/c10_intrusive_ptr_test 2025-12-04T09:42:30.3430980Z inflating: build/bin/c10_irange_test 2025-12-04T09:42:30.3501197Z inflating: build/bin/c10_lazy_test 2025-12-04T09:42:30.3575731Z inflating: build/bin/c10_logging_test 2025-12-04T09:42:30.3641421Z inflating: build/bin/c10_nofatal_test 2025-12-04T09:42:30.3737184Z inflating: build/bin/c10_optional_test 2025-12-04T09:42:30.3817836Z inflating: build/bin/c10_ordered_preserving_dict_test 2025-12-04T09:42:30.3887671Z inflating: build/bin/c10_registry_test 2025-12-04T09:42:30.4076454Z inflating: build/bin/c10_small_vector_test 2025-12-04T09:42:30.4144091Z inflating: build/bin/c10_ssize_test 2025-12-04T09:42:30.4217639Z inflating: build/bin/c10_string_util_test 2025-12-04T09:42:30.4275396Z inflating: build/bin/c10_intrusive_ptr_benchmark 2025-12-04T09:42:30.4341362Z inflating: build/bin/c10_tempfile_test 2025-12-04T09:42:30.4405870Z inflating: build/bin/c10_string_view_test 2025-12-04T09:42:30.4479163Z inflating: build/bin/c10_typeid_test 2025-12-04T09:42:30.4548783Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_catches_thread_and_block_and_device 2025-12-04T09:42:30.4618237Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_multiple_blocks 2025-12-04T09:42:30.4686032Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_from_2_processes 2025-12-04T09:42:30.4755266Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_blocks_and_threads 2025-12-04T09:42:30.4820061Z inflating: build/bin/c10_cuda_CUDATest 2025-12-04T09:42:30.4889211Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_catches_stream 2025-12-04T09:42:30.4958421Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_1_var_test 2025-12-04T09:42:30.5027652Z inflating: build/bin/c10_cuda_CUDAAssertionsTest_multiple_writes_from_same_block 2025-12-04T09:42:30.5729775Z inflating: build/bin/vec_test_all_types_DEFAULT 2025-12-04T09:42:30.6452818Z inflating: build/bin/vec_test_all_types_AVX512 2025-12-04T09:42:30.7187510Z inflating: build/bin/vec_test_all_types_AVX2 2025-12-04T09:42:30.7252328Z inflating: build/bin/test_vec_half_DEFAULT 2025-12-04T09:42:30.7375786Z inflating: build/bin/test_aoti_abi_check 2025-12-04T09:42:30.7441599Z inflating: build/bin/test_vec_half_AVX512 2025-12-04T09:42:30.7507402Z inflating: build/bin/test_vec_half_AVX2 2025-12-04T09:42:30.7576479Z inflating: build/bin/BackoffTest 2025-12-04T09:42:30.7646362Z inflating: build/bin/FileStoreTest 2025-12-04T09:42:30.7720758Z inflating: build/bin/TCPStoreTest 2025-12-04T09:42:30.7791004Z inflating: build/bin/HashStoreTest 2025-12-04T09:42:30.7807801Z inflating: build/bin/ProcessGroupMPITest 2025-12-04T09:42:30.7812179Z inflating: build/bin/torch_shm_manager 2025-12-04T09:42:30.7906717Z inflating: build/bin/Dict_test 2025-12-04T09:42:30.7975774Z inflating: build/bin/Dimname_test 2025-12-04T09:42:30.8060352Z inflating: build/bin/MaybeOwned_test 2025-12-04T09:42:30.8134539Z inflating: build/bin/NamedTensor_test 2025-12-04T09:42:30.8211282Z inflating: build/bin/apply_utils_test 2025-12-04T09:42:30.8287823Z inflating: build/bin/atest 2025-12-04T09:42:30.8370703Z inflating: build/bin/basic 2025-12-04T09:42:30.8441692Z inflating: build/bin/broadcast_test 2025-12-04T09:42:30.8508651Z inflating: build/bin/cpu_allocator_test 2025-12-04T09:42:30.8584278Z inflating: build/bin/cpu_generator_test 2025-12-04T09:42:30.8653447Z inflating: build/bin/cpu_profiling_allocator_test 2025-12-04T09:42:30.8770393Z inflating: build/bin/cpu_rng_test 2025-12-04T09:42:30.8837605Z inflating: build/bin/dlconvertor_test 2025-12-04T09:42:30.8912723Z inflating: build/bin/extension_backend_test 2025-12-04T09:42:30.8985546Z inflating: build/bin/half_test 2025-12-04T09:42:30.9109420Z inflating: build/bin/ivalue_test 2025-12-04T09:42:30.9174789Z inflating: build/bin/lazy_tensor_test 2025-12-04T09:42:30.9244774Z inflating: build/bin/math_kernel_test 2025-12-04T09:42:30.9314525Z inflating: build/bin/memory_format_test 2025-12-04T09:42:30.9384808Z inflating: build/bin/memory_overlapping_test 2025-12-04T09:42:30.9454910Z inflating: build/bin/mobile_memory_cleanup 2025-12-04T09:42:30.9528321Z inflating: build/bin/native_test 2025-12-04T09:42:30.9595220Z inflating: build/bin/operator_name_test 2025-12-04T09:42:30.9661775Z inflating: build/bin/operators_test 2025-12-04T09:42:30.9730599Z inflating: build/bin/packedtensoraccessor_test 2025-12-04T09:42:30.9818234Z inflating: build/bin/pow_test 2025-12-04T09:42:30.9891688Z inflating: build/bin/quantized_test 2025-12-04T09:42:30.9958189Z inflating: build/bin/reduce_ops_test 2025-12-04T09:42:31.0025331Z inflating: build/bin/reportMemoryUsage_test 2025-12-04T09:42:31.0098804Z inflating: build/bin/scalar_tensor_test 2025-12-04T09:42:31.0173777Z inflating: build/bin/scalar_test 2025-12-04T09:42:31.0241689Z inflating: build/bin/StorageUtils_test 2025-12-04T09:42:31.0310278Z inflating: build/bin/stride_properties_test 2025-12-04T09:42:31.0410917Z inflating: build/bin/tensor_iterator_test 2025-12-04T09:42:31.0481695Z inflating: build/bin/test_parallel 2025-12-04T09:42:31.0548626Z inflating: build/bin/thread_init_test 2025-12-04T09:42:31.0620674Z inflating: build/bin/type_ptr_test 2025-12-04T09:42:31.0698228Z inflating: build/bin/type_test 2025-12-04T09:42:31.0767243Z inflating: build/bin/undefined_tensor_test 2025-12-04T09:42:31.0833069Z inflating: build/bin/verify_api_visibility 2025-12-04T09:42:31.0925014Z inflating: build/bin/legacy_vmap_test 2025-12-04T09:42:31.0992021Z inflating: build/bin/weakref_test 2025-12-04T09:42:31.1059715Z inflating: build/bin/wrapdim_test 2025-12-04T09:42:31.1127053Z inflating: build/bin/xla_tensor_test 2025-12-04T09:42:31.1205051Z inflating: build/bin/IListRef_test 2025-12-04T09:42:31.1337743Z inflating: build/bin/List_test 2025-12-04T09:42:31.1423117Z inflating: build/bin/KernelFunction_test 2025-12-04T09:42:31.1573394Z inflating: build/bin/kernel_function_legacy_test 2025-12-04T09:42:31.1693723Z inflating: build/bin/kernel_function_test 2025-12-04T09:42:31.1851111Z inflating: build/bin/kernel_lambda_legacy_test 2025-12-04T09:42:31.1979123Z inflating: build/bin/kernel_lambda_test 2025-12-04T09:42:31.2057403Z inflating: build/bin/kernel_stackbased_test 2025-12-04T09:42:31.2178114Z inflating: build/bin/make_boxed_from_unboxed_functor_test 2025-12-04T09:42:31.2244960Z inflating: build/bin/CppSignature_test 2025-12-04T09:42:31.2317172Z inflating: build/bin/backend_fallback_test 2025-12-04T09:42:31.2381795Z inflating: build/bin/op_allowlist_test 2025-12-04T09:42:31.2755522Z inflating: build/bin/op_registration_test 2025-12-04T09:42:31.2841660Z inflating: build/bin/inline_container_test 2025-12-04T09:42:31.2911800Z inflating: build/bin/cuda_allocator_test 2025-12-04T09:42:31.2981111Z inflating: build/bin/cuda_apply_test 2025-12-04T09:42:31.3058680Z inflating: build/bin/cuda_atomic_ops_test 2025-12-04T09:42:31.3132346Z inflating: build/bin/cuda_caching_host_allocator_test 2025-12-04T09:42:31.3222027Z inflating: build/bin/cuda_complex_math_test 2025-12-04T09:42:31.3299713Z inflating: build/bin/cuda_complex_test 2025-12-04T09:42:31.3376144Z inflating: build/bin/cuda_cub_test 2025-12-04T09:42:31.3445560Z inflating: build/bin/cuda_cublas_handle_pool_test 2025-12-04T09:42:31.3511308Z inflating: build/bin/cuda_device_test 2025-12-04T09:42:31.3594171Z inflating: build/bin/cuda_distributions_test 2025-12-04T09:42:31.3662466Z inflating: build/bin/cuda_dlconvertor_test 2025-12-04T09:42:31.3732526Z inflating: build/bin/cuda_event_test 2025-12-04T09:42:31.3797633Z inflating: build/bin/cuda_exchange_device_test 2025-12-04T09:42:31.3871228Z inflating: build/bin/cuda_generator_test 2025-12-04T09:42:31.3936687Z inflating: build/bin/cuda_half_test 2025-12-04T09:42:31.4003900Z inflating: build/bin/cuda_integer_divider_test 2025-12-04T09:42:31.4068875Z inflating: build/bin/cuda_optional_test 2025-12-04T09:42:31.4137075Z inflating: build/bin/cuda_packedtensoraccessor_test 2025-12-04T09:42:31.4206011Z inflating: build/bin/cuda_reportMemoryUsage_test 2025-12-04T09:42:31.4271474Z inflating: build/bin/cuda_allocatorTraceTracker_test 2025-12-04T09:42:31.4351045Z inflating: build/bin/cuda_stream_test 2025-12-04T09:42:31.4420102Z inflating: build/bin/cuda_vectorized_test 2025-12-04T09:42:31.4485405Z inflating: build/bin/cuda_cudnn_test 2025-12-04T09:42:31.4906166Z inflating: build/bin/test_lazy 2025-12-04T09:42:31.4992497Z inflating: build/bin/ProcessGroupGlooTest 2025-12-04T09:42:31.5066225Z inflating: build/bin/ProcessGroupGlooAsyncTest 2025-12-04T09:42:31.6394670Z inflating: build/bin/test_jit 2025-12-04T09:42:31.6477707Z inflating: build/bin/ProcessGroupNCCLTest 2025-12-04T09:42:31.6557543Z inflating: build/bin/ProcessGroupNCCLErrorsTest 2025-12-04T09:42:31.6561149Z inflating: build/bin/example_allreduce 2025-12-04T09:42:31.6633237Z inflating: build/bin/test_dist_autograd 2025-12-04T09:42:31.6721756Z inflating: build/bin/test_cpp_rpc 2025-12-04T09:42:31.6724728Z inflating: build/bin/parallel_benchmark 2025-12-04T09:42:31.8145396Z inflating: build/bin/test_api 2025-12-04T09:42:31.8145804Z creating: .additional_ci_files/ 2025-12-04T09:42:31.8221690Z inflating: .additional_ci_files/test-times.json 2025-12-04T09:42:31.8500488Z inflating: .additional_ci_files/test-class-times.json 2025-12-04T09:42:31.8534956Z ##[group]Run rm artifacts.zip 2025-12-04T09:42:31.8535358Z rm artifacts.zip 2025-12-04T09:42:31.8544107Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:42:31.8544542Z env: 2025-12-04T09:42:31.8544796Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:31.8545109Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:31.8545653Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:31.8546084Z ##[endgroup] 2025-12-04T09:42:32.0395749Z ##[group]Run df -H 2025-12-04T09:42:32.0396279Z df -H 2025-12-04T09:42:32.0403030Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:42:32.0403459Z env: 2025-12-04T09:42:32.0403707Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:32.0404044Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:32.0404397Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:32.0404824Z ##[endgroup] 2025-12-04T09:42:32.0455237Z Filesystem Size Used Avail Use% Mounted on 2025-12-04T09:42:32.0455979Z devtmpfs 4.2M 0 4.2M 0% /dev 2025-12-04T09:42:32.0457028Z tmpfs 34G 0 34G 0% /dev/shm 2025-12-04T09:42:32.0457682Z tmpfs 14G 562k 14G 1% /run 2025-12-04T09:42:32.0458274Z /dev/nvme0n1p1 161G 51G 111G 32% / 2025-12-04T09:42:32.0458640Z tmpfs 34G 17k 34G 1% /tmp 2025-12-04T09:42:32.0459064Z /dev/nvme0n1p128 11M 1.4M 9.2M 13% /boot/efi 2025-12-04T09:42:32.0459493Z tmpfs 6.7G 0 6.7G 0% /run/user/0 2025-12-04T09:42:32.0498491Z Prepare all required actions 2025-12-04T09:42:32.0499388Z Getting action download info 2025-12-04T09:42:32.1910798Z ##[group]Run ./.github/actions/download-td-artifacts 2025-12-04T09:42:32.1911219Z with: 2025-12-04T09:42:32.1911453Z env: 2025-12-04T09:42:32.1911693Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:32.1912008Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:32.1912379Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:32.1912790Z ##[endgroup] 2025-12-04T09:42:32.1944090Z ##[group]Run seemethere/download-artifact-s3@v4 2025-12-04T09:42:32.1944514Z with: 2025-12-04T09:42:32.1944740Z name: td_results 2025-12-04T09:42:32.1945020Z s3-bucket: gha-artifacts 2025-12-04T09:42:32.1945331Z region: us-east-1 2025-12-04T09:42:32.1945577Z env: 2025-12-04T09:42:32.1945818Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:32.1946121Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:32.1946484Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:32.1946955Z ##[endgroup] 2025-12-04T09:42:32.7423595Z (node:68819) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-12-04T09:42:32.7424208Z 2025-12-04T09:42:32.7424427Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-12-04T09:42:32.7425063Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-12-04T09:42:32.7425721Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-12-04T09:42:32.8618133Z Found 1 objects with prefix pytorch/pytorch/19922826259/td_results/ 2025-12-04T09:42:32.8618963Z Starting download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/td_results.json 2025-12-04T09:42:32.9228740Z Finished download (1/1): /home/ec2-user/actions-runner/_work/pytorch/pytorch/td_results.json 2025-12-04T09:42:32.9234629Z Artifact download has finished successfully 2025-12-04T09:42:32.9409987Z ##[group]Run mkdir -p .additional_ci_files 2025-12-04T09:42:32.9410446Z mkdir -p .additional_ci_files 2025-12-04T09:42:32.9410963Z mv td_results.json .additional_ci_files/td_results.json || true 2025-12-04T09:42:32.9419111Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:42:32.9419537Z env: 2025-12-04T09:42:32.9419788Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:32.9420101Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:32.9420451Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:32.9420869Z ##[endgroup] 2025-12-04T09:42:32.9524336Z ##[group]Run .github/scripts/parse_ref.py 2025-12-04T09:42:32.9524787Z .github/scripts/parse_ref.py 2025-12-04T09:42:32.9531267Z shell: /usr/bin/bash -e {0} 2025-12-04T09:42:32.9531581Z env: 2025-12-04T09:42:32.9531826Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:32.9532122Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:32.9532486Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:32.9532905Z ##[endgroup] 2025-12-04T09:42:32.9776563Z Setting output branch=main 2025-12-04T09:42:32.9943009Z Prepare all required actions 2025-12-04T09:42:32.9943474Z Getting action download info 2025-12-04T09:42:33.1299013Z ##[group]Run ./.github/actions/filter-test-configs 2025-12-04T09:42:33.1299430Z with: 2025-12-04T09:42:33.1299881Z github-token: *** 2025-12-04T09:42:33.1306972Z test-matrix: {"include": [{"config": "legacy_nvidia_driver", "shard": 1, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 1, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 2, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 2, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 3, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 3, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 4, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 4, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 5, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 5, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}]} 2025-12-04T09:42:33.1314886Z job-name: linux-jammy-cuda12.4-py3.10-gcc11 / test (legacy_nvidia_driver, 3, 5, linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check, unstable) 2025-12-04T09:42:33.1315673Z env: 2025-12-04T09:42:33.1315918Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:33.1316215Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:33.1316581Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:33.1316996Z ##[endgroup] 2025-12-04T09:42:33.1357668Z ##[group]Run nick-fields/retry@v3.0.0 2025-12-04T09:42:33.1358017Z with: 2025-12-04T09:42:33.1358258Z shell: bash 2025-12-04T09:42:33.1358540Z timeout_minutes: 10 2025-12-04T09:42:33.1358827Z max_attempts: 5 2025-12-04T09:42:33.1359092Z retry_wait_seconds: 30 2025-12-04T09:42:33.1360034Z command: set -eux # PyYAML 6.0 doesn't work with MacOS x86 anymore # This must run on Python-3.7 (AmazonLinux2) so can't use request=3.32.2 python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-12-04T09:42:33.1361039Z polling_interval_seconds: 1 2025-12-04T09:42:33.1361356Z warning_on_retry: true 2025-12-04T09:42:33.1361657Z continue_on_error: false 2025-12-04T09:42:33.1361949Z env: 2025-12-04T09:42:33.1362177Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:33.1362476Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:33.1362835Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:33.1363454Z GITHUB_TOKEN: *** 2025-12-04T09:42:33.1363709Z ##[endgroup] 2025-12-04T09:42:33.2457159Z + python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-12-04T09:42:33.5204212Z Defaulting to user installation because normal site-packages is not writeable 2025-12-04T09:42:33.6495903Z Collecting requests==2.27.1 2025-12-04T09:42:33.6658981Z Downloading requests-2.27.1-py2.py3-none-any.whl (63 kB) 2025-12-04T09:42:33.8685305Z Collecting pyyaml==6.0.2 2025-12-04T09:42:33.8724651Z Downloading PyYAML-6.0.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (737 kB) 2025-12-04T09:42:34.3444827Z Collecting charset-normalizer~=2.0.0 2025-12-04T09:42:34.3484247Z Downloading charset_normalizer-2.0.12-py3-none-any.whl (39 kB) 2025-12-04T09:42:34.4074676Z Collecting certifi>=2017.4.17 2025-12-04T09:42:34.4112134Z Downloading certifi-2025.11.12-py3-none-any.whl (159 kB) 2025-12-04T09:42:34.4181527Z Requirement already satisfied: urllib3<1.27,>=1.21.1 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (1.25.10) 2025-12-04T09:42:34.4192770Z Requirement already satisfied: idna<4,>=2.5 in /usr/lib/python3.9/site-packages (from requests==2.27.1) (2.10) 2025-12-04T09:42:34.5153268Z Installing collected packages: charset-normalizer, certifi, requests, pyyaml 2025-12-04T09:42:34.6548869Z Successfully installed certifi-2025.11.12 charset-normalizer-2.0.12 pyyaml-6.0.2 requests-2.27.1 2025-12-04T09:42:35.2255355Z Command completed after 1 attempt(s). 2025-12-04T09:42:35.2307222Z ##[group]Run set -x 2025-12-04T09:42:35.2307512Z set -x 2025-12-04T09:42:35.2307772Z  2025-12-04T09:42:35.2308244Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-12-04T09:42:35.2308800Z # in runner workspace 2025-12-04T09:42:35.2309261Z python3 "${GITHUB_ACTION_PATH}/../../scripts/parse_ref.py" 2025-12-04T09:42:35.2316652Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:42:35.2317106Z env: 2025-12-04T09:42:35.2317361Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:35.2317664Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:35.2318035Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:35.2318457Z ##[endgroup] 2025-12-04T09:42:35.2347362Z + python3 /home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/filter-test-configs/../../scripts/parse_ref.py 2025-12-04T09:42:35.2557333Z Setting output branch=main 2025-12-04T09:42:35.2615381Z ##[group]Run echo "Workflow: ${GITHUB_WORKFLOW}" 2025-12-04T09:42:35.2615903Z echo "Workflow: ${GITHUB_WORKFLOW}" 2025-12-04T09:42:35.2616302Z echo "Job name: ${JOB_NAME}" 2025-12-04T09:42:35.2616644Z  2025-12-04T09:42:35.2617189Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-12-04T09:42:35.2617743Z # in runner workspace 2025-12-04T09:42:35.2618247Z python3 "${GITHUB_ACTION_PATH}/../../scripts/filter_test_configs.py" \ 2025-12-04T09:42:35.2618816Z  --workflow "${GITHUB_WORKFLOW}" \ 2025-12-04T09:42:35.2619204Z  --job-name "${JOB_NAME}" \ 2025-12-04T09:42:35.2626494Z  --test-matrix "{"include": [{"config": "legacy_nvidia_driver", "shard": 1, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 1, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 2, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 2, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 3, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 3, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 4, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 4, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 5, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 5, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}]}" \ 2025-12-04T09:42:35.2633888Z  --selected-test-configs "" \ 2025-12-04T09:42:35.2634281Z  --pr-number "${PR_NUMBER}" \ 2025-12-04T09:42:35.2634646Z  --tag "${TAG}" \ 2025-12-04T09:42:35.2634984Z  --event-name "${EVENT_NAME}" \ 2025-12-04T09:42:35.2635351Z  --schedule "${SCHEDULE}" \ 2025-12-04T09:42:35.2635842Z  --branch "${HEAD_BRANCH}" 2025-12-04T09:42:35.2642458Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:42:35.2642885Z env: 2025-12-04T09:42:35.2643135Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:35.2643444Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:35.2643815Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:35.2644499Z GITHUB_TOKEN: *** 2025-12-04T09:42:35.2645217Z JOB_NAME: linux-jammy-cuda12.4-py3.10-gcc11 / test (legacy_nvidia_driver, 3, 5, linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check, unstable) 2025-12-04T09:42:35.2646003Z PR_NUMBER: 2025-12-04T09:42:35.2646237Z TAG: 2025-12-04T09:42:35.2646477Z EVENT_NAME: schedule 2025-12-04T09:42:35.2646761Z SCHEDULE: 29 8 * * * 2025-12-04T09:42:35.2647028Z HEAD_BRANCH: main 2025-12-04T09:42:35.2647293Z ##[endgroup] 2025-12-04T09:42:35.2673916Z Workflow: periodic 2025-12-04T09:42:35.2674770Z Job name: linux-jammy-cuda12.4-py3.10-gcc11 / test (legacy_nvidia_driver, 3, 5, linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check, unstable) 2025-12-04T09:42:35.4769315Z Setting output keep-going=True 2025-12-04T09:42:35.4769748Z Setting output ci-verbose-test-logs=False 2025-12-04T09:42:35.4770145Z Setting output ci-test-showlocals=False 2025-12-04T09:42:35.4770541Z Setting output ci-no-test-timeout=False 2025-12-04T09:42:35.4770919Z Setting output ci-no-td=False 2025-12-04T09:42:35.4771287Z Setting output ci-td-distributed=False 2025-12-04T09:42:35.4771660Z Setting output is-unstable=True 2025-12-04T09:42:35.4772008Z Setting output reenabled-issues= 2025-12-04T09:42:35.4787974Z Setting output test-matrix={"include": [{"config": "legacy_nvidia_driver", "shard": 1, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 1, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "legacy_nvidia_driver", "shard": 1, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "legacy_nvidia_driver", "shard": 1, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 2, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 2, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "legacy_nvidia_driver", "shard": 2, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "legacy_nvidia_driver", "shard": 2, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 3, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 3, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "legacy_nvidia_driver", "shard": 3, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "legacy_nvidia_driver", "shard": 3, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 4, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 4, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "legacy_nvidia_driver", "shard": 4, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "legacy_nvidia_driver", "shard": 4, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 5, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 5, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "legacy_nvidia_driver", "shard": 5, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "legacy_nvidia_driver", "shard": 5, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}]} 2025-12-04T09:42:35.4804176Z Setting output is-test-matrix-empty=False 2025-12-04T09:42:35.4919088Z ##[group]Run echo "Filtered matrix:" 2025-12-04T09:42:35.4919541Z echo "Filtered matrix:" 2025-12-04T09:42:35.4935572Z echo "{"include": [{"config": "legacy_nvidia_driver", "shard": 1, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 1, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "legacy_nvidia_driver", "shard": 1, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "legacy_nvidia_driver", "shard": 1, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 2, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 2, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "legacy_nvidia_driver", "shard": 2, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "legacy_nvidia_driver", "shard": 2, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 3, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 3, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "legacy_nvidia_driver", "shard": 3, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "legacy_nvidia_driver", "shard": 3, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 4, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 4, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "legacy_nvidia_driver", "shard": 4, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "legacy_nvidia_driver", "shard": 4, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 5, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "legacy_nvidia_driver", "shard": 5, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "legacy_nvidia_driver", "shard": 5, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "legacy_nvidia_driver", "shard": 5, "num_shards": 5, "runner": "linux.g4dn.4xlarge.nvidia.gpu", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}]}" 2025-12-04T09:42:35.4951650Z  2025-12-04T09:42:35.4951892Z echo 2025-12-04T09:42:35.4952201Z echo "Is the current job unstable? True" 2025-12-04T09:42:35.4952586Z  2025-12-04T09:42:35.4952834Z echo 2025-12-04T09:42:35.4953127Z echo "Is keep-going label set? True" 2025-12-04T09:42:35.4953505Z  2025-12-04T09:42:35.4953726Z echo 2025-12-04T09:42:35.4954006Z echo "Reenabled issues? " 2025-12-04T09:42:35.4960807Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:42:35.4961275Z env: 2025-12-04T09:42:35.4961517Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:35.4961837Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:35.4962211Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:35.4962620Z ##[endgroup] 2025-12-04T09:42:35.4989917Z Filtered matrix: 2025-12-04T09:42:35.5009470Z {include: [{config: legacy_nvidia_driver, shard: 1, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check: mem_leak_check, unstable: unstable}, {config: legacy_nvidia_driver, shard: 1, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check: mem_leak_check, unstable: unstable, rerun_disabled_tests: rerun_disabled_tests}, {config: legacy_nvidia_driver, shard: 1, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable, mem_leak_check: mem_leak_check}, {config: legacy_nvidia_driver, shard: 1, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable}, {config: legacy_nvidia_driver, shard: 2, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check: mem_leak_check, unstable: unstable}, {config: legacy_nvidia_driver, shard: 2, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check: mem_leak_check, unstable: unstable, rerun_disabled_tests: rerun_disabled_tests}, {config: legacy_nvidia_driver, shard: 2, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable, mem_leak_check: mem_leak_check}, {config: legacy_nvidia_driver, shard: 2, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable}, {config: legacy_nvidia_driver, shard: 3, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check: mem_leak_check, unstable: unstable}, {config: legacy_nvidia_driver, shard: 3, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check: mem_leak_check, unstable: unstable, rerun_disabled_tests: rerun_disabled_tests}, {config: legacy_nvidia_driver, shard: 3, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable, mem_leak_check: mem_leak_check}, {config: legacy_nvidia_driver, shard: 3, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable}, {config: legacy_nvidia_driver, shard: 4, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check: mem_leak_check, unstable: unstable}, {config: legacy_nvidia_driver, shard: 4, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check: mem_leak_check, unstable: unstable, rerun_disabled_tests: rerun_disabled_tests}, {config: legacy_nvidia_driver, shard: 4, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable, mem_leak_check: mem_leak_check}, {config: legacy_nvidia_driver, shard: 4, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable}, {config: legacy_nvidia_driver, shard: 5, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check: mem_leak_check, unstable: unstable}, {config: legacy_nvidia_driver, shard: 5, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check: mem_leak_check, unstable: unstable, rerun_disabled_tests: rerun_disabled_tests}, {config: legacy_nvidia_driver, shard: 5, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable, mem_leak_check: mem_leak_check}, {config: legacy_nvidia_driver, shard: 5, num_shards: 5, runner: linux.g4dn.4xlarge.nvidia.gpu, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable}]} 2025-12-04T09:42:35.5025061Z 2025-12-04T09:42:35.5025204Z Is the current job unstable? True 2025-12-04T09:42:35.5025451Z 2025-12-04T09:42:35.5025571Z Is keep-going label set? True 2025-12-04T09:42:35.5025784Z 2025-12-04T09:42:35.5025908Z Reenabled issues? 2025-12-04T09:42:35.5062861Z ##[group]Run echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-12-04T09:42:35.5063479Z echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-12-04T09:42:35.5069991Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:42:35.5070421Z env: 2025-12-04T09:42:35.5070676Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:35.5070986Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:35.5071338Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:35.5071767Z JOB_TIMEOUT: 600 2025-12-04T09:42:35.5072039Z ##[endgroup] 2025-12-04T09:42:35.5124091Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:42:35.5124711Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:42:35.5125243Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:42:35.5131449Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:42:35.5131908Z env: 2025-12-04T09:42:35.5132154Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:35.5132452Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:35.5132815Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:35.5133229Z ##[endgroup] 2025-12-04T09:42:35.5239783Z ##[group]Run set -x 2025-12-04T09:42:35.5240174Z set -x 2025-12-04T09:42:35.5240436Z  2025-12-04T09:42:35.5240727Z if [[ $TEST_CONFIG == 'multigpu' ]]; then 2025-12-04T09:42:35.5241182Z  TEST_COMMAND=.ci/pytorch/multigpu-test.sh 2025-12-04T09:42:35.5241650Z elif [[ $BUILD_ENVIRONMENT == *onnx* ]]; then 2025-12-04T09:42:35.5242078Z  TEST_COMMAND=.ci/onnx/test.sh 2025-12-04T09:42:35.5242421Z else 2025-12-04T09:42:35.5242714Z  TEST_COMMAND=.ci/pytorch/test.sh 2025-12-04T09:42:35.5243077Z fi 2025-12-04T09:42:35.5243300Z  2025-12-04T09:42:35.5243608Z # Leaving 1GB for the runner and other things 2025-12-04T09:42:35.5244448Z TOTAL_AVAILABLE_MEMORY_IN_GB=$(awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo) 2025-12-04T09:42:35.5245487Z # https://docs.docker.com/engine/containers/resource_constraints/#--memory-swap-details, the 3GB swap 2025-12-04T09:42:35.5246330Z # comes from https://github.com/pytorch/test-infra/pull/6058 2025-12-04T09:42:35.5246965Z TOTAL_MEMORY_WITH_SWAP=$(("${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}" + 3)) 2025-12-04T09:42:35.5247466Z  2025-12-04T09:42:35.5247758Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-12-04T09:42:35.5248169Z  SHM_OPTS= 2025-12-04T09:42:35.5248456Z  JENKINS_USER= 2025-12-04T09:42:35.5248853Z  # ensure that docker container cleanly exits in 12 hours 2025-12-04T09:42:35.5249420Z  # if for some reason cleanup action doesn't stop container 2025-12-04T09:42:35.5249898Z  # when job is cancelled 2025-12-04T09:42:35.5250258Z  DOCKER_SHELL_CMD="sleep 12h" 2025-12-04T09:42:35.5250645Z  USED_IMAGE="${DOCKER_IMAGE_S390X}" 2025-12-04T09:42:35.5251014Z else 2025-12-04T09:42:35.5251306Z  SHM_OPTS="--shm-size=${SHM_SIZE}" 2025-12-04T09:42:35.5251694Z  JENKINS_USER="--user jenkins" 2025-12-04T09:42:35.5252066Z  DOCKER_SHELL_CMD= 2025-12-04T09:42:35.5252400Z  USED_IMAGE="${DOCKER_IMAGE}" 2025-12-04T09:42:35.5252734Z fi 2025-12-04T09:42:35.5252975Z  2025-12-04T09:42:35.5253369Z # detached container should get cleaned up by teardown_ec2_linux 2025-12-04T09:42:35.5254006Z # TODO: Stop building test binaries as part of the build phase 2025-12-04T09:42:35.5254724Z # Used for GPU_FLAG, SHM_OPTS, JENKINS_USER and DOCKER_SHELL_CMD since that doesn't play nice 2025-12-04T09:42:35.5255357Z # shellcheck disable=SC2086,SC2090 2025-12-04T09:42:35.5255752Z container_name=$(docker run \ 2025-12-04T09:42:35.5256111Z  ${GPU_FLAG:-} \ 2025-12-04T09:42:35.5256464Z  ${SCCACHE_SERVER_PORT_DOCKER_FLAG:-} \ 2025-12-04T09:42:35.5256984Z  -e BUILD_ENVIRONMENT \ 2025-12-04T09:42:35.5257341Z  -e PR_NUMBER \ 2025-12-04T09:42:35.5257653Z  -e GITHUB_ACTIONS \ 2025-12-04T09:42:35.5257999Z  -e GITHUB_REPOSITORY \ 2025-12-04T09:42:35.5258358Z  -e GITHUB_WORKFLOW \ 2025-12-04T09:42:35.5258678Z  -e GITHUB_JOB \ 2025-12-04T09:42:35.5258995Z  -e GITHUB_RUN_ID \ 2025-12-04T09:42:35.5259323Z  -e GITHUB_RUN_NUMBER \ 2025-12-04T09:42:35.5259660Z  -e GITHUB_RUN_ATTEMPT \ 2025-12-04T09:42:35.5260002Z  -e JOB_ID \ 2025-12-04T09:42:35.5260300Z  -e JOB_NAME \ 2025-12-04T09:42:35.5260601Z  -e BASE_SHA \ 2025-12-04T09:42:35.5260884Z  -e BRANCH \ 2025-12-04T09:42:35.5261175Z  -e SHA1 \ 2025-12-04T09:42:35.5261468Z  -e AWS_DEFAULT_REGION \ 2025-12-04T09:42:35.5261808Z  -e IN_WHEEL_TEST \ 2025-12-04T09:42:35.5262134Z  -e SHARD_NUMBER \ 2025-12-04T09:42:35.5262454Z  -e TEST_CONFIG \ 2025-12-04T09:42:35.5262764Z  -e NUM_TEST_SHARDS \ 2025-12-04T09:42:35.5263243Z  -e REENABLED_ISSUES \ 2025-12-04T09:42:35.5263604Z  -e CONTINUE_THROUGH_ERROR \ 2025-12-04T09:42:35.5263964Z  -e VERBOSE_TEST_LOGS \ 2025-12-04T09:42:35.5264319Z  -e TEST_SHOWLOCALS \ 2025-12-04T09:42:35.5264658Z  -e NO_TEST_TIMEOUT \ 2025-12-04T09:42:35.5264987Z  -e NO_TD \ 2025-12-04T09:42:35.5265277Z  -e TD_DISTRIBUTED \ 2025-12-04T09:42:35.5281166Z  -e PR_LABELS \ 2025-12-04T09:42:35.5281617Z  -e MAX_JOBS="$(nproc --ignore=2)" \ 2025-12-04T09:42:35.5282014Z  -e SCCACHE_BUCKET \ 2025-12-04T09:42:35.5282356Z  -e SCCACHE_REGION \ 2025-12-04T09:42:35.5282687Z  -e XLA_CUDA \ 2025-12-04T09:42:35.5283162Z  -e XLA_CLANG_CACHE_S3_BUCKET_NAME \ 2025-12-04T09:42:35.5283588Z  -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK \ 2025-12-04T09:42:35.5284031Z  -e PYTORCH_TEST_RERUN_DISABLED_TESTS \ 2025-12-04T09:42:35.5284479Z  -e SKIP_SCCACHE_INITIALIZATION=1 \ 2025-12-04T09:42:35.5284870Z  -e HUGGING_FACE_HUB_TOKEN \ 2025-12-04T09:42:35.5285260Z  -e VLLM_TEST_HUGGING_FACE_TOKEN \ 2025-12-04T09:42:35.5285670Z  -e SCRIBE_GRAPHQL_ACCESS_TOKEN \ 2025-12-04T09:42:35.5286036Z  -e DASHBOARD_TAG \ 2025-12-04T09:42:35.5286373Z  -e ARTIFACTS_FILE_SUFFIX \ 2025-12-04T09:42:35.5286802Z  --memory="${TOTAL_AVAILABLE_MEMORY_IN_GB%.*}g" \ 2025-12-04T09:42:35.5287280Z  --memory-swap="${TOTAL_MEMORY_WITH_SWAP}g" \ 2025-12-04T09:42:35.5287774Z  --env-file="/tmp/github_env_${GITHUB_RUN_ID}" \ 2025-12-04T09:42:35.5288237Z  --security-opt seccomp=unconfined \ 2025-12-04T09:42:35.5288636Z  --cap-add=SYS_PTRACE \ 2025-12-04T09:42:35.5288983Z  --ipc=host \ 2025-12-04T09:42:35.5289282Z  ${SHM_OPTS} \ 2025-12-04T09:42:35.5289574Z  --tty \ 2025-12-04T09:42:35.5289835Z  --detach \ 2025-12-04T09:42:35.5290146Z  --name="${container_name}" \ 2025-12-04T09:42:35.5290521Z  ${JENKINS_USER} \ 2025-12-04T09:42:35.5290921Z  -v "${GITHUB_WORKSPACE}:/var/lib/jenkins/workspace" \ 2025-12-04T09:42:35.5291389Z  -w /var/lib/jenkins/workspace \ 2025-12-04T09:42:35.5291762Z  "${USED_IMAGE}" \ 2025-12-04T09:42:35.5292085Z  ${DOCKER_SHELL_CMD} 2025-12-04T09:42:35.5292382Z ) 2025-12-04T09:42:35.5292772Z echo "DOCKER_CONTAINER_ID=${container_name}" >> "${GITHUB_ENV}" 2025-12-04T09:42:35.5293258Z  2025-12-04T09:42:35.5293552Z if [[ ${BUILD_ENVIRONMENT} == *"s390x"* ]]; then 2025-12-04T09:42:35.5294243Z  docker exec -t "${container_name}" sh -c "python3 -m pip install -r .ci/docker/requirements-ci.txt" 2025-12-04T09:42:35.5294867Z fi 2025-12-04T09:42:35.5295105Z  2025-12-04T09:42:35.5295678Z docker exec -t "${container_name}" sh -c "python3 -m pip install $(echo dist/*.whl)[opt-einsum] && ${TEST_COMMAND}" 2025-12-04T09:42:35.5302833Z shell: /usr/bin/bash -e {0} 2025-12-04T09:42:35.5303151Z env: 2025-12-04T09:42:35.5303392Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:42:35.5303704Z HAS_NVIDIA_GPU: true 2025-12-04T09:42:35.5304070Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:35.5304566Z BUILD_ENVIRONMENT: linux-jammy-cuda12.4-py3.10-gcc11 2025-12-04T09:42:35.5304985Z PR_NUMBER: 2025-12-04T09:42:35.5305264Z GITHUB_REPOSITORY: pytorch/pytorch 2025-12-04T09:42:35.5305615Z GITHUB_WORKFLOW: periodic 2025-12-04T09:42:35.5305925Z GITHUB_JOB: test 2025-12-04T09:42:35.5306198Z GITHUB_RUN_ID: 19922826259 2025-12-04T09:42:35.5306513Z GITHUB_RUN_NUMBER: 19107 2025-12-04T09:42:35.5306820Z GITHUB_RUN_ATTEMPT: 1 2025-12-04T09:42:35.5307104Z JOB_ID: 57119749282 2025-12-04T09:42:35.5307823Z JOB_NAME: linux-jammy-cuda12.4-py3.10-gcc11 / test (legacy_nvidia_driver, 3, 5, linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check, unstable) 2025-12-04T09:42:35.5308767Z BRANCH: main 2025-12-04T09:42:35.5309083Z SHA1: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:42:35.5309544Z BASE_SHA: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:42:35.5309970Z TEST_CONFIG: legacy_nvidia_driver 2025-12-04T09:42:35.5310297Z SHARD_NUMBER: 3 2025-12-04T09:42:35.5310569Z NUM_TEST_SHARDS: 5 2025-12-04T09:42:35.5310843Z EXTRA_FLAGS: 2025-12-04T09:42:35.5311096Z OP_BENCHMARK_TESTS: 2025-12-04T09:42:35.5311386Z REENABLED_ISSUES: 2025-12-04T09:42:35.5311680Z CONTINUE_THROUGH_ERROR: True 2025-12-04T09:42:35.5311993Z VERBOSE_TEST_LOGS: False 2025-12-04T09:42:35.5312302Z TEST_SHOWLOCALS: False 2025-12-04T09:42:35.5312600Z NO_TEST_TIMEOUT: False 2025-12-04T09:42:35.5312874Z NO_TD: False 2025-12-04T09:42:35.5313248Z TD_DISTRIBUTED: False 2025-12-04T09:42:35.5313604Z SCCACHE_BUCKET: ossci-compiler-cache-circleci-v2 2025-12-04T09:42:35.5314014Z SCCACHE_REGION: us-east-1 2025-12-04T09:42:35.5314321Z SHM_SIZE: 2g 2025-12-04T09:42:35.5315246Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:42:35.5316945Z DOCKER_IMAGE_S390X: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:42:35.5317967Z XLA_CUDA: 2025-12-04T09:42:35.5318386Z XLA_CLANG_CACHE_S3_BUCKET_NAME: ossci-compiler-clang-cache-circleci-xla 2025-12-04T09:42:35.5318925Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK: 1 2025-12-04T09:42:35.5319307Z PYTORCH_TEST_RERUN_DISABLED_TESTS: 0 2025-12-04T09:42:35.5319647Z DASHBOARD_TAG: 2025-12-04T09:42:35.5320161Z VLLM_TEST_HUGGING_FACE_TOKEN: *** 2025-12-04T09:42:35.5320647Z HUGGING_FACE_HUB_TOKEN: *** 2025-12-04T09:42:35.5321136Z SCRIBE_GRAPHQL_ACCESS_TOKEN: *** 2025-12-04T09:42:35.5321752Z ARTIFACTS_FILE_SUFFIX: test-legacy_nvidia_driver-3-5-linux.g4dn.4xlarge.nvidia.gpu_57119749282 2025-12-04T09:42:35.5322392Z ##[endgroup] 2025-12-04T09:42:35.5349867Z + [[ legacy_nvidia_driver == \m\u\l\t\i\g\p\u ]] 2025-12-04T09:42:35.5350341Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *onnx* ]] 2025-12-04T09:42:35.5350827Z + TEST_COMMAND=.ci/pytorch/test.sh 2025-12-04T09:42:35.5353638Z ++ awk '/MemTotal/ { printf "%.3f \n", $2/1024/1024 - 1 }' /proc/meminfo 2025-12-04T09:42:35.5378028Z + TOTAL_AVAILABLE_MEMORY_IN_GB='61.094 ' 2025-12-04T09:42:35.5378450Z + TOTAL_MEMORY_WITH_SWAP=64 2025-12-04T09:42:35.5378835Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *\s\3\9\0\x* ]] 2025-12-04T09:42:35.5379265Z + SHM_OPTS=--shm-size=2g 2025-12-04T09:42:35.5379580Z + JENKINS_USER='--user jenkins' 2025-12-04T09:42:35.5379892Z + DOCKER_SHELL_CMD= 2025-12-04T09:42:35.5380827Z + USED_IMAGE=308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:42:35.5388546Z +++ nproc --ignore=2 2025-12-04T09:42:35.5415460Z ++ docker run --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all -e BUILD_ENVIRONMENT -e PR_NUMBER -e GITHUB_ACTIONS -e GITHUB_REPOSITORY -e GITHUB_WORKFLOW -e GITHUB_JOB -e GITHUB_RUN_ID -e GITHUB_RUN_NUMBER -e GITHUB_RUN_ATTEMPT -e JOB_ID -e JOB_NAME -e BASE_SHA -e BRANCH -e SHA1 -e AWS_DEFAULT_REGION -e IN_WHEEL_TEST -e SHARD_NUMBER -e TEST_CONFIG -e NUM_TEST_SHARDS -e REENABLED_ISSUES -e CONTINUE_THROUGH_ERROR -e VERBOSE_TEST_LOGS -e TEST_SHOWLOCALS -e NO_TEST_TIMEOUT -e NO_TD -e TD_DISTRIBUTED -e PR_LABELS -e MAX_JOBS=14 -e SCCACHE_BUCKET -e SCCACHE_REGION -e XLA_CUDA -e XLA_CLANG_CACHE_S3_BUCKET_NAME -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK -e PYTORCH_TEST_RERUN_DISABLED_TESTS -e SKIP_SCCACHE_INITIALIZATION=1 -e HUGGING_FACE_HUB_TOKEN -e VLLM_TEST_HUGGING_FACE_TOKEN -e SCRIBE_GRAPHQL_ACCESS_TOKEN -e DASHBOARD_TAG -e ARTIFACTS_FILE_SUFFIX --memory=61g --memory-swap=64g --env-file=/tmp/github_env_19922826259 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --ipc=host --shm-size=2g --tty --detach --name= --user jenkins -v /home/ec2-user/actions-runner/_work/pytorch/pytorch:/var/lib/jenkins/workspace -w /var/lib/jenkins/workspace 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:42:43.9760706Z + container_name=0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T09:42:43.9761761Z + echo DOCKER_CONTAINER_ID=0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T09:42:43.9762582Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *\s\3\9\0\x* ]] 2025-12-04T09:42:43.9766766Z ++ echo dist/torch-2.10.0a0+gitffd9b0f-cp310-cp310-linux_x86_64.whl 2025-12-04T09:42:43.9769457Z + docker exec -t 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa sh -c 'python3 -m pip install dist/torch-2.10.0a0+gitffd9b0f-cp310-cp310-linux_x86_64.whl[opt-einsum] && .ci/pytorch/test.sh' 2025-12-04T09:42:44.4715289Z Processing ./dist/torch-2.10.0a0+gitffd9b0f-cp310-cp310-linux_x86_64.whl (from torch==2.10.0a0+gitffd9b0f) 2025-12-04T09:42:45.3499624Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.18.0) 2025-12-04T09:42:45.3504819Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (4.12.2) 2025-12-04T09:42:45.3510344Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (1.13.3) 2025-12-04T09:42:45.3515894Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (2.8.8) 2025-12-04T09:42:45.3520257Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.1.6) 2025-12-04T09:42:45.3525893Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (2025.10.0) 2025-12-04T09:42:45.3541510Z Requirement already satisfied: opt-einsum>=3.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.3.0) 2025-12-04T09:42:45.3972267Z Requirement already satisfied: numpy>=1.7 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from opt-einsum>=3.3->torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (1.22.4) 2025-12-04T09:42:45.3995898Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from sympy>=1.13.3->torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (1.3.0) 2025-12-04T09:42:45.4062550Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from jinja2->torch==2.10.0a0+gitffd9b0f->torch==2.10.0a0+gitffd9b0f) (3.0.3) 2025-12-04T09:42:45.8334221Z Installing collected packages: torch 2025-12-04T09:42:58.2304041Z Successfully installed torch-2.10.0a0+gitffd9b0f 2025-12-04T09:42:58.3025324Z + export TERM=vt100 2025-12-04T09:42:58.3025672Z + TERM=vt100 2025-12-04T09:42:58.3027727Z ++ dirname .ci/pytorch/test.sh 2025-12-04T09:42:58.3036373Z + source .ci/pytorch/common.sh 2025-12-04T09:42:58.3040120Z +++ dirname .ci/pytorch/common.sh 2025-12-04T09:42:58.3047740Z ++ source .ci/pytorch/common_utils.sh 2025-12-04T09:42:58.3049154Z +++ declare -f -t trap_add 2025-12-04T09:42:58.3055090Z ++ set -ex -o pipefail 2025-12-04T09:42:58.3055544Z ++ [[ linux-jammy-cuda12.4-py3.10-gcc11 == *rocm* ]] 2025-12-04T09:42:58.3055978Z ++ BUILD_TEST_LIBTORCH=0 2025-12-04T09:42:58.3059238Z ++ dirname .ci/pytorch/test.sh 2025-12-04T09:42:58.3066895Z + source .ci/pytorch/common-build.sh 2025-12-04T09:42:58.3068732Z ++ [[ linux-jammy-cuda12.4-py3.10-gcc11 != *win-* ]] 2025-12-04T09:42:58.3075024Z ++++ dirname .ci/pytorch/common-build.sh 2025-12-04T09:42:58.3083348Z +++ cd .ci/pytorch 2025-12-04T09:42:58.3083886Z +++ pwd -P 2025-12-04T09:42:58.3089534Z ++ script_dir=/var/lib/jenkins/workspace/.ci/pytorch 2025-12-04T09:42:58.3090102Z ++ [[ linux-jammy-cuda12.4-py3.10-gcc11 == *-pch* ]] 2025-12-04T09:42:58.3090511Z ++ which sccache 2025-12-04T09:42:58.3108819Z ++ [[ -z ossci-compiler-cache-circleci-v2 ]] 2025-12-04T09:42:58.3109272Z ++ sccache --stop-server 2025-12-04T09:42:58.3135019Z ++ true 2025-12-04T09:42:58.3135374Z ++ rm -f /var/lib/jenkins/sccache_error.log 2025-12-04T09:42:58.3145639Z ++ trap_add sccache_epilogue EXIT 2025-12-04T09:42:58.3146007Z ++ trap_add_cmd=sccache_epilogue 2025-12-04T09:42:58.3146329Z ++ shift 2025-12-04T09:42:58.3146581Z ++ for trap_add_name in "$@" 2025-12-04T09:42:58.3152825Z ++++ trap -p EXIT 2025-12-04T09:42:58.3155307Z +++ eval 'extract_trap_cmd ' 2025-12-04T09:42:58.3155871Z ++++ extract_trap_cmd 2025-12-04T09:42:58.3156324Z ++++ printf '%s\n' '' 2025-12-04T09:42:58.3156703Z +++ printf '%s\n' sccache_epilogue 2025-12-04T09:42:58.3158693Z ++ trap -- ' 2025-12-04T09:42:58.3159147Z sccache_epilogue' EXIT 2025-12-04T09:42:58.3159653Z ++ [[ -n 1 ]] 2025-12-04T09:42:58.3160390Z ++ echo 'Skipping sccache server initialization, setting environment variables' 2025-12-04T09:42:58.3161410Z Skipping sccache server initialization, setting environment variables 2025-12-04T09:42:58.3161940Z ++ export SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:42:58.3162280Z ++ SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:42:58.3162675Z ++ export SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:42:58.3163203Z ++ SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:42:58.3171170Z ++ export RUST_LOG=sccache::server=error 2025-12-04T09:42:58.3171596Z ++ RUST_LOG=sccache::server=error 2025-12-04T09:42:58.3171958Z ++ sccache --zero-stats 2025-12-04T09:42:58.5169135Z Statistics zeroed. 2025-12-04T09:42:58.5170748Z ++ which ccache 2025-12-04T09:42:58.5195614Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 != *rocm* ]] 2025-12-04T09:42:58.5196325Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 != *s390x* ]] 2025-12-04T09:42:58.5196764Z + [[ -d /var/lib/jenkins/workspace ]] 2025-12-04T09:42:58.5199166Z ++ stat -c %u /var/lib/jenkins/workspace 2025-12-04T09:42:58.5213145Z + WORKSPACE_ORIGINAL_OWNER_ID=1000 2025-12-04T09:42:58.5213542Z + trap_add cleanup_workspace EXIT 2025-12-04T09:42:58.5213904Z + trap_add_cmd=cleanup_workspace 2025-12-04T09:42:58.5214212Z + shift 2025-12-04T09:42:58.5214463Z + for trap_add_name in "$@" 2025-12-04T09:42:58.5221745Z +++ trap -p EXIT 2025-12-04T09:42:58.5224103Z ++ eval 'extract_trap_cmd trap -- '\'' 2025-12-04T09:42:58.5224504Z sccache_epilogue'\'' EXIT' 2025-12-04T09:42:58.5224896Z +++ extract_trap_cmd trap -- ' 2025-12-04T09:42:58.5225225Z sccache_epilogue' EXIT 2025-12-04T09:42:58.5225526Z +++ printf '%s\n' ' 2025-12-04T09:42:58.5225801Z sccache_epilogue' 2025-12-04T09:42:58.5226093Z ++ printf '%s\n' cleanup_workspace 2025-12-04T09:42:58.5226605Z + trap -- ' 2025-12-04T09:42:58.5226843Z sccache_epilogue 2025-12-04T09:42:58.5227129Z cleanup_workspace' EXIT 2025-12-04T09:42:58.5227490Z + sudo chown -R jenkins /var/lib/jenkins/workspace 2025-12-04T09:42:59.2444307Z + git config --global --add safe.directory /var/lib/jenkins/workspace 2025-12-04T09:42:59.2461493Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *cuda* ]] 2025-12-04T09:42:59.2464713Z ++ python -c 'import os;import numba.cuda; print(os.path.dirname(numba.cuda.__file__))' 2025-12-04T09:42:59.7331717Z + NUMBA_CUDA_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/numba/cuda 2025-12-04T09:42:59.7332470Z + '[' -n /opt/conda/envs/py_3.10/lib/python3.10/site-packages/numba/cuda ']' 2025-12-04T09:42:59.7337637Z +++ realpath .ci/pytorch/test.sh 2025-12-04T09:42:59.7348062Z ++ dirname /var/lib/jenkins/workspace/.ci/pytorch/test.sh 2025-12-04T09:42:59.7367228Z + NUMBA_PATCH=/var/lib/jenkins/workspace/.ci/pytorch/numba-cuda-13.patch 2025-12-04T09:42:59.7367892Z + pushd /opt/conda/envs/py_3.10/lib/python3.10/site-packages/numba/cuda 2025-12-04T09:42:59.7369831Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/numba/cuda ~/workspace 2025-12-04T09:42:59.7370439Z + patch -p4 2025-12-04T09:42:59.7383793Z patching file cudadrv/driver.py 2025-12-04T09:42:59.7384594Z Hunk #1 succeeded at 357 (offset -8 lines). 2025-12-04T09:42:59.7394555Z + popd 2025-12-04T09:42:59.7394817Z ~/workspace 2025-12-04T09:42:59.7395080Z + echo 'Environment variables:' 2025-12-04T09:42:59.7395421Z Environment variables: 2025-12-04T09:42:59.7395702Z + env 2025-12-04T09:42:59.7404803Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:42:59.7405538Z CONTINUE_THROUGH_ERROR=True 2025-12-04T09:42:59.7405992Z BUILD_ENVIRONMENT=linux-jammy-cuda12.4-py3.10-gcc11 2025-12-04T09:42:59.7406683Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-12-04T09:42:59.7407284Z HOSTNAME=0b946b67a9e0 2025-12-04T09:42:59.7407970Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_4a201f1c-8322-403a-9264-355aace8db6b 2025-12-04T09:42:59.7408724Z GITHUB_ACTION=__run_3 2025-12-04T09:42:59.7409038Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 2025-12-04T09:42:59.7409373Z GITHUB_RUN_NUMBER=19107 2025-12-04T09:42:59.7409679Z TEST_CONFIG=legacy_nvidia_driver 2025-12-04T09:42:59.7410028Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-12-04T09:42:59.7410395Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-12-04T09:42:59.7410802Z SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:42:59.7411261Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-12-04T09:42:59.7411591Z GITHUB_TRIGGERING_ACTOR=huydhn 2025-12-04T09:42:59.7411957Z GITHUB_REF_TYPE=branch 2025-12-04T09:42:59.7412298Z BASE_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:42:59.7412690Z XLA_CUDA= 2025-12-04T09:42:59.7412934Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-12-04T09:42:59.7413405Z HUGGING_FACE_HUB_TOKEN=*** 2025-12-04T09:42:59.7413927Z *** 2025-12-04T09:42:59.7414170Z GITHUB_REPOSITORY_ID=65600975 2025-12-04T09:42:59.7414479Z GITHUB_ACTIONS=true 2025-12-04T09:42:59.7414768Z NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:59.7415167Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:42:59.7415620Z SHA1=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:42:59.7416069Z GITHUB_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:42:59.7416703Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/periodic.yml@refs/heads/main 2025-12-04T09:42:59.7417340Z UCC_HOME=/usr 2025-12-04T09:42:59.7417586Z VERBOSE_TEST_LOGS=False 2025-12-04T09:42:59.7417889Z GITHUB_REF=refs/heads/main 2025-12-04T09:42:59.7418186Z SHARD_NUMBER=3 2025-12-04T09:42:59.7418441Z GITHUB_REF_PROTECTED=true 2025-12-04T09:42:59.7418745Z HOME=/var/lib/jenkins 2025-12-04T09:42:59.7419066Z GITHUB_API_URL=https://api.github.com 2025-12-04T09:42:59.7419435Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-12-04T09:42:59.7419849Z UCX_COMMIT=7836b165abdbe468a2f607e7254011c07d788152 2025-12-04T09:42:59.7420245Z USE_SYSTEM_NCCL=1 2025-12-04T09:42:59.7420497Z NUM_TEST_SHARDS=5 2025-12-04T09:42:59.7420754Z UCX_HOME=/usr 2025-12-04T09:42:59.7421421Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_4a201f1c-8322-403a-9264-355aace8db6b 2025-12-04T09:42:59.7422615Z JOB_NAME=linux-jammy-cuda12.4-py3.10-gcc11 / test (legacy_nvidia_driver, 3, 5, linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check, unstable) 2025-12-04T09:42:59.7423767Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_4a201f1c-8322-403a-9264-355aace8db6b 2025-12-04T09:42:59.7424725Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-12-04T09:42:59.7425321Z GITHUB_EVENT_NAME=schedule 2025-12-04T09:42:59.7425615Z DASHBOARD_TAG= 2025-12-04T09:42:59.7425879Z GITHUB_RUN_ID=19922826259 2025-12-04T09:42:59.7426180Z INSTALLED_OPENBLAS= 2025-12-04T09:42:59.7426893Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_4a201f1c-8322-403a-9264-355aace8db6b 2025-12-04T09:42:59.7427698Z GITHUB_ACTOR=huydhn 2025-12-04T09:42:59.7427957Z PR_NUMBER= 2025-12-04T09:42:59.7428195Z DESIRED_CUDA=12.4 2025-12-04T09:42:59.7428606Z GITHUB_RUN_ATTEMPT=1 2025-12-04T09:42:59.7428909Z ANACONDA_PYTHON_VERSION=3.10 2025-12-04T09:42:59.7429302Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-12-04T09:42:59.7429695Z TERM=vt100 2025-12-04T09:42:59.7429942Z INSTALLED_VISION=yes 2025-12-04T09:42:59.7430217Z BRANCH=main 2025-12-04T09:42:59.7430457Z SCCACHE_REGION=us-east-1 2025-12-04T09:42:59.7430772Z OPENSSL_ROOT_DIR=/opt/openssl 2025-12-04T09:42:59.7431098Z BUILD_AOT_INDUCTOR_TEST= 2025-12-04T09:42:59.7431387Z CUDA_PATH=/usr/local/cuda 2025-12-04T09:42:59.7432002Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-12-04T09:42:59.7432687Z GITHUB_SERVER_URL=https://github.com 2025-12-04T09:42:59.7433111Z UCC_COMMIT=430e241bf5d38cbc73fc7a6b89155397232e3f96 2025-12-04T09:42:59.7433580Z REENABLED_ISSUES= 2025-12-04T09:42:59.7433840Z DOCS= 2025-12-04T09:42:59.7434063Z SHLVL=1 2025-12-04T09:42:59.7434280Z MAX_JOBS=14 2025-12-04T09:42:59.7434531Z GITHUB_ACTOR_ID=475357 2025-12-04T09:42:59.7434935Z GITHUB_WORKFLOW_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:42:59.7435395Z GITHUB_REF_NAME=main 2025-12-04T09:42:59.7435833Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-12-04T09:42:59.7436341Z GITHUB_JOB=test 2025-12-04T09:42:59.7436611Z NO_TEST_TIMEOUT=False 2025-12-04T09:42:59.7436895Z TD_DISTRIBUTED=False 2025-12-04T09:42:59.7437204Z GITHUB_REPOSITORY=pytorch/pytorch 2025-12-04T09:42:59.7437558Z GITHUB_RETENTION_DAYS=90 2025-12-04T09:42:59.7437854Z OPENSSL_DIR=/opt/openssl 2025-12-04T09:42:59.7438163Z GITHUB_ACTION_REPOSITORY= 2025-12-04T09:42:59.7439095Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:42:59.7440053Z GITHUB_BASE_REF= 2025-12-04T09:42:59.7440320Z INSTALLED_ACL= 2025-12-04T09:42:59.7440861Z ARTIFACTS_FILE_SUFFIX=test-legacy_nvidia_driver-3-5-linux.g4dn.4xlarge.nvidia.gpu_57119749282 2025-12-04T09:42:59.7441485Z CI=true 2025-12-04T09:42:59.7441728Z GITHUB_REPOSITORY_OWNER=pytorch 2025-12-04T09:42:59.7442098Z RUST_LOG=sccache::server=error 2025-12-04T09:42:59.7442411Z JOB_ID=57119749282 2025-12-04T09:42:59.7442658Z GITHUB_HEAD_REF= 2025-12-04T09:42:59.7442917Z GITHUB_ACTION_REF= 2025-12-04T09:42:59.7443248Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-12-04T09:42:59.7443644Z TEST_SHOWLOCALS=False 2025-12-04T09:42:59.7443938Z GITHUB_WORKFLOW=periodic 2025-12-04T09:42:59.7444257Z DEBIAN_FRONTEND=noninteractive 2025-12-04T09:42:59.7444983Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_4a201f1c-8322-403a-9264-355aace8db6b 2025-12-04T09:42:59.7445739Z NO_TD=False 2025-12-04T09:42:59.7446012Z SKIP_SCCACHE_INITIALIZATION=1 2025-12-04T09:42:59.7446363Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-12-04T09:42:59.7446727Z _=/usr/bin/env 2025-12-04T09:42:59.7447138Z OLDPWD=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/numba/cuda 2025-12-04T09:42:59.7447763Z ++ python -c 'import site; print(site.getsitepackages()[0])' 2025-12-04T09:42:59.7552325Z + TORCH_INSTALL_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch 2025-12-04T09:42:59.7553050Z + TORCH_BIN_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/bin 2025-12-04T09:42:59.7553832Z + TORCH_LIB_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib 2025-12-04T09:42:59.7554524Z + TORCH_TEST_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/test 2025-12-04T09:42:59.7555057Z + BUILD_DIR=build 2025-12-04T09:42:59.7555344Z + BUILD_RENAMED_DIR=build_renamed 2025-12-04T09:42:59.7555686Z + BUILD_BIN_DIR=build/bin 2025-12-04T09:42:59.7555984Z + SHARD_NUMBER=3 2025-12-04T09:42:59.7556289Z + NUM_TEST_SHARDS=5 2025-12-04T09:42:59.7556595Z + export TORCH_SERIALIZATION_DEBUG=1 2025-12-04T09:42:59.7556948Z + TORCH_SERIALIZATION_DEBUG=1 2025-12-04T09:42:59.7557272Z + export VALGRIND=ON 2025-12-04T09:42:59.7557548Z + VALGRIND=ON 2025-12-04T09:42:59.7558025Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *clang9* ]] 2025-12-04T09:42:59.7558498Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *xpu* ]] 2025-12-04T09:42:59.7558908Z + detect_cuda_arch 2025-12-04T09:42:59.7559223Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *cuda* ]] 2025-12-04T09:42:59.7559638Z + command -v nvidia-smi 2025-12-04T09:42:59.7559930Z /usr/bin/nvidia-smi 2025-12-04T09:42:59.7563916Z ++ nvidia-smi --query-gpu=compute_cap --format=csv 2025-12-04T09:42:59.7564673Z ++ tail -n 1 2025-12-04T09:42:59.7793487Z + TORCH_CUDA_ARCH_LIST=7.5 2025-12-04T09:42:59.7793878Z + export TORCH_CUDA_ARCH_LIST 2025-12-04T09:42:59.7794259Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *s390x* ]] 2025-12-04T09:42:59.7794656Z + [[ 0 == \1 ]] 2025-12-04T09:42:59.7794908Z + [[ True == \1 ]] 2025-12-04T09:42:59.7795478Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 != *bazel* ]] 2025-12-04T09:42:59.7797780Z ++ realpath build/custom_test_artifacts 2025-12-04T09:42:59.7816663Z + CUSTOM_TEST_ARTIFACT_BUILD_DIR=/var/lib/jenkins/workspace/build/custom_test_artifacts 2025-12-04T09:42:59.7817318Z + [[ -n '' ]] 2025-12-04T09:42:59.7817578Z + echo 'Environment variables' 2025-12-04T09:42:59.7817906Z Environment variables 2025-12-04T09:42:59.7818175Z + env 2025-12-04T09:42:59.7836111Z GITHUB_WORKSPACE=/home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T09:42:59.7836757Z CONTINUE_THROUGH_ERROR=True 2025-12-04T09:42:59.7837215Z BUILD_ENVIRONMENT=linux-jammy-cuda12.4-py3.10-gcc11 2025-12-04T09:42:59.7838062Z VLLM_TEST_HUGGING_FACE_TOKEN=*** 2025-12-04T09:42:59.7838384Z HOSTNAME=0b946b67a9e0 2025-12-04T09:42:59.7839091Z GITHUB_PATH=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/add_path_4a201f1c-8322-403a-9264-355aace8db6b 2025-12-04T09:42:59.7839881Z GITHUB_ACTION=__run_3 2025-12-04T09:42:59.7840197Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 2025-12-04T09:42:59.7840547Z GITHUB_RUN_NUMBER=19107 2025-12-04T09:42:59.7840855Z TEST_CONFIG=legacy_nvidia_driver 2025-12-04T09:42:59.7841204Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-12-04T09:42:59.7841584Z TORCH_NVCC_FLAGS=-Xfatbin -compress-all 2025-12-04T09:42:59.7841951Z SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:42:59.7842451Z SCRIBE_GRAPHQL_ACCESS_TOKEN=*** 2025-12-04T09:42:59.7842794Z GITHUB_TRIGGERING_ACTOR=huydhn 2025-12-04T09:42:59.7843102Z GITHUB_REF_TYPE=branch 2025-12-04T09:42:59.7843387Z TORCH_CUDA_ARCH_LIST=7.5 2025-12-04T09:42:59.7843741Z BASE_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:42:59.7844119Z XLA_CUDA= 2025-12-04T09:42:59.7844377Z NCCL_LIB_DIR=/usr/local/cuda/lib64/ 2025-12-04T09:42:59.7844992Z HUGGING_FACE_HUB_TOKEN=*** 2025-12-04T09:42:59.7845431Z *** 2025-12-04T09:42:59.7845768Z GITHUB_REPOSITORY_ID=65600975 2025-12-04T09:42:59.7846128Z GITHUB_ACTIONS=true 2025-12-04T09:42:59.7846511Z NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T09:42:59.7846947Z SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:42:59.7847416Z SHA1=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:42:59.7847869Z GITHUB_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:42:59.7848508Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/periodic.yml@refs/heads/main 2025-12-04T09:42:59.7849072Z UCC_HOME=/usr 2025-12-04T09:42:59.7849343Z TORCH_SERIALIZATION_DEBUG=1 2025-12-04T09:42:59.7849667Z VERBOSE_TEST_LOGS=False 2025-12-04T09:42:59.7849956Z GITHUB_REF=refs/heads/main 2025-12-04T09:42:59.7850258Z SHARD_NUMBER=3 2025-12-04T09:42:59.7850531Z GITHUB_REF_PROTECTED=true 2025-12-04T09:42:59.7850822Z HOME=/var/lib/jenkins 2025-12-04T09:42:59.7851148Z GITHUB_API_URL=https://api.github.com 2025-12-04T09:42:59.7851538Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-12-04T09:42:59.7851926Z UCX_COMMIT=7836b165abdbe468a2f607e7254011c07d788152 2025-12-04T09:42:59.7852328Z USE_SYSTEM_NCCL=1 2025-12-04T09:42:59.7852594Z NUM_TEST_SHARDS=5 2025-12-04T09:42:59.7852842Z UCX_HOME=/usr 2025-12-04T09:42:59.7853506Z GITHUB_STATE=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/save_state_4a201f1c-8322-403a-9264-355aace8db6b 2025-12-04T09:42:59.7854925Z JOB_NAME=linux-jammy-cuda12.4-py3.10-gcc11 / test (legacy_nvidia_driver, 3, 5, linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check, unstable) 2025-12-04T09:42:59.7856092Z GITHUB_ENV=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_env_4a201f1c-8322-403a-9264-355aace8db6b 2025-12-04T09:42:59.7857120Z GITHUB_EVENT_PATH=/home/ec2-user/actions-runner/_work/_temp/_github_workflow/event.json 2025-12-04T09:42:59.7857720Z GITHUB_EVENT_NAME=schedule 2025-12-04T09:42:59.7858026Z DASHBOARD_TAG= 2025-12-04T09:42:59.7858277Z GITHUB_RUN_ID=19922826259 2025-12-04T09:42:59.7858578Z INSTALLED_OPENBLAS= 2025-12-04T09:42:59.7859300Z GITHUB_STEP_SUMMARY=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/step_summary_4a201f1c-8322-403a-9264-355aace8db6b 2025-12-04T09:42:59.7860108Z GITHUB_ACTOR=huydhn 2025-12-04T09:42:59.7860475Z PR_NUMBER= 2025-12-04T09:42:59.7860722Z DESIRED_CUDA=12.4 2025-12-04T09:42:59.7860995Z GITHUB_RUN_ATTEMPT=1 2025-12-04T09:42:59.7861261Z VALGRIND=ON 2025-12-04T09:42:59.7861528Z ANACONDA_PYTHON_VERSION=3.10 2025-12-04T09:42:59.7861928Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-12-04T09:42:59.7862324Z TERM=vt100 2025-12-04T09:42:59.7862568Z INSTALLED_VISION=yes 2025-12-04T09:42:59.7862844Z BRANCH=main 2025-12-04T09:42:59.7863084Z SCCACHE_REGION=us-east-1 2025-12-04T09:42:59.7863396Z OPENSSL_ROOT_DIR=/opt/openssl 2025-12-04T09:42:59.7863723Z BUILD_AOT_INDUCTOR_TEST= 2025-12-04T09:42:59.7864012Z CUDA_PATH=/usr/local/cuda 2025-12-04T09:42:59.7864630Z GITHUB_ACTION_PATH=/home/ec2-user/actions-runner/_work/pytorch/pytorch/./.github/actions/setup-linux 2025-12-04T09:42:59.7865329Z GITHUB_SERVER_URL=https://github.com 2025-12-04T09:42:59.7865732Z UCC_COMMIT=430e241bf5d38cbc73fc7a6b89155397232e3f96 2025-12-04T09:42:59.7866137Z REENABLED_ISSUES= 2025-12-04T09:42:59.7866398Z DOCS= 2025-12-04T09:42:59.7866626Z SHLVL=1 2025-12-04T09:42:59.7866844Z MAX_JOBS=14 2025-12-04T09:42:59.7867099Z GITHUB_ACTOR_ID=475357 2025-12-04T09:42:59.7867498Z GITHUB_WORKFLOW_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:42:59.7867949Z GITHUB_REF_NAME=main 2025-12-04T09:42:59.7868397Z XLA_CLANG_CACHE_S3_BUCKET_NAME=ossci-compiler-clang-cache-circleci-xla 2025-12-04T09:42:59.7868911Z GITHUB_JOB=test 2025-12-04T09:42:59.7869169Z NO_TEST_TIMEOUT=False 2025-12-04T09:42:59.7869459Z TD_DISTRIBUTED=False 2025-12-04T09:42:59.7869764Z GITHUB_REPOSITORY=pytorch/pytorch 2025-12-04T09:42:59.7870098Z GITHUB_RETENTION_DAYS=90 2025-12-04T09:42:59.7870406Z OPENSSL_DIR=/opt/openssl 2025-12-04T09:42:59.7870713Z GITHUB_ACTION_REPOSITORY= 2025-12-04T09:42:59.7871626Z PATH=/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:42:59.7872591Z GITHUB_BASE_REF= 2025-12-04T09:42:59.7872852Z INSTALLED_ACL= 2025-12-04T09:42:59.7873398Z ARTIFACTS_FILE_SUFFIX=test-legacy_nvidia_driver-3-5-linux.g4dn.4xlarge.nvidia.gpu_57119749282 2025-12-04T09:42:59.7874008Z CI=true 2025-12-04T09:42:59.7874265Z GITHUB_REPOSITORY_OWNER=pytorch 2025-12-04T09:42:59.7874646Z RUST_LOG=sccache::server=error 2025-12-04T09:42:59.7874947Z JOB_ID=57119749282 2025-12-04T09:42:59.7875211Z GITHUB_HEAD_REF= 2025-12-04T09:42:59.7875473Z GITHUB_ACTION_REF= 2025-12-04T09:42:59.7875797Z SCCACHE_BUCKET=ossci-compiler-cache-circleci-v2 2025-12-04T09:42:59.7876208Z TEST_SHOWLOCALS=False 2025-12-04T09:42:59.7876499Z GITHUB_WORKFLOW=periodic 2025-12-04T09:42:59.7876801Z DEBIAN_FRONTEND=noninteractive 2025-12-04T09:42:59.7877539Z GITHUB_OUTPUT=/home/ec2-user/actions-runner/_work/_temp/_runner_file_commands/set_output_4a201f1c-8322-403a-9264-355aace8db6b 2025-12-04T09:42:59.7878288Z NO_TD=False 2025-12-04T09:42:59.7878549Z SKIP_SCCACHE_INITIALIZATION=1 2025-12-04T09:42:59.7878895Z NCCL_INCLUDE_DIR=/usr/local/cuda/include/ 2025-12-04T09:42:59.7879427Z OLDPWD=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/numba/cuda 2025-12-04T09:42:59.7879932Z _=/usr/bin/env 2025-12-04T09:42:59.7880190Z + echo 'Testing pytorch' 2025-12-04T09:42:59.7880487Z Testing pytorch 2025-12-04T09:42:59.7880855Z + export LANG=C.UTF-8 2025-12-04T09:42:59.7881128Z + LANG=C.UTF-8 2025-12-04T09:42:59.7881399Z + PR_NUMBER= 2025-12-04T09:42:59.7881683Z + [[ legacy_nvidia_driver == \d\e\f\a\u\l\t ]] 2025-12-04T09:42:59.7882103Z + [[ legacy_nvidia_driver == \d\i\s\t\r\i\b\u\t\e\d ]] 2025-12-04T09:42:59.7882528Z + [[ legacy_nvidia_driver == \s\l\o\w ]] 2025-12-04T09:42:59.7882993Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *slow-gradcheck* ]] 2025-12-04T09:42:59.7883504Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *cuda* ]] 2025-12-04T09:42:59.7884046Z + export PYTORCH_TESTING_DEVICE_ONLY_FOR=cuda 2025-12-04T09:42:59.7884644Z + PYTORCH_TESTING_DEVICE_ONLY_FOR=cuda 2025-12-04T09:42:59.7885176Z + [[ legacy_nvidia_driver == *crossref* ]] 2025-12-04T09:42:59.7885891Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *rocm* ]] 2025-12-04T09:42:59.7886574Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *xpu* ]] 2025-12-04T09:42:59.7887069Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 != *-bazel-* ]] 2025-12-04T09:42:59.7887497Z + pip_install ninja==1.10.2 2025-12-04T09:42:59.7888005Z + pip_install_pkg='python3 -m pip install --progress-bar off' 2025-12-04T09:42:59.7888614Z + python3 -m pip install --progress-bar off ninja==1.10.2 2025-12-04T09:43:00.2236543Z Collecting ninja==1.10.2 2025-12-04T09:43:00.2493536Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl.metadata (5.0 kB) 2025-12-04T09:43:00.2606171Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl (108 kB) 2025-12-04T09:43:00.6886045Z Installing collected packages: ninja 2025-12-04T09:43:00.6886470Z Attempting uninstall: ninja 2025-12-04T09:43:00.6895501Z Found existing installation: ninja 1.11.1.4 2025-12-04T09:43:00.6919890Z Uninstalling ninja-1.11.1.4: 2025-12-04T09:43:00.6986366Z Successfully uninstalled ninja-1.11.1.4 2025-12-04T09:43:00.7369495Z Successfully installed ninja-1.10.2 2025-12-04T09:43:00.8083734Z + export PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:43:00.8085665Z + PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/usr/local/nvidia/bin:/usr/local/cuda/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:43:00.8086848Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *aarch64* ]] 2025-12-04T09:43:00.8087314Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *asan* ]] 2025-12-04T09:43:00.8087785Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *-debug* ]] 2025-12-04T09:43:00.8088262Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 != *-bazel-* ]] 2025-12-04T09:43:00.8088930Z + echo 'We are not in debug mode: linux-jammy-cuda12.4-py3.10-gcc11. Expect the assertion to pass' 2025-12-04T09:43:00.8089757Z We are not in debug mode: linux-jammy-cuda12.4-py3.10-gcc11. Expect the assertion to pass 2025-12-04T09:43:00.8090328Z + cd test 2025-12-04T09:43:00.8090735Z + python -c 'import torch; torch._C._crash_if_debug_asserts_fail(424242)' 2025-12-04T09:43:02.6376662Z + [[ legacy_nvidia_driver == \n\o\g\p\u\_\N\O\_\A\V\X\2 ]] 2025-12-04T09:43:02.6377240Z + [[ legacy_nvidia_driver == \n\o\g\p\u\_\A\V\X\5\1\2 ]] 2025-12-04T09:43:02.6377774Z + [[ legacy_nvidia_driver == \l\e\g\a\c\y\_\n\v\i\d\i\a\_\d\r\i\v\e\r ]] 2025-12-04T09:43:02.6379161Z + cd test 2025-12-04T09:43:02.6379861Z + python -c 'import torch; torch.rand(2, 2, device='\''cuda'\'')' 2025-12-04T09:43:07.5314603Z + export USE_LEGACY_DRIVER=1 2025-12-04T09:43:07.5314973Z + USE_LEGACY_DRIVER=1 2025-12-04T09:43:07.5320854Z + DYNAMO_BENCHMARK_FLAGS=() 2025-12-04T09:43:07.5322465Z + [[ legacy_nvidia_driver == *pr_time_benchmarks* ]] 2025-12-04T09:43:07.5322912Z + [[ legacy_nvidia_driver == *dynamo_eager* ]] 2025-12-04T09:43:07.5323344Z + [[ legacy_nvidia_driver == *aot_eager* ]] 2025-12-04T09:43:07.5323763Z + [[ legacy_nvidia_driver == *aot_inductor* ]] 2025-12-04T09:43:07.5324202Z + [[ legacy_nvidia_driver == *max_autotune_inductor* ]] 2025-12-04T09:43:07.5324879Z + [[ legacy_nvidia_driver == *inductor* ]] 2025-12-04T09:43:07.5325268Z + [[ legacy_nvidia_driver == *dynamic* ]] 2025-12-04T09:43:07.5325641Z + [[ legacy_nvidia_driver == *cpu* ]] 2025-12-04T09:43:07.5325992Z + [[ legacy_nvidia_driver == *xpu* ]] 2025-12-04T09:43:07.5326378Z + DYNAMO_BENCHMARK_FLAGS+=(--device cuda) 2025-12-04T09:43:07.5358835Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *libtorch* ]] 2025-12-04T09:43:07.5359322Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *-bazel-* ]] 2025-12-04T09:43:07.5362361Z + cd test 2025-12-04T09:43:07.5363171Z + python -c 'import torch; print(torch.__config__.show())' 2025-12-04T09:43:10.3845923Z PyTorch built with: 2025-12-04T09:43:10.3846255Z - GCC 11.4 2025-12-04T09:43:10.3846523Z - C++ Version: 201703 2025-12-04T09:43:10.3847539Z - Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-12-04T09:43:10.3848395Z - Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-12-04T09:43:10.3848939Z - OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-12-04T09:43:10.3849344Z - LAPACK is enabled (usually provided by MKL) 2025-12-04T09:43:10.3849725Z - NNPACK is enabled 2025-12-04T09:43:10.3850030Z - CPU capability usage: AVX512 2025-12-04T09:43:10.3850364Z - CUDA Runtime 12.4 2025-12-04T09:43:10.3850764Z - NVCC architecture flags: -gencode;arch=compute_75,code=sm_75 2025-12-04T09:43:10.3851214Z - CuDNN 90.1 2025-12-04T09:43:10.3857017Z - Build settings: BLAS_INFO=mkl, BUILD_TYPE=Release, COMMIT_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32, CUDA_VERSION=12.4, CUDNN_VERSION=9.1.0, CXX_COMPILER=/opt/cache/bin/c++, CXX_FLAGS= -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOROCTRACER -DLIBKINETO_NOXPUPTI=ON -DUSE_FBGEMM -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -DC10_NODEPRECATED -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=range-loop-construct -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-unknown-pragmas -Wno-unused-parameter -Wno-strict-overflow -Wno-strict-aliasing -Wno-stringop-overflow -Wsuggest-override -Wno-psabi -Wno-error=old-style-cast -faligned-new -Werror -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Wno-stringop-overflow, FORCE_FALLBACK_CUDA_MPI=1, LAPACK_INFO=mkl, PERF_WITH_AVX=1, PERF_WITH_AVX2=1, TORCH_VERSION=2.10.0, USE_CUDA=ON, USE_CUDNN=ON, USE_CUSPARSELT=ON, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_GLOO=ON, USE_MKL=ON, USE_MKLDNN=ON, USE_MPI=ON, USE_NCCL=ON, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=OFF, USE_ROCM_KERNEL_ASSERT=OFF, USE_XCCL=OFF, USE_XPU=OFF, 2025-12-04T09:43:10.3862955Z 2025-12-04T09:43:10.7824173Z + cd test 2025-12-04T09:43:10.7824658Z + python -c 'import torch; print(torch.__config__.parallel_info())' 2025-12-04T09:43:12.2396345Z ATen/Parallel: 2025-12-04T09:43:12.2396756Z at::get_num_threads() : 8 2025-12-04T09:43:12.2397123Z at::get_num_interop_threads() : 8 2025-12-04T09:43:12.2397491Z OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-12-04T09:43:12.2397852Z omp_get_max_threads() : 8 2025-12-04T09:43:12.2398525Z Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-12-04T09:43:12.2399240Z mkl_get_max_threads() : 8 2025-12-04T09:43:12.2399680Z Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-12-04T09:43:12.2400208Z std::thread::hardware_concurrency() : 16 2025-12-04T09:43:12.2400584Z Environment variables: 2025-12-04T09:43:12.2400876Z OMP_NUM_THREADS : [not set] 2025-12-04T09:43:12.2401200Z MKL_NUM_THREADS : [not set] 2025-12-04T09:43:12.2401527Z ATen parallel backend: OpenMP 2025-12-04T09:43:12.2401747Z 2025-12-04T09:43:12.5379779Z + [[ legacy_nvidia_driver == *numpy_2* ]] 2025-12-04T09:43:12.5380501Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *aarch64* ]] 2025-12-04T09:43:12.5380993Z + [[ legacy_nvidia_driver == *backward* ]] 2025-12-04T09:43:12.5381528Z + [[ legacy_nvidia_driver == *libtorch_agnostic_targetting* ]] 2025-12-04T09:43:12.5382294Z + [[ legacy_nvidia_driver == *xla* ]] 2025-12-04T09:43:12.5382668Z + [[ legacy_nvidia_driver == *vllm* ]] 2025-12-04T09:43:12.5383053Z + [[ legacy_nvidia_driver == *executorch* ]] 2025-12-04T09:43:12.5383472Z + [[ legacy_nvidia_driver == \j\i\t\_\l\e\g\a\c\y ]] 2025-12-04T09:43:12.5383911Z + [[ legacy_nvidia_driver == \q\u\a\n\t\i\z\a\t\i\o\n ]] 2025-12-04T09:43:12.5384383Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *libtorch* ]] 2025-12-04T09:43:12.5384828Z + [[ legacy_nvidia_driver == distributed ]] 2025-12-04T09:43:12.5385237Z + [[ legacy_nvidia_driver == *operator_benchmark* ]] 2025-12-04T09:43:12.5385731Z + [[ legacy_nvidia_driver == *operator_microbenchmark* ]] 2025-12-04T09:43:12.5386268Z + [[ legacy_nvidia_driver == *attention_microbenchmark* ]] 2025-12-04T09:43:12.5387129Z + [[ legacy_nvidia_driver == *inductor_distributed* ]] 2025-12-04T09:43:12.5387571Z + [[ legacy_nvidia_driver == *inductor-halide* ]] 2025-12-04T09:43:12.5388010Z + [[ legacy_nvidia_driver == *inductor-pallas* ]] 2025-12-04T09:43:12.5388474Z + [[ legacy_nvidia_driver == *inductor-triton-cpu* ]] 2025-12-04T09:43:12.5388959Z + [[ legacy_nvidia_driver == *inductor-micro-benchmark* ]] 2025-12-04T09:43:12.5389487Z + [[ legacy_nvidia_driver == *aoti_cross_compile_for_windows* ]] 2025-12-04T09:43:12.5389977Z + [[ legacy_nvidia_driver == *huggingface* ]] 2025-12-04T09:43:12.5390363Z + [[ legacy_nvidia_driver == *timm* ]] 2025-12-04T09:43:12.5390739Z + [[ legacy_nvidia_driver == cachebench ]] 2025-12-04T09:43:12.5391148Z + [[ legacy_nvidia_driver == verify_cachebench ]] 2025-12-04T09:43:12.5391569Z + [[ legacy_nvidia_driver == *torchbench* ]] 2025-12-04T09:43:12.5391990Z + [[ legacy_nvidia_driver == *inductor_cpp_wrapper* ]] 2025-12-04T09:43:12.5392438Z + [[ legacy_nvidia_driver == *inductor_core* ]] 2025-12-04T09:43:12.5392854Z + [[ legacy_nvidia_driver == *inductor* ]] 2025-12-04T09:43:12.5393223Z + [[ legacy_nvidia_driver == *einops* ]] 2025-12-04T09:43:12.5393617Z + [[ legacy_nvidia_driver == *dynamo_core* ]] 2025-12-04T09:43:12.5394036Z + [[ legacy_nvidia_driver == *dynamo_wrapped* ]] 2025-12-04T09:43:12.5394466Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *rocm* ]] 2025-12-04T09:43:12.5394867Z + [[ 3 == 1 ]] 2025-12-04T09:43:12.5395114Z + [[ 3 == 2 ]] 2025-12-04T09:43:12.5395358Z + [[ 3 -gt 2 ]] 2025-12-04T09:43:12.5395626Z + install_torchvision 2025-12-04T09:43:12.5395927Z + local orig_preload 2025-12-04T09:43:12.5396440Z + local commit 2025-12-04T09:43:12.5396715Z ++ get_pinned_commit vision 2025-12-04T09:43:12.5397050Z ++ cat .github/ci_commit_pins/vision.txt 2025-12-04T09:43:12.5401901Z + commit=617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:43:12.5402317Z + orig_preload= 2025-12-04T09:43:12.5402631Z + '[' -n '' ']' 2025-12-04T09:43:12.5403090Z + [[ linux-jammy-cuda12.4-py3.10-gcc11 == *cuda* ]] 2025-12-04T09:43:12.5403565Z + export FORCE_CUDA=1 2025-12-04T09:43:12.5403890Z + FORCE_CUDA=1 2025-12-04T09:43:12.5404150Z + export WITH_CUDA=1 2025-12-04T09:43:12.5404433Z + WITH_CUDA=1 2025-12-04T09:43:12.5405096Z + pip_build_and_install git+https://github.com/pytorch/vision.git@617079d944b0e72632311c30ae2bbdf1168b901e dist/vision 2025-12-04T09:43:12.5406159Z + local build_target=git+https://github.com/pytorch/vision.git@617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:43:12.5406846Z + local wheel_dir=dist/vision 2025-12-04T09:43:12.5407154Z + local found_whl=0 2025-12-04T09:43:12.5407439Z + for file in "${wheel_dir}"/*.whl 2025-12-04T09:43:12.5407795Z + [[ -f dist/vision/*.whl ]] 2025-12-04T09:43:12.5408083Z + '[' 0 == 0 ']' 2025-12-04T09:43:12.5408883Z + python3 -m pip wheel --no-build-isolation --no-deps -w dist/vision git+https://github.com/pytorch/vision.git@617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:43:12.9008136Z Collecting git+https://github.com/pytorch/vision.git@617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:43:12.9012853Z Cloning https://github.com/pytorch/vision.git (to revision 617079d944b0e72632311c30ae2bbdf1168b901e) to /tmp/pip-req-build-mum7chto 2025-12-04T09:43:12.9196686Z Running command git clone --filter=blob:none --quiet https://github.com/pytorch/vision.git /tmp/pip-req-build-mum7chto 2025-12-04T09:43:14.6129565Z Running command git rev-parse -q --verify 'sha^617079d944b0e72632311c30ae2bbdf1168b901e' 2025-12-04T09:43:14.6153792Z Running command git fetch -q https://github.com/pytorch/vision.git 617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:43:14.7277735Z Resolved https://github.com/pytorch/vision.git to commit 617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:43:18.3379622Z Preparing metadata (pyproject.toml) ... [?25l- \ | done 2025-12-04T09:43:18.3418249Z [?25hBuilding wheels for collected packages: torchvision 2025-12-04T09:44:50.1921541Z Building wheel for torchvision (pyproject.toml) ... [?25l- \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - done 2025-12-04T09:44:50.1986717Z [?25h Created wheel for torchvision: filename=torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl size=1821808 sha256=3c1cbed7b6bcbb7f402e992497e715e60454362a01b5415d26ccf387e6c97fcf 2025-12-04T09:44:50.1989474Z Stored in directory: /var/lib/jenkins/.cache/pip/wheels/12/b2/29/1f82685c5b5173629e1f36a9b93989ce92ce563e5fb91d27ac 2025-12-04T09:44:50.2031786Z Successfully built torchvision 2025-12-04T09:44:50.2923414Z + for file in "${wheel_dir}"/*.whl 2025-12-04T09:44:50.2924177Z + pip_install_whl dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:44:50.2925067Z + args=('dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl') 2025-12-04T09:44:50.2925688Z + local args 2025-12-04T09:44:50.2926180Z + [[ dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl == *\ * ]] 2025-12-04T09:44:50.2926844Z + for path in "${args[@]}" 2025-12-04T09:44:50.2927473Z + echo 'Installing dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl' 2025-12-04T09:44:50.2928473Z Installing dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:44:50.2929589Z + python3 -mpip install --no-index --no-deps dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:44:50.6625088Z Processing ./dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:44:50.6765516Z Installing collected packages: torchvision 2025-12-04T09:44:51.2034860Z Successfully installed torchvision-0.25.0a0+617079d 2025-12-04T09:44:51.2462009Z + '[' -n '' ']' 2025-12-04T09:44:51.2462280Z + test_python_shard 3 2025-12-04T09:44:51.2462555Z + [[ -z 5 ]] 2025-12-04T09:44:51.2463461Z + python test/run_test.py --exclude-jit-executor --exclude-distributed-tests --exclude-quantization-tests --shard 3 5 --verbose --upload-artifacts-while-running 2025-12-04T09:44:58.3357940Z Downloading https://ossci-metrics.s3.amazonaws.com/disabled-tests-condensed.json to /var/lib/jenkins/workspace/test/.pytorch-disabled-tests.json 2025-12-04T09:44:58.3895122Z Ignoring disabled issues: [''] 2025-12-04T09:44:58.4009823Z Found test times from artifacts 2025-12-04T09:44:58.4455856Z Found test times from artifacts 2025-12-04T09:44:58.4470946Z Running all tests 2025-12-04T09:44:58.5373863Z Running parallel tests on 1 processes 2025-12-04T09:44:58.5386053Z Name: tests to run (est. time: 288.63min) 2025-12-04T09:44:58.5386580Z Serial tests (136): 2025-12-04T09:44:58.5386902Z inductor/test_aot_inductor 3/6 2025-12-04T09:44:58.5387267Z inductor/test_torchinductor 2/2 2025-12-04T09:44:58.5387667Z inductor/test_torchinductor_dynamic_shapes 4/5 2025-12-04T09:44:58.5388178Z inductor/test_torchinductor_codegen_dynamic_shapes 4/4 2025-12-04T09:44:58.5388651Z inductor/test_torchinductor_opinfo 5/17 2025-12-04T09:44:58.5389085Z inductor/test_torchinductor_opinfo 10/17 2025-12-04T09:44:58.5389495Z inductor/test_torchinductor_opinfo 15/17 2025-12-04T09:44:58.5389876Z dynamo/test_logging 1/1 2025-12-04T09:44:58.5390180Z dynamo/test_repros 1/1 2025-12-04T09:44:58.5391405Z inductor/test_cuda_select_algorithm 5/5 2025-12-04T09:44:58.5391814Z inductor/test_deterministic 2/8 2025-12-04T09:44:58.5392168Z inductor/test_deterministic 7/8 2025-12-04T09:44:58.5392533Z inductor/test_native_matmul 2/2 2025-12-04T09:44:58.5392908Z inductor/test_decompose_mem_bound_mm 1/1 2025-12-04T09:44:58.5393299Z inductor/test_online_softmax 1/1 2025-12-04T09:44:58.5393656Z inductor/test_mix_order_reduction 2/2 2025-12-04T09:44:58.5394018Z test_matmul_cuda 1/1 2025-12-04T09:44:58.5394315Z test_dataloader 2/2 2025-12-04T09:44:58.5394589Z test_decomp 5/17 2025-12-04T09:44:58.5394864Z test_decomp 10/17 2025-12-04T09:44:58.5395144Z test_decomp 15/17 2025-12-04T09:44:58.5395423Z test_ci_sanity_check_fail 1/1 2025-12-04T09:44:58.5395905Z test_meta 1/5 2025-12-04T09:44:58.5396393Z test_ops_jit 1/2 2025-12-04T09:44:58.5396741Z test_nestedtensor 2/4 2025-12-04T09:44:58.5397057Z test_public_bindings 1/1 2025-12-04T09:44:58.5397374Z test_ops 1/11 2025-12-04T09:44:58.5397621Z test_ops 6/11 2025-12-04T09:44:58.5397881Z test_ops 11/11 2025-12-04T09:44:58.5398157Z functorch/test_ops 3/7 2025-12-04T09:44:58.5398454Z functorch/test_vmap 1/1 2025-12-04T09:44:58.5398768Z dynamo/test_after_aot 1/1 2025-12-04T09:44:58.5399101Z inductor/test_snode_runtime 1/1 2025-12-04T09:44:58.5399442Z inductor/test_minifier 1/1 2025-12-04T09:44:58.5399782Z inductor/test_compiled_autograd 1/1 2025-12-04T09:44:58.5400156Z inductor/test_layout_optim 1/1 2025-12-04T09:44:58.5400544Z dynamo/test_unspec 1/1 2025-12-04T09:44:58.5400966Z inductor/test_mmdecomp 1/1 2025-12-04T09:44:58.5401299Z dynamo/test_ctx_manager 1/1 2025-12-04T09:44:58.5401618Z dynamo/test_exc 1/1 2025-12-04T09:44:58.5401903Z dynamo/test_misc 1/1 2025-12-04T09:44:58.5402227Z inductor/test_aot_inductor_arrayref 2/2 2025-12-04T09:44:58.5402603Z inductor/test_halide 1/1 2025-12-04T09:44:58.5402913Z inductor/test_xpu_basic 1/1 2025-12-04T09:44:58.5403269Z inductor/test_provenance_tracing 1/1 2025-12-04T09:44:58.5403652Z dynamo/test_buffers_override 1/1 2025-12-04T09:44:58.5404005Z inductor/test_inplacing_pass 1/1 2025-12-04T09:44:58.5404391Z inductor/test_aot_inductor_custom_ops 1/1 2025-12-04T09:44:58.5404795Z inductor/test_split_cat_fx_passes 1/1 2025-12-04T09:44:58.5405162Z inductor/test_profiler 1/1 2025-12-04T09:44:58.5405505Z inductor/test_memory_planning 1/1 2025-12-04T09:44:58.5405873Z inductor/test_mem_estimation 1/1 2025-12-04T09:44:58.5406215Z dynamo/test_view 1/1 2025-12-04T09:44:58.5406508Z inductor/test_cutlass_evt 1/1 2025-12-04T09:44:58.5406857Z dynamo/test_reconstruct 1/1 2025-12-04T09:44:58.5407196Z dynamo/test_aot_autograd 1/1 2025-12-04T09:44:58.5407573Z export/test_cpp_serdes 1/1 2025-12-04T09:44:58.5407899Z inductor/test_block_analysis 1/1 2025-12-04T09:44:58.5408259Z dynamo/test_subgraphs 1/1 2025-12-04T09:44:58.5408597Z dynamo/test_pre_dispatch 1/1 2025-12-04T09:44:58.5408960Z inductor/test_custom_post_grad_passes 1/1 2025-12-04T09:44:58.5409354Z dynamo/test_fx_annotate 1/1 2025-12-04T09:44:58.5409677Z dynamo/test_pgo 1/1 2025-12-04T09:44:58.5409969Z export/test_export_opinfo 1/1 2025-12-04T09:44:58.5410318Z inductor/test_control_flow 2/4 2025-12-04T09:44:58.5410662Z dynamo/test_compile 1/1 2025-12-04T09:44:58.5410982Z dynamo/test_nested_graph_breaks 1/1 2025-12-04T09:44:58.5411362Z inductor/test_needs_exact_strides 1/1 2025-12-04T09:44:58.5411767Z inductor/test_split_cat_fx_aten_passes 1/1 2025-12-04T09:44:58.5412149Z dynamo/test_resume 1/1 2025-12-04T09:44:58.5412477Z dynamo/test_backward_higher_order_ops 1/1 2025-12-04T09:44:58.5412893Z inductor/test_custom_partitioner_fn 1/1 2025-12-04T09:44:58.5413283Z dynamo/test_debug_utils 1/1 2025-12-04T09:44:58.5413599Z dynamo/test_base_hop 1/1 2025-12-04T09:44:58.5413922Z dynamo/test_package 1/1 2025-12-04T09:44:58.5414432Z dynamo/test_aot_autograd_cache 1/1 2025-12-04T09:44:58.5414792Z inductor/test_mps_basic 1/1 2025-12-04T09:44:58.5415127Z dynamo/test_comptime 1/1 2025-12-04T09:44:58.5415451Z test_sort_and_select 1/1 2025-12-04T09:44:58.5415761Z functorch/test_rearrange 1/1 2025-12-04T09:44:58.5416110Z functorch/test_parsing 1/1 2025-12-04T09:44:58.5416444Z profiler/test_profiler 1/1 2025-12-04T09:44:58.5416857Z torch_np/test_binary_ufuncs 1/1 2025-12-04T09:44:58.5417218Z torch_np/test_unary_ufuncs 1/1 2025-12-04T09:44:58.5417559Z test_utils_filelock 1/1 2025-12-04T09:44:58.5417879Z test_extension_utils 1/1 2025-12-04T09:44:58.5418229Z test_rename_privateuse1_to_existing_device 1/1 2025-12-04T09:44:58.5431991Z nn/attention/test_fa4 1/1 2025-12-04T09:44:58.5432978Z typing/test_python_operators 1/1 2025-12-04T09:44:58.5433333Z test_functionalization 1/1 2025-12-04T09:44:58.5433662Z profiler/test_kineto 1/1 2025-12-04T09:44:58.5433976Z test_module_tracker 1/1 2025-12-04T09:44:58.5434328Z torch_np/numpy_tests/core/test_scalarinherit 1/1 2025-12-04T09:44:58.5434734Z test_tensorexpr_pybind 1/1 2025-12-04T09:44:58.5435058Z test_fx_experimental 1/1 2025-12-04T09:44:58.5435351Z test_fx_passes 1/1 2025-12-04T09:44:58.5435637Z functorch/test_logging 1/1 2025-12-04T09:44:58.5435944Z test_namedtensor 1/1 2025-12-04T09:44:58.5436246Z test_tensorexpr 1/1 2025-12-04T09:44:58.5436537Z functorch/test_minifier 1/1 2025-12-04T09:44:58.5436893Z higher_order_ops/test_invoke_quant 1/1 2025-12-04T09:44:58.5437268Z torch_np/test_basic 1/1 2025-12-04T09:44:58.5437565Z test_jiterator 1/1 2025-12-04T09:44:58.5437859Z test_native_functions 1/1 2025-12-04T09:44:58.5438175Z test_typing 1/1 2025-12-04T09:44:58.5438461Z lazy/test_functionalization 1/1 2025-12-04T09:44:58.5438815Z torch_np/test_random 1/1 2025-12-04T09:44:58.5439145Z nn/test_multihead_attention 1/1 2025-12-04T09:44:58.5439478Z test_legacy_vmap 1/1 2025-12-04T09:44:58.5439788Z lazy/test_bindings 1/1 2025-12-04T09:44:58.5440091Z test_utils 1/1 2025-12-04T09:44:58.5440343Z test_pytree 1/1 2025-12-04T09:44:58.5440632Z test_namedtuple_return_api 1/1 2025-12-04T09:44:58.5440995Z profiler/test_record_function 1/1 2025-12-04T09:44:58.5441352Z test_compile_benchmark_util 1/1 2025-12-04T09:44:58.5441730Z test_set_default_mobile_cpu_allocator 1/1 2025-12-04T09:44:58.5442110Z test_fake_tensor 1/1 2025-12-04T09:44:58.5442422Z higher_order_ops/test_print 1/1 2025-12-04T09:44:58.5442763Z test_per_overload_api 1/1 2025-12-04T09:44:58.5443121Z torch_np/numpy_tests/core/test_einsum 1/1 2025-12-04T09:44:58.5443514Z test_multiprocessing 1/1 2025-12-04T09:44:58.5443816Z test_modules 1/1 2025-12-04T09:44:58.5444128Z complex_tensor/test_complex_tensor 1/1 2025-12-04T09:44:58.5444537Z torch_np/numpy_tests/core/test_indexing 1/1 2025-12-04T09:44:58.5444905Z test_futures 1/1 2025-12-04T09:44:58.5445199Z test_tensor_creation_ops 1/1 2025-12-04T09:44:58.5445586Z torch_np/numpy_tests/core/test_scalarmath 1/1 2025-12-04T09:44:58.5445977Z test_serialization 1/1 2025-12-04T09:44:58.5446317Z torch_np/numpy_tests/core/test_dlpack 1/1 2025-12-04T09:44:58.5446717Z test_multiprocessing_spawn 1/1 2025-12-04T09:44:58.5447076Z test_cuda_nvml_based_avail 1/1 2025-12-04T09:44:58.5447410Z test_mobile_optimizer 1/1 2025-12-04T09:44:58.5447745Z torch_np/test_function_base 1/1 2025-12-04T09:44:58.5448096Z test_type_promotion 1/1 2025-12-04T09:44:58.5448397Z lazy/test_reuse_ir 1/1 2025-12-04T09:44:58.5448729Z test_functional_autograd_benchmark 1/1 2025-12-04T09:44:58.5449101Z Parallel tests (0): 2025-12-04T09:44:58.5449387Z Name: excluded (est. time: 0.0min) 2025-12-04T09:44:58.5449725Z Serial tests (0): 2025-12-04T09:44:58.5449999Z Parallel tests (0): 2025-12-04T09:44:58.5450470Z Running inductor/test_aot_inductor 3/6 ... [2025-12-04 09:44:58.539556][1882.229952104] 2025-12-04T09:44:58.5451046Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:44:58.5452456Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_aot_inductor.py', '--shard-id=3', '--num-shards=6', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 09:44:58.539994] 2025-12-04T09:54:48.0500187Z 2025-12-04T09:54:48.0501239Z PRINTING LOG FILE of inductor/test_aot_inductor 3/6 (test/test-reports/inductor.test_aot_inductor_3.6_01eb6db542665098_.log) 2025-12-04T09:54:48.0502600Z W1204 09:45:11.237000 1729 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T09:54:48.0504698Z Test results will be stored in test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-cf40a4f46675676c.xml 2025-12-04T09:54:48.0507471Z ============================= test session starts ============================== 2025-12-04T09:54:48.0508346Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:54:48.0509272Z cachedir: .pytest_cache 2025-12-04T09:54:48.0510478Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:54:48.0511828Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:54:48.0512428Z configfile: pytest.ini 2025-12-04T09:54:48.0513423Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:54:48.0514564Z collecting ... collected 934 items 2025-12-04T09:54:48.0515129Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:54:48.0602932Z Running 150 items in this shard: test/inductor/test_aot_inductor.py::TestAOTInductorConfig::test_compile_standalone_package_cpp_false_raises, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test__weight_int4pack_mm_m_32_n_64_q_group_64_num_groups_2_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_addmm_multiple_dynamic_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_runtime_asserts_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_assert_tensor_meta_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_bool_input_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_buffer_mutation_4_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_buffer_reuse_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_codegen_int_array_var_fix_memory_leak_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_composed_dynamic_size_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_cond_unbacked_symint_closure_dynamic_False_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_constant_folding_with_update_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_convolution_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_deconv_freezing_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_duplicate_constant_folding_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_embedding_bag_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_empty_graph_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_extract_constants_map_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_fallback_mem_leak_fix_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_load_package_multiple_gpus_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_multiple_output_alias_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_narrow_fallback_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_non_contiguous_output_alias_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_output_misaligned_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_rocm_triton_autotuning_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_runtime_checks_device_type_failed_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_runtime_checks_dtype_failed_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_runtime_checks_large_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_scatter_reduce_fallback_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_shifted_constraint_ranges_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_simple_embed_kernel_binary_False_max_autotune_True_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_size_with_unbacked_add_and_mul_expr_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_sympy_cpp_printer_min_max_minmax1_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_triton_kernel_bool_param_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_triton_kernel_dynamic_grid_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_triton_kernel_extern_kernel_arg_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_triton_kernel_grid_type_3_num_dims_2_dynamic_False_autotune_False_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_triton_kernel_multi_output_arg_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_triton_kernel_on_device_tma_dynamic_True_tma_version_old_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_triton_kernel_tma_descriptor_1d_dynamic_False_tma_version_new_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_triton_kernel_weird_param_order_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_unbacked_equals_input_size_runtime_assertion_mark_unbacked_False_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_update_inactive_constant_buffer_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_view_outputs_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_while_loop_with_conv_dynamic_True_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_while_loop_with_mixed_device_dynamic_False_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_while_loop_with_mixed_device_dynamic_True_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_while_loop_with_parameters_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_with_no_triton_profiler_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_zero_grid_with_backed_symbols_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_zero_size_weight_cpu, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__int_mm_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_with_scales_and_zeros_m_32_n_64_q_group_32_num_groups_2_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_with_scales_and_zeros_m_32_n_64_q_group_64_num_groups_1_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_aliased_buffer_reuse_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_aoti_debug_printer_sym_inputs_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_aoti_debug_printing_model_inputs_codegen_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_aoti_profiler_enable_kernel_profile_False_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_aoti_profiler_enable_kernel_profile_True_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_aoti_user_defined_triton_kernel_profiling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_autotune_int64_user_defined_triton_kernel_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_backward_no_op_logging_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_buffer_mutation_4_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_buffer_reuse_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_cond_cpu_predicate_cuda_operands_max_autotune_False_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_cond_predicate_on_cpu_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_cond_with_outer_code_before_after_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_foreach_multiple_dynamic_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_index_put_fallback_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_libtorch_free_so_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_load_package_multiple_gpus_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_misc_1_max_autotune_True_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_multiple_output_alias_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_nested_tensor_from_jagged_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_output_path_2_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_pad_fallback_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_quantized_linear_bias_none_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_user_defined_triton_kernel_embed_kernel_binary_True_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_return_view_constant_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_simple_embed_kernel_binary_False_max_autotune_False_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_simple_embed_kernel_binary_False_max_autotune_True_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_simple_embed_kernel_binary_True_max_autotune_True_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_simple_split_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_symbool_item_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_kernel_grid_type_1_num_dims_1_dynamic_False_autotune_True_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_kernel_grid_type_2_num_dims_1_dynamic_False_autotune_True_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_kernel_grid_type_3_num_dims_2_dynamic_False_autotune_False_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_kernel_grid_type_3_num_dims_2_dynamic_False_autotune_True_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_kernel_multi_output_arg_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_kernel_unbacked_symint_in_grid_dynamic_False_autotuning_False_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_kernel_unbacked_symint_in_grid_dynamic_False_autotuning_True_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_kernel_with_none_input_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_kernel_with_none_inputs_and_equal_to_1_arg_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_next_power_of_2_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_unbacked_equals_input_size_runtime_assertion_mark_unbacked_False_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_unbacked_expr_replacements_shift_k_1_use_static_size_True_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_update_inactive_constant_buffer_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_view_outputs_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_while_loop_with_outer_buffers_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_with_cudagraphs_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_1_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test__weight_int4pack_mm_with_scales_and_zeros_m_32_n_64_q_group_64_num_groups_2_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_aoti_runtime_asserts_backed_symint_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_assert_async_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_clamp_decomposition_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_cond_non_tensor_predicates_dynamic_True_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_cond_share_predicate_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_cond_unbacked_symint_closure_dynamic_False_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_copy_non_blocking_is_pinned_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_deconv_freezing_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_dup_unbacked_sym_decl_with_refinement_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_fft_c2c_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_fx_gm_return_tuple_validation_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_input_codegen_with_sympy_expr_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_large_dynamic_dim_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_large_mmaped_weights_on_disk_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_linear_freezing_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_multi_device_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_no_args_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_normal_functional_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_proxy_executor_abs_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_proxy_executor_squeeze_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_repeat_output_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_rocm_triton_autotuning_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_runtime_checks_shape_failed_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_simple_embed_kernel_binary_False_max_autotune_True_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_simple_embed_kernel_binary_True_max_autotune_False_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_size_with_unbacked_add_and_mul_expr_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_stft_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_symfloat_item_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_sympy_cpp_printer_min_max_minmax0_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_torchvision_transforms_functional_tensor_resize_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_bool_param_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_dynamic_shape_with_div_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_grid_type_1_num_dims_1_dynamic_True_autotune_False_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_grid_type_3_num_dims_1_dynamic_False_autotune_True_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_grid_type_3_num_dims_1_dynamic_True_autotune_False_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_grid_type_3_num_dims_2_dynamic_True_autotune_True_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_on_device_tma_dynamic_True_tma_version_new_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_tma_descriptor_1d_dynamic_False_tma_version_new_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_tma_descriptor_1d_dynamic_False_tma_version_old_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_tma_descriptor_1d_dynamic_True_tma_version_old_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_tma_descriptor_2d_dynamic_False_tma_version_old_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_tma_descriptor_2d_dynamic_True_tma_version_new_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_weird_param_order_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_unbacked_expr_replacements_shift_k_3_use_static_size_True_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_update_constant_buffer_mps, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_zero_size_weight_mps 2025-12-04T09:54:48.0689289Z 2025-12-04T09:54:48.0689870Z inductor/test_aot_inductor.py::TestAOTInductorConfig::test_compile_standalone_package_cpp_false_raises PASSED [0.0040s] [ 0%] 2025-12-04T09:54:48.0691285Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cpu SKIPPED [0.0029s] (requires GPU) [ 1%] 2025-12-04T09:54:48.0692855Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test__weight_int4pack_mm_m_32_n_64_q_group_64_num_groups_2_cpu SKIPPED [0.0028s] (requires GPU) [ 2%] 2025-12-04T09:54:48.0694542Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_addmm_multiple_dynamic_cpu SKIPPED [0.0012s] (Skipping triton backend only since not big GPU (not enough SM)) [ 2%] 2025-12-04T09:54:48.0697096Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_aoti_runtime_asserts_cpu <- test/inductor/test_torchinductor.py W1204 09:45:13.280000 1729 site-packages/torch/fx/experimental/symbolic_shapes.py:7357] propagate_real_tensors evaluate_expr(Ne(u1 - 3, 7)) -> False 2025-12-04T09:54:48.0698831Z W1204 09:45:13.481000 1729 site-packages/torch/export/_draft_export.py:524] 2025-12-04T09:54:48.0699720Z W1204 09:45:13.481000 1729 site-packages/torch/export/_draft_export.py:524] ################################################################################################### 2025-12-04T09:54:48.0700986Z W1204 09:45:13.481000 1729 site-packages/torch/export/_draft_export.py:524] WARNING: 2 issue(s) found during export, and it was not able to soundly produce a graph. 2025-12-04T09:54:48.0702461Z W1204 09:45:13.481000 1729 site-packages/torch/export/_draft_export.py:524] To view the report of failures in an html page, please run the command: 2025-12-04T09:54:48.0703812Z W1204 09:45:13.481000 1729 site-packages/torch/export/_draft_export.py:524] `tlparse /tmp/export_jenkins/dedicated_log_torch_trace_n12y8wqp.log --export` 2025-12-04T09:54:48.0705148Z W1204 09:45:13.481000 1729 site-packages/torch/export/_draft_export.py:524] Or, you can view the errors in python by inspecting `print(ep._report)`. 2025-12-04T09:54:48.0706112Z W1204 09:45:13.481000 1729 site-packages/torch/export/_draft_export.py:524] 2025-12-04T09:54:48.0707110Z W1204 09:45:13.481000 1729 site-packages/torch/export/_draft_export.py:524] While tracing we found 1 operator(s) which do not have a fake kernel registered. 2025-12-04T09:54:48.0708631Z W1204 09:45:13.481000 1729 site-packages/torch/export/_draft_export.py:524] If you intend to retrace the exported graph or run it with fake tensors, please run it under the 2025-12-04T09:54:48.0710080Z W1204 09:45:13.481000 1729 site-packages/torch/export/_draft_export.py:524] following context manager, which will register a fake kernel for those operators. 2025-12-04T09:54:48.0711119Z W1204 09:45:13.481000 1729 site-packages/torch/export/_draft_export.py:524] ``` 2025-12-04T09:54:48.0712171Z W1204 09:45:13.481000 1729 site-packages/torch/export/_draft_export.py:524] with torch._library.fake_profile.unsafe_generate_fake_kernels(ep._report.op_profiles): 2025-12-04T09:54:48.0713352Z W1204 09:45:13.481000 1729 site-packages/torch/export/_draft_export.py:524] # run with fake tensors 2025-12-04T09:54:48.0714137Z W1204 09:45:13.481000 1729 site-packages/torch/export/_draft_export.py:524] ``` 2025-12-04T09:54:48.0715014Z W1204 09:45:13.481000 1729 site-packages/torch/export/_draft_export.py:524] ################################################################################################# 2025-12-04T09:54:48.0715747Z Error: Expected u5 >= 10 but received 2 2025-12-04T09:54:48.0716111Z PASSED [13.5513s] [ 3%] 2025-12-04T09:54:48.0717011Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_assert_tensor_meta_cpu <- test/inductor/test_torchinductor.py PASSED [5.1081s] [ 4%] 2025-12-04T09:54:48.0718496Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_bool_input_cpu <- test/inductor/test_torchinductor.py PASSED [5.0474s] [ 4%] 2025-12-04T09:54:48.0720037Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_buffer_mutation_4_cpu <- test/inductor/test_torchinductor.py SKIPPED [0.0031s] (requires GPU) [ 5%] 2025-12-04T09:54:48.0721607Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_buffer_reuse_cpu <- test/inductor/test_torchinductor.py PASSED [7.0020s] [ 6%] 2025-12-04T09:54:48.0723298Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_codegen_int_array_var_fix_memory_leak_cpu <- test/inductor/test_torchinductor.py SKIPPED [0.0031s] (test is only for cuda) [ 6%] 2025-12-04T09:54:48.0725043Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_composed_dynamic_size_cpu <- test/inductor/test_torchinductor.py PASSED [5.1921s] [ 7%] 2025-12-04T09:54:48.0726501Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_cond_unbacked_symint_closure_dynamic_False_cpu PASSED [5.1353s] [ 8%] 2025-12-04T09:54:48.0728006Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_constant_folding_with_update_cpu <- test/inductor/test_torchinductor.py PASSED [5.2393s] [ 8%] 2025-12-04T09:54:48.0729671Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_convolution_cpu SKIPPED [0.0005s] (Skipping triton backend only since not big GPU (not enough SM)) [ 9%] 2025-12-04T09:54:48.0731083Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_deconv_freezing_cpu PASSED [21.3906s] [ 10%] 2025-12-04T09:54:48.0732466Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_duplicate_constant_folding_cpu <- test/inductor/test_torchinductor.py PASSED [5.2168s] [ 10%] 2025-12-04T09:54:48.0734075Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_embedding_bag_cpu <- test/inductor/test_torchinductor.py PASSED [5.0531s] [ 11%] 2025-12-04T09:54:48.0735538Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_empty_graph_cpu <- test/inductor/test_torchinductor.py PASSED [5.0094s] [ 12%] 2025-12-04T09:54:48.0737097Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_extract_constants_map_cpu <- test/inductor/test_torchinductor.py PASSED [5.1237s] [ 12%] 2025-12-04T09:54:48.0738535Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_fallback_mem_leak_fix_cpu SKIPPED [0.0031s] (requires GPU) [ 13%] 2025-12-04T09:54:48.0739951Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_load_package_multiple_gpus_cpu SKIPPED [0.0002s] (requires multiple cuda devices) [ 14%] 2025-12-04T09:54:48.0741646Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_multiple_output_alias_cpu <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (Skipped!) [ 14%] 2025-12-04T09:54:48.0743247Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_narrow_fallback_cpu <- test/inductor/test_torchinductor.py PASSED [5.1705s] [ 15%] 2025-12-04T09:54:48.0744913Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_non_contiguous_output_alias_cpu <- test/inductor/test_torchinductor.py PASSED [5.2729s] [ 16%] 2025-12-04T09:54:48.0746485Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_output_misaligned_cpu <- test/inductor/test_torchinductor.py PASSED [5.3138s] [ 16%] 2025-12-04T09:54:48.0748047Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_rocm_triton_autotuning_cpu SKIPPED [0.0031s] (test currently only works on the ROCm stack) [ 17%] 2025-12-04T09:54:48.0749662Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_runtime_checks_device_type_failed_cpu SKIPPED [0.0035s] (requires GPU) [ 18%] 2025-12-04T09:54:48.0751238Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_runtime_checks_dtype_failed_cpu Error: input_handles[0]: unmatched dtype, expected: 5(at::kHalf), but got: 6 2025-12-04T09:54:48.0752163Z 2025-12-04T09:54:48.0752283Z PASSED [5.1190s] [ 18%] 2025-12-04T09:54:48.0753243Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_runtime_checks_large_cpu SKIPPED [0.0006s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 19%] 2025-12-04T09:54:48.0754892Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_scatter_reduce_fallback_cpu <- test/inductor/test_torchinductor.py PASSED [5.1149s] [ 20%] 2025-12-04T09:54:48.0756476Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_shifted_constraint_ranges_cpu <- test/inductor/test_torchinductor.py PASSED [5.2303s] [ 20%] 2025-12-04T09:54:48.0758014Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_simple_embed_kernel_binary_False_max_autotune_True_cpu PASSED [12.2286s] [ 21%] 2025-12-04T09:54:48.0759474Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_size_with_unbacked_add_and_mul_expr_cpu SKIPPED [0.0031s] (requires GPU) [ 22%] 2025-12-04T09:54:48.0760891Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_sympy_cpp_printer_min_max_minmax1_cpu SKIPPED [0.0028s] (requires GPU) [ 22%] 2025-12-04T09:54:48.0762471Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_triton_kernel_bool_param_cpu <- test/inductor/test_torchinductor.py SKIPPED [0.0027s] (requires GPU) [ 23%] 2025-12-04T09:54:48.0764034Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_triton_kernel_dynamic_grid_cpu SKIPPED [0.0027s] (requires GPU) [ 24%] 2025-12-04T09:54:48.0765438Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_triton_kernel_extern_kernel_arg_cpu SKIPPED [0.0027s] (requires GPU) [ 24%] 2025-12-04T09:54:48.0767072Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_triton_kernel_grid_type_3_num_dims_2_dynamic_False_autotune_False_cpu SKIPPED [0.0027s] (requires GPU) [ 25%] 2025-12-04T09:54:48.0768834Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_triton_kernel_multi_output_arg_cpu <- test/inductor/test_torchinductor.py SKIPPED [0.0027s] (requires GPU) [ 26%] 2025-12-04T09:54:48.0770552Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_triton_kernel_on_device_tma_dynamic_True_tma_version_old_cpu SKIPPED [0.0027s] (requires GPU) [ 26%] 2025-12-04T09:54:48.0772227Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_triton_kernel_tma_descriptor_1d_dynamic_False_tma_version_new_cpu SKIPPED [0.0027s] (requires GPU) [ 27%] 2025-12-04T09:54:48.0773976Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_triton_kernel_weird_param_order_cpu <- test/inductor/test_torchinductor.py SKIPPED [0.0027s] (requires GPU) [ 28%] 2025-12-04T09:54:48.0776083Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_unbacked_equals_input_size_runtime_assertion_mark_unbacked_False_cpu W1204 09:47:29.867000 1729 site-packages/torch/_export/__init__.py:71] +============================+ 2025-12-04T09:54:48.0777709Z W1204 09:47:29.867000 1729 site-packages/torch/_export/__init__.py:72] | !!! WARNING !!! | 2025-12-04T09:54:48.0778560Z W1204 09:47:29.867000 1729 site-packages/torch/_export/__init__.py:73] +============================+ 2025-12-04T09:54:48.0780279Z W1204 09:47:29.868000 1729 site-packages/torch/_export/__init__.py:74] torch._export.aot_compile()/torch._export.aot_load() is being deprecated, please switch to directly calling torch._inductor.aoti_compile_and_package(torch.export.export())/torch._inductor.aoti_load_package() instead. 2025-12-04T09:54:48.0781774Z Error: Expected u0 >= 2 but received 0 2025-12-04T09:54:48.0782128Z PASSED [10.2924s] [ 28%] 2025-12-04T09:54:48.0783089Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_update_inactive_constant_buffer_cpu <- test/inductor/test_torchinductor.py PASSED [4.9429s] [ 29%] 2025-12-04T09:54:48.0784667Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_view_outputs_cpu <- test/inductor/test_torchinductor.py PASSED [4.9554s] [ 30%] 2025-12-04T09:54:48.0786680Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_while_loop_with_conv_dynamic_True_cpu W1204 09:47:45.133000 1729 site-packages/torch/export/dynamic_shapes.py:923] Using None as a dynamic shape dimension is deprecated. Please use Dim.STATIC instead 2025-12-04T09:54:48.0788121Z PASSED [5.8511s] [ 30%] 2025-12-04T09:54:48.0788924Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_while_loop_with_mixed_device_dynamic_False_cpu PASSED [5.6166s] [ 31%] 2025-12-04T09:54:48.0790923Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_while_loop_with_mixed_device_dynamic_True_cpu W1204 09:47:56.596000 1729 site-packages/torch/export/dynamic_shapes.py:923] Using None as a dynamic shape dimension is deprecated. Please use Dim.STATIC instead 2025-12-04T09:54:48.0792959Z W1204 09:47:56.596000 1729 site-packages/torch/export/dynamic_shapes.py:923] Using None as a dynamic shape dimension is deprecated. Please use Dim.STATIC instead 2025-12-04T09:54:48.0793850Z PASSED [5.5861s] [ 32%] 2025-12-04T09:54:48.0795210Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_while_loop_with_parameters_cpu W1204 09:48:02.214000 1729 site-packages/torch/export/dynamic_shapes.py:923] Using None as a dynamic shape dimension is deprecated. Please use Dim.STATIC instead 2025-12-04T09:54:48.0796932Z PASSED [5.6413s] [ 32%] 2025-12-04T09:54:48.0797836Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_with_no_triton_profiler_cpu <- test/inductor/test_torchinductor.py PASSED [4.9628s] [ 33%] 2025-12-04T09:54:48.0799527Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_zero_grid_with_backed_symbols_cpu <- test/inductor/test_torchinductor.py SKIPPED [0.0034s] (requires GPU) [ 34%] 2025-12-04T09:54:48.0801299Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleCpu::test_zero_size_weight_cpu <- test/inductor/test_torchinductor.py PASSED [5.2702s] [ 34%] 2025-12-04T09:54:48.0803188Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__int_mm_cuda <- test/inductor/test_torchinductor.py W1204 09:48:19.102000 1729 site-packages/torch/_inductor/utils.py:1703] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-12-04T09:54:48.0804523Z PASSED [11.2051s] [ 35%] 2025-12-04T09:54:48.0805456Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda ('RERUN', {'yellow': True}) [0.0339s] [ 36%] 2025-12-04T09:54:48.0807064Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda ('RERUN', {'yellow': True}) [0.0052s] [ 36%] 2025-12-04T09:54:48.0808736Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda FAILED [0.0052s] [ 36%] 2025-12-04T09:54:48.0809538Z 2025-12-04T09:54:48.0809686Z ==================================== RERUNS ==================================== 2025-12-04T09:54:48.0810413Z _ AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda _ 2025-12-04T09:54:48.0811116Z Traceback (most recent call last): 2025-12-04T09:54:48.0811834Z File "/var/lib/jenkins/workspace/test/inductor/test_aot_inductor.py", line 6893, in test__weight_int4pack_mm 2025-12-04T09:54:48.0812584Z self.check_model(model, (a,)) 2025-12-04T09:54:48.0813254Z File "/var/lib/jenkins/workspace/test/inductor/test_aot_inductor_utils.py", line 247, in check_model 2025-12-04T09:54:48.0813959Z ref_model = copy.deepcopy(model) 2025-12-04T09:54:48.0814475Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 172, in deepcopy 2025-12-04T09:54:48.0815009Z y = _reconstruct(x, memo, *rv) 2025-12-04T09:54:48.0815532Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 271, in _reconstruct 2025-12-04T09:54:48.0816074Z state = deepcopy(state, memo) 2025-12-04T09:54:48.0816576Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 146, in deepcopy 2025-12-04T09:54:48.0817158Z y = copier(x, memo) 2025-12-04T09:54:48.0817632Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 231, in _deepcopy_dict 2025-12-04T09:54:48.0818239Z y[deepcopy(key, memo)] = deepcopy(value, memo) 2025-12-04T09:54:48.0818796Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 153, in deepcopy 2025-12-04T09:54:48.0819307Z y = copier(memo) 2025-12-04T09:54:48.0819881Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_tensor.py", line 180, in __deepcopy__ 2025-12-04T09:54:48.0820607Z new_storage = self._typed_storage()._deepcopy(memo) 2025-12-04T09:54:48.0821312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 1139, in _deepcopy 2025-12-04T09:54:48.0822127Z return self._new_wrapped_storage(copy.deepcopy(self._untyped_storage, memo)) 2025-12-04T09:54:48.0822828Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 153, in deepcopy 2025-12-04T09:54:48.0823341Z y = copier(memo) 2025-12-04T09:54:48.0823927Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 243, in __deepcopy__ 2025-12-04T09:54:48.0824587Z new_storage = self.clone() 2025-12-04T09:54:48.0825180Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 257, in clone 2025-12-04T09:54:48.0825929Z return type(self)(self.nbytes(), device=self.device).copy_(self) 2025-12-04T09:54:48.0826504Z torch.AcceleratorError: CUDA error: invalid device function 2025-12-04T09:54:48.0827513Z Search for `cudaErrorInvalidDeviceFunction' in https://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__TYPES.html for more information. 2025-12-04T09:54:48.0828770Z CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. 2025-12-04T09:54:48.0829660Z For debugging consider passing CUDA_LAUNCH_BLOCKING=1 2025-12-04T09:54:48.0830203Z Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. 2025-12-04T09:54:48.0830584Z 2025-12-04T09:54:48.0830589Z 2025-12-04T09:54:48.0830806Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:48.0831956Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda 2025-12-04T09:54:48.0832869Z 2025-12-04T09:54:48.0833150Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:48.0833973Z _ AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda _ 2025-12-04T09:54:48.0834732Z Traceback (most recent call last): 2025-12-04T09:54:48.0835456Z File "/var/lib/jenkins/workspace/test/inductor/test_aot_inductor.py", line 6893, in test__weight_int4pack_mm 2025-12-04T09:54:48.0836195Z self.check_model(model, (a,)) 2025-12-04T09:54:48.0836847Z File "/var/lib/jenkins/workspace/test/inductor/test_aot_inductor_utils.py", line 247, in check_model 2025-12-04T09:54:48.0837553Z ref_model = copy.deepcopy(model) 2025-12-04T09:54:48.0838074Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 172, in deepcopy 2025-12-04T09:54:48.0838596Z y = _reconstruct(x, memo, *rv) 2025-12-04T09:54:48.0839119Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 271, in _reconstruct 2025-12-04T09:54:48.0839670Z state = deepcopy(state, memo) 2025-12-04T09:54:48.0840166Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 146, in deepcopy 2025-12-04T09:54:48.0840668Z y = copier(x, memo) 2025-12-04T09:54:48.0841151Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 231, in _deepcopy_dict 2025-12-04T09:54:48.0841759Z y[deepcopy(key, memo)] = deepcopy(value, memo) 2025-12-04T09:54:48.0842308Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 153, in deepcopy 2025-12-04T09:54:48.0842817Z y = copier(memo) 2025-12-04T09:54:48.0843404Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_tensor.py", line 180, in __deepcopy__ 2025-12-04T09:54:48.0844119Z new_storage = self._typed_storage()._deepcopy(memo) 2025-12-04T09:54:48.0844822Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 1139, in _deepcopy 2025-12-04T09:54:48.0845649Z return self._new_wrapped_storage(copy.deepcopy(self._untyped_storage, memo)) 2025-12-04T09:54:48.0846339Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 153, in deepcopy 2025-12-04T09:54:48.0846840Z y = copier(memo) 2025-12-04T09:54:48.0847429Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 243, in __deepcopy__ 2025-12-04T09:54:48.0848104Z new_storage = self.clone() 2025-12-04T09:54:48.0848682Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 257, in clone 2025-12-04T09:54:48.0849431Z return type(self)(self.nbytes(), device=self.device).copy_(self) 2025-12-04T09:54:48.0850021Z torch.AcceleratorError: CUDA error: invalid device function 2025-12-04T09:54:48.0851018Z Search for `cudaErrorInvalidDeviceFunction' in https://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__TYPES.html for more information. 2025-12-04T09:54:48.0852261Z CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. 2025-12-04T09:54:48.0853061Z For debugging consider passing CUDA_LAUNCH_BLOCKING=1 2025-12-04T09:54:48.0853614Z Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. 2025-12-04T09:54:48.0853981Z 2025-12-04T09:54:48.0853986Z 2025-12-04T09:54:48.0854212Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:48.0855355Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda 2025-12-04T09:54:48.0856351Z 2025-12-04T09:54:48.0856623Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:48.0857264Z =================================== FAILURES =================================== 2025-12-04T09:54:48.0857986Z _ AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda _ 2025-12-04T09:54:48.0858670Z Traceback (most recent call last): 2025-12-04T09:54:48.0859388Z File "/var/lib/jenkins/workspace/test/inductor/test_aot_inductor.py", line 6893, in test__weight_int4pack_mm 2025-12-04T09:54:48.0860129Z self.check_model(model, (a,)) 2025-12-04T09:54:48.0860782Z File "/var/lib/jenkins/workspace/test/inductor/test_aot_inductor_utils.py", line 247, in check_model 2025-12-04T09:54:48.0861553Z ref_model = copy.deepcopy(model) 2025-12-04T09:54:48.0862069Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 172, in deepcopy 2025-12-04T09:54:48.0862604Z y = _reconstruct(x, memo, *rv) 2025-12-04T09:54:48.0863115Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 271, in _reconstruct 2025-12-04T09:54:48.0863668Z state = deepcopy(state, memo) 2025-12-04T09:54:48.0864169Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 146, in deepcopy 2025-12-04T09:54:48.0864675Z y = copier(x, memo) 2025-12-04T09:54:48.0865165Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 231, in _deepcopy_dict 2025-12-04T09:54:48.0865762Z y[deepcopy(key, memo)] = deepcopy(value, memo) 2025-12-04T09:54:48.0866323Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 153, in deepcopy 2025-12-04T09:54:48.0866821Z y = copier(memo) 2025-12-04T09:54:48.0867408Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_tensor.py", line 180, in __deepcopy__ 2025-12-04T09:54:48.0868146Z new_storage = self._typed_storage()._deepcopy(memo) 2025-12-04T09:54:48.0868839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 1139, in _deepcopy 2025-12-04T09:54:48.0869672Z return self._new_wrapped_storage(copy.deepcopy(self._untyped_storage, memo)) 2025-12-04T09:54:48.0870366Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 153, in deepcopy 2025-12-04T09:54:48.0870885Z y = copier(memo) 2025-12-04T09:54:48.0871464Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 243, in __deepcopy__ 2025-12-04T09:54:48.0872141Z new_storage = self.clone() 2025-12-04T09:54:48.0872737Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 257, in clone 2025-12-04T09:54:48.0873474Z return type(self)(self.nbytes(), device=self.device).copy_(self) 2025-12-04T09:54:48.0874068Z torch.AcceleratorError: CUDA error: invalid device function 2025-12-04T09:54:48.0875075Z Search for `cudaErrorInvalidDeviceFunction' in https://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__TYPES.html for more information. 2025-12-04T09:54:48.0876338Z CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. 2025-12-04T09:54:48.0877134Z For debugging consider passing CUDA_LAUNCH_BLOCKING=1 2025-12-04T09:54:48.0877694Z Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. 2025-12-04T09:54:48.0878063Z 2025-12-04T09:54:48.0878068Z 2025-12-04T09:54:48.0878304Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:48.0879450Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda 2025-12-04T09:54:48.0880365Z 2025-12-04T09:54:48.0880636Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:48.0881809Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-cf40a4f46675676c.xml - 2025-12-04T09:54:48.0882951Z =========================== short test summary info ============================ 2025-12-04T09:54:48.0884187Z FAILED [0.0052s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda - torch.AcceleratorError: CUDA error: invalid device function 2025-12-04T09:54:48.0885842Z Search for `cudaErrorInvalidDeviceFunction' in https://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__TYPES.html for more information. 2025-12-04T09:54:48.0887104Z CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. 2025-12-04T09:54:48.0887909Z For debugging consider passing CUDA_LAUNCH_BLOCKING=1 2025-12-04T09:54:48.0888464Z Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. 2025-12-04T09:54:48.0888921Z 2025-12-04T09:54:48.0888926Z 2025-12-04T09:54:48.0889142Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:48.0890288Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda 2025-12-04T09:54:48.0891216Z 2025-12-04T09:54:48.0891484Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:48.0892076Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T09:54:48.0892616Z ======== 1 failed, 30 passed, 23 skipped, 2 rerun in 196.13s (0:03:16) ========= 2025-12-04T09:54:48.0893088Z Got exit code 1 2025-12-04T09:54:48.0893364Z Retrying single test... 2025-12-04T09:54:48.0893992Z W1204 09:48:40.427000 5162 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T09:54:48.0895136Z Test results will be stored in test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-3767af6d1470a6b5.xml 2025-12-04T09:54:48.0896202Z ============================= test session starts ============================== 2025-12-04T09:54:48.0896945Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:54:48.0897545Z cachedir: .pytest_cache 2025-12-04T09:54:48.0898260Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:54:48.0899048Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:54:48.0899403Z configfile: pytest.ini 2025-12-04T09:54:48.0900122Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:54:48.0901035Z collecting ... collected 934 items / 149 deselected / 785 selected 2025-12-04T09:54:48.0902267Z stepcurrent: skipping 53 already run items. Running only test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda 2025-12-04T09:54:48.0903385Z Running 1 items in this shard 2025-12-04T09:54:48.0903596Z 2025-12-04T09:54:48.0904741Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda [W1204 09:48:42.017243416 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T09:54:48.0906017Z 2025-12-04T09:54:48.0906148Z ('RERUN', {'yellow': True}) [15.7952s] [100%] 2025-12-04T09:54:48.0907535Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda [W1204 09:48:58.820374222 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T09:54:48.0908789Z 2025-12-04T09:54:48.0908935Z ('RERUN', {'yellow': True}) [0.0069s] [100%] 2025-12-04T09:54:48.0910444Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda [W1204 09:48:58.827767025 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T09:54:48.0911704Z 2025-12-04T09:54:48.0911810Z FAILED [0.0052s] [100%] 2025-12-04T09:54:48.0912002Z 2025-12-04T09:54:48.0912145Z ==================================== RERUNS ==================================== 2025-12-04T09:54:48.0912871Z _ AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda _ 2025-12-04T09:54:48.0913574Z Traceback (most recent call last): 2025-12-04T09:54:48.0914282Z File "/var/lib/jenkins/workspace/test/inductor/test_aot_inductor.py", line 6893, in test__weight_int4pack_mm 2025-12-04T09:54:48.0915019Z self.check_model(model, (a,)) 2025-12-04T09:54:48.0915680Z File "/var/lib/jenkins/workspace/test/inductor/test_aot_inductor_utils.py", line 247, in check_model 2025-12-04T09:54:48.0916460Z ref_model = copy.deepcopy(model) 2025-12-04T09:54:48.0916985Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 172, in deepcopy 2025-12-04T09:54:48.0917515Z y = _reconstruct(x, memo, *rv) 2025-12-04T09:54:48.0918045Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 271, in _reconstruct 2025-12-04T09:54:48.0918585Z state = deepcopy(state, memo) 2025-12-04T09:54:48.0919083Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 146, in deepcopy 2025-12-04T09:54:48.0919604Z y = copier(x, memo) 2025-12-04T09:54:48.0920075Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 231, in _deepcopy_dict 2025-12-04T09:54:48.0920680Z y[deepcopy(key, memo)] = deepcopy(value, memo) 2025-12-04T09:54:48.0921237Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 153, in deepcopy 2025-12-04T09:54:48.0921740Z y = copier(memo) 2025-12-04T09:54:48.0922330Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_tensor.py", line 180, in __deepcopy__ 2025-12-04T09:54:48.0923065Z new_storage = self._typed_storage()._deepcopy(memo) 2025-12-04T09:54:48.0923767Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 1139, in _deepcopy 2025-12-04T09:54:48.0924584Z return self._new_wrapped_storage(copy.deepcopy(self._untyped_storage, memo)) 2025-12-04T09:54:48.0925276Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 153, in deepcopy 2025-12-04T09:54:48.0925786Z y = copier(memo) 2025-12-04T09:54:48.0926356Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 243, in __deepcopy__ 2025-12-04T09:54:48.0927028Z new_storage = self.clone() 2025-12-04T09:54:48.0927615Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 257, in clone 2025-12-04T09:54:48.0928358Z return type(self)(self.nbytes(), device=self.device).copy_(self) 2025-12-04T09:54:48.0928931Z torch.AcceleratorError: CUDA error: invalid device function 2025-12-04T09:54:48.0929937Z Search for `cudaErrorInvalidDeviceFunction' in https://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__TYPES.html for more information. 2025-12-04T09:54:48.0931201Z CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. 2025-12-04T09:54:48.0932011Z For debugging consider passing CUDA_LAUNCH_BLOCKING=1 2025-12-04T09:54:48.0932554Z Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. 2025-12-04T09:54:48.0932934Z 2025-12-04T09:54:48.0933547Z Exception raised from copy_device_to_device at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/Copy.cu:337 (most recent call first): 2025-12-04T09:54:48.0934410Z C++ CapturedTraceback: 2025-12-04T09:54:48.0935919Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T09:54:48.0937996Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T09:54:48.0939290Z #6 c10::AcceleratorError::AcceleratorError(c10::SourceLocation, int, std::__cxx11::basic_string, std::allocator > const&) from :0 2025-12-04T09:54:48.0940462Z #7 c10::cuda::c10_cuda_check_implementation(int, char const*, char const*, unsigned int, bool) from ??:0 2025-12-04T09:54:48.0941247Z #8 at::native::copy_device_to_device(at::TensorIterator&, bool, bool) from ??:0 2025-12-04T09:54:48.0941953Z #9 at::native::copy_impl(at::Tensor&, at::Tensor const&, bool) [clone .isra.0] from Copy.cpp:0 2025-12-04T09:54:48.0942621Z #10 at::native::copy_(at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T09:54:48.0945131Z #11 c10::impl::wrap_kernel_functor_unboxed_, at::Tensor&, c10::guts::typelist::typelist >, at::Tensor& (c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool) from VariableTypeManual.cpp:0 2025-12-04T09:54:48.0948082Z #12 torch::autograd::VariableType::(anonymous namespace)::copy_(c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool) from VariableTypeManual.cpp:0 2025-12-04T09:54:48.0949056Z #13 at::_ops::copy_::call(at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T09:54:48.0949668Z #14 at::storage_copy(c10::Storage&, c10::Storage const&, bool) from ??:0 2025-12-04T09:54:48.0950296Z #15 THPStorage_copy_(_object*, _object*, _object*) from StorageMethods.cpp:0 2025-12-04T09:54:48.0951134Z #16 method_vectorcall_VARARGS_KEYWORDS from /usr/local/src/conda/python-3.10.14/Objects/descrobject.c:344 2025-12-04T09:54:48.0952127Z #17 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.0953076Z #18 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0954002Z #19 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.0954942Z #20 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0955877Z #21 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.0956795Z #22 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0957733Z #23 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.0958676Z #24 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0959614Z #25 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.0960534Z #26 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0961467Z #27 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.0962395Z #28 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0963326Z #29 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.0964240Z #30 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0965167Z #31 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.0966099Z #32 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0967050Z #33 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.0968032Z #34 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0968818Z #35 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.0969600Z #36 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0970515Z #37 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.0971442Z #38 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0972368Z #39 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.0973292Z #40 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0974154Z #41 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.0974860Z #42 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.0975639Z #43 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0976452Z #44 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.0977223Z #45 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.0978008Z #46 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0978824Z #47 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.0979523Z #48 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.0980285Z #49 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0981223Z #50 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.0982163Z #51 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0982931Z #52 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.0983706Z #53 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0984634Z #54 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.0985562Z #55 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0986477Z #56 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.0987399Z #57 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0988326Z #58 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.0989259Z #59 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0990176Z #60 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.0991107Z #61 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0991923Z #62 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.0992609Z #63 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.0993389Z #64 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0994267Z #65 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.0995080Z #66 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.0995816Z #67 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.0996944Z #68 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.0997820Z #69 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.0998757Z #70 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.0999680Z #71 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1000612Z #72 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1001397Z #73 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1002269Z #74 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1003186Z #75 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1004121Z #76 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1005051Z #77 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1005967Z #78 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1006840Z #79 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1007649Z #80 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1008395Z #81 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1009095Z #82 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T09:54:48.1009776Z #83 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1010560Z #84 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1011488Z #85 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1012405Z #86 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1013332Z #87 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1014261Z #88 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1015040Z #89 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1015808Z #90 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1016733Z #91 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1017720Z #92 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1018638Z #93 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1019573Z #94 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1020362Z #95 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1021252Z #96 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1022176Z #97 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1023100Z #98 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1024034Z #99 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1024977Z #100 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1026018Z #101 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1026845Z #102 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1027604Z #103 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1028352Z #104 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1029229Z #105 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1030182Z #106 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1031049Z #107 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1031831Z #108 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1032789Z #109 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1033742Z #110 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1034689Z #111 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1035626Z #112 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1036524Z #113 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1037347Z #114 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1038114Z #115 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1038867Z #116 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1039753Z #117 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1040707Z #118 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1041647Z #119 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1042598Z #120 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1043553Z #121 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1044501Z #122 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1045284Z #123 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1046085Z #124 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1047037Z #125 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1047980Z #126 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1048915Z #127 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1049858Z #128 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1050751Z #129 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1051570Z #130 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1052315Z #131 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1053080Z #132 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1054018Z #133 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1054952Z #134 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1055893Z #135 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1056919Z #136 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1057865Z #137 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1058795Z #138 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1059743Z #139 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1060751Z #140 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1061701Z #141 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1062634Z #142 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1063457Z #143 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T09:54:48.1064211Z #144 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T09:54:48.1064944Z #145 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T09:54:48.1065634Z #146 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T09:54:48.1066424Z #147 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T09:54:48.1067259Z #148 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T09:54:48.1068019Z #149 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T09:54:48.1068739Z #150 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T09:54:48.1069423Z #151 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T09:54:48.1070031Z #152 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T09:54:48.1070457Z #153 _start from ??:0 2025-12-04T09:54:48.1070762Z #154 from ??:0 2025-12-04T09:54:48.1070999Z 2025-12-04T09:54:48.1071004Z 2025-12-04T09:54:48.1071233Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:48.1072389Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda 2025-12-04T09:54:48.1073311Z 2025-12-04T09:54:48.1073583Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:48.1074221Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:48.1075722Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py:257: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T09:54:48.1077215Z return type(self)(self.nbytes(), device=self.device).copy_(self) 2025-12-04T09:54:48.1077998Z _ AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda _ 2025-12-04T09:54:48.1078697Z Traceback (most recent call last): 2025-12-04T09:54:48.1079413Z File "/var/lib/jenkins/workspace/test/inductor/test_aot_inductor.py", line 6893, in test__weight_int4pack_mm 2025-12-04T09:54:48.1080141Z self.check_model(model, (a,)) 2025-12-04T09:54:48.1080814Z File "/var/lib/jenkins/workspace/test/inductor/test_aot_inductor_utils.py", line 247, in check_model 2025-12-04T09:54:48.1081521Z ref_model = copy.deepcopy(model) 2025-12-04T09:54:48.1082040Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 172, in deepcopy 2025-12-04T09:54:48.1082658Z y = _reconstruct(x, memo, *rv) 2025-12-04T09:54:48.1083185Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 271, in _reconstruct 2025-12-04T09:54:48.1083739Z state = deepcopy(state, memo) 2025-12-04T09:54:48.1084225Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 146, in deepcopy 2025-12-04T09:54:48.1084744Z y = copier(x, memo) 2025-12-04T09:54:48.1085233Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 231, in _deepcopy_dict 2025-12-04T09:54:48.1085840Z y[deepcopy(key, memo)] = deepcopy(value, memo) 2025-12-04T09:54:48.1086388Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 153, in deepcopy 2025-12-04T09:54:48.1086899Z y = copier(memo) 2025-12-04T09:54:48.1087487Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_tensor.py", line 180, in __deepcopy__ 2025-12-04T09:54:48.1088271Z new_storage = self._typed_storage()._deepcopy(memo) 2025-12-04T09:54:48.1088971Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 1139, in _deepcopy 2025-12-04T09:54:48.1089797Z return self._new_wrapped_storage(copy.deepcopy(self._untyped_storage, memo)) 2025-12-04T09:54:48.1090487Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 153, in deepcopy 2025-12-04T09:54:48.1090981Z y = copier(memo) 2025-12-04T09:54:48.1091566Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 243, in __deepcopy__ 2025-12-04T09:54:48.1092237Z new_storage = self.clone() 2025-12-04T09:54:48.1092813Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 257, in clone 2025-12-04T09:54:48.1093555Z return type(self)(self.nbytes(), device=self.device).copy_(self) 2025-12-04T09:54:48.1094150Z torch.AcceleratorError: CUDA error: invalid device function 2025-12-04T09:54:48.1095151Z Search for `cudaErrorInvalidDeviceFunction' in https://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__TYPES.html for more information. 2025-12-04T09:54:48.1096795Z CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. 2025-12-04T09:54:48.1097970Z For debugging consider passing CUDA_LAUNCH_BLOCKING=1 2025-12-04T09:54:48.1098529Z Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. 2025-12-04T09:54:48.1098899Z 2025-12-04T09:54:48.1099528Z Exception raised from copy_device_to_device at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/Copy.cu:337 (most recent call first): 2025-12-04T09:54:48.1100383Z C++ CapturedTraceback: 2025-12-04T09:54:48.1101896Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T09:54:48.1103836Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T09:54:48.1105116Z #6 c10::AcceleratorError::AcceleratorError(c10::SourceLocation, int, std::__cxx11::basic_string, std::allocator > const&) from :0 2025-12-04T09:54:48.1106270Z #7 c10::cuda::c10_cuda_check_implementation(int, char const*, char const*, unsigned int, bool) from ??:0 2025-12-04T09:54:48.1107048Z #8 at::native::copy_device_to_device(at::TensorIterator&, bool, bool) from ??:0 2025-12-04T09:54:48.1107765Z #9 at::native::copy_impl(at::Tensor&, at::Tensor const&, bool) [clone .isra.0] from Copy.cpp:0 2025-12-04T09:54:48.1108447Z #10 at::native::copy_(at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T09:54:48.1111085Z #11 c10::impl::wrap_kernel_functor_unboxed_, at::Tensor&, c10::guts::typelist::typelist >, at::Tensor& (c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool) from VariableTypeManual.cpp:0 2025-12-04T09:54:48.1113995Z #12 torch::autograd::VariableType::(anonymous namespace)::copy_(c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool) from VariableTypeManual.cpp:0 2025-12-04T09:54:48.1114976Z #13 at::_ops::copy_::call(at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T09:54:48.1115574Z #14 at::storage_copy(c10::Storage&, c10::Storage const&, bool) from ??:0 2025-12-04T09:54:48.1116220Z #15 THPStorage_copy_(_object*, _object*, _object*) from StorageMethods.cpp:0 2025-12-04T09:54:48.1117140Z #16 method_vectorcall_VARARGS_KEYWORDS from /usr/local/src/conda/python-3.10.14/Objects/descrobject.c:344 2025-12-04T09:54:48.1118129Z #17 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1119077Z #18 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1120001Z #19 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1120942Z #20 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1121879Z #21 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1122814Z #22 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1123734Z #23 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1124680Z #24 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1125609Z #25 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1126547Z #26 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1127466Z #27 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1128393Z #28 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1129320Z #29 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1130253Z #30 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1131171Z #31 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1132100Z #32 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1133034Z #33 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1133962Z #34 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1134730Z #35 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1135507Z #36 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1136434Z #37 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1137425Z #38 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1138352Z #39 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1139294Z #40 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1140117Z #41 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1140894Z #42 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1141684Z #43 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1142501Z #44 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1143205Z #45 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1143979Z #46 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1144795Z #47 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1145492Z #48 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1146317Z #49 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1147247Z #50 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1148180Z #51 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1148962Z #52 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1149729Z #53 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1150665Z #54 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1151600Z #55 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1152529Z #56 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1153449Z #57 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1154380Z #58 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1155315Z #59 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1156241Z #60 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1157154Z #61 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1157969Z #62 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1158670Z #63 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1159432Z #64 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1160302Z #65 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1161112Z #66 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1161867Z #67 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1162685Z #68 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1163551Z #69 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1164477Z #70 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1165404Z #71 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1166324Z #72 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1167103Z #73 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1167882Z #74 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1168889Z #75 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1169807Z #76 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1170733Z #77 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1171661Z #78 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1172519Z #79 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1173324Z #80 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1174073Z #81 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1174854Z #82 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T09:54:48.1175516Z #83 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1176299Z #84 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1177310Z #85 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1178238Z #86 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1179155Z #87 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1180086Z #88 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1180865Z #89 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1181636Z #90 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1182579Z #91 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1183514Z #92 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1184445Z #93 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1185359Z #94 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1186142Z #95 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1186919Z #96 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1187845Z #97 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1188758Z #98 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1189692Z #99 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1190631Z #100 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1191520Z #101 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1192329Z #102 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1193092Z #103 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1193852Z #104 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1194718Z #105 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1195670Z #106 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1196678Z #107 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1197480Z #108 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1198545Z #109 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1199499Z #110 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1200448Z #111 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1201398Z #112 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1202278Z #113 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1203108Z #114 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1203981Z #115 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1204742Z #116 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1205610Z #117 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1206558Z #118 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1227353Z #119 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1228335Z #120 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1229298Z #121 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1230241Z #122 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1231061Z #123 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1231861Z #124 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1232802Z #125 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1233751Z #126 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1234696Z #127 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1235650Z #128 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1236527Z #129 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1237352Z #130 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1238123Z #131 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1238887Z #132 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1239750Z #133 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1240699Z #134 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1241643Z #135 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1242581Z #136 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1243516Z #137 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1244463Z #138 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1245408Z #139 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1246362Z #140 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1247452Z #141 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1248403Z #142 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1249227Z #143 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T09:54:48.1249969Z #144 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T09:54:48.1250701Z #145 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T09:54:48.1251410Z #146 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T09:54:48.1252204Z #147 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T09:54:48.1253101Z #148 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T09:54:48.1253876Z #149 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T09:54:48.1254604Z #150 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T09:54:48.1255297Z #151 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T09:54:48.1255903Z #152 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T09:54:48.1256350Z #153 _start from ??:0 2025-12-04T09:54:48.1256656Z #154 from ??:0 2025-12-04T09:54:48.1257002Z 2025-12-04T09:54:48.1257007Z 2025-12-04T09:54:48.1257229Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:48.1258392Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda 2025-12-04T09:54:48.1259332Z 2025-12-04T09:54:48.1259604Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:48.1260246Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:48.1261729Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py:257: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T09:54:48.1263224Z return type(self)(self.nbytes(), device=self.device).copy_(self) 2025-12-04T09:54:48.1263737Z =================================== FAILURES =================================== 2025-12-04T09:54:48.1264471Z _ AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda _ 2025-12-04T09:54:48.1265163Z Traceback (most recent call last): 2025-12-04T09:54:48.1265885Z File "/var/lib/jenkins/workspace/test/inductor/test_aot_inductor.py", line 6893, in test__weight_int4pack_mm 2025-12-04T09:54:48.1266632Z self.check_model(model, (a,)) 2025-12-04T09:54:48.1267284Z File "/var/lib/jenkins/workspace/test/inductor/test_aot_inductor_utils.py", line 247, in check_model 2025-12-04T09:54:48.1267991Z ref_model = copy.deepcopy(model) 2025-12-04T09:54:48.1268518Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 172, in deepcopy 2025-12-04T09:54:48.1269051Z y = _reconstruct(x, memo, *rv) 2025-12-04T09:54:48.1269562Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 271, in _reconstruct 2025-12-04T09:54:48.1270122Z state = deepcopy(state, memo) 2025-12-04T09:54:48.1270632Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 146, in deepcopy 2025-12-04T09:54:48.1271135Z y = copier(x, memo) 2025-12-04T09:54:48.1271621Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 231, in _deepcopy_dict 2025-12-04T09:54:48.1272226Z y[deepcopy(key, memo)] = deepcopy(value, memo) 2025-12-04T09:54:48.1272800Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 153, in deepcopy 2025-12-04T09:54:48.1273304Z y = copier(memo) 2025-12-04T09:54:48.1273895Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_tensor.py", line 180, in __deepcopy__ 2025-12-04T09:54:48.1274711Z new_storage = self._typed_storage()._deepcopy(memo) 2025-12-04T09:54:48.1275403Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 1139, in _deepcopy 2025-12-04T09:54:48.1276229Z return self._new_wrapped_storage(copy.deepcopy(self._untyped_storage, memo)) 2025-12-04T09:54:48.1276923Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 153, in deepcopy 2025-12-04T09:54:48.1277436Z y = copier(memo) 2025-12-04T09:54:48.1278009Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 243, in __deepcopy__ 2025-12-04T09:54:48.1278682Z new_storage = self.clone() 2025-12-04T09:54:48.1279269Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 257, in clone 2025-12-04T09:54:48.1280069Z return type(self)(self.nbytes(), device=self.device).copy_(self) 2025-12-04T09:54:48.1280655Z torch.AcceleratorError: CUDA error: invalid device function 2025-12-04T09:54:48.1281671Z Search for `cudaErrorInvalidDeviceFunction' in https://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__TYPES.html for more information. 2025-12-04T09:54:48.1282932Z CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. 2025-12-04T09:54:48.1283728Z For debugging consider passing CUDA_LAUNCH_BLOCKING=1 2025-12-04T09:54:48.1284288Z Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. 2025-12-04T09:54:48.1284656Z 2025-12-04T09:54:48.1285286Z Exception raised from copy_device_to_device at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/Copy.cu:337 (most recent call first): 2025-12-04T09:54:48.1286152Z C++ CapturedTraceback: 2025-12-04T09:54:48.1287762Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T09:54:48.1289708Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T09:54:48.1291000Z #6 c10::AcceleratorError::AcceleratorError(c10::SourceLocation, int, std::__cxx11::basic_string, std::allocator > const&) from :0 2025-12-04T09:54:48.1292175Z #7 c10::cuda::c10_cuda_check_implementation(int, char const*, char const*, unsigned int, bool) from ??:0 2025-12-04T09:54:48.1293038Z #8 at::native::copy_device_to_device(at::TensorIterator&, bool, bool) from ??:0 2025-12-04T09:54:48.1293764Z #9 at::native::copy_impl(at::Tensor&, at::Tensor const&, bool) [clone .isra.0] from Copy.cpp:0 2025-12-04T09:54:48.1294439Z #10 at::native::copy_(at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T09:54:48.1297229Z #11 c10::impl::wrap_kernel_functor_unboxed_, at::Tensor&, c10::guts::typelist::typelist >, at::Tensor& (c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool) from VariableTypeManual.cpp:0 2025-12-04T09:54:48.1300130Z #12 torch::autograd::VariableType::(anonymous namespace)::copy_(c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool) from VariableTypeManual.cpp:0 2025-12-04T09:54:48.1301096Z #13 at::_ops::copy_::call(at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T09:54:48.1301696Z #14 at::storage_copy(c10::Storage&, c10::Storage const&, bool) from ??:0 2025-12-04T09:54:48.1302356Z #15 THPStorage_copy_(_object*, _object*, _object*) from StorageMethods.cpp:0 2025-12-04T09:54:48.1303334Z #16 method_vectorcall_VARARGS_KEYWORDS from /usr/local/src/conda/python-3.10.14/Objects/descrobject.c:344 2025-12-04T09:54:48.1304491Z #17 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1305439Z #18 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1306383Z #19 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1307322Z #20 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1308243Z #21 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1309168Z #22 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1310207Z #23 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1311145Z #24 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1312067Z #25 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1312998Z #26 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1313932Z #27 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1314852Z #28 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1315779Z #29 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1316708Z #30 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1317643Z #31 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1318560Z #32 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1319489Z #33 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1320418Z #34 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1321200Z #35 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1321968Z #36 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1322896Z #37 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1323823Z #38 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1324754Z #39 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1325675Z #40 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1326493Z #41 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1327193Z #42 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1327957Z #43 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1328770Z #44 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1329467Z #45 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1330244Z #46 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1331048Z #47 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1331749Z #48 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1332599Z #49 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1333541Z #50 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1334461Z #51 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1335240Z #52 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1336018Z #53 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1337027Z #54 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1337946Z #55 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1338971Z #56 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1339905Z #57 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1340821Z #58 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1341753Z #59 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1342686Z #60 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1343625Z #61 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1344433Z #62 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1345132Z #63 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1345919Z #64 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1346788Z #65 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1347589Z #66 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1348336Z #67 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1349076Z #68 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1349943Z #69 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1350862Z #70 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1351787Z #71 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1352712Z #72 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1353491Z #73 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1354267Z #74 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1355195Z #75 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1356126Z #76 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1357043Z #77 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1357967Z #78 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1358835Z #79 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1359635Z #80 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1360373Z #81 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1361143Z #82 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T09:54:48.1361822Z #83 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1362590Z #84 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1363511Z #85 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1364438Z #86 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1365366Z #87 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1366280Z #88 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1367125Z #89 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1367900Z #90 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1368821Z #91 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1369740Z #92 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1370660Z #93 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1371578Z #94 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1372348Z #95 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1373116Z #96 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1374044Z #97 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1374969Z #98 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1375898Z #99 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1376882Z #100 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1377770Z #101 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1378585Z #102 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1379330Z #103 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1380085Z #104 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1380951Z #105 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1381902Z #106 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1382689Z #107 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1383475Z #108 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1384419Z #109 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1385360Z #110 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1386292Z #111 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1387230Z #112 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1388105Z #113 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1388916Z #114 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1389746Z #115 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1390509Z #116 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1391380Z #117 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1392313Z #118 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1393256Z #119 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1394199Z #120 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1395147Z #121 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1396334Z #122 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1397130Z #123 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1397922Z #124 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1398871Z #125 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1399800Z #126 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1400742Z #127 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1401681Z #128 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1402556Z #129 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1403361Z #130 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1404114Z #131 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1404870Z #132 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1405729Z #133 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1406675Z #134 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1407611Z #135 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1408554Z #136 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1409481Z #137 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1410421Z #138 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1411356Z #139 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1412293Z #140 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1413223Z #141 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1414159Z #142 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1414973Z #143 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T09:54:48.1415715Z #144 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T09:54:48.1416431Z #145 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T09:54:48.1417189Z #146 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T09:54:48.1417982Z #147 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T09:54:48.1418924Z #148 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T09:54:48.1419694Z #149 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T09:54:48.1420407Z #150 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T09:54:48.1421092Z #151 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T09:54:48.1421692Z #152 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T09:54:48.1422125Z #153 _start from ??:0 2025-12-04T09:54:48.1422423Z #154 from ??:0 2025-12-04T09:54:48.1422659Z 2025-12-04T09:54:48.1422665Z 2025-12-04T09:54:48.1422884Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:48.1424122Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda 2025-12-04T09:54:48.1425054Z 2025-12-04T09:54:48.1425328Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:48.1425966Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:48.1427448Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py:257: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T09:54:48.1428932Z return type(self)(self.nbytes(), device=self.device).copy_(self) 2025-12-04T09:54:48.1430047Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-3767af6d1470a6b5.xml - 2025-12-04T09:54:48.1431107Z =========================== short test summary info ============================ 2025-12-04T09:54:48.1432343Z FAILED [0.0052s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda - torch.AcceleratorError: CUDA error: invalid device function 2025-12-04T09:54:48.1434008Z Search for `cudaErrorInvalidDeviceFunction' in https://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__TYPES.html for more information. 2025-12-04T09:54:48.1435262Z CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. 2025-12-04T09:54:48.1436076Z For debugging consider passing CUDA_LAUNCH_BLOCKING=1 2025-12-04T09:54:48.1436632Z Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. 2025-12-04T09:54:48.1436999Z 2025-12-04T09:54:48.1437609Z Exception raised from copy_device_to_device at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/Copy.cu:337 (most recent call first): 2025-12-04T09:54:48.1438478Z C++ CapturedTraceback: 2025-12-04T09:54:48.1439988Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T09:54:48.1441917Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T09:54:48.1443188Z #6 c10::AcceleratorError::AcceleratorError(c10::SourceLocation, int, std::__cxx11::basic_string, std::allocator > const&) from :0 2025-12-04T09:54:48.1444365Z #7 c10::cuda::c10_cuda_check_implementation(int, char const*, char const*, unsigned int, bool) from ??:0 2025-12-04T09:54:48.1445152Z #8 at::native::copy_device_to_device(at::TensorIterator&, bool, bool) from ??:0 2025-12-04T09:54:48.1445870Z #9 at::native::copy_impl(at::Tensor&, at::Tensor const&, bool) [clone .isra.0] from Copy.cpp:0 2025-12-04T09:54:48.1446535Z #10 at::native::copy_(at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T09:54:48.1449128Z #11 c10::impl::wrap_kernel_functor_unboxed_, at::Tensor&, c10::guts::typelist::typelist >, at::Tensor& (c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool) from VariableTypeManual.cpp:0 2025-12-04T09:54:48.1452008Z #12 torch::autograd::VariableType::(anonymous namespace)::copy_(c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool) from VariableTypeManual.cpp:0 2025-12-04T09:54:48.1452981Z #13 at::_ops::copy_::call(at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T09:54:48.1453643Z #14 at::storage_copy(c10::Storage&, c10::Storage const&, bool) from ??:0 2025-12-04T09:54:48.1454272Z #15 THPStorage_copy_(_object*, _object*, _object*) from StorageMethods.cpp:0 2025-12-04T09:54:48.1455111Z #16 method_vectorcall_VARARGS_KEYWORDS from /usr/local/src/conda/python-3.10.14/Objects/descrobject.c:344 2025-12-04T09:54:48.1456088Z #17 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1457089Z #18 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1458014Z #19 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1458940Z #20 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1459871Z #21 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1460786Z #22 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1461727Z #23 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1462665Z #24 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1463595Z #25 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1464508Z #26 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1465440Z #27 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1466368Z #28 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1467300Z #29 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1468217Z #30 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1469146Z #31 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1470077Z #32 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1471002Z #33 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1471914Z #34 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1472699Z #35 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1473485Z #36 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1474403Z #37 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1475332Z #38 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1476268Z #39 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1477279Z #40 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1478091Z #41 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1478790Z #42 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1479573Z #43 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1480390Z #44 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1481077Z #45 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1481860Z #46 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1482734Z #47 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1483435Z #48 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1484203Z #49 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1485135Z #50 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1486067Z #51 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1486835Z #52 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1487615Z #53 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1488547Z #54 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1489475Z #55 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1490395Z #56 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1491330Z #57 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1492255Z #58 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1493184Z #59 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1494104Z #60 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1495030Z #61 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1495844Z #62 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1496729Z #63 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1497558Z #64 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1498436Z #65 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1499241Z #66 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1499978Z #67 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1500729Z #68 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1501590Z #69 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1502519Z #70 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1503439Z #71 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1504376Z #72 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1505161Z #73 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1506166Z #74 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1507090Z #75 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1508019Z #76 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1508953Z #77 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1509866Z #78 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1510740Z #79 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1511630Z #80 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1512378Z #81 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1513083Z #82 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T09:54:48.1513760Z #83 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1514546Z #84 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1515486Z #85 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1516406Z #86 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1517340Z #87 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1518277Z #88 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1519066Z #89 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1519837Z #90 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1520779Z #91 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1521717Z #92 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1522637Z #93 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1523568Z #94 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1524354Z #95 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1525140Z #96 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1526065Z #97 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1527003Z #98 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1527935Z #99 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1528877Z #100 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1529758Z #101 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1530586Z #102 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1531346Z #103 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1532106Z #104 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1532971Z #105 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1533914Z #106 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1534775Z #107 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1535558Z #108 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1536500Z #109 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1537515Z #110 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1538463Z #111 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1539391Z #112 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1540346Z #113 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1541168Z #114 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1541936Z #115 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1542685Z #116 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1543563Z #117 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1544511Z #118 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1545456Z #119 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1546389Z #120 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1547331Z #121 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1548276Z #122 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1549064Z #123 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1549854Z #124 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1550792Z #125 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1551735Z #126 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1552665Z #127 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1553610Z #128 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1554491Z #129 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1555319Z #130 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1556069Z #131 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1556826Z #132 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1557702Z #133 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1558654Z #134 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1559586Z #135 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1560535Z #136 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1561477Z #137 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1562412Z #138 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1563460Z #139 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1564415Z #140 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1565364Z #141 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1566296Z #142 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1567123Z #143 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T09:54:48.1567879Z #144 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T09:54:48.1568614Z #145 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T09:54:48.1569363Z #146 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T09:54:48.1570154Z #147 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T09:54:48.1570992Z #148 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T09:54:48.1571753Z #149 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T09:54:48.1572467Z #150 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T09:54:48.1573152Z #151 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T09:54:48.1573769Z #152 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T09:54:48.1574195Z #153 _start from ??:0 2025-12-04T09:54:48.1574496Z #154 from ??:0 2025-12-04T09:54:48.1574731Z 2025-12-04T09:54:48.1574737Z 2025-12-04T09:54:48.1574962Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:48.1576118Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda 2025-12-04T09:54:48.1577101Z 2025-12-04T09:54:48.1577375Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:48.1577968Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T09:54:48.1578500Z ================= 1 failed, 149 deselected, 2 rerun in 15.90s ================== 2025-12-04T09:54:48.1578937Z Got exit code 1 2025-12-04T09:54:48.1579210Z Retrying single test... 2025-12-04T09:54:48.1579842Z W1204 09:49:08.506000 5283 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T09:54:48.1580999Z Test results will be stored in test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-bf1966c65cc0ce25.xml 2025-12-04T09:54:48.1581867Z ============================= test session starts ============================== 2025-12-04T09:54:48.1582545Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:54:48.1583157Z cachedir: .pytest_cache 2025-12-04T09:54:48.1583875Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:54:48.1584656Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:54:48.1585007Z configfile: pytest.ini 2025-12-04T09:54:48.1585735Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:54:48.1586634Z collecting ... collected 934 items / 149 deselected / 785 selected 2025-12-04T09:54:48.1587873Z stepcurrent: skipping 53 already run items. Running only test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda 2025-12-04T09:54:48.1588999Z Running 1 items in this shard 2025-12-04T09:54:48.1589212Z 2025-12-04T09:54:48.1590446Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda [W1204 09:49:10.086860829 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T09:54:48.1591717Z 2025-12-04T09:54:48.1591865Z ('RERUN', {'yellow': True}) [16.2162s] [100%] 2025-12-04T09:54:48.1593248Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda [W1204 09:49:26.311584384 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T09:54:48.1594522Z 2025-12-04T09:54:48.1594654Z ('RERUN', {'yellow': True}) [0.0074s] [100%] 2025-12-04T09:54:48.1596238Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda [W1204 09:49:26.319693569 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T09:54:48.1597608Z 2025-12-04T09:54:48.1597734Z FAILED [0.0058s] [100%] 2025-12-04T09:54:48.1597918Z 2025-12-04T09:54:48.1598061Z ==================================== RERUNS ==================================== 2025-12-04T09:54:48.1598786Z _ AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda _ 2025-12-04T09:54:48.1599495Z Traceback (most recent call last): 2025-12-04T09:54:48.1600221Z File "/var/lib/jenkins/workspace/test/inductor/test_aot_inductor.py", line 6893, in test__weight_int4pack_mm 2025-12-04T09:54:48.1600950Z self.check_model(model, (a,)) 2025-12-04T09:54:48.1601617Z File "/var/lib/jenkins/workspace/test/inductor/test_aot_inductor_utils.py", line 247, in check_model 2025-12-04T09:54:48.1602325Z ref_model = copy.deepcopy(model) 2025-12-04T09:54:48.1602839Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 172, in deepcopy 2025-12-04T09:54:48.1603385Z y = _reconstruct(x, memo, *rv) 2025-12-04T09:54:48.1603914Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 271, in _reconstruct 2025-12-04T09:54:48.1604478Z state = deepcopy(state, memo) 2025-12-04T09:54:48.1604967Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 146, in deepcopy 2025-12-04T09:54:48.1605487Z y = copier(x, memo) 2025-12-04T09:54:48.1605975Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 231, in _deepcopy_dict 2025-12-04T09:54:48.1606569Z y[deepcopy(key, memo)] = deepcopy(value, memo) 2025-12-04T09:54:48.1607126Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 153, in deepcopy 2025-12-04T09:54:48.1607639Z y = copier(memo) 2025-12-04T09:54:48.1608226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_tensor.py", line 180, in __deepcopy__ 2025-12-04T09:54:48.1608948Z new_storage = self._typed_storage()._deepcopy(memo) 2025-12-04T09:54:48.1609653Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 1139, in _deepcopy 2025-12-04T09:54:48.1610483Z return self._new_wrapped_storage(copy.deepcopy(self._untyped_storage, memo)) 2025-12-04T09:54:48.1611167Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 153, in deepcopy 2025-12-04T09:54:48.1611678Z y = copier(memo) 2025-12-04T09:54:48.1612259Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 243, in __deepcopy__ 2025-12-04T09:54:48.1612928Z new_storage = self.clone() 2025-12-04T09:54:48.1613502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 257, in clone 2025-12-04T09:54:48.1614245Z return type(self)(self.nbytes(), device=self.device).copy_(self) 2025-12-04T09:54:48.1614842Z torch.AcceleratorError: CUDA error: invalid device function 2025-12-04T09:54:48.1615842Z Search for `cudaErrorInvalidDeviceFunction' in https://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__TYPES.html for more information. 2025-12-04T09:54:48.1617169Z CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. 2025-12-04T09:54:48.1618081Z For debugging consider passing CUDA_LAUNCH_BLOCKING=1 2025-12-04T09:54:48.1618642Z Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. 2025-12-04T09:54:48.1619015Z 2025-12-04T09:54:48.1619631Z Exception raised from copy_device_to_device at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/Copy.cu:337 (most recent call first): 2025-12-04T09:54:48.1620508Z C++ CapturedTraceback: 2025-12-04T09:54:48.1622021Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T09:54:48.1624072Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T09:54:48.1625379Z #6 c10::AcceleratorError::AcceleratorError(c10::SourceLocation, int, std::__cxx11::basic_string, std::allocator > const&) from :0 2025-12-04T09:54:48.1626544Z #7 c10::cuda::c10_cuda_check_implementation(int, char const*, char const*, unsigned int, bool) from ??:0 2025-12-04T09:54:48.1627335Z #8 at::native::copy_device_to_device(at::TensorIterator&, bool, bool) from ??:0 2025-12-04T09:54:48.1628052Z #9 at::native::copy_impl(at::Tensor&, at::Tensor const&, bool) [clone .isra.0] from Copy.cpp:0 2025-12-04T09:54:48.1628717Z #10 at::native::copy_(at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T09:54:48.1631221Z #11 c10::impl::wrap_kernel_functor_unboxed_, at::Tensor&, c10::guts::typelist::typelist >, at::Tensor& (c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool) from VariableTypeManual.cpp:0 2025-12-04T09:54:48.1634121Z #12 torch::autograd::VariableType::(anonymous namespace)::copy_(c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool) from VariableTypeManual.cpp:0 2025-12-04T09:54:48.1635093Z #13 at::_ops::copy_::call(at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T09:54:48.1635692Z #14 at::storage_copy(c10::Storage&, c10::Storage const&, bool) from ??:0 2025-12-04T09:54:48.1636334Z #15 THPStorage_copy_(_object*, _object*, _object*) from StorageMethods.cpp:0 2025-12-04T09:54:48.1637152Z #16 method_vectorcall_VARARGS_KEYWORDS from /usr/local/src/conda/python-3.10.14/Objects/descrobject.c:344 2025-12-04T09:54:48.1638137Z #17 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1639077Z #18 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1640018Z #19 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1640937Z #20 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1641867Z #21 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1642795Z #22 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1643713Z #23 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1644641Z #24 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1645569Z #25 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1646500Z #26 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1647490Z #27 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1648420Z #28 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1649353Z #29 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1650289Z #30 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1651213Z #31 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1652146Z #32 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1653076Z #33 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1654091Z #34 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1654862Z #35 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1655644Z #36 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1656582Z #37 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1657570Z #38 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1658505Z #39 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1659439Z #40 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1660260Z #41 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1660979Z #42 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1661753Z #43 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1662573Z #44 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1663278Z #45 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1664046Z #46 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1664863Z #47 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1665567Z #48 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1666347Z #49 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1667271Z #50 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1668222Z #51 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1669013Z #52 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1669796Z #53 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1670717Z #54 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1671655Z #55 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1672590Z #56 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1673533Z #57 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1674453Z #58 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1675387Z #59 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1676397Z #60 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1677319Z #61 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1678133Z #62 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1678844Z #63 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1679625Z #64 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1680489Z #65 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1681296Z #66 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1682113Z #67 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1682860Z #68 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1683714Z #69 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1684642Z #70 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1685580Z #71 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1686498Z #72 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1687281Z #73 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1688055Z #74 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1688988Z #75 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1689905Z #76 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1690839Z #77 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1691765Z #78 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1692643Z #79 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1693439Z #80 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1694185Z #81 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1694896Z #82 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T09:54:48.1695580Z #83 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1696577Z #84 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1697577Z #85 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1698510Z #86 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1699431Z #87 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1700374Z #88 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1701158Z #89 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1701940Z #90 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1702865Z #91 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1703801Z #92 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1704858Z #93 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1705798Z #94 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1706571Z #95 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1707359Z #96 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1708295Z #97 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1709231Z #98 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1710153Z #99 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1711211Z #100 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1712116Z #101 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1712922Z #102 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1713696Z #103 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1714461Z #104 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1715340Z #105 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1716279Z #106 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1717075Z #107 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1717867Z #108 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1718809Z #109 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1719748Z #110 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1720695Z #111 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1721638Z #112 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1722527Z #113 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1723333Z #114 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1724097Z #115 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1724858Z #116 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1725725Z #117 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1726677Z #118 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1727624Z #119 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1728569Z #120 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1729503Z #121 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1730450Z #122 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1731246Z #123 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1732045Z #124 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1732979Z #125 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1733988Z #126 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1734946Z #127 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1735894Z #128 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1736764Z #129 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1737651Z #130 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1738409Z #131 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1739155Z #132 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1740106Z #133 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1741055Z #134 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1742010Z #135 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1742954Z #136 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1743901Z #137 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1744850Z #138 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1745802Z #139 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1746736Z #140 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1747690Z #141 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1748644Z #142 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1749473Z #143 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T09:54:48.1750218Z #144 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T09:54:48.1750956Z #145 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T09:54:48.1751664Z #146 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T09:54:48.1752444Z #147 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T09:54:48.1753283Z #148 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T09:54:48.1754060Z #149 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T09:54:48.1754786Z #150 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T09:54:48.1755476Z #151 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T09:54:48.1756092Z #152 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T09:54:48.1756534Z #153 _start from ??:0 2025-12-04T09:54:48.1756827Z #154 from ??:0 2025-12-04T09:54:48.1757076Z 2025-12-04T09:54:48.1757081Z 2025-12-04T09:54:48.1757298Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:48.1758454Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda 2025-12-04T09:54:48.1759373Z 2025-12-04T09:54:48.1759655Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:48.1760279Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:48.1761836Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py:257: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T09:54:48.1763332Z return type(self)(self.nbytes(), device=self.device).copy_(self) 2025-12-04T09:54:48.1764125Z _ AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda _ 2025-12-04T09:54:48.1764810Z Traceback (most recent call last): 2025-12-04T09:54:48.1765530Z File "/var/lib/jenkins/workspace/test/inductor/test_aot_inductor.py", line 6893, in test__weight_int4pack_mm 2025-12-04T09:54:48.1766273Z self.check_model(model, (a,)) 2025-12-04T09:54:48.1766939Z File "/var/lib/jenkins/workspace/test/inductor/test_aot_inductor_utils.py", line 247, in check_model 2025-12-04T09:54:48.1767628Z ref_model = copy.deepcopy(model) 2025-12-04T09:54:48.1768238Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 172, in deepcopy 2025-12-04T09:54:48.1768771Z y = _reconstruct(x, memo, *rv) 2025-12-04T09:54:48.1769284Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 271, in _reconstruct 2025-12-04T09:54:48.1769837Z state = deepcopy(state, memo) 2025-12-04T09:54:48.1770338Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 146, in deepcopy 2025-12-04T09:54:48.1770856Z y = copier(x, memo) 2025-12-04T09:54:48.1771329Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 231, in _deepcopy_dict 2025-12-04T09:54:48.1771936Z y[deepcopy(key, memo)] = deepcopy(value, memo) 2025-12-04T09:54:48.1772492Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 153, in deepcopy 2025-12-04T09:54:48.1772990Z y = copier(memo) 2025-12-04T09:54:48.1773582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_tensor.py", line 180, in __deepcopy__ 2025-12-04T09:54:48.1774317Z new_storage = self._typed_storage()._deepcopy(memo) 2025-12-04T09:54:48.1775021Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 1139, in _deepcopy 2025-12-04T09:54:48.1775837Z return self._new_wrapped_storage(copy.deepcopy(self._untyped_storage, memo)) 2025-12-04T09:54:48.1776529Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 153, in deepcopy 2025-12-04T09:54:48.1777127Z y = copier(memo) 2025-12-04T09:54:48.1777707Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 243, in __deepcopy__ 2025-12-04T09:54:48.1778388Z new_storage = self.clone() 2025-12-04T09:54:48.1778980Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 257, in clone 2025-12-04T09:54:48.1779733Z return type(self)(self.nbytes(), device=self.device).copy_(self) 2025-12-04T09:54:48.1780315Z torch.AcceleratorError: CUDA error: invalid device function 2025-12-04T09:54:48.1781318Z Search for `cudaErrorInvalidDeviceFunction' in https://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__TYPES.html for more information. 2025-12-04T09:54:48.1782590Z CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. 2025-12-04T09:54:48.1783383Z For debugging consider passing CUDA_LAUNCH_BLOCKING=1 2025-12-04T09:54:48.1783939Z Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. 2025-12-04T09:54:48.1784317Z 2025-12-04T09:54:48.1784935Z Exception raised from copy_device_to_device at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/Copy.cu:337 (most recent call first): 2025-12-04T09:54:48.1785811Z C++ CapturedTraceback: 2025-12-04T09:54:48.1787312Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T09:54:48.1789236Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T09:54:48.1790621Z #6 c10::AcceleratorError::AcceleratorError(c10::SourceLocation, int, std::__cxx11::basic_string, std::allocator > const&) from :0 2025-12-04T09:54:48.1791801Z #7 c10::cuda::c10_cuda_check_implementation(int, char const*, char const*, unsigned int, bool) from ??:0 2025-12-04T09:54:48.1792594Z #8 at::native::copy_device_to_device(at::TensorIterator&, bool, bool) from ??:0 2025-12-04T09:54:48.1793296Z #9 at::native::copy_impl(at::Tensor&, at::Tensor const&, bool) [clone .isra.0] from Copy.cpp:0 2025-12-04T09:54:48.1793969Z #10 at::native::copy_(at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T09:54:48.1796643Z #11 c10::impl::wrap_kernel_functor_unboxed_, at::Tensor&, c10::guts::typelist::typelist >, at::Tensor& (c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool) from VariableTypeManual.cpp:0 2025-12-04T09:54:48.1799631Z #12 torch::autograd::VariableType::(anonymous namespace)::copy_(c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool) from VariableTypeManual.cpp:0 2025-12-04T09:54:48.1800608Z #13 at::_ops::copy_::call(at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T09:54:48.1801199Z #14 at::storage_copy(c10::Storage&, c10::Storage const&, bool) from ??:0 2025-12-04T09:54:48.1801842Z #15 THPStorage_copy_(_object*, _object*, _object*) from StorageMethods.cpp:0 2025-12-04T09:54:48.1802680Z #16 method_vectorcall_VARARGS_KEYWORDS from /usr/local/src/conda/python-3.10.14/Objects/descrobject.c:344 2025-12-04T09:54:48.1803661Z #17 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1804601Z #18 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1805547Z #19 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1806478Z #20 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1807400Z #21 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1808328Z #22 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1809255Z #23 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1810187Z #24 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1811113Z #25 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1812051Z #26 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1812981Z #27 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1813912Z #28 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1814834Z #29 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1815761Z #30 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1816694Z #31 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1817670Z #32 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1818613Z #33 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1819637Z #34 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1820421Z #35 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1821191Z #36 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1822115Z #37 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1823048Z #38 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1823992Z #39 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1824908Z #40 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1825796Z #41 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1826508Z #42 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1827295Z #43 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1828101Z #44 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1828810Z #45 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1829596Z #46 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1830400Z #47 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1831105Z #48 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1831889Z #49 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1832834Z #50 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1833762Z #51 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1834545Z #52 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1835329Z #53 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1836267Z #54 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1837188Z #55 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1838118Z #56 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1839051Z #57 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1839987Z #58 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1840907Z #59 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1841835Z #60 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1842765Z #61 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1843569Z #62 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.1844266Z #63 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1845045Z #64 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1845918Z #65 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1846716Z #66 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1847463Z #67 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1848280Z #68 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1849149Z #69 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1850070Z #70 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1851016Z #71 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1851946Z #72 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1852719Z #73 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1853500Z #74 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1854494Z #75 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1855429Z #76 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1856346Z #77 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1857365Z #78 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1858231Z #79 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1859042Z #80 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1859779Z #81 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1860492Z #82 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T09:54:48.1861173Z #83 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1861937Z #84 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1862870Z #85 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1863803Z #86 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1864736Z #87 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1865650Z #88 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1866432Z #89 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1867213Z #90 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1868144Z #91 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1869063Z #92 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1870001Z #93 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1870927Z #94 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1871708Z #95 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1872475Z #96 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1873402Z #97 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1874335Z #98 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1875265Z #99 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1876198Z #100 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1877191Z #101 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1878011Z #102 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1878762Z #103 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1879526Z #104 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1880407Z #105 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1881359Z #106 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1882145Z #107 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1882999Z #108 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1883950Z #109 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1884897Z #110 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1885826Z #111 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1886774Z #112 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1887660Z #113 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1888470Z #114 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1889232Z #115 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1889998Z #116 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1890872Z #117 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1891817Z #118 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1892766Z #119 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1893713Z #120 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1894653Z #121 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1895586Z #122 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1896540Z #123 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.1897410Z #124 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1898362Z #125 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1899294Z #126 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1900238Z #127 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1901180Z #128 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1902066Z #129 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.1902871Z #130 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.1903630Z #131 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.1904392Z #132 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.1905261Z #133 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.1906330Z #134 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1907283Z #135 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1908231Z #136 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1909169Z #137 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1910117Z #138 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1911063Z #139 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1912016Z #140 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1913030Z #141 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1913988Z #142 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1914817Z #143 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T09:54:48.1915582Z #144 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T09:54:48.1916302Z #145 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T09:54:48.1917011Z #146 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T09:54:48.1917803Z #147 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T09:54:48.1918622Z #148 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T09:54:48.1919405Z #149 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T09:54:48.1920126Z #150 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T09:54:48.1920825Z #151 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T09:54:48.1921430Z #152 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T09:54:48.1921870Z #153 _start from ??:0 2025-12-04T09:54:48.1922175Z #154 from ??:0 2025-12-04T09:54:48.1922411Z 2025-12-04T09:54:48.1922415Z 2025-12-04T09:54:48.1922632Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:48.1923781Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda 2025-12-04T09:54:48.1924710Z 2025-12-04T09:54:48.1924978Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:48.1925619Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:48.1927102Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py:257: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T09:54:48.1928587Z return type(self)(self.nbytes(), device=self.device).copy_(self) 2025-12-04T09:54:48.1929102Z =================================== FAILURES =================================== 2025-12-04T09:54:48.1929832Z _ AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda _ 2025-12-04T09:54:48.1930519Z Traceback (most recent call last): 2025-12-04T09:54:48.1931237Z File "/var/lib/jenkins/workspace/test/inductor/test_aot_inductor.py", line 6893, in test__weight_int4pack_mm 2025-12-04T09:54:48.1931977Z self.check_model(model, (a,)) 2025-12-04T09:54:48.1932644Z File "/var/lib/jenkins/workspace/test/inductor/test_aot_inductor_utils.py", line 247, in check_model 2025-12-04T09:54:48.1933331Z ref_model = copy.deepcopy(model) 2025-12-04T09:54:48.1933916Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 172, in deepcopy 2025-12-04T09:54:48.1934454Z y = _reconstruct(x, memo, *rv) 2025-12-04T09:54:48.1934963Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 271, in _reconstruct 2025-12-04T09:54:48.1935513Z state = deepcopy(state, memo) 2025-12-04T09:54:48.1936010Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 146, in deepcopy 2025-12-04T09:54:48.1936527Z y = copier(x, memo) 2025-12-04T09:54:48.1937069Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 231, in _deepcopy_dict 2025-12-04T09:54:48.1937672Z y[deepcopy(key, memo)] = deepcopy(value, memo) 2025-12-04T09:54:48.1957422Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 153, in deepcopy 2025-12-04T09:54:48.1958066Z y = copier(memo) 2025-12-04T09:54:48.1958919Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_tensor.py", line 180, in __deepcopy__ 2025-12-04T09:54:48.1959652Z new_storage = self._typed_storage()._deepcopy(memo) 2025-12-04T09:54:48.1960350Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 1139, in _deepcopy 2025-12-04T09:54:48.1961154Z return self._new_wrapped_storage(copy.deepcopy(self._untyped_storage, memo)) 2025-12-04T09:54:48.1961842Z File "/opt/conda/envs/py_3.10/lib/python3.10/copy.py", line 153, in deepcopy 2025-12-04T09:54:48.1962342Z y = copier(memo) 2025-12-04T09:54:48.1962929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 243, in __deepcopy__ 2025-12-04T09:54:48.1963601Z new_storage = self.clone() 2025-12-04T09:54:48.1964183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py", line 257, in clone 2025-12-04T09:54:48.1964898Z return type(self)(self.nbytes(), device=self.device).copy_(self) 2025-12-04T09:54:48.1965481Z torch.AcceleratorError: CUDA error: invalid device function 2025-12-04T09:54:48.1966482Z Search for `cudaErrorInvalidDeviceFunction' in https://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__TYPES.html for more information. 2025-12-04T09:54:48.1967728Z CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. 2025-12-04T09:54:48.1968529Z For debugging consider passing CUDA_LAUNCH_BLOCKING=1 2025-12-04T09:54:48.1969067Z Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. 2025-12-04T09:54:48.1969431Z 2025-12-04T09:54:48.1970045Z Exception raised from copy_device_to_device at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/Copy.cu:337 (most recent call first): 2025-12-04T09:54:48.1970888Z C++ CapturedTraceback: 2025-12-04T09:54:48.1972381Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T09:54:48.1974311Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T09:54:48.1975588Z #6 c10::AcceleratorError::AcceleratorError(c10::SourceLocation, int, std::__cxx11::basic_string, std::allocator > const&) from :0 2025-12-04T09:54:48.1976756Z #7 c10::cuda::c10_cuda_check_implementation(int, char const*, char const*, unsigned int, bool) from ??:0 2025-12-04T09:54:48.1977620Z #8 at::native::copy_device_to_device(at::TensorIterator&, bool, bool) from ??:0 2025-12-04T09:54:48.1978316Z #9 at::native::copy_impl(at::Tensor&, at::Tensor const&, bool) [clone .isra.0] from Copy.cpp:0 2025-12-04T09:54:48.1978959Z #10 at::native::copy_(at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T09:54:48.1981549Z #11 c10::impl::wrap_kernel_functor_unboxed_, at::Tensor&, c10::guts::typelist::typelist >, at::Tensor& (c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool) from VariableTypeManual.cpp:0 2025-12-04T09:54:48.1984411Z #12 torch::autograd::VariableType::(anonymous namespace)::copy_(c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool) from VariableTypeManual.cpp:0 2025-12-04T09:54:48.1985363Z #13 at::_ops::copy_::call(at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T09:54:48.1985940Z #14 at::storage_copy(c10::Storage&, c10::Storage const&, bool) from ??:0 2025-12-04T09:54:48.1986558Z #15 THPStorage_copy_(_object*, _object*, _object*) from StorageMethods.cpp:0 2025-12-04T09:54:48.1987429Z #16 method_vectorcall_VARARGS_KEYWORDS from /usr/local/src/conda/python-3.10.14/Objects/descrobject.c:344 2025-12-04T09:54:48.1988401Z #17 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1989320Z #18 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1990237Z #19 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1991146Z #20 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1992054Z #21 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1992957Z #22 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1993862Z #23 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1994762Z #24 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1995671Z #25 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1996722Z #26 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1997629Z #27 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.1998534Z #28 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.1999440Z #29 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2000351Z #30 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2001262Z #31 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2002172Z #32 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2003103Z #33 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2004023Z #34 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2004794Z #35 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2005575Z #36 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2006509Z #37 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2007438Z #38 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2008356Z #39 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2009293Z #40 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2010257Z #41 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.2010966Z #42 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2011736Z #43 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2012553Z #44 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.2013263Z #45 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2014031Z #46 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2014844Z #47 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.2015545Z #48 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2016406Z #49 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2017385Z #50 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2018314Z #51 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2019097Z #52 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2019875Z #53 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2020791Z #54 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2021720Z #55 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2022647Z #56 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2023586Z #57 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2024506Z #58 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2025437Z #59 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2026366Z #60 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2027283Z #61 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2028102Z #62 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.2028808Z #63 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2029587Z #64 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2030457Z #65 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.2031268Z #66 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.2032033Z #67 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.2032784Z #68 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.2033640Z #69 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.2034573Z #70 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2035505Z #71 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2036435Z #72 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2037211Z #73 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2037991Z #74 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2039055Z #75 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2039977Z #76 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2040908Z #77 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2041841Z #78 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2042717Z #79 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.2043514Z #80 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.2044270Z #81 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.2045051Z #82 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T09:54:48.2045730Z #83 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2046497Z #84 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2047429Z #85 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2048359Z #86 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2049278Z #87 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2050207Z #88 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2050994Z #89 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2051778Z #90 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2052703Z #91 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2053635Z #92 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2054569Z #93 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2055499Z #94 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2056267Z #95 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2057121Z #96 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2058050Z #97 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2058985Z #98 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2059897Z #99 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2060844Z #100 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2061738Z #101 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.2062548Z #102 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.2063312Z #103 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.2064069Z #104 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.2064947Z #105 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.2065879Z #106 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2066683Z #107 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2067550Z #108 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2068502Z #109 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2069438Z #110 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2070386Z #111 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2071325Z #112 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2072223Z #113 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.2073031Z #114 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.2073853Z #115 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.2074614Z #116 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.2075472Z #117 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.2076417Z #118 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2077364Z #119 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2078307Z #120 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2079243Z #121 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2080182Z #122 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2080980Z #123 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2081773Z #124 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2082192Z #125 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2082570Z #126 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2082997Z #127 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2083375Z #128 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2083748Z #129 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.2084060Z #130 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.2084365Z #131 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.2084687Z #132 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.2085108Z #133 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.2085499Z #134 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2085912Z #135 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2086292Z #136 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2086720Z #137 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2087099Z #138 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2087517Z #139 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2087905Z #140 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2088368Z #141 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2088758Z #142 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2089049Z #143 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T09:54:48.2089358Z #144 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T09:54:48.2089640Z #145 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T09:54:48.2089925Z #146 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T09:54:48.2090290Z #147 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T09:54:48.2090695Z #148 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T09:54:48.2090991Z #149 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T09:54:48.2091275Z #150 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T09:54:48.2091544Z #151 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T09:54:48.2091744Z #152 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T09:54:48.2091861Z #153 _start from ??:0 2025-12-04T09:54:48.2091985Z #154 from ??:0 2025-12-04T09:54:48.2091992Z 2025-12-04T09:54:48.2091997Z 2025-12-04T09:54:48.2092232Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:48.2093032Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda 2025-12-04T09:54:48.2093043Z 2025-12-04T09:54:48.2093315Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:48.2093557Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:48.2094683Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/storage.py:257: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T09:54:48.2094924Z return type(self)(self.nbytes(), device=self.device).copy_(self) 2025-12-04T09:54:48.2095677Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-bf1966c65cc0ce25.xml - 2025-12-04T09:54:48.2095863Z =========================== short test summary info ============================ 2025-12-04T09:54:48.2097048Z FAILED [0.0058s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda - torch.AcceleratorError: CUDA error: invalid device function 2025-12-04T09:54:48.2097704Z Search for `cudaErrorInvalidDeviceFunction' in https://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__TYPES.html for more information. 2025-12-04T09:54:48.2098196Z CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect. 2025-12-04T09:54:48.2098384Z For debugging consider passing CUDA_LAUNCH_BLOCKING=1 2025-12-04T09:54:48.2098640Z Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. 2025-12-04T09:54:48.2098646Z 2025-12-04T09:54:48.2099258Z Exception raised from copy_device_to_device at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/Copy.cu:337 (most recent call first): 2025-12-04T09:54:48.2099373Z C++ CapturedTraceback: 2025-12-04T09:54:48.2100804Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T09:54:48.2101299Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T09:54:48.2101977Z #6 c10::AcceleratorError::AcceleratorError(c10::SourceLocation, int, std::__cxx11::basic_string, std::allocator > const&) from :0 2025-12-04T09:54:48.2102349Z #7 c10::cuda::c10_cuda_check_implementation(int, char const*, char const*, unsigned int, bool) from ??:0 2025-12-04T09:54:48.2102618Z #8 at::native::copy_device_to_device(at::TensorIterator&, bool, bool) from ??:0 2025-12-04T09:54:48.2102939Z #9 at::native::copy_impl(at::Tensor&, at::Tensor const&, bool) [clone .isra.0] from Copy.cpp:0 2025-12-04T09:54:48.2103154Z #10 at::native::copy_(at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T09:54:48.2105387Z #11 c10::impl::wrap_kernel_functor_unboxed_, at::Tensor&, c10::guts::typelist::typelist >, at::Tensor& (c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool) from VariableTypeManual.cpp:0 2025-12-04T09:54:48.2106001Z #12 torch::autograd::VariableType::(anonymous namespace)::copy_(c10::DispatchKeySet, at::Tensor&, at::Tensor const&, bool) from VariableTypeManual.cpp:0 2025-12-04T09:54:48.2106236Z #13 at::_ops::copy_::call(at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T09:54:48.2106469Z #14 at::storage_copy(c10::Storage&, c10::Storage const&, bool) from ??:0 2025-12-04T09:54:48.2106739Z #15 THPStorage_copy_(_object*, _object*, _object*) from StorageMethods.cpp:0 2025-12-04T09:54:48.2107178Z #16 method_vectorcall_VARARGS_KEYWORDS from /usr/local/src/conda/python-3.10.14/Objects/descrobject.c:344 2025-12-04T09:54:48.2107598Z #17 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2107976Z #18 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2108401Z #19 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2108776Z #20 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2109200Z #21 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2109572Z #22 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2109983Z #23 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2110367Z #24 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2110778Z #25 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2111159Z #26 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2111567Z #27 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2111938Z #28 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2112357Z #29 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2112729Z #30 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2113151Z #31 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2113521Z #32 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2113985Z #33 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2114370Z #34 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2114632Z #35 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2115016Z #36 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2115423Z #37 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2115792Z #38 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2116264Z #39 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2116633Z #40 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2116932Z #41 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.2117203Z #42 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2117574Z #43 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2117881Z #44 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.2118139Z #45 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2118509Z #46 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2118813Z #47 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.2119075Z #48 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2119456Z #49 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2119873Z #50 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2120241Z #51 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2120505Z #52 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2120874Z #53 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2121285Z #54 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2121656Z #55 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2122062Z #56 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2122443Z #57 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2122854Z #58 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2123222Z #59 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2123640Z #60 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2124008Z #61 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2124306Z #62 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T09:54:48.2124561Z #63 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2124929Z #64 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2125292Z #65 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.2125655Z #66 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.2125956Z #67 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.2126259Z #68 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.2126666Z #69 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.2127043Z #70 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2127449Z #71 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2127831Z #72 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2128148Z #73 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2128516Z #74 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2128931Z #75 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2129299Z #76 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2129702Z #77 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2130079Z #78 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2130430Z #79 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.2130744Z #80 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.2131043Z #81 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.2131310Z #82 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T09:54:48.2131582Z #83 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2131954Z #84 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2132372Z #85 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2132740Z #86 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2133142Z #87 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2133518Z #88 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2133773Z #89 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2134158Z #90 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2134562Z #91 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2134934Z #92 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2135352Z #93 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2135717Z #94 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2135972Z #95 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2136348Z #96 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2136750Z #97 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2137211Z #98 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2137683Z #99 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2138068Z #100 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2138434Z #101 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.2138739Z #102 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.2139048Z #103 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.2139351Z #104 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.2139760Z #105 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.2140229Z #106 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2140488Z #107 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2140868Z #108 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2141295Z #109 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2141670Z #110 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2142094Z #111 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2142556Z #112 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2142911Z #113 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.2143223Z #114 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.2143531Z #115 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.2143839Z #116 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.2144258Z #117 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.2144631Z #118 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2145042Z #119 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2145425Z #120 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2145836Z #121 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2146213Z #122 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2146494Z #123 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T09:54:48.2146874Z #124 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2147291Z #125 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2147663Z #126 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2148076Z #127 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2148460Z #128 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2148813Z #129 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T09:54:48.2149132Z #130 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T09:54:48.2149431Z #131 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T09:54:48.2149795Z #132 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T09:54:48.2150219Z #133 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T09:54:48.2150594Z #134 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2151015Z #135 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2151390Z #136 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2151801Z #137 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2152188Z #138 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2152660Z #139 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2153038Z #140 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2153464Z #141 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T09:54:48.2153840Z #142 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T09:54:48.2154141Z #143 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T09:54:48.2154447Z #144 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T09:54:48.2154714Z #145 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T09:54:48.2155011Z #146 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T09:54:48.2155366Z #147 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T09:54:48.2155700Z #148 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T09:54:48.2155995Z #149 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T09:54:48.2156263Z #150 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T09:54:48.2156541Z #151 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T09:54:48.2156739Z #152 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T09:54:48.2156839Z #153 _start from ??:0 2025-12-04T09:54:48.2156973Z #154 from ??:0 2025-12-04T09:54:48.2156978Z 2025-12-04T09:54:48.2156984Z 2025-12-04T09:54:48.2157200Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:48.2157997Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda 2025-12-04T09:54:48.2158008Z 2025-12-04T09:54:48.2158277Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:48.2158459Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T09:54:48.2158674Z ================= 1 failed, 149 deselected, 2 rerun in 16.32s ================== 2025-12-04T09:54:48.2158774Z Got exit code 1 2025-12-04T09:54:48.2159487Z FAILED CONSISTENTLY: test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda 2025-12-04T09:54:48.2159900Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T09:54:48.2160337Z W1204 09:49:37.220000 5404 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T09:54:48.2160919Z Test results will be stored in test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-4d0cbc1f4fa8e643.xml 2025-12-04T09:54:48.2161085Z ============================= test session starts ============================== 2025-12-04T09:54:48.2161506Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:54:48.2161615Z cachedir: .pytest_cache 2025-12-04T09:54:48.2162134Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:54:48.2162262Z rootdir: /var/lib/jenkins/workspace 2025-12-04T09:54:48.2162369Z configfile: pytest.ini 2025-12-04T09:54:48.2162914Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:54:48.2163149Z collecting ... collected 934 items / 54 deselected / 880 selected 2025-12-04T09:54:48.2163295Z stepcurrent: skipping 54 already run items. 2025-12-04T09:54:48.2163482Z Running 96 items in this shard 2025-12-04T09:54:48.2163488Z 2025-12-04T09:54:48.2164334Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_with_scales_and_zeros_m_32_n_64_q_group_32_num_groups_2_cuda SKIPPED [0.0042s] (requires Intel GPU) [ 1%] 2025-12-04T09:54:48.2165173Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_with_scales_and_zeros_m_32_n_64_q_group_64_num_groups_1_cuda SKIPPED [0.0029s] (requires Intel GPU) [ 2%] 2025-12-04T09:54:48.2166339Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_aliased_buffer_reuse_cuda <- test/inductor/test_torchinductor.py W1204 09:49:40.887000 5404 site-packages/torch/_inductor/utils.py:1703] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-12-04T09:54:48.2166446Z PASSED [10.1833s] [ 3%] 2025-12-04T09:54:48.2167200Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_aoti_debug_printer_sym_inputs_cuda <- test/inductor/test_torchinductor.py PASSED [6.3218s] [ 4%] 2025-12-04T09:54:48.2168000Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_aoti_debug_printing_model_inputs_codegen_cuda <- test/inductor/test_torchinductor.py PASSED [6.1084s] [ 5%] 2025-12-04T09:54:48.2168635Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_aoti_profiler_enable_kernel_profile_False_cuda PASSED [11.4398s] [ 6%] 2025-12-04T09:54:48.2169246Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_aoti_profiler_enable_kernel_profile_True_cuda PASSED [11.8252s] [ 7%] 2025-12-04T09:54:48.2170048Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_aoti_user_defined_triton_kernel_profiling_cuda <- test/inductor/test_torchinductor.py PASSED [6.4015s] [ 8%] 2025-12-04T09:54:48.2170667Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_autotune_int64_user_defined_triton_kernel_cuda PASSED [7.5688s] [ 9%] 2025-12-04T09:54:48.2171186Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_backward_no_op_logging_cuda PASSED [0.0065s] [ 10%] 2025-12-04T09:54:48.2172241Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_buffer_mutation_4_cuda <- test/inductor/test_torchinductor.py W1204 09:50:38.995000 5404 site-packages/torch/_export/__init__.py:71] +============================+ 2025-12-04T09:54:48.2172590Z W1204 09:50:38.995000 5404 site-packages/torch/_export/__init__.py:72] | !!! WARNING !!! | 2025-12-04T09:54:48.2172930Z W1204 09:50:38.996000 5404 site-packages/torch/_export/__init__.py:73] +============================+ 2025-12-04T09:54:48.2174170Z W1204 09:50:38.996000 5404 site-packages/torch/_export/__init__.py:74] torch._export.aot_compile()/torch._export.aot_load() is being deprecated, please switch to directly calling torch._inductor.aoti_compile_and_package(torch.export.export())/torch._inductor.aoti_load_package() instead. 2025-12-04T09:54:48.2174270Z PASSED [5.7600s] [ 11%] 2025-12-04T09:54:48.2174954Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_buffer_reuse_cuda <- test/inductor/test_torchinductor.py PASSED [6.1311s] [ 12%] 2025-12-04T09:54:48.2175676Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_cond_cpu_predicate_cuda_operands_max_autotune_False_cuda PASSED [6.3817s] [ 13%] 2025-12-04T09:54:48.2176394Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_cond_predicate_on_cpu_cuda <- test/inductor/test_torchinductor.py PASSED [5.9075s] [ 14%] 2025-12-04T09:54:48.2177877Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_cond_with_outer_code_before_after_cuda <- test/inductor/test_torchinductor.py W1204 09:51:03.255000 5404 site-packages/torch/export/dynamic_shapes.py:923] Using None as a dynamic shape dimension is deprecated. Please use Dim.STATIC instead 2025-12-04T09:54:48.2178530Z W1204 09:51:03.255000 5404 site-packages/torch/export/dynamic_shapes.py:923] Using None as a dynamic shape dimension is deprecated. Please use Dim.STATIC instead 2025-12-04T09:54:48.2178954Z PASSED [6.5947s] [ 15%] 2025-12-04T09:54:48.2179491Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_foreach_multiple_dynamic_cuda PASSED [5.7604s] [ 16%] 2025-12-04T09:54:48.2180195Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_index_put_fallback_cuda <- test/inductor/test_torchinductor.py PASSED [5.5426s] [ 17%] 2025-12-04T09:54:48.2180871Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_libtorch_free_so_cuda <- test/inductor/test_torchinductor.py PASSED [9.6178s] [ 18%] 2025-12-04T09:54:48.2181583Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_load_package_multiple_gpus_cuda SKIPPED [0.0003s] (requires multiple cuda devices) [ 19%] 2025-12-04T09:54:48.2182115Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_misc_1_max_autotune_True_cuda PASSED [6.6044s] [ 20%] 2025-12-04T09:54:48.2182823Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_multiple_output_alias_cuda <- test/inductor/test_torchinductor.py PASSED [5.5407s] [ 21%] 2025-12-04T09:54:48.2184821Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_nested_tensor_from_jagged_cuda <- test/inductor/test_torchinductor.py W1204 09:51:48.804000 5404 site-packages/torch/fx/_symbolic_trace.py:53] is_fx_tracing will return true for both fx.symbolic_trace and torch.export. Please use is_fx_tracing_symbolic_tracing() for specifically fx.symbolic_trace or torch.compiler.is_compiling() for specifically torch.export/compile. 2025-12-04T09:54:48.2184926Z PASSED [18.5612s] [ 22%] 2025-12-04T09:54:48.2185595Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_output_path_2_cuda <- test/inductor/test_torchinductor.py PASSED [5.5993s] [ 23%] 2025-12-04T09:54:48.2186259Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_pad_fallback_cuda <- test/inductor/test_torchinductor.py PASSED [5.8814s] [ 25%] 2025-12-04T09:54:48.2186797Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_quantized_linear_bias_none_cuda XFAIL [0.0332s] [ 26%] 2025-12-04T09:54:48.2187525Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_user_defined_triton_kernel_embed_kernel_binary_True_cuda PASSED [10.7708s] [ 27%] 2025-12-04T09:54:48.2188224Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_return_view_constant_cuda <- test/inductor/test_torchinductor.py PASSED [5.0460s] [ 28%] 2025-12-04T09:54:48.2188891Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_simple_embed_kernel_binary_False_max_autotune_False_cuda PASSED [6.7776s] [ 29%] 2025-12-04T09:54:48.2189543Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_simple_embed_kernel_binary_False_max_autotune_True_cuda PASSED [6.8405s] [ 30%] 2025-12-04T09:54:48.2190201Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_simple_embed_kernel_binary_True_max_autotune_True_cuda PASSED [6.7874s] [ 31%] 2025-12-04T09:54:48.2190677Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_simple_split_cuda PASSED [5.5865s] [ 32%] 2025-12-04T09:54:48.2191406Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_symbool_item_cuda <- test/inductor/test_torchinductor.py PASSED [5.0112s] [ 33%] 2025-12-04T09:54:48.2192132Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_kernel_grid_type_1_num_dims_1_dynamic_False_autotune_True_cuda PASSED [6.4816s] [ 34%] 2025-12-04T09:54:48.2192837Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_kernel_grid_type_2_num_dims_1_dynamic_False_autotune_True_cuda PASSED [5.9798s] [ 35%] 2025-12-04T09:54:48.2193556Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_kernel_grid_type_3_num_dims_2_dynamic_False_autotune_False_cuda PASSED [5.8711s] [ 36%] 2025-12-04T09:54:48.2194261Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_kernel_grid_type_3_num_dims_2_dynamic_False_autotune_True_cuda PASSED [9.1043s] [ 37%] 2025-12-04T09:54:48.2195059Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_kernel_multi_output_arg_cuda <- test/inductor/test_torchinductor.py PASSED [6.1743s] [ 38%] 2025-12-04T09:54:48.2195813Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_kernel_unbacked_symint_in_grid_dynamic_False_autotuning_False_cuda PASSED [5.8243s] [ 39%] 2025-12-04T09:54:48.2196744Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_kernel_unbacked_symint_in_grid_dynamic_False_autotuning_True_cuda PASSED [5.8908s] [ 40%] 2025-12-04T09:54:48.2197493Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_kernel_with_none_input_cuda <- test/inductor/test_torchinductor.py PASSED [6.2110s] [ 41%] 2025-12-04T09:54:48.2198315Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_kernel_with_none_inputs_and_equal_to_1_arg_cuda <- test/inductor/test_torchinductor.py PASSED [5.9634s] [ 42%] 2025-12-04T09:54:48.2199029Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_triton_next_power_of_2_cuda <- test/inductor/test_torchinductor.py PASSED [6.5415s] [ 43%] 2025-12-04T09:54:48.2199834Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_unbacked_equals_input_size_runtime_assertion_mark_unbacked_False_cuda Error: Expected u0 >= 2 but received 0 2025-12-04T09:54:48.2199941Z PASSED [11.2446s] [ 44%] 2025-12-04T09:54:48.2200639Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_unbacked_expr_replacements_shift_k_1_use_static_size_True_cuda PASSED [7.8550s] [ 45%] 2025-12-04T09:54:48.2201396Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_update_inactive_constant_buffer_cuda <- test/inductor/test_torchinductor.py PASSED [5.2282s] [ 46%] 2025-12-04T09:54:48.2202065Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_view_outputs_cuda <- test/inductor/test_torchinductor.py PASSED [5.6873s] [ 47%] 2025-12-04T09:54:48.2202803Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_while_loop_with_outer_buffers_cuda <- test/inductor/test_torchinductor.py PASSED [6.6564s] [ 48%] 2025-12-04T09:54:48.2203486Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_with_cudagraphs_cuda <- test/inductor/test_torchinductor.py PASSED [5.2262s] [ 50%] 2025-12-04T09:54:48.2204260Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_1_mps SKIPPED [0.0004s] (No MPS backend available) [ 51%] 2025-12-04T09:54:48.2205124Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test__weight_int4pack_mm_with_scales_and_zeros_m_32_n_64_q_group_64_num_groups_2_mps SKIPPED [0.0002s] (No MPS backend available) [ 52%] 2025-12-04T09:54:48.2205838Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_aoti_runtime_asserts_backed_symint_mps SKIPPED [0.0002s] (No MPS backend available) [ 53%] 2025-12-04T09:54:48.2206440Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_assert_async_mps SKIPPED [0.0002s] (No MPS backend available) [ 54%] 2025-12-04T09:54:48.2207384Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_clamp_decomposition_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0005s] (No MPS backend available) [ 55%] 2025-12-04T09:54:48.2208111Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_cond_non_tensor_predicates_dynamic_True_mps SKIPPED [0.0002s] (No MPS backend available) [ 56%] 2025-12-04T09:54:48.2208942Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_cond_share_predicate_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (No MPS backend available) [ 57%] 2025-12-04T09:54:48.2209703Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_cond_unbacked_symint_closure_dynamic_False_mps SKIPPED [0.0002s] (No MPS backend available) [ 58%] 2025-12-04T09:54:48.2210557Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_copy_non_blocking_is_pinned_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (No MPS backend available) [ 59%] 2025-12-04T09:54:48.2211270Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_deconv_freezing_mps SKIPPED [0.0002s] (No MPS backend available) [ 60%] 2025-12-04T09:54:48.2212172Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_dup_unbacked_sym_decl_with_refinement_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (No MPS backend available) [ 61%] 2025-12-04T09:54:48.2212940Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_fft_c2c_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (No MPS backend available) [ 62%] 2025-12-04T09:54:48.2213807Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_fx_gm_return_tuple_validation_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (No MPS backend available) [ 63%] 2025-12-04T09:54:48.2214669Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_input_codegen_with_sympy_expr_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0004s] (No MPS backend available) [ 64%] 2025-12-04T09:54:48.2215308Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_large_dynamic_dim_mps SKIPPED [0.0002s] (No MPS backend available) [ 65%] 2025-12-04T09:54:48.2216168Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_large_mmaped_weights_on_disk_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (No MPS backend available) [ 66%] 2025-12-04T09:54:48.2217040Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_linear_freezing_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (No MPS backend available) [ 67%] 2025-12-04T09:54:48.2217831Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_multi_device_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (No MPS backend available) [ 68%] 2025-12-04T09:54:48.2218606Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_no_args_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0016s] (No MPS backend available) [ 69%] 2025-12-04T09:54:48.2219431Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_normal_functional_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (No MPS backend available) [ 70%] 2025-12-04T09:54:48.2220246Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_proxy_executor_abs_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (No MPS backend available) [ 71%] 2025-12-04T09:54:48.2221094Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_proxy_executor_squeeze_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (No MPS backend available) [ 72%] 2025-12-04T09:54:48.2221889Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_repeat_output_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (No MPS backend available) [ 73%] 2025-12-04T09:54:48.2222552Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_rocm_triton_autotuning_mps SKIPPED [0.0002s] (No MPS backend available) [ 75%] 2025-12-04T09:54:48.2223280Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_runtime_checks_shape_failed_mps SKIPPED [0.0002s] (No MPS backend available) [ 76%] 2025-12-04T09:54:48.2224077Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_simple_embed_kernel_binary_False_max_autotune_True_mps SKIPPED [0.0002s] (No MPS backend available) [ 77%] 2025-12-04T09:54:48.2224853Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_simple_embed_kernel_binary_True_max_autotune_False_mps SKIPPED [0.0002s] (No MPS backend available) [ 78%] 2025-12-04T09:54:48.2225544Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_size_with_unbacked_add_and_mul_expr_mps SKIPPED [0.0002s] (No MPS backend available) [ 79%] 2025-12-04T09:54:48.2226303Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_stft_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (No MPS backend available) [ 80%] 2025-12-04T09:54:48.2227148Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_symfloat_item_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0004s] (No MPS backend available) [ 81%] 2025-12-04T09:54:48.2227863Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_sympy_cpp_printer_min_max_minmax0_mps SKIPPED [0.0002s] (No MPS backend available) [ 82%] 2025-12-04T09:54:48.2228842Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_torchvision_transforms_functional_tensor_resize_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (No MPS backend available) [ 83%] 2025-12-04T09:54:48.2229699Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_bool_param_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (No MPS backend available) [ 84%] 2025-12-04T09:54:48.2230602Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_dynamic_shape_with_div_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (No MPS backend available) [ 85%] 2025-12-04T09:54:48.2231431Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_grid_type_1_num_dims_1_dynamic_True_autotune_False_mps SKIPPED [0.0004s] (No MPS backend available) [ 86%] 2025-12-04T09:54:48.2232276Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_grid_type_3_num_dims_1_dynamic_False_autotune_True_mps SKIPPED [0.0002s] (No MPS backend available) [ 87%] 2025-12-04T09:54:48.2233101Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_grid_type_3_num_dims_1_dynamic_True_autotune_False_mps SKIPPED [0.0002s] (No MPS backend available) [ 88%] 2025-12-04T09:54:48.2233928Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_grid_type_3_num_dims_2_dynamic_True_autotune_True_mps SKIPPED [0.0002s] (No MPS backend available) [ 89%] 2025-12-04T09:54:48.2234724Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_on_device_tma_dynamic_True_tma_version_new_mps SKIPPED [0.0002s] (No MPS backend available) [ 90%] 2025-12-04T09:54:48.2235557Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_tma_descriptor_1d_dynamic_False_tma_version_new_mps SKIPPED [0.0002s] (No MPS backend available) [ 91%] 2025-12-04T09:54:48.2236374Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_tma_descriptor_1d_dynamic_False_tma_version_old_mps SKIPPED [0.0002s] (No MPS backend available) [ 92%] 2025-12-04T09:54:48.2237195Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_tma_descriptor_1d_dynamic_True_tma_version_old_mps SKIPPED [0.0002s] (No MPS backend available) [ 93%] 2025-12-04T09:54:48.2238020Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_tma_descriptor_2d_dynamic_False_tma_version_old_mps SKIPPED [0.0002s] (No MPS backend available) [ 94%] 2025-12-04T09:54:48.2238906Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_tma_descriptor_2d_dynamic_True_tma_version_new_mps SKIPPED [0.0002s] (No MPS backend available) [ 95%] 2025-12-04T09:54:48.2239797Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_triton_kernel_weird_param_order_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (No MPS backend available) [ 96%] 2025-12-04T09:54:48.2240603Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_unbacked_expr_replacements_shift_k_3_use_static_size_True_mps SKIPPED [0.0002s] (No MPS backend available) [ 97%] 2025-12-04T09:54:48.2241454Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_update_constant_buffer_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (No MPS backend available) [ 98%] 2025-12-04T09:54:48.2242260Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleMps::test_zero_size_weight_mps <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (No MPS backend available) [100%] 2025-12-04T09:54:48.2242393Z 2025-12-04T09:54:48.2243152Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-4d0cbc1f4fa8e643.xml - 2025-12-04T09:54:48.2243412Z ===== 44 passed, 51 skipped, 54 deselected, 1 xfailed in 306.77s (0:05:06) ===== 2025-12-04T09:54:48.2244209Z The following tests failed consistently: ['test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cuda'] 2025-12-04T09:54:48.2244215Z 2025-12-04T09:54:48.2244787Z FINISHED PRINTING LOG FILE of inductor/test_aot_inductor 3/6 (test/test-reports/inductor.test_aot_inductor_3.6_01eb6db542665098_.log) 2025-12-04T09:54:48.2244793Z 2025-12-04T09:54:48.2245134Z Finished inductor/test_aot_inductor 3/6 ... [2025-12-04 09:54:48.051498][2471.741884823], took 9.83min 2025-12-04T09:54:48.2245954Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-cf40a4f46675676c.xml 2025-12-04T09:54:48.3774027Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-3767af6d1470a6b5.xml 2025-12-04T09:54:48.4131799Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-bf1966c65cc0ce25.xml 2025-12-04T09:54:48.4477321Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-4d0cbc1f4fa8e643.xml 2025-12-04T09:54:48.6539422Z Uploading logs for 57119749282 to S3 2025-12-04T09:54:48.6882112Z Uploading artifacts took 0.20 seconds 2025-12-04T09:54:48.6882565Z inductor/test_aot_inductor 3/6 failed! 2025-12-04T09:54:48.6887140Z Running inductor/test_torchinductor 2/2 ... [2025-12-04 09:54:48.688538][2472.378931164] 2025-12-04T09:54:48.6887731Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T09:54:48.6892018Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor.py', '--shard-id=2', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 09:54:48.688970] 2025-12-04T10:05:23.2422527Z 2025-12-04T10:05:23.2423683Z PRINTING LOG FILE of inductor/test_torchinductor 2/2 (test/test-reports/inductor.test_torchinductor_2.2_198ec367be5c71a3_.log) 2025-12-04T10:05:23.2425496Z W1204 09:54:57.626000 13298 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:05:23.2427410Z Test results will be stored in test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-6ce11cbc918a017c.xml 2025-12-04T10:05:23.2428824Z ============================= test session starts ============================== 2025-12-04T10:05:23.2429923Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T10:05:23.2430859Z cachedir: .pytest_cache 2025-12-04T10:05:23.2432505Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T10:05:23.2433708Z rootdir: /var/lib/jenkins/workspace 2025-12-04T10:05:23.2434231Z configfile: pytest.ini 2025-12-04T10:05:23.2435206Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T10:05:23.2436152Z collecting ... collected 981 items 2025-12-04T10:05:23.2436807Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T10:05:23.2647489Z Running 518 items in this shard: test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast1_broadcast1, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast1_broadcast3, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast1_dense, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast1_double, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast1_int, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast2_broadcast1, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast2_broadcast3, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast2_dense, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast2_strided, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast3_broadcast1, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast3_broadcast2, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast3_broadcast3, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast3_double, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast3_strided, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast3_transposed, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_dense_broadcast2, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_dense_int, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_dense_transposed, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_double_broadcast1, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_double_broadcast3, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_double_dense, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_int_broadcast1, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_int_broadcast2, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_int_dense, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_int_double, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_int_int, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_strided_broadcast2, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_strided_dense, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_strided_double, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_strided_int, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_transposed_broadcast1, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_transposed_broadcast3, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_transposed_dense, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_transposed_strided, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_transposed_transposed, test/inductor/test_torchinductor.py::GPUTests::test__dyn_quant_matmul_4bit_bf16_input_cuda, test/inductor/test_torchinductor.py::GPUTests::test__dyn_quant_matmul_4bit_fp32_input_cuda, test/inductor/test_torchinductor.py::GPUTests::test__dyn_quant_pack_4bit_weight_fp32_cuda, test/inductor/test_torchinductor.py::GPUTests::test__unsafe_masked_index_cuda, test/inductor/test_torchinductor.py::GPUTests::test_adaptive_avg_pool2d2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_adaptive_avg_pool2d_low_prec_cuda, test/inductor/test_torchinductor.py::GPUTests::test_adaptive_avg_pool_errors_with_long_cuda, test/inductor/test_torchinductor.py::GPUTests::test_adaptive_max_pool2d2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_add_complex3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_add_complex7_cuda, test/inductor/test_torchinductor.py::GPUTests::test_add_complex_cuda, test/inductor/test_torchinductor.py::GPUTests::test_add_complex_strided_fallback_cuda, test/inductor/test_torchinductor.py::GPUTests::test_add_const_int_cuda, test/inductor/test_torchinductor.py::GPUTests::test_adding_tensor_offsets_cuda, test/inductor/test_torchinductor.py::GPUTests::test_addmv_cuda, test/inductor/test_torchinductor.py::GPUTests::test_aliased_buffer_reuse_cuda, test/inductor/test_torchinductor.py::GPUTests::test_allow_reuse_disable_if_exceed_peak_cuda, test/inductor/test_torchinductor.py::GPUTests::test_aoti_eager_cache_hit_cuda, test/inductor/test_torchinductor.py::GPUTests::test_aoti_eager_dtype_device_layout_cuda, test/inductor/test_torchinductor.py::GPUTests::test_aoti_eager_override_registration_cuda, test/inductor/test_torchinductor.py::GPUTests::test_aoti_eager_support_out_cuda, test/inductor/test_torchinductor.py::GPUTests::test_aoti_eager_support_str_cuda, test/inductor/test_torchinductor.py::GPUTests::test_arange3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_arange4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_arange5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_argmax_argmin3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_argmax_argmin_with_duplicates_cuda, test/inductor/test_torchinductor.py::GPUTests::test_argmax_argmin_with_nan_cuda, test/inductor/test_torchinductor.py::GPUTests::test_argmax_min_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_argmax_to_float_cuda, test/inductor/test_torchinductor.py::GPUTests::test_as_strided_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_assert_alignment_op_name_pass_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool2d1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool2d2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool2d3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool2d5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool2d_backward2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool2d_backward3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool2d_backward4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool2d_backward_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool3d_backward2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool3d_backward3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool3d_backward_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool_errors_with_uint_cuda, test/inductor/test_torchinductor.py::GPUTests::test_baddbmm_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bernoulli1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bernoulli2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bfloat16_to_int16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bitwise3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bitwise_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bmm1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bool_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_broadcast_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int16_int16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int16_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int16_int64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int16_int8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int16_uint8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int32_int64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int32_int8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int8_int16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int8_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int8_int64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_uint8_int16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_uint8_uint8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_nd_tiling_False_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_nd_tiling_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_buffer_batch_norm_cuda, test/inductor/test_torchinductor.py::GPUTests::test_buffer_copied_in_graph_cuda, test/inductor/test_torchinductor.py::GPUTests::test_buffer_copied_in_graph_with_different_shapes_cuda, test/inductor/test_torchinductor.py::GPUTests::test_buffer_use_after_remove_cuda, test/inductor/test_torchinductor.py::GPUTests::test_builtins_round_cuda, test/inductor/test_torchinductor.py::GPUTests::test_builtins_round_float_ndigits_pos_cuda, test/inductor/test_torchinductor.py::GPUTests::test_builtins_round_float_ndigits_zero_cuda, test/inductor/test_torchinductor.py::GPUTests::test_builtins_round_int_ndigits_zero_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cat_extern_kernel_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cat_inplace_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cat_negative_dim_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cat_of_loops_and_extern_kernel_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cat_uint8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cat_unbacked_2d_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cat_unbacked_legacy_empty_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cat_upcasting_cuda, test/inductor/test_torchinductor.py::GPUTests::test_check_stack_no_cycles_cuda, test/inductor/test_torchinductor.py::GPUTests::test_clamp_cuda, test/inductor/test_torchinductor.py::GPUTests::test_config_option_dont_assume_alignment_cuda, test/inductor/test_torchinductor.py::GPUTests::test_consecutive_split_cumsum_cuda, test/inductor/test_torchinductor.py::GPUTests::test_constant_pad_1d_cuda, test/inductor/test_torchinductor.py::GPUTests::test_constant_pad_2d_cuda, test/inductor/test_torchinductor.py::GPUTests::test_constant_pad_float64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_conv1d_depthwise_cuda, test/inductor/test_torchinductor.py::GPUTests::test_conv1d_with_permute_cuda, test/inductor/test_torchinductor.py::GPUTests::test_conv3d_channels_last_use_block_ptr_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_conv_backward_cuda, test/inductor/test_torchinductor.py::GPUTests::test_conv_bn_fuse_cuda, test/inductor/test_torchinductor.py::GPUTests::test_conv_functional_bn_fuse_cuda, test/inductor/test_torchinductor.py::GPUTests::test_conv_inference_heuristics_cuda, test/inductor/test_torchinductor.py::GPUTests::test_conv_with_as_strided_cuda, test/inductor/test_torchinductor.py::GPUTests::test_convolution2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_convolution4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_convolution5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_copy_non_blocking_is_pinned_use_cat_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cos_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cpu_scalar_with_gpu_tensor_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cpu_scalar_with_gpu_tensor_dynamic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cpu_tensor_with_gpu_tensor_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cumsum_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cumsum_no_mask_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cumsum_zero_dim_cuda, test/inductor/test_torchinductor.py::GPUTests::test_custom_op_1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_custom_op_2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_custom_op_default_layout_constraint_cuda, test/inductor/test_torchinductor.py::GPUTests::test_custom_op_fixed_layout_channels_last_cuda, test/inductor/test_torchinductor.py::GPUTests::test_custom_op_fixed_layout_sequential_cuda, test/inductor/test_torchinductor.py::GPUTests::test_custom_scan_op_compiled_cuda, test/inductor/test_torchinductor.py::GPUTests::test_custom_scan_op_cuda, test/inductor/test_torchinductor.py::GPUTests::test_custom_scan_op_multi_input_cuda, test/inductor/test_torchinductor.py::GPUTests::test_custom_scan_would_split_cuda, test/inductor/test_torchinductor.py::GPUTests::test_deterministic_codegen_cuda, test/inductor/test_torchinductor.py::GPUTests::test_diagonal_copy_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dist_bf16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dist_cuda, test/inductor/test_torchinductor.py::GPUTests::test_div5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_div6_cuda, test/inductor/test_torchinductor.py::GPUTests::test_div8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_div9_cuda, test/inductor/test_torchinductor.py::GPUTests::test_div_prim_cuda, test/inductor/test_torchinductor.py::GPUTests::test_div_softmax_symfloat_cuda, test/inductor/test_torchinductor.py::GPUTests::test_div_zero_dim_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_trivial_0_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtype_sympy_expr_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float16_float16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float16_float64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float16_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float16_int64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float16_uint8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float32_float16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float32_float32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float32_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float32_int8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float32_uint8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float64_float16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float64_int16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int16_float16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int16_float32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int16_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int16_int64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int16_int8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_float16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_float32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_float64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_int16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_int64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_int8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int64_float32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int64_float64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int64_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int64_int64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int64_int8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int64_uint8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int8_float32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int8_float64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int8_int16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int8_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int8_int64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_uint8_float16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_uint8_float32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_uint8_float64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_uint8_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_embedding_cuda, test/inductor/test_torchinductor.py::GPUTests::test_empty1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_empty2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_emulate_precision_triton_fp_fusion_cuda, test/inductor/test_torchinductor.py::GPUTests::test_erfc_cuda, test/inductor/test_torchinductor.py::GPUTests::test_erfinv_cuda, test/inductor/test_torchinductor.py::GPUTests::test_exact_stride_cuda, test/inductor/test_torchinductor.py::GPUTests::test_expm1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fallback_mutable_op_list_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fallback_mutable_op_no_mutated_tensors_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fft_real_input_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fill2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_flexible_layout_immutable_free_symbols_cuda, test/inductor/test_torchinductor.py::GPUTests::test_flip_cat_cuda, test/inductor/test_torchinductor.py::GPUTests::test_flip_cuda, test/inductor/test_torchinductor.py::GPUTests::test_float_index_expression_cuda, test/inductor/test_torchinductor.py::GPUTests::test_float_repr_dynamic_shapes_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fmin_fmax_cuda, test/inductor/test_torchinductor.py::GPUTests::test_forced_buffer_realize_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fractional_max_pool2d1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fractional_max_pool2d2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fractional_max_pool2d3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fractional_max_pool2d4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fractional_max_pool2d5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_full_like_transposed_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fuse_tiled_cuda, test/inductor/test_torchinductor.py::GPUTests::test_gather2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_gather_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_getitem_cuda, test/inductor/test_torchinductor.py::GPUTests::test_gpu_scalar_with_cpu_tensor_cuda, test/inductor/test_torchinductor.py::GPUTests::test_graph_partition_arange1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_graph_partition_constant_tensor2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_graph_partition_misaligned_input_cuda, test/inductor/test_torchinductor.py::GPUTests::test_graph_partition_no_inputs_cuda, test/inductor/test_torchinductor.py::GPUTests::test_graph_partition_unbacked_symint_as_output_cuda, test/inductor/test_torchinductor.py::GPUTests::test_hardsigmoid_cuda, test/inductor/test_torchinductor.py::GPUTests::test_hardswish_cuda, test/inductor/test_torchinductor.py::GPUTests::test_horizonal_fusion2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_dynamic_shapes_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_propagation_abs_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_propagation_device_assert_masked_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_propagation_nested_indirect_indexing_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_put1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_put2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_put4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_put_deterministic_fallback_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_put_index_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_put_reinplace_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_remainder_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_select_cuda, test/inductor/test_torchinductor.py::GPUTests::test_inductor_multiple_specializations_cuda, test/inductor/test_torchinductor.py::GPUTests::test_inf_cuda, test/inductor/test_torchinductor.py::GPUTests::test_inner_fn_str_and_stride_cuda, test/inductor/test_torchinductor.py::GPUTests::test_inner_reduction_detection_cuda, test/inductor/test_torchinductor.py::GPUTests::test_inplace_activations_cuda, test/inductor/test_torchinductor.py::GPUTests::test_inplace_add_cuda, test/inductor/test_torchinductor.py::GPUTests::test_inplace_flip_cuda, test/inductor/test_torchinductor.py::GPUTests::test_input_mutation1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_input_mutation3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_input_mutation4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_input_mutation5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_insignificant_strides_cuda, test/inductor/test_torchinductor.py::GPUTests::test_isinf_cuda, test/inductor/test_torchinductor.py::GPUTests::test_issue102546_cuda, test/inductor/test_torchinductor.py::GPUTests::test_kernel_names_cuda, test/inductor/test_torchinductor.py::GPUTests::test_kwargs_cuda, test/inductor/test_torchinductor.py::GPUTests::test_l1_loss_cuda, test/inductor/test_torchinductor.py::GPUTests::test_large_broadcast_reduction_cuda, test/inductor/test_torchinductor.py::GPUTests::test_large_grid_use_block_ptr_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_large_pointwise_cuda, test/inductor/test_torchinductor.py::GPUTests::test_large_strided_reduction_cuda, test/inductor/test_torchinductor.py::GPUTests::test_large_tensor_reduction_cuda, test/inductor/test_torchinductor.py::GPUTests::test_leaky_relu_cuda, test/inductor/test_torchinductor.py::GPUTests::test_lgamma_cuda, test/inductor/test_torchinductor.py::GPUTests::test_like_rands3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_like_rands_sliced_cuda, test/inductor/test_torchinductor.py::GPUTests::test_linalg_eig_stride_consistency_cuda, test/inductor/test_torchinductor.py::GPUTests::test_linear1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_linear2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_linear_mixed_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_linspace1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_linspace2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_linspace4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_lite_mode_fallback_cuda, test/inductor/test_torchinductor.py::GPUTests::test_lite_regional_compile_flex_attention_cuda, test/inductor/test_torchinductor.py::GPUTests::test_lite_regional_compile_repeated_blocks_cuda, test/inductor/test_torchinductor.py::GPUTests::test_lite_triton_kernel_wrapper_functional_cuda, test/inductor/test_torchinductor.py::GPUTests::test_log_fp64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_log_softmax_cuda, test/inductor/test_torchinductor.py::GPUTests::test_logaddexp_cuda, test/inductor/test_torchinductor.py::GPUTests::test_logcumsumexp_cuda, test/inductor/test_torchinductor.py::GPUTests::test_logcumsumexp_zero_dim_cuda, test/inductor/test_torchinductor.py::GPUTests::test_low_memory_max_pool_dilation_2_dim_2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_low_memory_max_pool_dilation_2_dim_3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_masked_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_matmul_layer_norm_cuda, test/inductor/test_torchinductor.py::GPUTests::test_max_min_cuda, test/inductor/test_torchinductor.py::GPUTests::test_max_pool2d3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_max_pool2d5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_max_pool2d_with_indices_backward2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_max_pool2d_with_indices_backward3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_max_pool2d_with_indices_backward4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_max_pool2d_with_indices_backward6_cuda, test/inductor/test_torchinductor.py::GPUTests::test_min_max_reduction_nan_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mix_device_index_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mixed_mm2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mixed_mm3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mixed_mm_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mm_mixed_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mm_views_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mul_index_expr_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mul_softmax_symfloat_cuda, test/inductor/test_torchinductor.py::GPUTests::test_multi_gpu_device_cuda, test/inductor/test_torchinductor.py::GPUTests::test_multi_gpu_recompile_on_index_cuda, test/inductor/test_torchinductor.py::GPUTests::test_multilayer_var_cuda, test/inductor/test_torchinductor.py::GPUTests::test_multilayer_var_lowp_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mutable_custom_op_fixed_layout2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mutable_custom_op_fixed_layout_cuda, test/inductor/test_torchinductor.py::GPUTests::test_nan_sort_stable_False_descending_False_cuda, test/inductor/test_torchinductor.py::GPUTests::test_nan_sort_stable_False_descending_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_nan_to_num_cuda, test/inductor/test_torchinductor.py::GPUTests::test_neg_index_cuda, test/inductor/test_torchinductor.py::GPUTests::test_neg_max_uint8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_nll_loss_backward_cuda, test/inductor/test_torchinductor.py::GPUTests::test_nll_loss_forward_cuda, test/inductor/test_torchinductor.py::GPUTests::test_no_mega_fusion_during_lowering_cuda, test/inductor/test_torchinductor.py::GPUTests::test_no_op_reduction_cuda, test/inductor/test_torchinductor.py::GPUTests::test_no_specization_over_symbolic_value_cuda, test/inductor/test_torchinductor.py::GPUTests::test_nonzero_unbacked_refinement_cuda, test/inductor/test_torchinductor.py::GPUTests::test_norm_constant_overflow_cuda, test/inductor/test_torchinductor.py::GPUTests::test_output_strides_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pattern_matcher_multi_user_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pattern_matcher_unbacked_cuda, test/inductor/test_torchinductor.py::GPUTests::test_permute1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_permute2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_bessel_j0_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_bessel_j1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_chebyshev_polynomial_u_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_chebyshev_polynomial_v_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_chebyshev_polynomial_w_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_digamma_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_entr_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_erfc_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_expit_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_expm1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_gammainc_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_gammaincc_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_hermite_polynomial_he_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_i0_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_i0e_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_i1e_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_logit_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_modified_bessel_k0_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_ndtr_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_polygamma_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_scaled_modified_bessel_k1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_shifted_chebyshev_polynomial_t_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_shifted_chebyshev_polynomial_u_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_sinc_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_xlog1py_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_xlogy_cuda, test/inductor/test_torchinductor.py::GPUTests::test_polar_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pow1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pow2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pow3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pow_int_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pow_symfloat_cuda, test/inductor/test_torchinductor.py::GPUTests::test_prepare_softmax_with_fast_math_cuda, test/inductor/test_torchinductor.py::GPUTests::test_profiler_mark_wrapper_call_cuda, test/inductor/test_torchinductor.py::GPUTests::test_randint_cuda, test/inductor/test_torchinductor.py::GPUTests::test_randint_distribution_cuda, test/inductor/test_torchinductor.py::GPUTests::test_randint_int64_mod_cuda, test/inductor/test_torchinductor.py::GPUTests::test_randn_generator_cuda, test/inductor/test_torchinductor.py::GPUTests::test_randn_like_empty_cuda, test/inductor/test_torchinductor.py::GPUTests::test_reduction1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_reduction2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_reduction3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_reduction5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_reduction_config_limit_cuda, test/inductor/test_torchinductor.py::GPUTests::test_reflection_pad2d_backward_cuda, test/inductor/test_torchinductor.py::GPUTests::test_reinterpret_dtypeview_cuda, test/inductor/test_torchinductor.py::GPUTests::test_relu_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_no_ops_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_repeat_interleave_2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_repeat_interleave_Tensor_decomp_int32_nd_1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_replication_pad_errors_with_bool_cuda, test/inductor/test_torchinductor.py::GPUTests::test_require_stride_expanded_cuda, test/inductor/test_torchinductor.py::GPUTests::test_resize_as_cuda, test/inductor/test_torchinductor.py::GPUTests::test_resize_cuda, test/inductor/test_torchinductor.py::GPUTests::test_reuse_buffers_with_aliasing_cuda, test/inductor/test_torchinductor.py::GPUTests::test_roll_cuda, test/inductor/test_torchinductor.py::GPUTests::test_rsqrt_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scalar_output_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scaled_dot_product_attention_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scatter1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scatter2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scatter3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scatter4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scatter5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scatter_add1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scatter_add3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scatter_reduce1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sdpa_prefer_nd_tiling_True_use_block_ptr_False_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sdpa_unaligned_mask_freezing_cuda, test/inductor/test_torchinductor.py::GPUTests::test_searchsorted_cuda, test/inductor/test_torchinductor.py::GPUTests::test_select_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sgn_extremal_cuda, test/inductor/test_torchinductor.py::GPUTests::test_shape_padding_cuda, test/inductor/test_torchinductor.py::GPUTests::test_should_pad_bench_for_bmm_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sigmoid_cuda, test/inductor/test_torchinductor.py::GPUTests::test_signbit_cuda, test/inductor/test_torchinductor.py::GPUTests::test_silu_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sin_cuda, test/inductor/test_torchinductor.py::GPUTests::test_single_elem_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice_mutation1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice_mutation2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice_mutation3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice_scatter2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice_scatter4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice_scatter5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice_scatter_dtype_consistency_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice_scatter_reinplace_cuda, test/inductor/test_torchinductor.py::GPUTests::test_softmax_backward_data_cuda, test/inductor/test_torchinductor.py::GPUTests::test_softmax_one_kernel_loop_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sort_bool_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sort_transpose_cuda, test/inductor/test_torchinductor.py::GPUTests::test_special_polygamma_cuda, test/inductor/test_torchinductor.py::GPUTests::test_split_cuda, test/inductor/test_torchinductor.py::GPUTests::test_split_cumprod_low_prec_cuda, test/inductor/test_torchinductor.py::GPUTests::test_split_cumsum_cuda, test/inductor/test_torchinductor.py::GPUTests::test_split_cumsum_index_cuda, test/inductor/test_torchinductor.py::GPUTests::test_split_cumsum_low_prec_cuda, test/inductor/test_torchinductor.py::GPUTests::test_split_reduction_dynamic_shape_cuda, test/inductor/test_torchinductor.py::GPUTests::test_split_with_integer_cuda, test/inductor/test_torchinductor.py::GPUTests::test_squeeze2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_squeeze_varargs_cuda, test/inductor/test_torchinductor.py::GPUTests::test_std_cuda, test/inductor/test_torchinductor.py::GPUTests::test_stride_preservation_with_stride_modifying_fx_pass_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sum1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sum2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sum3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sum5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_tensor2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_tensor_index_slice_cuda, test/inductor/test_torchinductor.py::GPUTests::test_tmp_not_defined_issue1_use_block_ptr_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_tmp_not_defined_issue2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_to_device_constant_cuda, test/inductor/test_torchinductor.py::GPUTests::test_to_device_cuda, test/inductor/test_torchinductor.py::GPUTests::test_to_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_to_memory_format_cuda, test/inductor/test_torchinductor.py::GPUTests::test_transposed_propagates_cuda, test/inductor/test_torchinductor.py::GPUTests::test_triton_argmin_argmax_transpose_logical_index_cuda, test/inductor/test_torchinductor.py::GPUTests::test_triu_cuda, test/inductor/test_torchinductor.py::GPUTests::test_uint4x2_mixed_mm_cuda, test/inductor/test_torchinductor.py::GPUTests::test_unbacked_floordiv_simplify_cuda, test/inductor/test_torchinductor.py::GPUTests::test_unbacked_floordiv_simplify_errors_cuda, test/inductor/test_torchinductor.py::GPUTests::test_unbind_cuda, test/inductor/test_torchinductor.py::GPUTests::test_unfold_zero_dimension_tensor_cuda, test/inductor/test_torchinductor.py::GPUTests::test_unsigned_constant_tensors_cuda, test/inductor/test_torchinductor.py::GPUTests::test_unspec_inputs_float32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_unspec_inputs_float64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_unspec_inputs_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_unspec_inputs_uint8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_upsample_bilinear2d_a_cuda, test/inductor/test_torchinductor.py::GPUTests::test_upsample_bilinear2d_b_cuda, test/inductor/test_torchinductor.py::GPUTests::test_upsample_nearest1d_cuda, test/inductor/test_torchinductor.py::GPUTests::test_upsample_nearest2d_cuda, test/inductor/test_torchinductor.py::GPUTests::test_upsample_nearest3d_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_correction_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_False_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_vectorized_ops_masked_cuda, test/inductor/test_torchinductor.py::GPUTests::test_vectorized_ops_masked_var_novec_cuda, test/inductor/test_torchinductor.py::GPUTests::test_view_as_complex_cuda, test/inductor/test_torchinductor.py::GPUTests::test_view_detach_cuda, test/inductor/test_torchinductor.py::GPUTests::test_view_on_aliased_cuda, test/inductor/test_torchinductor.py::GPUTests::test_view_uint8_through_differing_bitwidths_cuda, test/inductor/test_torchinductor.py::GPUTests::test_views4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_views5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_views6_cuda, test/inductor/test_torchinductor.py::GPUTests::test_views7_cuda, test/inductor/test_torchinductor.py::GPUTests::test_weight_norm_bwd_cuda, test/inductor/test_torchinductor.py::GPUTests::test_xblock_divides_xnumel_cuda, test/inductor/test_torchinductor.py::GPUTests::test_zero_dim_reductions_cuda, test/inductor/test_torchinductor.py::GPUTests::test_zero_element_mutation_cuda, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_cant_optimize_compute, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_ctr_not_moved_to_cuda_when_used_in_index_put, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_donated_buffer_inplace, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_evict_last_non_coalesced_loads_block_ptr, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_graph_partition_default_device_context, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_grouped_mm, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_has_constant_mask_block_multiple_False_ynumel_exceed_ygrid_size_False, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_has_constant_mask_block_multiple_True_ynumel_exceed_ygrid_size_False, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_has_constant_mask_block_multiple_True_ynumel_exceed_ygrid_size_True, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_inductor_detach_view_backend_inductor, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_inductor_sequence_nr, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_not_materialize_pointwise_reduction, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_numpy_on_gpu, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_optimize_compute, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_optimize_indexing_assert, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_optimize_indexing_dtype_with_constraint, test/inductor/test_torchinductor.py::RNNTest::test_rnn_compile_safe, test/inductor/test_torchinductor.py::NanCheckerTest::test_nan_checker_pass 2025-12-04T10:05:23.2845677Z 2025-12-04T10:05:23.2846165Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast1_broadcast1 PASSED [3.1502s] [ 0%] 2025-12-04T10:05:23.2847222Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast1_broadcast3 PASSED [0.1727s] [ 0%] 2025-12-04T10:05:23.2848238Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast1_dense PASSED [0.1884s] [ 0%] 2025-12-04T10:05:23.2849373Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast1_double PASSED [0.1773s] [ 0%] 2025-12-04T10:05:23.2850353Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast1_int PASSED [0.1691s] [ 0%] 2025-12-04T10:05:23.2851356Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast2_broadcast1 PASSED [0.1893s] [ 1%] 2025-12-04T10:05:23.2852395Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast2_broadcast3 PASSED [0.1696s] [ 1%] 2025-12-04T10:05:23.2853408Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast2_dense PASSED [0.1875s] [ 1%] 2025-12-04T10:05:23.2854394Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast2_strided PASSED [0.2136s] [ 1%] 2025-12-04T10:05:23.2855407Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast3_broadcast1 PASSED [0.1682s] [ 1%] 2025-12-04T10:05:23.2856449Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast3_broadcast2 PASSED [0.1697s] [ 2%] 2025-12-04T10:05:23.2857524Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast3_broadcast3 PASSED [0.1600s] [ 2%] 2025-12-04T10:05:23.2858548Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast3_double PASSED [0.1720s] [ 2%] 2025-12-04T10:05:23.2859548Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast3_strided PASSED [0.1769s] [ 2%] 2025-12-04T10:05:23.2860566Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast3_transposed PASSED [0.1705s] [ 2%] 2025-12-04T10:05:23.2861564Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_dense_broadcast2 PASSED [0.1857s] [ 3%] 2025-12-04T10:05:23.2862523Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_dense_int PASSED [0.1734s] [ 3%] 2025-12-04T10:05:23.2863482Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_dense_transposed PASSED [0.2109s] [ 3%] 2025-12-04T10:05:23.2864479Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_double_broadcast1 PASSED [0.4785s] [ 3%] 2025-12-04T10:05:23.2865470Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_double_broadcast3 PASSED [0.4726s] [ 3%] 2025-12-04T10:05:23.2866446Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_double_dense PASSED [0.4703s] [ 4%] 2025-12-04T10:05:23.2867406Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_int_broadcast1 PASSED [0.4655s] [ 4%] 2025-12-04T10:05:23.2868372Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_int_broadcast2 PASSED [0.4871s] [ 4%] 2025-12-04T10:05:23.2869305Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_int_dense PASSED [0.4716s] [ 4%] 2025-12-04T10:05:23.2870228Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_int_double PASSED [0.4751s] [ 4%] 2025-12-04T10:05:23.2871141Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_int_int PASSED [0.4655s] [ 5%] 2025-12-04T10:05:23.2872095Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_strided_broadcast2 PASSED [0.5016s] [ 5%] 2025-12-04T10:05:23.2873203Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_strided_dense PASSED [0.6706s] [ 5%] 2025-12-04T10:05:23.2874179Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_strided_double PASSED [0.4771s] [ 5%] 2025-12-04T10:05:23.2875141Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_strided_int PASSED [0.4811s] [ 5%] 2025-12-04T10:05:23.2876123Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_transposed_broadcast1 PASSED [0.4887s] [ 5%] 2025-12-04T10:05:23.2877166Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_transposed_broadcast3 PASSED [0.4699s] [ 6%] 2025-12-04T10:05:23.2878189Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_transposed_dense PASSED [0.5013s] [ 6%] 2025-12-04T10:05:23.2879193Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_transposed_strided PASSED [0.4798s] [ 6%] 2025-12-04T10:05:23.2880275Z inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_transposed_transposed PASSED [0.1684s] [ 6%] 2025-12-04T10:05:23.2881525Z inductor/test_torchinductor.py::GPUTests::test__dyn_quant_matmul_4bit_bf16_input_cuda SKIPPED [0.0032s] (No _dyn_quant_matmul_4bit implementation on CUDA) [ 6%] 2025-12-04T10:05:23.2882965Z inductor/test_torchinductor.py::GPUTests::test__dyn_quant_matmul_4bit_fp32_input_cuda SKIPPED [0.0029s] (No _dyn_quant_matmul_4bit implementation on CUDA) [ 7%] 2025-12-04T10:05:23.2884430Z inductor/test_torchinductor.py::GPUTests::test__dyn_quant_pack_4bit_weight_fp32_cuda SKIPPED [0.0028s] (No _dyn_quant_pack_4bit_weight implementation on CUDA) [ 7%] 2025-12-04T10:05:23.2885619Z inductor/test_torchinductor.py::GPUTests::test__unsafe_masked_index_cuda PASSED [0.4880s] [ 7%] 2025-12-04T10:05:23.2886502Z inductor/test_torchinductor.py::GPUTests::test_adaptive_avg_pool2d2_cuda PASSED [0.1180s] [ 7%] 2025-12-04T10:05:23.2887443Z inductor/test_torchinductor.py::GPUTests::test_adaptive_avg_pool2d_low_prec_cuda PASSED [0.3446s] [ 7%] 2025-12-04T10:05:23.2888451Z inductor/test_torchinductor.py::GPUTests::test_adaptive_avg_pool_errors_with_long_cuda PASSED [0.3124s] [ 8%] 2025-12-04T10:05:23.2889401Z inductor/test_torchinductor.py::GPUTests::test_adaptive_max_pool2d2_cuda PASSED [0.1184s] [ 8%] 2025-12-04T10:05:23.2890260Z inductor/test_torchinductor.py::GPUTests::test_add_complex3_cuda PASSED [0.4052s] [ 8%] 2025-12-04T10:05:23.2891082Z inductor/test_torchinductor.py::GPUTests::test_add_complex7_cuda PASSED [0.3058s] [ 8%] 2025-12-04T10:05:23.2891892Z inductor/test_torchinductor.py::GPUTests::test_add_complex_cuda PASSED [0.3001s] [ 8%] 2025-12-04T10:05:23.2892771Z inductor/test_torchinductor.py::GPUTests::test_add_complex_strided_fallback_cuda PASSED [0.1145s] [ 9%] 2025-12-04T10:05:23.2893668Z inductor/test_torchinductor.py::GPUTests::test_add_const_int_cuda PASSED [0.8110s] [ 9%] 2025-12-04T10:05:23.2894551Z inductor/test_torchinductor.py::GPUTests::test_adding_tensor_offsets_cuda PASSED [0.2632s] [ 9%] 2025-12-04T10:05:23.2895383Z inductor/test_torchinductor.py::GPUTests::test_addmv_cuda PASSED [4.6746s] [ 9%] 2025-12-04T10:05:23.2896936Z inductor/test_torchinductor.py::GPUTests::test_aliased_buffer_reuse_cuda W1204 09:55:23.027000 13298 site-packages/torch/_inductor/utils.py:1703] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:05:23.2898009Z PASSED [1.0831s] [ 9%] 2025-12-04T10:05:23.2898640Z inductor/test_torchinductor.py::GPUTests::test_allow_reuse_disable_if_exceed_peak_cuda PASSED [0.6909s] [ 10%] 2025-12-04T10:05:23.2899675Z inductor/test_torchinductor.py::GPUTests::test_aoti_eager_cache_hit_cuda SKIPPED [0.0032s] (Requires sm80) [ 10%] 2025-12-04T10:05:23.2900759Z inductor/test_torchinductor.py::GPUTests::test_aoti_eager_dtype_device_layout_cuda SKIPPED [0.0029s] (Requires sm80) [ 10%] 2025-12-04T10:05:23.2901917Z inductor/test_torchinductor.py::GPUTests::test_aoti_eager_override_registration_cuda SKIPPED [0.0029s] (Requires sm80) [ 10%] 2025-12-04T10:05:23.2903049Z inductor/test_torchinductor.py::GPUTests::test_aoti_eager_support_out_cuda SKIPPED [0.0028s] (Requires sm80) [ 10%] 2025-12-04T10:05:23.2904238Z inductor/test_torchinductor.py::GPUTests::test_aoti_eager_support_str_cuda SKIPPED [0.0028s] (Requires sm80) [ 11%] 2025-12-04T10:05:23.2905151Z inductor/test_torchinductor.py::GPUTests::test_arange3_cuda PASSED [0.3359s] [ 11%] 2025-12-04T10:05:23.2905921Z inductor/test_torchinductor.py::GPUTests::test_arange4_cuda PASSED [0.3536s] [ 11%] 2025-12-04T10:05:23.2906691Z inductor/test_torchinductor.py::GPUTests::test_arange5_cuda PASSED [0.2994s] [ 11%] 2025-12-04T10:05:23.2907478Z inductor/test_torchinductor.py::GPUTests::test_argmax_argmin3_cuda PASSED [1.6283s] [ 11%] 2025-12-04T10:05:23.2908389Z inductor/test_torchinductor.py::GPUTests::test_argmax_argmin_with_duplicates_cuda PASSED [1.7534s] [ 11%] 2025-12-04T10:05:23.2909339Z inductor/test_torchinductor.py::GPUTests::test_argmax_argmin_with_nan_cuda PASSED [3.3626s] [ 12%] 2025-12-04T10:05:23.2910310Z inductor/test_torchinductor.py::GPUTests::test_argmax_min_int32_cuda PASSED [0.3079s] [ 12%] 2025-12-04T10:05:23.2911148Z inductor/test_torchinductor.py::GPUTests::test_argmax_to_float_cuda PASSED [0.8281s] [ 12%] 2025-12-04T10:05:23.2912011Z inductor/test_torchinductor.py::GPUTests::test_as_strided_scatter_cuda PASSED [0.4530s] [ 12%] 2025-12-04T10:05:23.2912944Z inductor/test_torchinductor.py::GPUTests::test_assert_alignment_op_name_pass_cuda PASSED [0.0033s] [ 12%] 2025-12-04T10:05:23.2913848Z inductor/test_torchinductor.py::GPUTests::test_avg_pool2d1_cuda PASSED [0.5971s] [ 13%] 2025-12-04T10:05:23.2914642Z inductor/test_torchinductor.py::GPUTests::test_avg_pool2d2_cuda PASSED [0.7514s] [ 13%] 2025-12-04T10:05:23.2915444Z inductor/test_torchinductor.py::GPUTests::test_avg_pool2d3_cuda PASSED [1.5708s] [ 13%] 2025-12-04T10:05:23.2916247Z inductor/test_torchinductor.py::GPUTests::test_avg_pool2d5_cuda PASSED [1.1620s] [ 13%] 2025-12-04T10:05:23.2917091Z inductor/test_torchinductor.py::GPUTests::test_avg_pool2d_backward2_cuda PASSED [10.1999s] [ 13%] 2025-12-04T10:05:23.2917990Z inductor/test_torchinductor.py::GPUTests::test_avg_pool2d_backward3_cuda PASSED [2.0159s] [ 14%] 2025-12-04T10:05:23.2918886Z inductor/test_torchinductor.py::GPUTests::test_avg_pool2d_backward4_cuda PASSED [0.1268s] [ 14%] 2025-12-04T10:05:23.2919773Z inductor/test_torchinductor.py::GPUTests::test_avg_pool2d_backward_cuda PASSED [1.7204s] [ 14%] 2025-12-04T10:05:23.2920918Z inductor/test_torchinductor.py::GPUTests::test_avg_pool3d_backward2_cuda SKIPPED [0.0005s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 14%] 2025-12-04T10:05:23.2922086Z inductor/test_torchinductor.py::GPUTests::test_avg_pool3d_backward3_cuda PASSED [2.8164s] [ 14%] 2025-12-04T10:05:23.2922970Z inductor/test_torchinductor.py::GPUTests::test_avg_pool3d_backward_cuda PASSED [2.6073s] [ 15%] 2025-12-04T10:05:23.2924452Z inductor/test_torchinductor.py::GPUTests::test_avg_pool_errors_with_uint_cuda E1204 09:55:56.852000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/0] failed while attempting to run meta for aten.avg_pool2d.default 2025-12-04T10:05:23.2926034Z E1204 09:55:56.852000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/0] Traceback (most recent call last): 2025-12-04T10:05:23.2927520Z E1204 09:55:56.852000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_subclasses/fake_tensor.py", line 2823, in _dispatch_impl 2025-12-04T10:05:23.2928978Z E1204 09:55:56.852000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/0] r = func(*args, **kwargs) 2025-12-04T10:05:23.2930290Z E1204 09:55:56.852000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_ops.py", line 836, in __call__ 2025-12-04T10:05:23.2931631Z E1204 09:55:56.852000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/0] return self._op(*args, **kwargs) 2025-12-04T10:05:23.2933170Z E1204 09:55:56.852000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_meta_registrations.py", line 2803, in meta_avg_pool2d 2025-12-04T10:05:23.2934531Z E1204 09:55:56.852000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/0] torch._check( 2025-12-04T10:05:23.2935800Z E1204 09:55:56.852000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1734, in _check 2025-12-04T10:05:23.2937424Z E1204 09:55:56.852000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/0] _check_with(RuntimeError, cond, message) # pyrefly: ignore [bad-argument-type] 2025-12-04T10:05:23.2939008Z E1204 09:55:56.852000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1716, in _check_with 2025-12-04T10:05:23.2940445Z E1204 09:55:56.852000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/0] raise error_type(message_evaluated) 2025-12-04T10:05:23.2941645Z E1204 09:55:56.852000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/0] RuntimeError: "avg_pool2d" not implemented for 'torch.uint8' 2025-12-04T10:05:23.2942959Z E1204 09:55:56.867000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/1] failed while attempting to run meta for aten.avg_pool2d.default 2025-12-04T10:05:23.2944141Z E1204 09:55:56.867000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/1] Traceback (most recent call last): 2025-12-04T10:05:23.2945625Z E1204 09:55:56.867000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/1] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_subclasses/fake_tensor.py", line 2823, in _dispatch_impl 2025-12-04T10:05:23.2947059Z E1204 09:55:56.867000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/1] r = func(*args, **kwargs) 2025-12-04T10:05:23.2948381Z E1204 09:55:56.867000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/1] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_ops.py", line 836, in __call__ 2025-12-04T10:05:23.2949712Z E1204 09:55:56.867000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/1] return self._op(*args, **kwargs) 2025-12-04T10:05:23.2951174Z E1204 09:55:56.867000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/1] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_meta_registrations.py", line 2803, in meta_avg_pool2d 2025-12-04T10:05:23.2952534Z E1204 09:55:56.867000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/1] torch._check( 2025-12-04T10:05:23.2953798Z E1204 09:55:56.867000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/1] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1734, in _check 2025-12-04T10:05:23.2955366Z E1204 09:55:56.867000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/1] _check_with(RuntimeError, cond, message) # pyrefly: ignore [bad-argument-type] 2025-12-04T10:05:23.2956948Z E1204 09:55:56.867000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/1] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1716, in _check_with 2025-12-04T10:05:23.2958332Z E1204 09:55:56.867000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/1] raise error_type(message_evaluated) 2025-12-04T10:05:23.2959521Z E1204 09:55:56.867000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/1] RuntimeError: "avg_pool2d" not implemented for 'torch.uint16' 2025-12-04T10:05:23.2960833Z E1204 09:55:56.880000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/2] failed while attempting to run meta for aten.avg_pool2d.default 2025-12-04T10:05:23.2962021Z E1204 09:55:56.880000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/2] Traceback (most recent call last): 2025-12-04T10:05:23.2963574Z E1204 09:55:56.880000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/2] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_subclasses/fake_tensor.py", line 2823, in _dispatch_impl 2025-12-04T10:05:23.2965009Z E1204 09:55:56.880000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/2] r = func(*args, **kwargs) 2025-12-04T10:05:23.2966320Z E1204 09:55:56.880000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/2] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_ops.py", line 836, in __call__ 2025-12-04T10:05:23.2967648Z E1204 09:55:56.880000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/2] return self._op(*args, **kwargs) 2025-12-04T10:05:23.2969112Z E1204 09:55:56.880000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/2] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_meta_registrations.py", line 2803, in meta_avg_pool2d 2025-12-04T10:05:23.2970564Z E1204 09:55:56.880000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/2] torch._check( 2025-12-04T10:05:23.2971828Z E1204 09:55:56.880000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/2] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1734, in _check 2025-12-04T10:05:23.2973387Z E1204 09:55:56.880000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/2] _check_with(RuntimeError, cond, message) # pyrefly: ignore [bad-argument-type] 2025-12-04T10:05:23.2974966Z E1204 09:55:56.880000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/2] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1716, in _check_with 2025-12-04T10:05:23.2976350Z E1204 09:55:56.880000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/2] raise error_type(message_evaluated) 2025-12-04T10:05:23.2977590Z E1204 09:55:56.880000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/2] RuntimeError: "avg_pool2d" not implemented for 'torch.uint32' 2025-12-04T10:05:23.2978910Z E1204 09:55:56.893000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/3] failed while attempting to run meta for aten.avg_pool2d.default 2025-12-04T10:05:23.2980095Z E1204 09:55:56.893000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/3] Traceback (most recent call last): 2025-12-04T10:05:23.2981658Z E1204 09:55:56.893000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/3] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_subclasses/fake_tensor.py", line 2823, in _dispatch_impl 2025-12-04T10:05:23.2983111Z E1204 09:55:56.893000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/3] r = func(*args, **kwargs) 2025-12-04T10:05:23.2984403Z E1204 09:55:56.893000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/3] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_ops.py", line 836, in __call__ 2025-12-04T10:05:23.2985748Z E1204 09:55:56.893000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/3] return self._op(*args, **kwargs) 2025-12-04T10:05:23.2987215Z E1204 09:55:56.893000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/3] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_meta_registrations.py", line 2803, in meta_avg_pool2d 2025-12-04T10:05:23.2988591Z E1204 09:55:56.893000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/3] torch._check( 2025-12-04T10:05:23.2989848Z E1204 09:55:56.893000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/3] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1734, in _check 2025-12-04T10:05:23.2991401Z E1204 09:55:56.893000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/3] _check_with(RuntimeError, cond, message) # pyrefly: ignore [bad-argument-type] 2025-12-04T10:05:23.2992972Z E1204 09:55:56.893000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/3] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1716, in _check_with 2025-12-04T10:05:23.2994430Z E1204 09:55:56.893000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/3] raise error_type(message_evaluated) 2025-12-04T10:05:23.2995671Z E1204 09:55:56.893000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/3] RuntimeError: "avg_pool2d" not implemented for 'torch.uint64' 2025-12-04T10:05:23.2997153Z E1204 09:55:56.930000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/4] failed while attempting to run meta for aten.avg_pool2d.default 2025-12-04T10:05:23.2998347Z E1204 09:55:56.930000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/4] Traceback (most recent call last): 2025-12-04T10:05:23.2999824Z E1204 09:55:56.930000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/4] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_subclasses/fake_tensor.py", line 2823, in _dispatch_impl 2025-12-04T10:05:23.3001458Z E1204 09:55:56.930000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/4] r = func(*args, **kwargs) 2025-12-04T10:05:23.3002761Z E1204 09:55:56.930000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/4] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_ops.py", line 836, in __call__ 2025-12-04T10:05:23.3004107Z E1204 09:55:56.930000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/4] return self._op(*args, **kwargs) 2025-12-04T10:05:23.3005564Z E1204 09:55:56.930000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/4] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_meta_registrations.py", line 2803, in meta_avg_pool2d 2025-12-04T10:05:23.3006939Z E1204 09:55:56.930000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/4] torch._check( 2025-12-04T10:05:23.3008206Z E1204 09:55:56.930000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/4] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1734, in _check 2025-12-04T10:05:23.3009752Z E1204 09:55:56.930000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/4] _check_with(RuntimeError, cond, message) # pyrefly: ignore [bad-argument-type] 2025-12-04T10:05:23.3011329Z E1204 09:55:56.930000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/4] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1716, in _check_with 2025-12-04T10:05:23.3012709Z E1204 09:55:56.930000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/4] raise error_type(message_evaluated) 2025-12-04T10:05:23.3013891Z E1204 09:55:56.930000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/4] RuntimeError: "avg_pool2d" not implemented for 'torch.uint8' 2025-12-04T10:05:23.3015212Z E1204 09:55:56.949000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/5] failed while attempting to run meta for aten.avg_pool2d.default 2025-12-04T10:05:23.3016405Z E1204 09:55:56.949000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/5] Traceback (most recent call last): 2025-12-04T10:05:23.3017984Z E1204 09:55:56.949000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/5] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_subclasses/fake_tensor.py", line 2823, in _dispatch_impl 2025-12-04T10:05:23.3019443Z E1204 09:55:56.949000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/5] r = func(*args, **kwargs) 2025-12-04T10:05:23.3020738Z E1204 09:55:56.949000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/5] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_ops.py", line 836, in __call__ 2025-12-04T10:05:23.3022078Z E1204 09:55:56.949000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/5] return self._op(*args, **kwargs) 2025-12-04T10:05:23.3023539Z E1204 09:55:56.949000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/5] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_meta_registrations.py", line 2803, in meta_avg_pool2d 2025-12-04T10:05:23.3025007Z E1204 09:55:56.949000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/5] torch._check( 2025-12-04T10:05:23.3026267Z E1204 09:55:56.949000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/5] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1734, in _check 2025-12-04T10:05:23.3027823Z E1204 09:55:56.949000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/5] _check_with(RuntimeError, cond, message) # pyrefly: ignore [bad-argument-type] 2025-12-04T10:05:23.3029405Z E1204 09:55:56.949000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/5] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1716, in _check_with 2025-12-04T10:05:23.3030791Z E1204 09:55:56.949000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/5] raise error_type(message_evaluated) 2025-12-04T10:05:23.3032062Z E1204 09:55:56.949000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/5] RuntimeError: "avg_pool2d" not implemented for 'torch.uint16' 2025-12-04T10:05:23.3033359Z E1204 09:55:56.967000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/6] failed while attempting to run meta for aten.avg_pool2d.default 2025-12-04T10:05:23.3034550Z E1204 09:55:56.967000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/6] Traceback (most recent call last): 2025-12-04T10:05:23.3036032Z E1204 09:55:56.967000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/6] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_subclasses/fake_tensor.py", line 2823, in _dispatch_impl 2025-12-04T10:05:23.3037481Z E1204 09:55:56.967000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/6] r = func(*args, **kwargs) 2025-12-04T10:05:23.3038770Z E1204 09:55:56.967000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/6] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_ops.py", line 836, in __call__ 2025-12-04T10:05:23.3040120Z E1204 09:55:56.967000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/6] return self._op(*args, **kwargs) 2025-12-04T10:05:23.3041581Z E1204 09:55:56.967000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/6] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_meta_registrations.py", line 2803, in meta_avg_pool2d 2025-12-04T10:05:23.3042959Z E1204 09:55:56.967000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/6] torch._check( 2025-12-04T10:05:23.3044215Z E1204 09:55:56.967000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/6] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1734, in _check 2025-12-04T10:05:23.3045745Z E1204 09:55:56.967000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/6] _check_with(RuntimeError, cond, message) # pyrefly: ignore [bad-argument-type] 2025-12-04T10:05:23.3047325Z E1204 09:55:56.967000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/6] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1716, in _check_with 2025-12-04T10:05:23.3048713Z E1204 09:55:56.967000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/6] raise error_type(message_evaluated) 2025-12-04T10:05:23.3049904Z E1204 09:55:56.967000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/6] RuntimeError: "avg_pool2d" not implemented for 'torch.uint32' 2025-12-04T10:05:23.3051197Z E1204 09:55:56.984000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/7] failed while attempting to run meta for aten.avg_pool2d.default 2025-12-04T10:05:23.3052384Z E1204 09:55:56.984000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/7] Traceback (most recent call last): 2025-12-04T10:05:23.3053914Z E1204 09:55:56.984000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/7] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_subclasses/fake_tensor.py", line 2823, in _dispatch_impl 2025-12-04T10:05:23.3055430Z E1204 09:55:56.984000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/7] r = func(*args, **kwargs) 2025-12-04T10:05:23.3056741Z E1204 09:55:56.984000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/7] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_ops.py", line 836, in __call__ 2025-12-04T10:05:23.3058124Z E1204 09:55:56.984000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/7] return self._op(*args, **kwargs) 2025-12-04T10:05:23.3059590Z E1204 09:55:56.984000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/7] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_meta_registrations.py", line 2803, in meta_avg_pool2d 2025-12-04T10:05:23.3060964Z E1204 09:55:56.984000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/7] torch._check( 2025-12-04T10:05:23.3062303Z E1204 09:55:56.984000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/7] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1734, in _check 2025-12-04T10:05:23.3063840Z E1204 09:55:56.984000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/7] _check_with(RuntimeError, cond, message) # pyrefly: ignore [bad-argument-type] 2025-12-04T10:05:23.3065422Z E1204 09:55:56.984000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/7] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1716, in _check_with 2025-12-04T10:05:23.3066813Z E1204 09:55:56.984000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/7] raise error_type(message_evaluated) 2025-12-04T10:05:23.3068009Z E1204 09:55:56.984000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/7] RuntimeError: "avg_pool2d" not implemented for 'torch.uint64' 2025-12-04T10:05:23.3069315Z E1204 09:55:57.016000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/8] failed while attempting to run meta for aten.avg_pool3d.default 2025-12-04T10:05:23.3070494Z E1204 09:55:57.016000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/8] Traceback (most recent call last): 2025-12-04T10:05:23.3071979Z E1204 09:55:57.016000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/8] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_subclasses/fake_tensor.py", line 2823, in _dispatch_impl 2025-12-04T10:05:23.3073417Z E1204 09:55:57.016000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/8] r = func(*args, **kwargs) 2025-12-04T10:05:23.3074714Z E1204 09:55:57.016000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/8] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_ops.py", line 836, in __call__ 2025-12-04T10:05:23.3076038Z E1204 09:55:57.016000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/8] return self._op(*args, **kwargs) 2025-12-04T10:05:23.3077461Z E1204 09:55:57.016000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/8] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_prims_common/wrappers.py", line 315, in _fn 2025-12-04T10:05:23.3078854Z E1204 09:55:57.016000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/8] result = fn(*args, **kwargs) 2025-12-04T10:05:23.3080303Z E1204 09:55:57.016000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/8] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_meta_registrations.py", line 3001, in meta_avg_pool3d 2025-12-04T10:05:23.3081676Z E1204 09:55:57.016000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/8] torch._check( 2025-12-04T10:05:23.3082929Z E1204 09:55:57.016000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/8] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1734, in _check 2025-12-04T10:05:23.3084488Z E1204 09:55:57.016000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/8] _check_with(RuntimeError, cond, message) # pyrefly: ignore [bad-argument-type] 2025-12-04T10:05:23.3086140Z E1204 09:55:57.016000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/8] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1716, in _check_with 2025-12-04T10:05:23.3087528Z E1204 09:55:57.016000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/8] raise error_type(message_evaluated) 2025-12-04T10:05:23.3088741Z E1204 09:55:57.016000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/8] RuntimeError: "avg_pool3d" not implemented for 'torch.uint8' 2025-12-04T10:05:23.3090043Z E1204 09:55:57.036000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/9] failed while attempting to run meta for aten.avg_pool3d.default 2025-12-04T10:05:23.3091222Z E1204 09:55:57.036000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/9] Traceback (most recent call last): 2025-12-04T10:05:23.3092785Z E1204 09:55:57.036000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/9] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_subclasses/fake_tensor.py", line 2823, in _dispatch_impl 2025-12-04T10:05:23.3094229Z E1204 09:55:57.036000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/9] r = func(*args, **kwargs) 2025-12-04T10:05:23.3095528Z E1204 09:55:57.036000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/9] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_ops.py", line 836, in __call__ 2025-12-04T10:05:23.3097145Z E1204 09:55:57.036000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/9] return self._op(*args, **kwargs) 2025-12-04T10:05:23.3098567Z E1204 09:55:57.036000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/9] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_prims_common/wrappers.py", line 315, in _fn 2025-12-04T10:05:23.3099963Z E1204 09:55:57.036000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/9] result = fn(*args, **kwargs) 2025-12-04T10:05:23.3101402Z E1204 09:55:57.036000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/9] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_meta_registrations.py", line 3001, in meta_avg_pool3d 2025-12-04T10:05:23.3102774Z E1204 09:55:57.036000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/9] torch._check( 2025-12-04T10:05:23.3104035Z E1204 09:55:57.036000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/9] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1734, in _check 2025-12-04T10:05:23.3105586Z E1204 09:55:57.036000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/9] _check_with(RuntimeError, cond, message) # pyrefly: ignore [bad-argument-type] 2025-12-04T10:05:23.3107161Z E1204 09:55:57.036000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/9] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1716, in _check_with 2025-12-04T10:05:23.3108542Z E1204 09:55:57.036000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/9] raise error_type(message_evaluated) 2025-12-04T10:05:23.3109737Z E1204 09:55:57.036000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/9] RuntimeError: "avg_pool3d" not implemented for 'torch.uint16' 2025-12-04T10:05:23.3111159Z E1204 09:55:57.055000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/10] failed while attempting to run meta for aten.avg_pool3d.default 2025-12-04T10:05:23.3112358Z E1204 09:55:57.055000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/10] Traceback (most recent call last): 2025-12-04T10:05:23.3113837Z E1204 09:55:57.055000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/10] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_subclasses/fake_tensor.py", line 2823, in _dispatch_impl 2025-12-04T10:05:23.3115306Z E1204 09:55:57.055000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/10] r = func(*args, **kwargs) 2025-12-04T10:05:23.3116748Z E1204 09:55:57.055000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/10] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_ops.py", line 836, in __call__ 2025-12-04T10:05:23.3118095Z E1204 09:55:57.055000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/10] return self._op(*args, **kwargs) 2025-12-04T10:05:23.3119511Z E1204 09:55:57.055000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/10] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_prims_common/wrappers.py", line 315, in _fn 2025-12-04T10:05:23.3120901Z E1204 09:55:57.055000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/10] result = fn(*args, **kwargs) 2025-12-04T10:05:23.3122483Z E1204 09:55:57.055000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/10] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_meta_registrations.py", line 3001, in meta_avg_pool3d 2025-12-04T10:05:23.3123972Z E1204 09:55:57.055000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/10] torch._check( 2025-12-04T10:05:23.3125254Z E1204 09:55:57.055000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/10] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1734, in _check 2025-12-04T10:05:23.3126804Z E1204 09:55:57.055000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/10] _check_with(RuntimeError, cond, message) # pyrefly: ignore [bad-argument-type] 2025-12-04T10:05:23.3128394Z E1204 09:55:57.055000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/10] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1716, in _check_with 2025-12-04T10:05:23.3129791Z E1204 09:55:57.055000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/10] raise error_type(message_evaluated) 2025-12-04T10:05:23.3131008Z E1204 09:55:57.055000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/10] RuntimeError: "avg_pool3d" not implemented for 'torch.uint32' 2025-12-04T10:05:23.3132329Z E1204 09:55:57.073000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/11] failed while attempting to run meta for aten.avg_pool3d.default 2025-12-04T10:05:23.3133507Z E1204 09:55:57.073000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/11] Traceback (most recent call last): 2025-12-04T10:05:23.3155638Z E1204 09:55:57.073000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/11] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_subclasses/fake_tensor.py", line 2823, in _dispatch_impl 2025-12-04T10:05:23.3157134Z E1204 09:55:57.073000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/11] r = func(*args, **kwargs) 2025-12-04T10:05:23.3158452Z E1204 09:55:57.073000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/11] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_ops.py", line 836, in __call__ 2025-12-04T10:05:23.3159827Z E1204 09:55:57.073000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/11] return self._op(*args, **kwargs) 2025-12-04T10:05:23.3161263Z E1204 09:55:57.073000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/11] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_prims_common/wrappers.py", line 315, in _fn 2025-12-04T10:05:23.3162673Z E1204 09:55:57.073000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/11] result = fn(*args, **kwargs) 2025-12-04T10:05:23.3164129Z E1204 09:55:57.073000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/11] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_meta_registrations.py", line 3001, in meta_avg_pool3d 2025-12-04T10:05:23.3165507Z E1204 09:55:57.073000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/11] torch._check( 2025-12-04T10:05:23.3166929Z E1204 09:55:57.073000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/11] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1734, in _check 2025-12-04T10:05:23.3168498Z E1204 09:55:57.073000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/11] _check_with(RuntimeError, cond, message) # pyrefly: ignore [bad-argument-type] 2025-12-04T10:05:23.3170082Z E1204 09:55:57.073000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/11] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1716, in _check_with 2025-12-04T10:05:23.3171474Z E1204 09:55:57.073000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/11] raise error_type(message_evaluated) 2025-12-04T10:05:23.3172668Z E1204 09:55:57.073000 13298 site-packages/torch/_subclasses/fake_tensor.py:2827] [0/11] RuntimeError: "avg_pool3d" not implemented for 'torch.uint64' 2025-12-04T10:05:23.3173614Z PASSED [0.2405s] [ 15%] 2025-12-04T10:05:23.3174141Z inductor/test_torchinductor.py::GPUTests::test_baddbmm_cuda PASSED [0.8518s] [ 15%] 2025-12-04T10:05:23.3174928Z inductor/test_torchinductor.py::GPUTests::test_bernoulli1_cuda PASSED [1.3087s] [ 15%] 2025-12-04T10:05:23.3175728Z inductor/test_torchinductor.py::GPUTests::test_bernoulli2_cuda PASSED [1.1459s] [ 15%] 2025-12-04T10:05:23.3176745Z inductor/test_torchinductor.py::GPUTests::test_bfloat16_to_int16_cuda SKIPPED [0.0032s] (uses bfloat16 which requires SM >= 80) [ 16%] 2025-12-04T10:05:23.3177837Z inductor/test_torchinductor.py::GPUTests::test_bitwise3_cuda PASSED [0.3375s] [ 16%] 2025-12-04T10:05:23.3178603Z inductor/test_torchinductor.py::GPUTests::test_bitwise_cuda PASSED [0.2857s] [ 16%] 2025-12-04T10:05:23.3179371Z inductor/test_torchinductor.py::GPUTests::test_bmm1_cuda PASSED [0.7487s] [ 16%] 2025-12-04T10:05:23.3180127Z inductor/test_torchinductor.py::GPUTests::test_bool_cuda PASSED [0.4258s] [ 16%] 2025-12-04T10:05:23.3180960Z inductor/test_torchinductor.py::GPUTests::test_bucketize_broadcast_cuda PASSED [0.3439s] [ 16%] 2025-12-04T10:05:23.3181871Z inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int16_int16_cuda PASSED [0.7874s] [ 17%] 2025-12-04T10:05:23.3182819Z inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int16_int32_cuda PASSED [0.7905s] [ 17%] 2025-12-04T10:05:23.3183763Z inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int16_int64_cuda PASSED [0.7939s] [ 17%] 2025-12-04T10:05:23.3184691Z inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int16_int8_cuda PASSED [0.8143s] [ 17%] 2025-12-04T10:05:23.3185632Z inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int16_uint8_cuda PASSED [0.8000s] [ 17%] 2025-12-04T10:05:23.3186570Z inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int32_int64_cuda PASSED [0.7820s] [ 18%] 2025-12-04T10:05:23.3187508Z inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int32_int8_cuda PASSED [0.7925s] [ 18%] 2025-12-04T10:05:23.3188428Z inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int8_int16_cuda PASSED [0.7700s] [ 18%] 2025-12-04T10:05:23.3189359Z inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int8_int32_cuda PASSED [0.7774s] [ 18%] 2025-12-04T10:05:23.3190291Z inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int8_int64_cuda PASSED [0.7746s] [ 18%] 2025-12-04T10:05:23.3191227Z inductor/test_torchinductor.py::GPUTests::test_bucketize_int_uint8_int16_cuda PASSED [0.7756s] [ 19%] 2025-12-04T10:05:23.3192154Z inductor/test_torchinductor.py::GPUTests::test_bucketize_int_uint8_uint8_cuda PASSED [0.7598s] [ 19%] 2025-12-04T10:05:23.3193095Z inductor/test_torchinductor.py::GPUTests::test_bucketize_nd_tiling_False_cuda PASSED [0.3959s] [ 19%] 2025-12-04T10:05:23.3194034Z inductor/test_torchinductor.py::GPUTests::test_bucketize_nd_tiling_True_cuda PASSED [0.3992s] [ 19%] 2025-12-04T10:05:23.3194933Z inductor/test_torchinductor.py::GPUTests::test_buffer_batch_norm_cuda PASSED [1.2194s] [ 19%] 2025-12-04T10:05:23.3195815Z inductor/test_torchinductor.py::GPUTests::test_buffer_copied_in_graph_cuda PASSED [0.2305s] [ 20%] 2025-12-04T10:05:23.3197196Z inductor/test_torchinductor.py::GPUTests::test_buffer_copied_in_graph_with_different_shapes_cuda PASSED [0.4332s] [ 20%] 2025-12-04T10:05:23.3198232Z inductor/test_torchinductor.py::GPUTests::test_buffer_use_after_remove_cuda PASSED [2.7236s] [ 20%] 2025-12-04T10:05:23.3199117Z inductor/test_torchinductor.py::GPUTests::test_builtins_round_cuda PASSED [0.3458s] [ 20%] 2025-12-04T10:05:23.3200038Z inductor/test_torchinductor.py::GPUTests::test_builtins_round_float_ndigits_pos_cuda PASSED [0.2594s] [ 20%] 2025-12-04T10:05:23.3201064Z inductor/test_torchinductor.py::GPUTests::test_builtins_round_float_ndigits_zero_cuda PASSED [0.2313s] [ 21%] 2025-12-04T10:05:23.3202080Z inductor/test_torchinductor.py::GPUTests::test_builtins_round_int_ndigits_zero_cuda PASSED [0.1747s] [ 21%] 2025-12-04T10:05:23.3203016Z inductor/test_torchinductor.py::GPUTests::test_cat_extern_kernel_cuda PASSED [0.2501s] [ 21%] 2025-12-04T10:05:23.3203928Z inductor/test_torchinductor.py::GPUTests::test_cat_inplace_cuda PASSED [0.7329s] [ 21%] 2025-12-04T10:05:23.3204767Z inductor/test_torchinductor.py::GPUTests::test_cat_negative_dim_cuda PASSED [1.0190s] [ 21%] 2025-12-04T10:05:23.3205680Z inductor/test_torchinductor.py::GPUTests::test_cat_of_loops_and_extern_kernel_cuda PASSED [1.1208s] [ 22%] 2025-12-04T10:05:23.3206552Z inductor/test_torchinductor.py::GPUTests::test_cat_uint8_cuda PASSED [0.2779s] [ 22%] 2025-12-04T10:05:23.3207372Z inductor/test_torchinductor.py::GPUTests::test_cat_unbacked_2d_cuda PASSED [0.7193s] [ 22%] 2025-12-04T10:05:23.3208263Z inductor/test_torchinductor.py::GPUTests::test_cat_unbacked_legacy_empty_cuda PASSED [0.0279s] [ 22%] 2025-12-04T10:05:23.3209151Z inductor/test_torchinductor.py::GPUTests::test_cat_upcasting_cuda PASSED [0.4255s] [ 22%] 2025-12-04T10:05:23.3210009Z inductor/test_torchinductor.py::GPUTests::test_check_stack_no_cycles_cuda PASSED [0.1734s] [ 22%] 2025-12-04T10:05:23.3210848Z inductor/test_torchinductor.py::GPUTests::test_clamp_cuda PASSED [0.4963s] [ 23%] 2025-12-04T10:05:23.3211752Z inductor/test_torchinductor.py::GPUTests::test_config_option_dont_assume_alignment_cuda PASSED [0.5434s] [ 23%] 2025-12-04T10:05:23.3212748Z inductor/test_torchinductor.py::GPUTests::test_consecutive_split_cumsum_cuda PASSED [0.3851s] [ 23%] 2025-12-04T10:05:23.3213629Z inductor/test_torchinductor.py::GPUTests::test_constant_pad_1d_cuda PASSED [0.4418s] [ 23%] 2025-12-04T10:05:23.3214476Z inductor/test_torchinductor.py::GPUTests::test_constant_pad_2d_cuda PASSED [0.4675s] [ 23%] 2025-12-04T10:05:23.3215347Z inductor/test_torchinductor.py::GPUTests::test_constant_pad_float64_cuda PASSED [0.2871s] [ 24%] 2025-12-04T10:05:23.3216209Z inductor/test_torchinductor.py::GPUTests::test_conv1d_depthwise_cuda PASSED [0.2563s] [ 24%] 2025-12-04T10:05:23.3217168Z inductor/test_torchinductor.py::GPUTests::test_conv1d_with_permute_cuda PASSED [1.7348s] [ 24%] 2025-12-04T10:05:23.3218340Z inductor/test_torchinductor.py::GPUTests::test_conv3d_channels_last_use_block_ptr_True_cuda SKIPPED [0.0033s] (only support cpu conv3d channels_last) [ 24%] 2025-12-04T10:05:23.3219484Z inductor/test_torchinductor.py::GPUTests::test_conv_backward_cuda PASSED [0.3526s] [ 24%] 2025-12-04T10:05:23.3220564Z inductor/test_torchinductor.py::GPUTests::test_conv_bn_fuse_cuda SKIPPED [0.0033s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 25%] 2025-12-04T10:05:23.3221852Z inductor/test_torchinductor.py::GPUTests::test_conv_functional_bn_fuse_cuda SKIPPED [0.0030s] (only support cpu conv bn test) [ 25%] 2025-12-04T10:05:23.3222943Z inductor/test_torchinductor.py::GPUTests::test_conv_inference_heuristics_cuda PASSED [0.4970s] [ 25%] 2025-12-04T10:05:23.3223871Z inductor/test_torchinductor.py::GPUTests::test_conv_with_as_strided_cuda PASSED [0.9252s] [ 25%] 2025-12-04T10:05:23.3224716Z inductor/test_torchinductor.py::GPUTests::test_convolution2_cuda PASSED [0.2272s] [ 25%] 2025-12-04T10:05:23.3225546Z inductor/test_torchinductor.py::GPUTests::test_convolution4_cuda PASSED [0.5162s] [ 26%] 2025-12-04T10:05:23.3226363Z inductor/test_torchinductor.py::GPUTests::test_convolution5_cuda PASSED [0.4025s] [ 26%] 2025-12-04T10:05:23.3227758Z inductor/test_torchinductor.py::GPUTests::test_copy_non_blocking_is_pinned_use_cat_True_cuda W1204 09:56:32.322000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3229120Z W1204 09:56:32.323000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3230052Z W1204 09:56:32.323000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3230979Z W1204 09:56:32.324000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3231908Z W1204 09:56:32.325000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3232880Z W1204 09:56:32.325000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3233828Z W1204 09:56:32.326000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3234758Z W1204 09:56:32.327000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3235699Z W1204 09:56:32.327000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3236613Z W1204 09:56:32.328000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3237534Z W1204 09:56:32.329000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3238457Z W1204 09:56:32.330000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3239382Z W1204 09:56:32.330000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3240299Z W1204 09:56:32.331000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3241235Z W1204 09:56:32.332000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3242153Z W1204 09:56:32.332000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3243058Z W1204 09:56:32.333000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3243980Z W1204 09:56:32.334000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3244905Z W1204 09:56:32.334000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3245815Z W1204 09:56:32.335000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3246740Z W1204 09:56:32.336000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3247661Z W1204 09:56:32.336000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3248590Z W1204 09:56:32.337000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3249501Z W1204 09:56:32.338000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3250438Z W1204 09:56:32.338000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3251361Z W1204 09:56:32.339000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3252406Z W1204 09:56:32.340000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3253322Z W1204 09:56:32.341000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3254247Z W1204 09:56:32.341000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3255249Z W1204 09:56:32.342000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3256176Z W1204 09:56:32.343000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3257157Z W1204 09:56:32.343000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3258088Z W1204 09:56:32.344000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3259008Z W1204 09:56:32.345000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3259932Z W1204 09:56:32.345000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3260908Z W1204 09:56:32.346000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3261834Z W1204 09:56:32.347000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3262945Z W1204 09:56:32.347000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3263872Z W1204 09:56:32.348000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3264783Z W1204 09:56:32.349000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3265705Z W1204 09:56:32.350000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3266629Z W1204 09:56:32.350000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3267551Z W1204 09:56:32.351000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3268465Z W1204 09:56:32.352000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3269391Z W1204 09:56:32.352000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3270311Z W1204 09:56:32.353000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3271234Z W1204 09:56:32.354000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3272146Z W1204 09:56:32.354000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3273063Z W1204 09:56:32.355000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3273981Z W1204 09:56:32.356000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3274892Z W1204 09:56:32.356000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3275806Z W1204 09:56:32.357000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3276729Z W1204 09:56:32.358000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3277651Z W1204 09:56:32.358000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3278560Z W1204 09:56:32.359000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3279482Z W1204 09:56:32.360000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3280406Z W1204 09:56:32.361000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3281331Z W1204 09:56:32.361000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3282245Z W1204 09:56:32.362000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3283279Z W1204 09:56:32.363000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3284208Z W1204 09:56:32.363000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3285132Z W1204 09:56:32.364000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3286043Z W1204 09:56:32.365000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3286965Z W1204 09:56:32.365000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3287882Z W1204 09:56:32.366000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3288804Z W1204 09:56:32.367000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3289784Z W1204 09:56:32.367000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3290707Z W1204 09:56:32.368000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3291631Z W1204 09:56:32.369000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3292561Z W1204 09:56:32.369000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3293474Z W1204 09:56:32.370000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3294401Z W1204 09:56:32.371000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3295325Z W1204 09:56:32.372000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3296954Z W1204 09:56:32.372000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3298026Z W1204 09:56:32.373000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3298957Z W1204 09:56:32.374000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3299882Z W1204 09:56:32.374000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3300797Z W1204 09:56:32.375000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3301701Z W1204 09:56:32.376000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3302614Z W1204 09:56:32.376000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3303528Z W1204 09:56:32.377000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3304530Z W1204 09:56:32.378000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3305455Z W1204 09:56:32.379000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3306367Z W1204 09:56:32.379000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3307278Z W1204 09:56:32.380000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3308237Z W1204 09:56:32.381000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3309152Z W1204 09:56:32.381000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3310074Z W1204 09:56:32.382000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3310992Z W1204 09:56:32.383000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3312177Z W1204 09:56:32.383000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3313101Z W1204 09:56:32.384000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3314014Z W1204 09:56:32.385000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3314929Z W1204 09:56:32.386000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3315827Z W1204 09:56:32.386000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3316741Z W1204 09:56:32.387000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3317657Z W1204 09:56:32.388000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3318669Z W1204 09:56:32.388000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3319578Z W1204 09:56:32.389000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3320495Z W1204 09:56:32.390000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3321405Z W1204 09:56:32.390000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3322041Z PASSED [3.8472s] [ 26%] 2025-12-04T10:05:23.3322536Z inductor/test_torchinductor.py::GPUTests::test_cos_cuda PASSED [0.5836s] [ 26%] 2025-12-04T10:05:23.3323753Z inductor/test_torchinductor.py::GPUTests::test_cpu_scalar_with_gpu_tensor_cuda W1204 09:56:35.248000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3324772Z PASSED [0.1712s] [ 26%] 2025-12-04T10:05:23.3325749Z inductor/test_torchinductor.py::GPUTests::test_cpu_scalar_with_gpu_tensor_dynamic_cuda W1204 09:56:35.447000 13298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:05:23.3326792Z PASSED [0.2096s] [ 27%] 2025-12-04T10:05:23.3327363Z inductor/test_torchinductor.py::GPUTests::test_cpu_tensor_with_gpu_tensor_cuda PASSED [0.0172s] [ 27%] 2025-12-04T10:05:23.3328211Z inductor/test_torchinductor.py::GPUTests::test_cumsum_cuda PASSED [3.3145s] [ 27%] 2025-12-04T10:05:23.3328995Z inductor/test_torchinductor.py::GPUTests::test_cumsum_no_mask_cuda PASSED [0.7974s] [ 27%] 2025-12-04T10:05:23.3329827Z inductor/test_torchinductor.py::GPUTests::test_cumsum_zero_dim_cuda PASSED [0.3122s] [ 27%] 2025-12-04T10:05:23.3330644Z inductor/test_torchinductor.py::GPUTests::test_custom_op_1_cuda PASSED [0.2631s] [ 27%] 2025-12-04T10:05:23.3331436Z inductor/test_torchinductor.py::GPUTests::test_custom_op_2_cuda PASSED [0.1919s] [ 28%] 2025-12-04T10:05:23.3332350Z inductor/test_torchinductor.py::GPUTests::test_custom_op_default_layout_constraint_cuda PASSED [0.2084s] [ 28%] 2025-12-04T10:05:23.3333384Z inductor/test_torchinductor.py::GPUTests::test_custom_op_fixed_layout_channels_last_cuda PASSED [0.4995s] [ 28%] 2025-12-04T10:05:23.3334403Z inductor/test_torchinductor.py::GPUTests::test_custom_op_fixed_layout_sequential_cuda PASSED [0.3127s] [ 28%] 2025-12-04T10:05:23.3335365Z inductor/test_torchinductor.py::GPUTests::test_custom_scan_op_compiled_cuda PASSED [1.2233s] [ 28%] 2025-12-04T10:05:23.3336232Z inductor/test_torchinductor.py::GPUTests::test_custom_scan_op_cuda PASSED [0.2106s] [ 29%] 2025-12-04T10:05:23.3337185Z inductor/test_torchinductor.py::GPUTests::test_custom_scan_op_multi_input_cuda PASSED [0.1583s] [ 29%] 2025-12-04T10:05:23.3338123Z inductor/test_torchinductor.py::GPUTests::test_custom_scan_would_split_cuda PASSED [0.4541s] [ 29%] 2025-12-04T10:05:23.3339028Z inductor/test_torchinductor.py::GPUTests::test_deterministic_codegen_cuda PASSED [3.8727s] [ 29%] 2025-12-04T10:05:23.3339904Z inductor/test_torchinductor.py::GPUTests::test_diagonal_copy_cuda PASSED [0.8877s] [ 29%] 2025-12-04T10:05:23.3340860Z inductor/test_torchinductor.py::GPUTests::test_dist_bf16_cuda SKIPPED [0.0032s] (Requires sm80) [ 30%] 2025-12-04T10:05:23.3341703Z inductor/test_torchinductor.py::GPUTests::test_dist_cuda PASSED [0.6232s] [ 30%] 2025-12-04T10:05:23.3342442Z inductor/test_torchinductor.py::GPUTests::test_div5_cuda PASSED [0.3246s] [ 30%] 2025-12-04T10:05:23.3343184Z inductor/test_torchinductor.py::GPUTests::test_div6_cuda PASSED [0.3529s] [ 30%] 2025-12-04T10:05:23.3343926Z inductor/test_torchinductor.py::GPUTests::test_div8_cuda PASSED [0.7834s] [ 30%] 2025-12-04T10:05:23.3344667Z inductor/test_torchinductor.py::GPUTests::test_div9_cuda PASSED [0.3736s] [ 31%] 2025-12-04T10:05:23.3345409Z inductor/test_torchinductor.py::GPUTests::test_div_prim_cuda PASSED [0.5868s] [ 31%] 2025-12-04T10:05:23.3346237Z inductor/test_torchinductor.py::GPUTests::test_div_softmax_symfloat_cuda PASSED [2.7055s] [ 31%] 2025-12-04T10:05:23.3347148Z inductor/test_torchinductor.py::GPUTests::test_div_zero_dim_cuda PASSED [1.7924s] [ 31%] 2025-12-04T10:05:23.3347933Z inductor/test_torchinductor.py::GPUTests::test_dropout2_cuda PASSED [0.8098s] [ 31%] 2025-12-04T10:05:23.3350198Z inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda SKIPPED [0.0007s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/133025 for platform(s) inductor, linux. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 32%] 2025-12-04T10:05:23.3352486Z inductor/test_torchinductor.py::GPUTests::test_dropout_trivial_0_cuda PASSED [0.3418s] [ 32%] 2025-12-04T10:05:23.3353338Z inductor/test_torchinductor.py::GPUTests::test_dtype_sympy_expr_cuda PASSED [2.1910s] [ 32%] 2025-12-04T10:05:23.3354419Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_float16_float16_cuda SKIPPED [0.0034s] (uses bfloat16 which requires SM >= 80) [ 32%] 2025-12-04T10:05:23.3355732Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_float16_float64_cuda SKIPPED [0.0031s] (uses bfloat16 which requires SM >= 80) [ 32%] 2025-12-04T10:05:23.3357032Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_float16_int32_cuda SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 33%] 2025-12-04T10:05:23.3358312Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_float16_int64_cuda SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 33%] 2025-12-04T10:05:23.3359593Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_float16_uint8_cuda SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 33%] 2025-12-04T10:05:23.3360876Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_float32_float16_cuda SKIPPED [0.0032s] (uses bfloat16 which requires SM >= 80) [ 33%] 2025-12-04T10:05:23.3362250Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_float32_float32_cuda SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 33%] 2025-12-04T10:05:23.3363543Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_float32_int32_cuda SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 33%] 2025-12-04T10:05:23.3364817Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_float32_int8_cuda SKIPPED [0.0030s] (uses bfloat16 which requires SM >= 80) [ 34%] 2025-12-04T10:05:23.3366100Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_float32_uint8_cuda SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 34%] 2025-12-04T10:05:23.3367384Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_float64_float16_cuda SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 34%] 2025-12-04T10:05:23.3368677Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_float64_int16_cuda SKIPPED [0.0031s] (uses bfloat16 which requires SM >= 80) [ 34%] 2025-12-04T10:05:23.3370011Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int16_float16_cuda SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 34%] 2025-12-04T10:05:23.3371365Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int16_float32_cuda SKIPPED [0.0028s] (uses bfloat16 which requires SM >= 80) [ 35%] 2025-12-04T10:05:23.3372622Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int16_int32_cuda SKIPPED [0.0031s] (uses bfloat16 which requires SM >= 80) [ 35%] 2025-12-04T10:05:23.3373881Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int16_int64_cuda SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 35%] 2025-12-04T10:05:23.3375135Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int16_int8_cuda SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 35%] 2025-12-04T10:05:23.3376401Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_float16_cuda SKIPPED [0.0032s] (uses bfloat16 which requires SM >= 80) [ 35%] 2025-12-04T10:05:23.3377760Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_float32_cuda SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 36%] 2025-12-04T10:05:23.3379123Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_float64_cuda SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 36%] 2025-12-04T10:05:23.3380392Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_int16_cuda SKIPPED [0.0028s] (uses bfloat16 which requires SM >= 80) [ 36%] 2025-12-04T10:05:23.3381646Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_int32_cuda SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 36%] 2025-12-04T10:05:23.3382894Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_int64_cuda SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 36%] 2025-12-04T10:05:23.3384144Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_int8_cuda SKIPPED [0.0031s] (uses bfloat16 which requires SM >= 80) [ 37%] 2025-12-04T10:05:23.3385412Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int64_float32_cuda SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 37%] 2025-12-04T10:05:23.3386699Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int64_float64_cuda SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 37%] 2025-12-04T10:05:23.3387990Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int64_int32_cuda SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 37%] 2025-12-04T10:05:23.3389244Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int64_int64_cuda SKIPPED [0.0028s] (uses bfloat16 which requires SM >= 80) [ 37%] 2025-12-04T10:05:23.3390511Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int64_int8_cuda SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 38%] 2025-12-04T10:05:23.3391775Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int64_uint8_cuda SKIPPED [0.0031s] (uses bfloat16 which requires SM >= 80) [ 38%] 2025-12-04T10:05:23.3393049Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int8_float32_cuda SKIPPED [0.0028s] (uses bfloat16 which requires SM >= 80) [ 38%] 2025-12-04T10:05:23.3394325Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int8_float64_cuda SKIPPED [0.0028s] (uses bfloat16 which requires SM >= 80) [ 38%] 2025-12-04T10:05:23.3395610Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int8_int16_cuda SKIPPED [0.0028s] (uses bfloat16 which requires SM >= 80) [ 38%] 2025-12-04T10:05:23.3397200Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int8_int32_cuda SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 38%] 2025-12-04T10:05:23.3398458Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_int8_int64_cuda SKIPPED [0.0028s] (uses bfloat16 which requires SM >= 80) [ 39%] 2025-12-04T10:05:23.3399725Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_uint8_float16_cuda SKIPPED [0.0031s] (uses bfloat16 which requires SM >= 80) [ 39%] 2025-12-04T10:05:23.3401023Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_uint8_float32_cuda SKIPPED [0.0028s] (uses bfloat16 which requires SM >= 80) [ 39%] 2025-12-04T10:05:23.3402319Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_uint8_float64_cuda SKIPPED [0.0028s] (uses bfloat16 which requires SM >= 80) [ 39%] 2025-12-04T10:05:23.3403728Z inductor/test_torchinductor.py::GPUTests::test_dtypeview_uint8_int32_cuda SKIPPED [0.0028s] (uses bfloat16 which requires SM >= 80) [ 39%] 2025-12-04T10:05:23.3404761Z inductor/test_torchinductor.py::GPUTests::test_embedding_cuda PASSED [0.6488s] [ 40%] 2025-12-04T10:05:23.3405550Z inductor/test_torchinductor.py::GPUTests::test_empty1_cuda PASSED [0.1542s] [ 40%] 2025-12-04T10:05:23.3406328Z inductor/test_torchinductor.py::GPUTests::test_empty2_cuda PASSED [0.1521s] [ 40%] 2025-12-04T10:05:23.3407239Z inductor/test_torchinductor.py::GPUTests::test_emulate_precision_triton_fp_fusion_cuda PASSED [0.1824s] [ 40%] 2025-12-04T10:05:23.3408132Z inductor/test_torchinductor.py::GPUTests::test_erfc_cuda PASSED [0.6455s] [ 40%] 2025-12-04T10:05:23.3408894Z inductor/test_torchinductor.py::GPUTests::test_erfinv_cuda PASSED [0.6288s] [ 41%] 2025-12-04T10:05:23.3409785Z inductor/test_torchinductor.py::GPUTests::test_exact_stride_cuda PASSED [0.4715s] [ 41%] 2025-12-04T10:05:23.3410587Z inductor/test_torchinductor.py::GPUTests::test_expm1_cuda PASSED [3.3392s] [ 41%] 2025-12-04T10:05:23.3411428Z inductor/test_torchinductor.py::GPUTests::test_fallback_mutable_op_list_cuda PASSED [0.1864s] [ 41%] 2025-12-04T10:05:23.3412439Z inductor/test_torchinductor.py::GPUTests::test_fallback_mutable_op_no_mutated_tensors_cuda PASSED [0.0328s] [ 41%] 2025-12-04T10:05:23.3413408Z inductor/test_torchinductor.py::GPUTests::test_fft_real_input_cuda PASSED [0.2165s] [ 42%] 2025-12-04T10:05:23.3414204Z inductor/test_torchinductor.py::GPUTests::test_fill2_cuda PASSED [0.3625s] [ 42%] 2025-12-04T10:05:23.3415192Z inductor/test_torchinductor.py::GPUTests::test_flexible_layout_immutable_free_symbols_cuda PASSED [0.0037s] [ 42%] 2025-12-04T10:05:23.3416127Z inductor/test_torchinductor.py::GPUTests::test_flip_cat_cuda PASSED [0.5448s] [ 42%] 2025-12-04T10:05:23.3416955Z inductor/test_torchinductor.py::GPUTests::test_flip_cuda PASSED [0.4412s] [ 42%] 2025-12-04T10:05:23.3417785Z inductor/test_torchinductor.py::GPUTests::test_float_index_expression_cuda PASSED [0.0036s] [ 43%] 2025-12-04T10:05:23.3418871Z inductor/test_torchinductor.py::GPUTests::test_float_repr_dynamic_shapes_cuda PASSED [2.6135s] [ 43%] 2025-12-04T10:05:23.3419982Z inductor/test_torchinductor.py::GPUTests::test_fmin_fmax_cuda PASSED [0.5992s] [ 43%] 2025-12-04T10:05:23.3420846Z inductor/test_torchinductor.py::GPUTests::test_forced_buffer_realize_cuda PASSED [0.3169s] [ 43%] 2025-12-04T10:05:23.3421740Z inductor/test_torchinductor.py::GPUTests::test_fractional_max_pool2d1_cuda PASSED [1.1298s] [ 43%] 2025-12-04T10:05:23.3422668Z inductor/test_torchinductor.py::GPUTests::test_fractional_max_pool2d2_cuda PASSED [1.4301s] [ 44%] 2025-12-04T10:05:23.3423587Z inductor/test_torchinductor.py::GPUTests::test_fractional_max_pool2d3_cuda PASSED [0.3842s] [ 44%] 2025-12-04T10:05:23.3424515Z inductor/test_torchinductor.py::GPUTests::test_fractional_max_pool2d4_cuda PASSED [1.4173s] [ 44%] 2025-12-04T10:05:23.3425429Z inductor/test_torchinductor.py::GPUTests::test_fractional_max_pool2d5_cuda PASSED [0.9577s] [ 44%] 2025-12-04T10:05:23.3426338Z inductor/test_torchinductor.py::GPUTests::test_full_like_transposed_cuda PASSED [0.3326s] [ 44%] 2025-12-04T10:05:23.3427188Z inductor/test_torchinductor.py::GPUTests::test_fuse_tiled_cuda PASSED [0.4027s] [ 44%] 2025-12-04T10:05:23.3427966Z inductor/test_torchinductor.py::GPUTests::test_gather2_cuda PASSED [0.0036s] [ 45%] 2025-12-04T10:05:23.3428772Z inductor/test_torchinductor.py::GPUTests::test_gather_scatter_cuda PASSED [0.3441s] [ 45%] 2025-12-04T10:05:23.3429583Z inductor/test_torchinductor.py::GPUTests::test_getitem_cuda PASSED [0.0238s] [ 45%] 2025-12-04T10:05:23.3430441Z inductor/test_torchinductor.py::GPUTests::test_gpu_scalar_with_cpu_tensor_cuda PASSED [0.0184s] [ 45%] 2025-12-04T10:05:23.3431369Z inductor/test_torchinductor.py::GPUTests::test_graph_partition_arange1_cuda PASSED [1.4498s] [ 45%] 2025-12-04T10:05:23.3432354Z inductor/test_torchinductor.py::GPUTests::test_graph_partition_constant_tensor2_cuda PASSED [0.1747s] [ 46%] 2025-12-04T10:05:23.3433481Z inductor/test_torchinductor.py::GPUTests::test_graph_partition_misaligned_input_cuda PASSED [1.2450s] [ 46%] 2025-12-04T10:05:23.3434468Z inductor/test_torchinductor.py::GPUTests::test_graph_partition_no_inputs_cuda PASSED [0.7803s] [ 46%] 2025-12-04T10:05:23.3435483Z inductor/test_torchinductor.py::GPUTests::test_graph_partition_unbacked_symint_as_output_cuda PASSED [0.4385s] [ 46%] 2025-12-04T10:05:23.3436452Z inductor/test_torchinductor.py::GPUTests::test_hardsigmoid_cuda PASSED [0.5388s] [ 46%] 2025-12-04T10:05:23.3437255Z inductor/test_torchinductor.py::GPUTests::test_hardswish_cuda PASSED [0.5591s] [ 47%] 2025-12-04T10:05:23.3438082Z inductor/test_torchinductor.py::GPUTests::test_horizonal_fusion2_cuda PASSED [0.4327s] [ 47%] 2025-12-04T10:05:23.3438956Z inductor/test_torchinductor.py::GPUTests::test_index1_cuda PASSED [1.0464s] [ 47%] 2025-12-04T10:05:23.3439782Z inductor/test_torchinductor.py::GPUTests::test_index_dynamic_shapes_cuda PASSED [1.0635s] [ 47%] 2025-12-04T10:05:23.3440688Z inductor/test_torchinductor.py::GPUTests::test_index_propagation_abs_cuda PASSED [0.2058s] [ 47%] 2025-12-04T10:05:23.3441671Z inductor/test_torchinductor.py::GPUTests::test_index_propagation_device_assert_masked_cuda PASSED [0.5223s] [ 48%] 2025-12-04T10:05:23.3442792Z inductor/test_torchinductor.py::GPUTests::test_index_propagation_nested_indirect_indexing_cuda PASSED [0.3754s] [ 48%] 2025-12-04T10:05:23.3443757Z inductor/test_torchinductor.py::GPUTests::test_index_put1_cuda PASSED [1.4962s] [ 48%] 2025-12-04T10:05:23.3444549Z inductor/test_torchinductor.py::GPUTests::test_index_put2_cuda PASSED [0.3467s] [ 48%] 2025-12-04T10:05:23.3445329Z inductor/test_torchinductor.py::GPUTests::test_index_put4_cuda PASSED [0.3868s] [ 48%] 2025-12-04T10:05:23.3446230Z inductor/test_torchinductor.py::GPUTests::test_index_put_deterministic_fallback_cuda PASSED [0.1815s] [ 49%] 2025-12-04T10:05:23.3447168Z inductor/test_torchinductor.py::GPUTests::test_index_put_index_cuda PASSED [0.5174s] [ 49%] 2025-12-04T10:05:23.3448042Z inductor/test_torchinductor.py::GPUTests::test_index_put_reinplace_cuda PASSED [0.4578s] [ 49%] 2025-12-04T10:05:23.3448894Z inductor/test_torchinductor.py::GPUTests::test_index_remainder_cuda PASSED [0.4667s] [ 49%] 2025-12-04T10:05:23.3449725Z inductor/test_torchinductor.py::GPUTests::test_index_select_cuda PASSED [1.2571s] [ 49%] 2025-12-04T10:05:23.3450978Z inductor/test_torchinductor.py::GPUTests::test_inductor_multiple_specializations_cuda SKIPPED [0.0003s] (Skipping triton backend only since not big GPU (not enough SM)) [ 50%] 2025-12-04T10:05:23.3452195Z inductor/test_torchinductor.py::GPUTests::test_inf_cuda PASSED [0.3773s] [ 50%] 2025-12-04T10:05:23.3453017Z inductor/test_torchinductor.py::GPUTests::test_inner_fn_str_and_stride_cuda PASSED [0.1758s] [ 50%] 2025-12-04T10:05:23.3453972Z inductor/test_torchinductor.py::GPUTests::test_inner_reduction_detection_cuda PASSED [0.2807s] [ 50%] 2025-12-04T10:05:23.3454898Z inductor/test_torchinductor.py::GPUTests::test_inplace_activations_cuda PASSED [0.8857s] [ 50%] 2025-12-04T10:05:23.3455752Z inductor/test_torchinductor.py::GPUTests::test_inplace_add_cuda PASSED [0.1821s] [ 50%] 2025-12-04T10:05:23.3456550Z inductor/test_torchinductor.py::GPUTests::test_inplace_flip_cuda PASSED [1.2131s] [ 51%] 2025-12-04T10:05:23.3457467Z inductor/test_torchinductor.py::GPUTests::test_input_mutation1_cuda PASSED [0.2328s] [ 51%] 2025-12-04T10:05:23.3458313Z inductor/test_torchinductor.py::GPUTests::test_input_mutation3_cuda PASSED [0.2456s] [ 51%] 2025-12-04T10:05:23.3459146Z inductor/test_torchinductor.py::GPUTests::test_input_mutation4_cuda PASSED [0.6558s] [ 51%] 2025-12-04T10:05:23.3459995Z inductor/test_torchinductor.py::GPUTests::test_input_mutation5_cuda PASSED [0.1659s] [ 51%] 2025-12-04T10:05:23.3460881Z inductor/test_torchinductor.py::GPUTests::test_insignificant_strides_cuda PASSED [0.1719s] [ 52%] 2025-12-04T10:05:23.3461811Z inductor/test_torchinductor.py::GPUTests::test_isinf_cuda ('RERUN', {'yellow': True}) [0.8071s] [ 52%] 2025-12-04T10:05:23.3462816Z inductor/test_torchinductor.py::GPUTests::test_isinf_cuda ('RERUN', {'yellow': True}) [0.7713s] [ 52%] 2025-12-04T10:05:23.3463663Z inductor/test_torchinductor.py::GPUTests::test_isinf_cuda FAILED [0.7696s] [ 52%] 2025-12-04T10:05:23.3464124Z 2025-12-04T10:05:23.3464266Z ==================================== RERUNS ==================================== 2025-12-04T10:05:23.3464787Z ___________________________ GPUTests.test_isinf_cuda ___________________________ 2025-12-04T10:05:23.3465264Z Traceback (most recent call last): 2025-12-04T10:05:23.3465910Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T10:05:23.3466560Z return value(self) 2025-12-04T10:05:23.3467134Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 8264, in test_isinf 2025-12-04T10:05:23.3467845Z with ctx: 2025-12-04T10:05:23.3468312Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T10:05:23.3468948Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T10:05:23.3469585Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T10:05:23.3470223Z raise self.test_case.failureException(msg) 2025-12-04T10:05:23.3470638Z AssertionError: TypeError not raised 2025-12-04T10:05:23.3470883Z 2025-12-04T10:05:23.3471100Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.3471867Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_isinf_cuda 2025-12-04T10:05:23.3472431Z 2025-12-04T10:05:23.3472698Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.3473331Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.3473856Z stats [('calls_captured', 10), ('unique_graphs', 3)] 2025-12-04T10:05:23.3474512Z aot_autograd [('total', 5), ('autograd_cache_miss', 5), ('autograd_cache_saved', 3), ('ok', 3), ('not_ok', 2)] 2025-12-04T10:05:23.3475692Z inductor [('triton_bundler_save_kernel', 24), ('async_compile_cache_miss', 6), ('fxgraph_cache_miss', 5), ('async_compile_cache_hit', 3), ('triton_bundler_save_static_autotuner', 3)] 2025-12-04T10:05:23.3476582Z graph_break [] 2025-12-04T10:05:23.3476943Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3478039Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3479003Z warnings.warn( 2025-12-04T10:05:23.3479873Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3480828Z warnings.warn( 2025-12-04T10:05:23.3481202Z ___________________________ GPUTests.test_isinf_cuda ___________________________ 2025-12-04T10:05:23.3481678Z Traceback (most recent call last): 2025-12-04T10:05:23.3482301Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T10:05:23.3482930Z return value(self) 2025-12-04T10:05:23.3483509Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 8264, in test_isinf 2025-12-04T10:05:23.3484133Z with ctx: 2025-12-04T10:05:23.3484579Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T10:05:23.3485202Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T10:05:23.3485838Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T10:05:23.3486448Z raise self.test_case.failureException(msg) 2025-12-04T10:05:23.3486851Z AssertionError: TypeError not raised 2025-12-04T10:05:23.3487087Z 2025-12-04T10:05:23.3487306Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.3488129Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_isinf_cuda 2025-12-04T10:05:23.3488682Z 2025-12-04T10:05:23.3488944Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.3489564Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.3490083Z stats [('calls_captured', 10), ('unique_graphs', 3)] 2025-12-04T10:05:23.3490723Z aot_autograd [('total', 5), ('autograd_cache_miss', 5), ('autograd_cache_saved', 3), ('ok', 3), ('not_ok', 2)] 2025-12-04T10:05:23.3491877Z inductor [('triton_bundler_save_kernel', 24), ('async_compile_cache_miss', 6), ('fxgraph_cache_miss', 5), ('async_compile_cache_hit', 3), ('triton_bundler_save_static_autotuner', 3)] 2025-12-04T10:05:23.3492775Z graph_break [] 2025-12-04T10:05:23.3493200Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3494278Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3495242Z warnings.warn( 2025-12-04T10:05:23.3496290Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3497312Z warnings.warn( 2025-12-04T10:05:23.3497679Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.3498198Z stats [('calls_captured', 10), ('unique_graphs', 3)] 2025-12-04T10:05:23.3498840Z aot_autograd [('total', 5), ('autograd_cache_miss', 5), ('autograd_cache_saved', 3), ('ok', 3), ('not_ok', 2)] 2025-12-04T10:05:23.3499997Z inductor [('triton_bundler_save_kernel', 24), ('async_compile_cache_miss', 6), ('fxgraph_cache_miss', 5), ('async_compile_cache_hit', 3), ('triton_bundler_save_static_autotuner', 3)] 2025-12-04T10:05:23.3500895Z graph_break [] 2025-12-04T10:05:23.3501269Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3502348Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3503296Z warnings.warn( 2025-12-04T10:05:23.3504171Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3505126Z warnings.warn( 2025-12-04T10:05:23.3505418Z =================================== FAILURES =================================== 2025-12-04T10:05:23.3505924Z ___________________________ GPUTests.test_isinf_cuda ___________________________ 2025-12-04T10:05:23.3506403Z Traceback (most recent call last): 2025-12-04T10:05:23.3507047Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T10:05:23.3507675Z return value(self) 2025-12-04T10:05:23.3508268Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 8264, in test_isinf 2025-12-04T10:05:23.3508912Z with ctx: 2025-12-04T10:05:23.3509359Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T10:05:23.3509990Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T10:05:23.3510630Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T10:05:23.3511264Z raise self.test_case.failureException(msg) 2025-12-04T10:05:23.3511660Z AssertionError: TypeError not raised 2025-12-04T10:05:23.3511910Z 2025-12-04T10:05:23.3512125Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.3512898Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_isinf_cuda 2025-12-04T10:05:23.3513449Z 2025-12-04T10:05:23.3513725Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.3514517Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.3515051Z stats [('calls_captured', 10), ('unique_graphs', 3)] 2025-12-04T10:05:23.3515710Z aot_autograd [('total', 5), ('autograd_cache_miss', 5), ('autograd_cache_saved', 3), ('ok', 3), ('not_ok', 2)] 2025-12-04T10:05:23.3516862Z inductor [('triton_bundler_save_kernel', 24), ('async_compile_cache_miss', 6), ('fxgraph_cache_miss', 5), ('async_compile_cache_hit', 3), ('triton_bundler_save_static_autotuner', 3)] 2025-12-04T10:05:23.3517759Z graph_break [] 2025-12-04T10:05:23.3518137Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3519234Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3520272Z warnings.warn( 2025-12-04T10:05:23.3521164Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3522130Z warnings.warn( 2025-12-04T10:05:23.3522508Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.3523022Z stats [('calls_captured', 10), ('unique_graphs', 3)] 2025-12-04T10:05:23.3523682Z aot_autograd [('total', 5), ('autograd_cache_miss', 5), ('autograd_cache_saved', 3), ('ok', 3), ('not_ok', 2)] 2025-12-04T10:05:23.3524844Z inductor [('triton_bundler_save_kernel', 24), ('async_compile_cache_miss', 6), ('fxgraph_cache_miss', 5), ('async_compile_cache_hit', 3), ('triton_bundler_save_static_autotuner', 3)] 2025-12-04T10:05:23.3525727Z graph_break [] 2025-12-04T10:05:23.3526108Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3527193Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3528167Z warnings.warn( 2025-12-04T10:05:23.3529037Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3529996Z warnings.warn( 2025-12-04T10:05:23.3530375Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.3530892Z stats [('calls_captured', 10), ('unique_graphs', 3)] 2025-12-04T10:05:23.3531531Z aot_autograd [('total', 5), ('autograd_cache_miss', 5), ('autograd_cache_saved', 3), ('ok', 3), ('not_ok', 2)] 2025-12-04T10:05:23.3532694Z inductor [('triton_bundler_save_kernel', 24), ('async_compile_cache_miss', 6), ('fxgraph_cache_miss', 5), ('async_compile_cache_hit', 3), ('triton_bundler_save_static_autotuner', 3)] 2025-12-04T10:05:23.3533595Z graph_break [] 2025-12-04T10:05:23.3533960Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3535056Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3536021Z warnings.warn( 2025-12-04T10:05:23.3536968Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3537916Z warnings.warn( 2025-12-04T10:05:23.3538853Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-6ce11cbc918a017c.xml - 2025-12-04T10:05:23.3539935Z =========================== short test summary info ============================ 2025-12-04T10:05:23.3540718Z FAILED [0.7696s] inductor/test_torchinductor.py::GPUTests::test_isinf_cuda - AssertionError: TypeError not raised 2025-12-04T10:05:23.3541312Z 2025-12-04T10:05:23.3541527Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.3542361Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_isinf_cuda 2025-12-04T10:05:23.3542910Z 2025-12-04T10:05:23.3543189Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.3543780Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T10:05:23.3544323Z ======== 1 failed, 215 passed, 55 skipped, 2 rerun in 158.93s (0:02:38) ======== 2025-12-04T10:05:23.3544803Z Got exit code 1 2025-12-04T10:05:23.3545073Z Retrying single test... 2025-12-04T10:05:23.3545703Z W1204 09:57:50.274000 18147 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:05:23.3546887Z Test results will be stored in test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-c850c8e21fb6d04c.xml 2025-12-04T10:05:23.3547854Z ============================= test session starts ============================== 2025-12-04T10:05:23.3548524Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T10:05:23.3549110Z cachedir: .pytest_cache 2025-12-04T10:05:23.3549823Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T10:05:23.3550606Z rootdir: /var/lib/jenkins/workspace 2025-12-04T10:05:23.3550950Z configfile: pytest.ini 2025-12-04T10:05:23.3551683Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T10:05:23.3552600Z collecting ... collected 981 items / 517 deselected / 464 selected 2025-12-04T10:05:23.3553472Z stepcurrent: skipping 270 already run items. Running only test/inductor/test_torchinductor.py::GPUTests::test_isinf_cuda 2025-12-04T10:05:23.3554226Z Running 1 items in this shard 2025-12-04T10:05:23.3554447Z 2025-12-04T10:05:23.3554847Z inductor/test_torchinductor.py::GPUTests::test_isinf_cuda ('RERUN', {'yellow': True}) [3.7883s] [100%] 2025-12-04T10:05:23.3555780Z inductor/test_torchinductor.py::GPUTests::test_isinf_cuda ('RERUN', {'yellow': True}) [0.7789s] [100%] 2025-12-04T10:05:23.3556633Z inductor/test_torchinductor.py::GPUTests::test_isinf_cuda FAILED [0.7883s] [100%] 2025-12-04T10:05:23.3557084Z 2025-12-04T10:05:23.3557224Z ==================================== RERUNS ==================================== 2025-12-04T10:05:23.3557736Z ___________________________ GPUTests.test_isinf_cuda ___________________________ 2025-12-04T10:05:23.3558223Z Traceback (most recent call last): 2025-12-04T10:05:23.3558852Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T10:05:23.3559555Z return value(self) 2025-12-04T10:05:23.3560145Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 8264, in test_isinf 2025-12-04T10:05:23.3560792Z with ctx: 2025-12-04T10:05:23.3561241Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T10:05:23.3561866Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T10:05:23.3562513Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T10:05:23.3563135Z raise self.test_case.failureException(msg) 2025-12-04T10:05:23.3563545Z AssertionError: TypeError not raised 2025-12-04T10:05:23.3563798Z 2025-12-04T10:05:23.3564014Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.3564791Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_isinf_cuda 2025-12-04T10:05:23.3565335Z 2025-12-04T10:05:23.3565604Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.3566235Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.3566764Z stats [('calls_captured', 10), ('unique_graphs', 3)] 2025-12-04T10:05:23.3567687Z inductor [('triton_bundler_save_kernel', 24), ('fxgraph_cache_miss', 5), ('async_compile_cache_miss', 3), ('triton_bundler_save_static_autotuner', 3)] 2025-12-04T10:05:23.3568721Z aot_autograd [('total', 5), ('autograd_cache_miss', 5), ('autograd_cache_saved', 3), ('ok', 3), ('not_ok', 2)] 2025-12-04T10:05:23.3569321Z graph_break [] 2025-12-04T10:05:23.3569693Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3570779Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3571809Z warnings.warn( 2025-12-04T10:05:23.3572693Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3573720Z warnings.warn( 2025-12-04T10:05:23.3574089Z ___________________________ GPUTests.test_isinf_cuda ___________________________ 2025-12-04T10:05:23.3574571Z Traceback (most recent call last): 2025-12-04T10:05:23.3575212Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T10:05:23.3575846Z return value(self) 2025-12-04T10:05:23.3576431Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 8264, in test_isinf 2025-12-04T10:05:23.3577156Z with ctx: 2025-12-04T10:05:23.3577617Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T10:05:23.3578240Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T10:05:23.3578887Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T10:05:23.3579526Z raise self.test_case.failureException(msg) 2025-12-04T10:05:23.3579940Z AssertionError: TypeError not raised 2025-12-04T10:05:23.3580194Z 2025-12-04T10:05:23.3580409Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.3581176Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_isinf_cuda 2025-12-04T10:05:23.3581724Z 2025-12-04T10:05:23.3582002Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.3582614Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.3583136Z stats [('calls_captured', 10), ('unique_graphs', 3)] 2025-12-04T10:05:23.3583977Z inductor [('triton_bundler_save_kernel', 24), ('fxgraph_cache_miss', 5), ('async_compile_cache_miss', 3), ('triton_bundler_save_static_autotuner', 3)] 2025-12-04T10:05:23.3585015Z aot_autograd [('total', 5), ('autograd_cache_miss', 5), ('autograd_cache_saved', 3), ('ok', 3), ('not_ok', 2)] 2025-12-04T10:05:23.3585610Z graph_break [] 2025-12-04T10:05:23.3585979Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3587088Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3588040Z warnings.warn( 2025-12-04T10:05:23.3588927Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3589887Z warnings.warn( 2025-12-04T10:05:23.3590261Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.3590768Z stats [('calls_captured', 10), ('unique_graphs', 3)] 2025-12-04T10:05:23.3591416Z aot_autograd [('total', 5), ('autograd_cache_miss', 5), ('autograd_cache_saved', 3), ('ok', 3), ('not_ok', 2)] 2025-12-04T10:05:23.3592458Z inductor [('triton_bundler_save_kernel', 24), ('fxgraph_cache_miss', 5), ('async_compile_cache_miss', 3), ('triton_bundler_save_static_autotuner', 3)] 2025-12-04T10:05:23.3593244Z graph_break [] 2025-12-04T10:05:23.3593601Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3594770Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3595743Z warnings.warn( 2025-12-04T10:05:23.3596769Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3597730Z warnings.warn( 2025-12-04T10:05:23.3598042Z =================================== FAILURES =================================== 2025-12-04T10:05:23.3598557Z ___________________________ GPUTests.test_isinf_cuda ___________________________ 2025-12-04T10:05:23.3599025Z Traceback (most recent call last): 2025-12-04T10:05:23.3599770Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T10:05:23.3600416Z return value(self) 2025-12-04T10:05:23.3600990Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 8264, in test_isinf 2025-12-04T10:05:23.3601630Z with ctx: 2025-12-04T10:05:23.3602083Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T10:05:23.3602716Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T10:05:23.3603357Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T10:05:23.3603987Z raise self.test_case.failureException(msg) 2025-12-04T10:05:23.3604392Z AssertionError: TypeError not raised 2025-12-04T10:05:23.3604635Z 2025-12-04T10:05:23.3604849Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.3605606Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_isinf_cuda 2025-12-04T10:05:23.3606167Z 2025-12-04T10:05:23.3606432Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.3607062Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.3607573Z stats [('calls_captured', 10), ('unique_graphs', 3)] 2025-12-04T10:05:23.3608421Z inductor [('triton_bundler_save_kernel', 24), ('fxgraph_cache_miss', 5), ('async_compile_cache_miss', 3), ('triton_bundler_save_static_autotuner', 3)] 2025-12-04T10:05:23.3609477Z aot_autograd [('total', 5), ('autograd_cache_miss', 5), ('autograd_cache_saved', 3), ('ok', 3), ('not_ok', 2)] 2025-12-04T10:05:23.3610082Z graph_break [] 2025-12-04T10:05:23.3610449Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3611548Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3612532Z warnings.warn( 2025-12-04T10:05:23.3613427Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3614398Z warnings.warn( 2025-12-04T10:05:23.3614782Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.3615313Z stats [('calls_captured', 10), ('unique_graphs', 3)] 2025-12-04T10:05:23.3615964Z aot_autograd [('total', 5), ('autograd_cache_miss', 5), ('autograd_cache_saved', 3), ('ok', 3), ('not_ok', 2)] 2025-12-04T10:05:23.3617112Z inductor [('triton_bundler_save_kernel', 24), ('fxgraph_cache_miss', 5), ('async_compile_cache_miss', 3), ('triton_bundler_save_static_autotuner', 3)] 2025-12-04T10:05:23.3617912Z graph_break [] 2025-12-04T10:05:23.3618278Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3619442Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3620420Z warnings.warn( 2025-12-04T10:05:23.3621421Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3622375Z warnings.warn( 2025-12-04T10:05:23.3622758Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.3623287Z stats [('calls_captured', 10), ('unique_graphs', 3)] 2025-12-04T10:05:23.3623975Z aot_autograd [('total', 5), ('autograd_cache_miss', 5), ('autograd_cache_saved', 3), ('ok', 3), ('not_ok', 2)] 2025-12-04T10:05:23.3625006Z inductor [('triton_bundler_save_kernel', 24), ('fxgraph_cache_miss', 5), ('async_compile_cache_miss', 3), ('triton_bundler_save_static_autotuner', 3)] 2025-12-04T10:05:23.3625796Z graph_break [] 2025-12-04T10:05:23.3626237Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3627320Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3628281Z warnings.warn( 2025-12-04T10:05:23.3629161Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3630122Z warnings.warn( 2025-12-04T10:05:23.3631044Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-c850c8e21fb6d04c.xml - 2025-12-04T10:05:23.3632130Z =========================== short test summary info ============================ 2025-12-04T10:05:23.3632913Z FAILED [0.7883s] inductor/test_torchinductor.py::GPUTests::test_isinf_cuda - AssertionError: TypeError not raised 2025-12-04T10:05:23.3633507Z 2025-12-04T10:05:23.3633739Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.3634493Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_isinf_cuda 2025-12-04T10:05:23.3635060Z 2025-12-04T10:05:23.3635328Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.3635922Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T10:05:23.3636451Z ================== 1 failed, 517 deselected, 2 rerun in 5.44s ================== 2025-12-04T10:05:23.3636886Z Got exit code 1 2025-12-04T10:05:23.3637161Z Retrying single test... 2025-12-04T10:05:23.3637797Z W1204 09:58:09.424000 18417 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:05:23.3638959Z Test results will be stored in test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-17e608243d035d18.xml 2025-12-04T10:05:23.3639861Z ============================= test session starts ============================== 2025-12-04T10:05:23.3640525Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T10:05:23.3641125Z cachedir: .pytest_cache 2025-12-04T10:05:23.3641826Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T10:05:23.3642614Z rootdir: /var/lib/jenkins/workspace 2025-12-04T10:05:23.3642968Z configfile: pytest.ini 2025-12-04T10:05:23.3643690Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T10:05:23.3644599Z collecting ... collected 981 items / 517 deselected / 464 selected 2025-12-04T10:05:23.3645472Z stepcurrent: skipping 270 already run items. Running only test/inductor/test_torchinductor.py::GPUTests::test_isinf_cuda 2025-12-04T10:05:23.3646234Z Running 1 items in this shard 2025-12-04T10:05:23.3646456Z 2025-12-04T10:05:23.3646858Z inductor/test_torchinductor.py::GPUTests::test_isinf_cuda ('RERUN', {'yellow': True}) [3.7674s] [100%] 2025-12-04T10:05:23.3647870Z inductor/test_torchinductor.py::GPUTests::test_isinf_cuda ('RERUN', {'yellow': True}) [0.7762s] [100%] 2025-12-04T10:05:23.3648730Z inductor/test_torchinductor.py::GPUTests::test_isinf_cuda FAILED [0.7771s] [100%] 2025-12-04T10:05:23.3649177Z 2025-12-04T10:05:23.3649335Z ==================================== RERUNS ==================================== 2025-12-04T10:05:23.3649834Z ___________________________ GPUTests.test_isinf_cuda ___________________________ 2025-12-04T10:05:23.3650324Z Traceback (most recent call last): 2025-12-04T10:05:23.3650970Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T10:05:23.3651622Z return value(self) 2025-12-04T10:05:23.3652193Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 8264, in test_isinf 2025-12-04T10:05:23.3652897Z with ctx: 2025-12-04T10:05:23.3653358Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T10:05:23.3653977Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T10:05:23.3654631Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T10:05:23.3655266Z raise self.test_case.failureException(msg) 2025-12-04T10:05:23.3655668Z AssertionError: TypeError not raised 2025-12-04T10:05:23.3655920Z 2025-12-04T10:05:23.3656135Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.3656977Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_isinf_cuda 2025-12-04T10:05:23.3657526Z 2025-12-04T10:05:23.3657808Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.3658424Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.3658956Z stats [('calls_captured', 10), ('unique_graphs', 3)] 2025-12-04T10:05:23.3659807Z inductor [('triton_bundler_save_kernel', 24), ('fxgraph_cache_miss', 5), ('async_compile_cache_miss', 3), ('triton_bundler_save_static_autotuner', 3)] 2025-12-04T10:05:23.3660857Z aot_autograd [('total', 5), ('autograd_cache_miss', 5), ('autograd_cache_saved', 3), ('ok', 3), ('not_ok', 2)] 2025-12-04T10:05:23.3661446Z graph_break [] 2025-12-04T10:05:23.3661819Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3662915Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3663890Z warnings.warn( 2025-12-04T10:05:23.3664765Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3665730Z warnings.warn( 2025-12-04T10:05:23.3666122Z ___________________________ GPUTests.test_isinf_cuda ___________________________ 2025-12-04T10:05:23.3666593Z Traceback (most recent call last): 2025-12-04T10:05:23.3667239Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T10:05:23.3667936Z return value(self) 2025-12-04T10:05:23.3668510Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 8264, in test_isinf 2025-12-04T10:05:23.3669152Z with ctx: 2025-12-04T10:05:23.3669609Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T10:05:23.3670240Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T10:05:23.3670875Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T10:05:23.3671506Z raise self.test_case.failureException(msg) 2025-12-04T10:05:23.3671918Z AssertionError: TypeError not raised 2025-12-04T10:05:23.3672165Z 2025-12-04T10:05:23.3672379Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.3673144Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_isinf_cuda 2025-12-04T10:05:23.3673787Z 2025-12-04T10:05:23.3674632Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.3675267Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.3675784Z stats [('calls_captured', 10), ('unique_graphs', 3)] 2025-12-04T10:05:23.3676637Z inductor [('triton_bundler_save_kernel', 24), ('fxgraph_cache_miss', 5), ('async_compile_cache_miss', 3), ('triton_bundler_save_static_autotuner', 3)] 2025-12-04T10:05:23.3677684Z aot_autograd [('total', 5), ('autograd_cache_miss', 5), ('autograd_cache_saved', 3), ('ok', 3), ('not_ok', 2)] 2025-12-04T10:05:23.3678297Z graph_break [] 2025-12-04T10:05:23.3678659Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3679828Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3680810Z warnings.warn( 2025-12-04T10:05:23.3681710Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3682661Z warnings.warn( 2025-12-04T10:05:23.3683065Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.3683592Z stats [('calls_captured', 10), ('unique_graphs', 3)] 2025-12-04T10:05:23.3684233Z aot_autograd [('total', 5), ('autograd_cache_miss', 5), ('autograd_cache_saved', 3), ('ok', 3), ('not_ok', 2)] 2025-12-04T10:05:23.3685286Z inductor [('triton_bundler_save_kernel', 24), ('fxgraph_cache_miss', 5), ('async_compile_cache_miss', 3), ('triton_bundler_save_static_autotuner', 3)] 2025-12-04T10:05:23.3686076Z graph_break [] 2025-12-04T10:05:23.3686452Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3687532Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3688496Z warnings.warn( 2025-12-04T10:05:23.3689383Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3690344Z warnings.warn( 2025-12-04T10:05:23.3690640Z =================================== FAILURES =================================== 2025-12-04T10:05:23.3691147Z ___________________________ GPUTests.test_isinf_cuda ___________________________ 2025-12-04T10:05:23.3691670Z Traceback (most recent call last): 2025-12-04T10:05:23.3692300Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T10:05:23.3692951Z return value(self) 2025-12-04T10:05:23.3693530Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 8264, in test_isinf 2025-12-04T10:05:23.3694167Z with ctx: 2025-12-04T10:05:23.3694618Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T10:05:23.3695252Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T10:05:23.3695897Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T10:05:23.3696734Z raise self.test_case.failureException(msg) 2025-12-04T10:05:23.3697221Z AssertionError: TypeError not raised 2025-12-04T10:05:23.3697462Z 2025-12-04T10:05:23.3697691Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.3698445Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_isinf_cuda 2025-12-04T10:05:23.3699009Z 2025-12-04T10:05:23.3699277Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.3699912Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.3700441Z stats [('calls_captured', 10), ('unique_graphs', 3)] 2025-12-04T10:05:23.3701396Z inductor [('triton_bundler_save_kernel', 24), ('fxgraph_cache_miss', 5), ('async_compile_cache_miss', 3), ('triton_bundler_save_static_autotuner', 3)] 2025-12-04T10:05:23.3702445Z aot_autograd [('total', 5), ('autograd_cache_miss', 5), ('autograd_cache_saved', 3), ('ok', 3), ('not_ok', 2)] 2025-12-04T10:05:23.3703084Z graph_break [] 2025-12-04T10:05:23.3703461Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3704543Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3705521Z warnings.warn( 2025-12-04T10:05:23.3706412Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3707520Z warnings.warn( 2025-12-04T10:05:23.3707899Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.3708430Z stats [('calls_captured', 10), ('unique_graphs', 3)] 2025-12-04T10:05:23.3709094Z aot_autograd [('total', 5), ('autograd_cache_miss', 5), ('autograd_cache_saved', 3), ('ok', 3), ('not_ok', 2)] 2025-12-04T10:05:23.3710139Z inductor [('triton_bundler_save_kernel', 24), ('fxgraph_cache_miss', 5), ('async_compile_cache_miss', 3), ('triton_bundler_save_static_autotuner', 3)] 2025-12-04T10:05:23.3710925Z graph_break [] 2025-12-04T10:05:23.3711302Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3712392Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3713376Z warnings.warn( 2025-12-04T10:05:23.3714264Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3715228Z warnings.warn( 2025-12-04T10:05:23.3715606Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.3716114Z stats [('calls_captured', 10), ('unique_graphs', 3)] 2025-12-04T10:05:23.3716768Z aot_autograd [('total', 5), ('autograd_cache_miss', 5), ('autograd_cache_saved', 3), ('ok', 3), ('not_ok', 2)] 2025-12-04T10:05:23.3717810Z inductor [('triton_bundler_save_kernel', 24), ('fxgraph_cache_miss', 5), ('async_compile_cache_miss', 3), ('triton_bundler_save_static_autotuner', 3)] 2025-12-04T10:05:23.3718580Z graph_break [] 2025-12-04T10:05:23.3718948Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3720034Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3721001Z warnings.warn( 2025-12-04T10:05:23.3721879Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.3722847Z warnings.warn( 2025-12-04T10:05:23.3723777Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-17e608243d035d18.xml - 2025-12-04T10:05:23.3724846Z =========================== short test summary info ============================ 2025-12-04T10:05:23.3725613Z FAILED [0.7771s] inductor/test_torchinductor.py::GPUTests::test_isinf_cuda - AssertionError: TypeError not raised 2025-12-04T10:05:23.3726416Z 2025-12-04T10:05:23.3726692Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.3727478Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_isinf_cuda 2025-12-04T10:05:23.3728027Z 2025-12-04T10:05:23.3728384Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.3728962Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T10:05:23.3729488Z ================== 1 failed, 517 deselected, 2 rerun in 5.41s ================== 2025-12-04T10:05:23.3729930Z Got exit code 1 2025-12-04T10:05:23.3730412Z FAILED CONSISTENTLY: test/inductor/test_torchinductor.py::GPUTests::test_isinf_cuda 2025-12-04T10:05:23.3731293Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T10:05:23.3732296Z W1204 09:58:28.393000 18687 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:05:23.3733468Z Test results will be stored in test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-7d6e32f4c213ad66.xml 2025-12-04T10:05:23.3734486Z ============================= test session starts ============================== 2025-12-04T10:05:23.3735149Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T10:05:23.3735755Z cachedir: .pytest_cache 2025-12-04T10:05:23.3736457Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T10:05:23.3737313Z rootdir: /var/lib/jenkins/workspace 2025-12-04T10:05:23.3737668Z configfile: pytest.ini 2025-12-04T10:05:23.3738396Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T10:05:23.3739303Z collecting ... collected 981 items / 271 deselected / 710 selected 2025-12-04T10:05:23.3739810Z stepcurrent: skipping 271 already run items. 2025-12-04T10:05:23.3740201Z Running 247 items in this shard 2025-12-04T10:05:23.3740418Z 2025-12-04T10:05:23.3740763Z inductor/test_torchinductor.py::GPUTests::test_issue102546_cuda PASSED [3.2376s] [ 0%] 2025-12-04T10:05:23.3741570Z inductor/test_torchinductor.py::GPUTests::test_kernel_names_cuda PASSED [0.1980s] [ 0%] 2025-12-04T10:05:23.3742519Z inductor/test_torchinductor.py::GPUTests::test_kwargs_cuda SKIPPED [0.0032s] (histogramdd only supports cpu) [ 1%] 2025-12-04T10:05:23.3743445Z inductor/test_torchinductor.py::GPUTests::test_l1_loss_cuda PASSED [0.2979s] [ 1%] 2025-12-04T10:05:23.3744294Z inductor/test_torchinductor.py::GPUTests::test_large_broadcast_reduction_cuda PASSED [0.5578s] [ 2%] 2025-12-04T10:05:23.3745265Z inductor/test_torchinductor.py::GPUTests::test_large_grid_use_block_ptr_True_cuda PASSED [0.5293s] [ 2%] 2025-12-04T10:05:23.3746184Z inductor/test_torchinductor.py::GPUTests::test_large_pointwise_cuda PASSED [0.4523s] [ 2%] 2025-12-04T10:05:23.3747077Z inductor/test_torchinductor.py::GPUTests::test_large_strided_reduction_cuda PASSED [0.8577s] [ 3%] 2025-12-04T10:05:23.3747996Z inductor/test_torchinductor.py::GPUTests::test_large_tensor_reduction_cuda PASSED [0.7854s] [ 3%] 2025-12-04T10:05:23.3748859Z inductor/test_torchinductor.py::GPUTests::test_leaky_relu_cuda PASSED [0.8494s] [ 4%] 2025-12-04T10:05:23.3749642Z inductor/test_torchinductor.py::GPUTests::test_lgamma_cuda PASSED [1.6621s] [ 4%] 2025-12-04T10:05:23.3750431Z inductor/test_torchinductor.py::GPUTests::test_like_rands3_cuda PASSED [1.5886s] [ 4%] 2025-12-04T10:05:23.3751259Z inductor/test_torchinductor.py::GPUTests::test_like_rands_sliced_cuda PASSED [0.2737s] [ 5%] 2025-12-04T10:05:23.3752281Z inductor/test_torchinductor.py::GPUTests::test_linalg_eig_stride_consistency_cuda ('RERUN', {'yellow': True}) [0.0333s] [ 5%] 2025-12-04T10:05:23.3753456Z inductor/test_torchinductor.py::GPUTests::test_linalg_eig_stride_consistency_cuda ('RERUN', {'yellow': True}) [0.0049s] [ 5%] 2025-12-04T10:05:23.3754543Z inductor/test_torchinductor.py::GPUTests::test_linalg_eig_stride_consistency_cuda FAILED [0.0045s] [ 5%] 2025-12-04T10:05:23.3755106Z 2025-12-04T10:05:23.3755249Z ==================================== RERUNS ==================================== 2025-12-04T10:05:23.3755872Z _______________ GPUTests.test_linalg_eig_stride_consistency_cuda _______________ 2025-12-04T10:05:23.3756407Z Traceback (most recent call last): 2025-12-04T10:05:23.3757238Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 5999, in test_linalg_eig_stride_consistency 2025-12-04T10:05:23.3758017Z self.common( 2025-12-04T10:05:23.3758468Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T10:05:23.3759008Z return func(*args, **kwds) 2025-12-04T10:05:23.3759646Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T10:05:23.3760323Z check_model( 2025-12-04T10:05:23.3760891Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 497, in check_model 2025-12-04T10:05:23.3761640Z correct = ref_model(*ref_inputs, **ref_kwargs) 2025-12-04T10:05:23.3762285Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 5994, in fn 2025-12-04T10:05:23.3762943Z eigenvals, eigenvecs = torch.linalg.eig(x) 2025-12-04T10:05:23.3763978Z RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T10:05:23.3764869Z 2025-12-04T10:05:23.3765083Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.3765971Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_linalg_eig_stride_consistency_cuda 2025-12-04T10:05:23.3766650Z 2025-12-04T10:05:23.3766915Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.3767584Z _______________ GPUTests.test_linalg_eig_stride_consistency_cuda _______________ 2025-12-04T10:05:23.3768106Z Traceback (most recent call last): 2025-12-04T10:05:23.3768891Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 5999, in test_linalg_eig_stride_consistency 2025-12-04T10:05:23.3769662Z self.common( 2025-12-04T10:05:23.3770128Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T10:05:23.3770666Z return func(*args, **kwds) 2025-12-04T10:05:23.3771314Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T10:05:23.3771991Z check_model( 2025-12-04T10:05:23.3772542Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 497, in check_model 2025-12-04T10:05:23.3773239Z correct = ref_model(*ref_inputs, **ref_kwargs) 2025-12-04T10:05:23.3773893Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 5994, in fn 2025-12-04T10:05:23.3774533Z eigenvals, eigenvecs = torch.linalg.eig(x) 2025-12-04T10:05:23.3775582Z RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T10:05:23.3776518Z 2025-12-04T10:05:23.3776737Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.3777692Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_linalg_eig_stride_consistency_cuda 2025-12-04T10:05:23.3778361Z 2025-12-04T10:05:23.3778628Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.3779190Z =================================== FAILURES =================================== 2025-12-04T10:05:23.3779746Z _______________ GPUTests.test_linalg_eig_stride_consistency_cuda _______________ 2025-12-04T10:05:23.3780272Z Traceback (most recent call last): 2025-12-04T10:05:23.3781040Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 5999, in test_linalg_eig_stride_consistency 2025-12-04T10:05:23.3781827Z self.common( 2025-12-04T10:05:23.3782271Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T10:05:23.3782792Z return func(*args, **kwds) 2025-12-04T10:05:23.3783510Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T10:05:23.3784191Z check_model( 2025-12-04T10:05:23.3784765Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 497, in check_model 2025-12-04T10:05:23.3785452Z correct = ref_model(*ref_inputs, **ref_kwargs) 2025-12-04T10:05:23.3786102Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 5994, in fn 2025-12-04T10:05:23.3786765Z eigenvals, eigenvecs = torch.linalg.eig(x) 2025-12-04T10:05:23.3787794Z RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T10:05:23.3788758Z 2025-12-04T10:05:23.3788974Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.3789858Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_linalg_eig_stride_consistency_cuda 2025-12-04T10:05:23.3790536Z 2025-12-04T10:05:23.3790812Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.3791996Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-7d6e32f4c213ad66.xml - 2025-12-04T10:05:23.3793072Z =========================== short test summary info ============================ 2025-12-04T10:05:23.3794639Z FAILED [0.0045s] inductor/test_torchinductor.py::GPUTests::test_linalg_eig_stride_consistency_cuda - RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T10:05:23.3796212Z 2025-12-04T10:05:23.3796433Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.3797324Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_linalg_eig_stride_consistency_cuda 2025-12-04T10:05:23.3797997Z 2025-12-04T10:05:23.3798261Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.3798858Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T10:05:23.3799427Z ====== 1 failed, 12 passed, 1 skipped, 271 deselected, 2 rerun in 11.45s ======= 2025-12-04T10:05:23.3799917Z Got exit code 1 2025-12-04T10:05:23.3800174Z Retrying single test... 2025-12-04T10:05:23.3800811Z W1204 09:58:52.615000 19259 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:05:23.3801993Z Test results will be stored in test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-e7bcb2cd86408aa6.xml 2025-12-04T10:05:23.3802887Z ============================= test session starts ============================== 2025-12-04T10:05:23.3803551Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T10:05:23.3804154Z cachedir: .pytest_cache 2025-12-04T10:05:23.3804869Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T10:05:23.3805637Z rootdir: /var/lib/jenkins/workspace 2025-12-04T10:05:23.3805987Z configfile: pytest.ini 2025-12-04T10:05:23.3806717Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T10:05:23.3807613Z collecting ... collected 981 items / 517 deselected / 464 selected 2025-12-04T10:05:23.3808610Z stepcurrent: skipping 284 already run items. Running only test/inductor/test_torchinductor.py::GPUTests::test_linalg_eig_stride_consistency_cuda 2025-12-04T10:05:23.3808746Z Running 1 items in this shard 2025-12-04T10:05:23.3808752Z 2025-12-04T10:05:23.3809789Z inductor/test_torchinductor.py::GPUTests::test_linalg_eig_stride_consistency_cuda [W1204 09:58:54.233341934 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T10:05:23.3809796Z 2025-12-04T10:05:23.3810318Z [W1204 09:59:10.198972866 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T10:05:23.3810336Z 2025-12-04T10:05:23.3810470Z ('RERUN', {'yellow': True}) [16.1889s] [100%] 2025-12-04T10:05:23.3811394Z inductor/test_torchinductor.py::GPUTests::test_linalg_eig_stride_consistency_cuda [W1204 09:59:10.427013371 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T10:05:23.3811399Z 2025-12-04T10:05:23.3811933Z [W1204 09:59:10.427877946 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T10:05:23.3812035Z 2025-12-04T10:05:23.3812166Z ('RERUN', {'yellow': True}) [0.0055s] [100%] 2025-12-04T10:05:23.3813108Z inductor/test_torchinductor.py::GPUTests::test_linalg_eig_stride_consistency_cuda [W1204 09:59:10.433756314 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T10:05:23.3813113Z 2025-12-04T10:05:23.3813628Z [W1204 09:59:10.434542797 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T10:05:23.3813633Z 2025-12-04T10:05:23.3813748Z FAILED [0.0050s] [100%] 2025-12-04T10:05:23.3813753Z 2025-12-04T10:05:23.3813896Z ==================================== RERUNS ==================================== 2025-12-04T10:05:23.3814151Z _______________ GPUTests.test_linalg_eig_stride_consistency_cuda _______________ 2025-12-04T10:05:23.3814290Z Traceback (most recent call last): 2025-12-04T10:05:23.3814838Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 5999, in test_linalg_eig_stride_consistency 2025-12-04T10:05:23.3814938Z self.common( 2025-12-04T10:05:23.3815235Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T10:05:23.3815347Z return func(*args, **kwds) 2025-12-04T10:05:23.3815798Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T10:05:23.3815894Z check_model( 2025-12-04T10:05:23.3816298Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 497, in check_model 2025-12-04T10:05:23.3816460Z correct = ref_model(*ref_inputs, **ref_kwargs) 2025-12-04T10:05:23.3816818Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 5994, in fn 2025-12-04T10:05:23.3817031Z eigenvals, eigenvecs = torch.linalg.eig(x) 2025-12-04T10:05:23.3817810Z RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T10:05:23.3818584Z Exception raised from apply_magma_eig at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebra.cpp:2024 (most recent call first): 2025-12-04T10:05:23.3818712Z C++ CapturedTraceback: 2025-12-04T10:05:23.3820027Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T10:05:23.3820526Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T10:05:23.3820864Z #6 c10::detail::torchCheckFail(char const*, char const*, unsigned int, char const*) from ??:0 2025-12-04T10:05:23.3821679Z #7 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool)::{lambda()#1}::operator()() const::{lambda()#1}::operator()() const [clone .constprop.0] from BatchLinearAlgebra.cpp:0 2025-12-04T10:05:23.3822192Z #8 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T10:05:23.3822633Z #9 at::native::linalg_eig_out_info(at::Tensor const&, at::Tensor&, at::Tensor&, at::Tensor&, bool) from BatchLinearAlgebra.cpp:0 2025-12-04T10:05:23.3822921Z #10 at::native::linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T10:05:23.3823527Z #11 at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.3825922Z #12 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&, at::Tensor&, at::Tensor&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&, at::Tensor&, at::Tensor&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.3826277Z #13 at::_ops::linalg_eig_out::call(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T10:05:23.3826444Z #14 at::native::linalg_eig(at::Tensor const&) from ??:0 2025-12-04T10:05:23.3828486Z #15 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA__linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.3828787Z #16 at::_ops::linalg_eig::redispatch(c10::DispatchKeySet, at::Tensor const&) from ??:0 2025-12-04T10:05:23.3829352Z #17 torch::autograd::VariableType::(anonymous namespace)::linalg_eig(c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T10:05:23.3831606Z #18 c10::impl::wrap_kernel_functor_unboxed_ (c10::DispatchKeySet, at::Tensor const&), &torch::autograd::VariableType::(anonymous namespace)::linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (c10::DispatchKeySet, at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T10:05:23.3831808Z #19 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T10:05:23.3832217Z #20 torch::autograd::THPVariable_linalg_eig(_object*, _object*, _object*) from python_linalg_functions.cpp:0 2025-12-04T10:05:23.3832563Z #21 cfunction_call from /usr/local/src/conda/python-3.10.14/Objects/methodobject.c:543 2025-12-04T10:05:23.3832871Z #22 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.3833284Z #23 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.3833678Z #24 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3833943Z #25 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3834320Z #26 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3834744Z #27 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3835122Z #28 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3835522Z #29 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T10:05:23.3835787Z #30 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3836160Z #31 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3836583Z #32 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3836955Z #33 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3837228Z #34 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3837597Z #35 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3838060Z #36 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3838444Z #37 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3838705Z #38 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3839076Z #39 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3839495Z #40 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3839868Z #41 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3840285Z #42 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3840653Z #43 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3841063Z #44 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3841444Z #45 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3841852Z #46 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3842231Z #47 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3842521Z #48 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T10:05:23.3842777Z #49 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3843160Z #50 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3843510Z #51 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.3843827Z #52 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.3844125Z #53 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.3844431Z #54 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.3844852Z #55 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.3845221Z #56 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3845627Z #57 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3846006Z #58 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3846266Z #59 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3846704Z #60 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3847112Z #61 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3847482Z #62 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3847954Z #63 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3848327Z #64 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3848688Z #65 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.3848995Z #66 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.3849286Z #67 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.3849565Z #68 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T10:05:23.3849820Z #69 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3850265Z #70 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3850675Z #71 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3851044Z #72 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3868688Z #73 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3869078Z #74 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3869359Z #75 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3869737Z #76 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3870148Z #77 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3870558Z #78 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3870979Z #79 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3871368Z #80 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3871630Z #81 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3872000Z #82 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3872421Z #83 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3872792Z #84 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3873240Z #85 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3873615Z #86 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3873972Z #87 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.3874334Z #88 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.3874629Z #89 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.3874932Z #90 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.3875354Z #91 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.3875723Z #92 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3875993Z #93 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3876369Z #94 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3876773Z #95 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3877313Z #96 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3877719Z #97 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3878103Z #98 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3878449Z #99 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.3878761Z #100 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.3879070Z #101 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.3879377Z #102 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.3879880Z #103 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.3880262Z #104 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3880674Z #105 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3881066Z #106 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3881480Z #107 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3881859Z #108 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3882135Z #109 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3882512Z #110 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3882942Z #111 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3883325Z #112 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3883739Z #113 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3884132Z #114 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3884490Z #115 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.3884816Z #116 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.3885114Z #117 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.3885422Z #118 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.3885851Z #119 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.3886233Z #120 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3886662Z #121 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3887043Z #122 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3887455Z #123 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3887848Z #124 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3888262Z #125 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3888639Z #126 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3889069Z #127 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3889499Z #128 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3889808Z #129 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T10:05:23.3890117Z #130 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T10:05:23.3890389Z #131 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T10:05:23.3890684Z #132 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T10:05:23.3891038Z #133 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T10:05:23.3891377Z #134 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T10:05:23.3891669Z #135 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T10:05:23.3892000Z #136 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T10:05:23.3892287Z #137 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T10:05:23.3892486Z #138 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T10:05:23.3892592Z #139 _start from ??:0 2025-12-04T10:05:23.3892730Z #140 from ??:0 2025-12-04T10:05:23.3892737Z 2025-12-04T10:05:23.3892742Z 2025-12-04T10:05:23.3892965Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.3893524Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_linalg_eig_stride_consistency_cuda 2025-12-04T10:05:23.3893530Z 2025-12-04T10:05:23.3893804Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.3894030Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3895182Z /var/lib/jenkins/workspace/test/inductor/test_torchinductor.py:5994: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T10:05:23.3895334Z eigenvals, eigenvecs = torch.linalg.eig(x) 2025-12-04T10:05:23.3895603Z _______________ GPUTests.test_linalg_eig_stride_consistency_cuda _______________ 2025-12-04T10:05:23.3895726Z Traceback (most recent call last): 2025-12-04T10:05:23.3896472Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 5999, in test_linalg_eig_stride_consistency 2025-12-04T10:05:23.3896592Z self.common( 2025-12-04T10:05:23.3896937Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T10:05:23.3897072Z return func(*args, **kwds) 2025-12-04T10:05:23.3897509Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T10:05:23.3897617Z check_model( 2025-12-04T10:05:23.3898033Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 497, in check_model 2025-12-04T10:05:23.3898192Z correct = ref_model(*ref_inputs, **ref_kwargs) 2025-12-04T10:05:23.3898558Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 5994, in fn 2025-12-04T10:05:23.3898722Z eigenvals, eigenvecs = torch.linalg.eig(x) 2025-12-04T10:05:23.3899486Z RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T10:05:23.3900256Z Exception raised from apply_magma_eig at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebra.cpp:2024 (most recent call first): 2025-12-04T10:05:23.3900367Z C++ CapturedTraceback: 2025-12-04T10:05:23.3901695Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T10:05:23.3902316Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T10:05:23.3902653Z #6 c10::detail::torchCheckFail(char const*, char const*, unsigned int, char const*) from ??:0 2025-12-04T10:05:23.3903477Z #7 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool)::{lambda()#1}::operator()() const::{lambda()#1}::operator()() const [clone .constprop.0] from BatchLinearAlgebra.cpp:0 2025-12-04T10:05:23.3903882Z #8 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T10:05:23.3904334Z #9 at::native::linalg_eig_out_info(at::Tensor const&, at::Tensor&, at::Tensor&, at::Tensor&, bool) from BatchLinearAlgebra.cpp:0 2025-12-04T10:05:23.3904688Z #10 at::native::linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T10:05:23.3905305Z #11 at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.3907725Z #12 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&, at::Tensor&, at::Tensor&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&, at::Tensor&, at::Tensor&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.3908012Z #13 at::_ops::linalg_eig_out::call(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T10:05:23.3908193Z #14 at::native::linalg_eig(at::Tensor const&) from ??:0 2025-12-04T10:05:23.3910468Z #15 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA__linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.3910780Z #16 at::_ops::linalg_eig::redispatch(c10::DispatchKeySet, at::Tensor const&) from ??:0 2025-12-04T10:05:23.3911325Z #17 torch::autograd::VariableType::(anonymous namespace)::linalg_eig(c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T10:05:23.3913633Z #18 c10::impl::wrap_kernel_functor_unboxed_ (c10::DispatchKeySet, at::Tensor const&), &torch::autograd::VariableType::(anonymous namespace)::linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (c10::DispatchKeySet, at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T10:05:23.3913815Z #19 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T10:05:23.3914219Z #20 torch::autograd::THPVariable_linalg_eig(_object*, _object*, _object*) from python_linalg_functions.cpp:0 2025-12-04T10:05:23.3914558Z #21 cfunction_call from /usr/local/src/conda/python-3.10.14/Objects/methodobject.c:543 2025-12-04T10:05:23.3914863Z #22 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.3915293Z #23 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.3915669Z #24 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3916035Z #25 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3916426Z #26 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3916834Z #27 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3917217Z #28 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3917512Z #29 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T10:05:23.3917770Z #30 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3918219Z #31 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3918686Z #32 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3919061Z #33 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3919330Z #34 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3919699Z #35 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3920128Z #36 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3920500Z #37 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3920754Z #38 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3921131Z #39 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3921542Z #40 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3921925Z #41 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3922329Z #42 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3922696Z #43 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3923112Z #44 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3923483Z #45 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3923901Z #46 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3924271Z #47 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3924567Z #48 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T10:05:23.3924839Z #49 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3925212Z #50 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3925564Z #51 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.3925880Z #52 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.3926174Z #53 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.3926486Z #54 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.3926894Z #55 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.3927263Z #56 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3927685Z #57 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3928132Z #58 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3928405Z #59 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3928773Z #60 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3929180Z #61 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3929559Z #62 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3929963Z #63 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3930346Z #64 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3930752Z #65 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.3931060Z #66 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.3931365Z #67 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.3931632Z #68 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T10:05:23.3931888Z #69 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3932316Z #70 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3932724Z #71 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3933109Z #72 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3933518Z #73 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3933890Z #74 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3934170Z #75 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3934542Z #76 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3934962Z #77 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3935329Z #78 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3935734Z #79 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3936117Z #80 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3936379Z #81 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3936756Z #82 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3937240Z #83 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3937613Z #84 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3938031Z #85 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3938398Z #86 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3938756Z #87 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.3939059Z #88 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.3939350Z #89 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.3939695Z #90 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.3940164Z #91 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.3940548Z #92 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3940806Z #93 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3941177Z #94 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3941595Z #95 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3941962Z #96 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3942377Z #97 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3942804Z #98 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3943155Z #99 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.3943485Z #100 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.3943784Z #101 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.3944090Z #102 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.3944520Z #103 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.3944901Z #104 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3945325Z #105 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3945704Z #106 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3946120Z #107 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3946513Z #108 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3946779Z #109 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3947172Z #110 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3947583Z #111 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3947962Z #112 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3948389Z #113 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3948767Z #114 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3949134Z #115 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.3949450Z #116 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.3949748Z #117 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.3950064Z #118 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.3950475Z #119 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.3950852Z #120 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3951273Z #121 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3951646Z #122 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3952128Z #123 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3952570Z #124 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3952982Z #125 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3953371Z #126 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3953783Z #127 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3954171Z #128 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3954461Z #129 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T10:05:23.3954769Z #130 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T10:05:23.3955106Z #131 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T10:05:23.3955394Z #132 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T10:05:23.3955757Z #133 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T10:05:23.3956082Z #134 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T10:05:23.3956371Z #135 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T10:05:23.3956655Z #136 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T10:05:23.3956922Z #137 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T10:05:23.3957119Z #138 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T10:05:23.3957232Z #139 _start from ??:0 2025-12-04T10:05:23.3957359Z #140 from ??:0 2025-12-04T10:05:23.3957366Z 2025-12-04T10:05:23.3957371Z 2025-12-04T10:05:23.3957604Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.3958146Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_linalg_eig_stride_consistency_cuda 2025-12-04T10:05:23.3958152Z 2025-12-04T10:05:23.3958416Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.3958657Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.3959782Z /var/lib/jenkins/workspace/test/inductor/test_torchinductor.py:5994: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T10:05:23.3959981Z eigenvals, eigenvecs = torch.linalg.eig(x) 2025-12-04T10:05:23.3960125Z =================================== FAILURES =================================== 2025-12-04T10:05:23.3960379Z _______________ GPUTests.test_linalg_eig_stride_consistency_cuda _______________ 2025-12-04T10:05:23.3960521Z Traceback (most recent call last): 2025-12-04T10:05:23.3961080Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 5999, in test_linalg_eig_stride_consistency 2025-12-04T10:05:23.3961176Z self.common( 2025-12-04T10:05:23.3961468Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T10:05:23.3961580Z return func(*args, **kwds) 2025-12-04T10:05:23.3962022Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T10:05:23.3962154Z check_model( 2025-12-04T10:05:23.3962548Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 497, in check_model 2025-12-04T10:05:23.3962712Z correct = ref_model(*ref_inputs, **ref_kwargs) 2025-12-04T10:05:23.3963072Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 5994, in fn 2025-12-04T10:05:23.3963221Z eigenvals, eigenvecs = torch.linalg.eig(x) 2025-12-04T10:05:23.3964050Z RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T10:05:23.3964823Z Exception raised from apply_magma_eig at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebra.cpp:2024 (most recent call first): 2025-12-04T10:05:23.3964944Z C++ CapturedTraceback: 2025-12-04T10:05:23.3966257Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T10:05:23.3966752Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T10:05:23.3967159Z #6 c10::detail::torchCheckFail(char const*, char const*, unsigned int, char const*) from ??:0 2025-12-04T10:05:23.3967970Z #7 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool)::{lambda()#1}::operator()() const::{lambda()#1}::operator()() const [clone .constprop.0] from BatchLinearAlgebra.cpp:0 2025-12-04T10:05:23.3968383Z #8 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T10:05:23.3968820Z #9 at::native::linalg_eig_out_info(at::Tensor const&, at::Tensor&, at::Tensor&, at::Tensor&, bool) from BatchLinearAlgebra.cpp:0 2025-12-04T10:05:23.3969103Z #10 at::native::linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T10:05:23.3969710Z #11 at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.3972106Z #12 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&, at::Tensor&, at::Tensor&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&, at::Tensor&, at::Tensor&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.3972385Z #13 at::_ops::linalg_eig_out::call(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T10:05:23.3972551Z #14 at::native::linalg_eig(at::Tensor const&) from ??:0 2025-12-04T10:05:23.3974567Z #15 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA__linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.3974861Z #16 at::_ops::linalg_eig::redispatch(c10::DispatchKeySet, at::Tensor const&) from ??:0 2025-12-04T10:05:23.3975423Z #17 torch::autograd::VariableType::(anonymous namespace)::linalg_eig(c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T10:05:23.3977786Z #18 c10::impl::wrap_kernel_functor_unboxed_ (c10::DispatchKeySet, at::Tensor const&), &torch::autograd::VariableType::(anonymous namespace)::linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (c10::DispatchKeySet, at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T10:05:23.3978050Z #19 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T10:05:23.3978456Z #20 torch::autograd::THPVariable_linalg_eig(_object*, _object*, _object*) from python_linalg_functions.cpp:0 2025-12-04T10:05:23.3978786Z #21 cfunction_call from /usr/local/src/conda/python-3.10.14/Objects/methodobject.c:543 2025-12-04T10:05:23.3979088Z #22 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.3979492Z #23 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.3979879Z #24 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3980136Z #25 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3980569Z #26 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3980987Z #27 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3981360Z #28 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3981663Z #29 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T10:05:23.3981920Z #30 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3982287Z #31 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3982702Z #32 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3983073Z #33 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3983342Z #34 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3983715Z #35 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3984120Z #36 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3984497Z #37 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3984752Z #38 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3985137Z #39 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3985539Z #40 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3985903Z #41 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3986318Z #42 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3986690Z #43 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3987103Z #44 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3987471Z #45 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3987870Z #46 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3988245Z #47 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3988535Z #48 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T10:05:23.3988785Z #49 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3989163Z #50 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3989514Z #51 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.3989824Z #52 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.3990172Z #53 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.3990472Z #54 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.3990881Z #55 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.3991252Z #56 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3991665Z #57 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3992032Z #58 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3992414Z #59 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3992791Z #60 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3993199Z #61 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3993565Z #62 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3993979Z #63 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3994349Z #64 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3994707Z #65 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.3995008Z #66 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.3995298Z #67 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.3995573Z #68 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T10:05:23.3995832Z #69 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3996377Z #70 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3996780Z #71 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3997149Z #72 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3997560Z #73 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3997928Z #74 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3998193Z #75 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.3998562Z #76 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3998965Z #77 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.3999384Z #78 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.3999783Z #79 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4000150Z #80 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4000414Z #81 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4000778Z #82 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4001192Z #83 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4001559Z #84 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4001957Z #85 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4003324Z #86 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4003679Z #87 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4003997Z #88 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4004286Z #89 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4004585Z #90 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4005001Z #91 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4005366Z #92 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4005715Z #93 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4006088Z #94 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4006489Z #95 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4006869Z #96 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4007271Z #97 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4007633Z #98 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4007991Z #99 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4008301Z #100 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4008609Z #101 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4008919Z #102 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4009326Z #103 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4009714Z #104 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4010128Z #105 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4010515Z #106 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4010931Z #107 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4011307Z #108 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4011581Z #109 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4011957Z #110 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4012380Z #111 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4012755Z #112 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4013164Z #113 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4013546Z #114 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4013901Z #115 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4014210Z #116 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4014515Z #117 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4014817Z #118 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4015298Z #119 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4015673Z #120 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4016083Z #121 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4016464Z #122 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4016936Z #123 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4017323Z #124 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4017815Z #125 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4018190Z #126 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4018611Z #127 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4018984Z #128 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4019283Z #129 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T10:05:23.4019589Z #130 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T10:05:23.4019856Z #131 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T10:05:23.4020152Z #132 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T10:05:23.4020503Z #133 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T10:05:23.4020836Z #134 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T10:05:23.4021141Z #135 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T10:05:23.4021409Z #136 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T10:05:23.4021686Z #137 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T10:05:23.4021879Z #138 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T10:05:23.4021980Z #139 _start from ??:0 2025-12-04T10:05:23.4022109Z #140 from ??:0 2025-12-04T10:05:23.4022116Z 2025-12-04T10:05:23.4022121Z 2025-12-04T10:05:23.4022338Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.4022887Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_linalg_eig_stride_consistency_cuda 2025-12-04T10:05:23.4022896Z 2025-12-04T10:05:23.4023160Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.4023379Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4024516Z /var/lib/jenkins/workspace/test/inductor/test_torchinductor.py:5994: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T10:05:23.4024663Z eigenvals, eigenvecs = torch.linalg.eig(x) 2025-12-04T10:05:23.4025448Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-e7bcb2cd86408aa6.xml - 2025-12-04T10:05:23.4025617Z =========================== short test summary info ============================ 2025-12-04T10:05:23.4026852Z FAILED [0.0050s] inductor/test_torchinductor.py::GPUTests::test_linalg_eig_stride_consistency_cuda - RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T10:05:23.4027677Z Exception raised from apply_magma_eig at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebra.cpp:2024 (most recent call first): 2025-12-04T10:05:23.4027789Z C++ CapturedTraceback: 2025-12-04T10:05:23.4029110Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T10:05:23.4029598Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T10:05:23.4029942Z #6 c10::detail::torchCheckFail(char const*, char const*, unsigned int, char const*) from ??:0 2025-12-04T10:05:23.4030812Z #7 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool)::{lambda()#1}::operator()() const::{lambda()#1}::operator()() const [clone .constprop.0] from BatchLinearAlgebra.cpp:0 2025-12-04T10:05:23.4031213Z #8 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T10:05:23.4031659Z #9 at::native::linalg_eig_out_info(at::Tensor const&, at::Tensor&, at::Tensor&, at::Tensor&, bool) from BatchLinearAlgebra.cpp:0 2025-12-04T10:05:23.4031931Z #10 at::native::linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T10:05:23.4032552Z #11 at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.4034945Z #12 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&, at::Tensor&, at::Tensor&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&, at::Tensor&, at::Tensor&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.4035238Z #13 at::_ops::linalg_eig_out::call(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T10:05:23.4035399Z #14 at::native::linalg_eig(at::Tensor const&) from ??:0 2025-12-04T10:05:23.4037777Z #15 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA__linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.4038092Z #16 at::_ops::linalg_eig::redispatch(c10::DispatchKeySet, at::Tensor const&) from ??:0 2025-12-04T10:05:23.4038638Z #17 torch::autograd::VariableType::(anonymous namespace)::linalg_eig(c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T10:05:23.4040912Z #18 c10::impl::wrap_kernel_functor_unboxed_ (c10::DispatchKeySet, at::Tensor const&), &torch::autograd::VariableType::(anonymous namespace)::linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (c10::DispatchKeySet, at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T10:05:23.4041092Z #19 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T10:05:23.4041583Z #20 torch::autograd::THPVariable_linalg_eig(_object*, _object*, _object*) from python_linalg_functions.cpp:0 2025-12-04T10:05:23.4041905Z #21 cfunction_call from /usr/local/src/conda/python-3.10.14/Objects/methodobject.c:543 2025-12-04T10:05:23.4042208Z #22 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4042623Z #23 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4042996Z #24 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4043267Z #25 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4043634Z #26 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4044100Z #27 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4044481Z #28 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4044773Z #29 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T10:05:23.4045046Z #30 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4045415Z #31 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4045816Z #32 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4046195Z #33 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4046447Z #34 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4046826Z #35 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4047235Z #36 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4047605Z #37 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4047868Z #38 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4048233Z #39 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4048636Z #40 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4049014Z #41 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4049421Z #42 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4049793Z #43 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4050198Z #44 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4050572Z #45 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4050978Z #46 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4051344Z #47 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4051638Z #48 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T10:05:23.4051893Z #49 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4052260Z #50 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4052610Z #51 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4052917Z #52 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4053218Z #53 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4053578Z #54 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4053983Z #55 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4054361Z #56 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4054764Z #57 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4055127Z #58 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4055393Z #59 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4055758Z #60 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4056226Z #61 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4056599Z #62 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4057072Z #63 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4057453Z #64 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4057794Z #65 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4058109Z #66 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4058403Z #67 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4058665Z #68 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T10:05:23.4058938Z #69 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4059307Z #70 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4059716Z #71 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4060096Z #72 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4060497Z #73 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4060874Z #74 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4061128Z #75 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4061492Z #76 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4061907Z #77 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4062271Z #78 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4062687Z #79 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4063053Z #80 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4063303Z #81 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4063684Z #82 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4064084Z #83 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4064461Z #84 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4064856Z #85 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4065229Z #86 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4065665Z #87 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4065968Z #88 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4066260Z #89 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4066569Z #90 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4066974Z #91 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4067350Z #92 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4067608Z #93 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4068033Z #94 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4068446Z #95 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4068813Z #96 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4069225Z #97 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4069592Z #98 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4069941Z #99 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4070257Z #100 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4070556Z #101 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4070875Z #102 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4071291Z #103 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4071685Z #104 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4072107Z #105 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4072482Z #106 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4072897Z #107 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4073274Z #108 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4073536Z #109 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4073926Z #110 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4074337Z #111 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4074744Z #112 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4075158Z #113 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4075535Z #114 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4075898Z #115 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4076207Z #116 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4076502Z #117 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4076811Z #118 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4077228Z #119 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4077667Z #120 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4078080Z #121 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4078456Z #122 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4078871Z #123 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4079246Z #124 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4079657Z #125 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4080097Z #126 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4080506Z #127 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4080895Z #128 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4081181Z #129 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T10:05:23.4081487Z #130 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T10:05:23.4081759Z #131 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T10:05:23.4082044Z #132 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T10:05:23.4082399Z #133 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T10:05:23.4082721Z #134 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T10:05:23.4083016Z #135 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T10:05:23.4083287Z #136 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T10:05:23.4083555Z #137 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T10:05:23.4083752Z #138 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T10:05:23.4083858Z #139 _start from ??:0 2025-12-04T10:05:23.4083978Z #140 from ??:0 2025-12-04T10:05:23.4083985Z 2025-12-04T10:05:23.4083991Z 2025-12-04T10:05:23.4084212Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.4084751Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_linalg_eig_stride_consistency_cuda 2025-12-04T10:05:23.4084757Z 2025-12-04T10:05:23.4085022Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.4085208Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T10:05:23.4085410Z ================= 1 failed, 517 deselected, 2 rerun in 16.29s ================== 2025-12-04T10:05:23.4085523Z Got exit code 1 2025-12-04T10:05:23.4085636Z Retrying single test... 2025-12-04T10:05:23.4086090Z W1204 09:59:20.915000 19380 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:05:23.4086698Z Test results will be stored in test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-54ad0f704656a459.xml 2025-12-04T10:05:23.4086862Z ============================= test session starts ============================== 2025-12-04T10:05:23.4087227Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T10:05:23.4087337Z cachedir: .pytest_cache 2025-12-04T10:05:23.4087857Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T10:05:23.4087999Z rootdir: /var/lib/jenkins/workspace 2025-12-04T10:05:23.4088107Z configfile: pytest.ini 2025-12-04T10:05:23.4088713Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T10:05:23.4088958Z collecting ... collected 981 items / 517 deselected / 464 selected 2025-12-04T10:05:23.4089590Z stepcurrent: skipping 284 already run items. Running only test/inductor/test_torchinductor.py::GPUTests::test_linalg_eig_stride_consistency_cuda 2025-12-04T10:05:23.4089718Z Running 1 items in this shard 2025-12-04T10:05:23.4089724Z 2025-12-04T10:05:23.4090661Z inductor/test_torchinductor.py::GPUTests::test_linalg_eig_stride_consistency_cuda [W1204 09:59:22.478944774 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T10:05:23.4090667Z 2025-12-04T10:05:23.4091185Z [W1204 09:59:38.420845626 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T10:05:23.4091257Z 2025-12-04T10:05:23.4091390Z ('RERUN', {'yellow': True}) [16.1608s] [100%] 2025-12-04T10:05:23.4092326Z inductor/test_torchinductor.py::GPUTests::test_linalg_eig_stride_consistency_cuda [W1204 09:59:38.644373398 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T10:05:23.4092331Z 2025-12-04T10:05:23.4092857Z [W1204 09:59:38.645225291 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T10:05:23.4092862Z 2025-12-04T10:05:23.4093053Z ('RERUN', {'yellow': True}) [0.0054s] [100%] 2025-12-04T10:05:23.4093997Z inductor/test_torchinductor.py::GPUTests::test_linalg_eig_stride_consistency_cuda [W1204 09:59:38.650962217 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T10:05:23.4094006Z 2025-12-04T10:05:23.4094525Z [W1204 09:59:38.651734546 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T10:05:23.4094530Z 2025-12-04T10:05:23.4094651Z FAILED [0.0049s] [100%] 2025-12-04T10:05:23.4094656Z 2025-12-04T10:05:23.4094802Z ==================================== RERUNS ==================================== 2025-12-04T10:05:23.4095060Z _______________ GPUTests.test_linalg_eig_stride_consistency_cuda _______________ 2025-12-04T10:05:23.4095197Z Traceback (most recent call last): 2025-12-04T10:05:23.4095759Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 5999, in test_linalg_eig_stride_consistency 2025-12-04T10:05:23.4095860Z self.common( 2025-12-04T10:05:23.4096334Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T10:05:23.4096454Z return func(*args, **kwds) 2025-12-04T10:05:23.4096961Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T10:05:23.4097069Z check_model( 2025-12-04T10:05:23.4097471Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 497, in check_model 2025-12-04T10:05:23.4097644Z correct = ref_model(*ref_inputs, **ref_kwargs) 2025-12-04T10:05:23.4098002Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 5994, in fn 2025-12-04T10:05:23.4098150Z eigenvals, eigenvecs = torch.linalg.eig(x) 2025-12-04T10:05:23.4098926Z RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T10:05:23.4099692Z Exception raised from apply_magma_eig at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebra.cpp:2024 (most recent call first): 2025-12-04T10:05:23.4099819Z C++ CapturedTraceback: 2025-12-04T10:05:23.4101244Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T10:05:23.4101739Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T10:05:23.4102089Z #6 c10::detail::torchCheckFail(char const*, char const*, unsigned int, char const*) from ??:0 2025-12-04T10:05:23.4102894Z #7 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool)::{lambda()#1}::operator()() const::{lambda()#1}::operator()() const [clone .constprop.0] from BatchLinearAlgebra.cpp:0 2025-12-04T10:05:23.4103312Z #8 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T10:05:23.4103825Z #9 at::native::linalg_eig_out_info(at::Tensor const&, at::Tensor&, at::Tensor&, at::Tensor&, bool) from BatchLinearAlgebra.cpp:0 2025-12-04T10:05:23.4104111Z #10 at::native::linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T10:05:23.4104726Z #11 at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.4107106Z #12 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&, at::Tensor&, at::Tensor&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&, at::Tensor&, at::Tensor&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.4107407Z #13 at::_ops::linalg_eig_out::call(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T10:05:23.4107575Z #14 at::native::linalg_eig(at::Tensor const&) from ??:0 2025-12-04T10:05:23.4109607Z #15 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA__linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.4109901Z #16 at::_ops::linalg_eig::redispatch(c10::DispatchKeySet, at::Tensor const&) from ??:0 2025-12-04T10:05:23.4110460Z #17 torch::autograd::VariableType::(anonymous namespace)::linalg_eig(c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T10:05:23.4112738Z #18 c10::impl::wrap_kernel_functor_unboxed_ (c10::DispatchKeySet, at::Tensor const&), &torch::autograd::VariableType::(anonymous namespace)::linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (c10::DispatchKeySet, at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T10:05:23.4112930Z #19 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T10:05:23.4113338Z #20 torch::autograd::THPVariable_linalg_eig(_object*, _object*, _object*) from python_linalg_functions.cpp:0 2025-12-04T10:05:23.4113660Z #21 cfunction_call from /usr/local/src/conda/python-3.10.14/Objects/methodobject.c:543 2025-12-04T10:05:23.4113978Z #22 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4114392Z #23 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4114853Z #24 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4115115Z #25 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4115488Z #26 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4115906Z #27 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4116277Z #28 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4116585Z #29 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T10:05:23.4116848Z #30 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4117274Z #31 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4117696Z #32 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4118070Z #33 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4118341Z #34 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4118711Z #35 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4119120Z #36 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4119503Z #37 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4119760Z #38 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4120130Z #39 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4120554Z #40 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4120928Z #41 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4121343Z #42 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4121713Z #43 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4122118Z #44 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4122499Z #45 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4122906Z #46 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4123290Z #47 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4123589Z #48 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T10:05:23.4123851Z #49 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4124235Z #50 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4124583Z #51 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4124889Z #52 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4125195Z #53 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4125501Z #54 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4125920Z #55 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4126296Z #56 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4126703Z #57 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4127143Z #58 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4127405Z #59 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4127789Z #60 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4128195Z #61 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4128565Z #62 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4128986Z #63 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4129410Z #64 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4129772Z #65 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4130081Z #66 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4130375Z #67 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4130654Z #68 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T10:05:23.4130910Z #69 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4131304Z #70 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4131721Z #71 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4132091Z #72 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4132515Z #73 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4132889Z #74 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4133150Z #75 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4133529Z #76 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4133933Z #77 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4134312Z #78 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4134719Z #79 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4135088Z #80 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4135361Z #81 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4135733Z #82 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4136154Z #83 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4136522Z #84 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4136994Z #85 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4137378Z #86 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4137725Z #87 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4138030Z #88 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4138335Z #89 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4138637Z #90 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4139120Z #91 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4139491Z #92 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4139746Z #93 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4140128Z #94 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4140531Z #95 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4140914Z #96 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4141318Z #97 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4141742Z #98 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4142109Z #99 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4142418Z #100 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4142727Z #101 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4143037Z #102 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4143453Z #103 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4143842Z #104 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4144257Z #105 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4144637Z #106 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4145066Z #107 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4145446Z #108 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4145723Z #109 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4146100Z #110 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4146510Z #111 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4146896Z #112 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4147307Z #113 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4147700Z #114 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4148059Z #115 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4148367Z #116 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4148671Z #117 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4148979Z #118 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4149404Z #119 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4149780Z #120 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4150192Z #121 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4150580Z #122 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4150997Z #123 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4151427Z #124 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4151846Z #125 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4152220Z #126 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4152640Z #127 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4153019Z #128 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4153314Z #129 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T10:05:23.4153702Z #130 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T10:05:23.4153969Z #131 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T10:05:23.4154271Z #132 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T10:05:23.4154622Z #133 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T10:05:23.4154944Z #134 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T10:05:23.4155240Z #135 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T10:05:23.4155509Z #136 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T10:05:23.4155773Z #137 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T10:05:23.4155976Z #138 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T10:05:23.4156081Z #139 _start from ??:0 2025-12-04T10:05:23.4156212Z #140 from ??:0 2025-12-04T10:05:23.4156218Z 2025-12-04T10:05:23.4156223Z 2025-12-04T10:05:23.4156440Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.4156987Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_linalg_eig_stride_consistency_cuda 2025-12-04T10:05:23.4156993Z 2025-12-04T10:05:23.4157270Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.4157492Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4158627Z /var/lib/jenkins/workspace/test/inductor/test_torchinductor.py:5994: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T10:05:23.4158772Z eigenvals, eigenvecs = torch.linalg.eig(x) 2025-12-04T10:05:23.4159027Z _______________ GPUTests.test_linalg_eig_stride_consistency_cuda _______________ 2025-12-04T10:05:23.4159165Z Traceback (most recent call last): 2025-12-04T10:05:23.4159706Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 5999, in test_linalg_eig_stride_consistency 2025-12-04T10:05:23.4159818Z self.common( 2025-12-04T10:05:23.4160093Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T10:05:23.4160204Z return func(*args, **kwds) 2025-12-04T10:05:23.4160639Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T10:05:23.4160732Z check_model( 2025-12-04T10:05:23.4161128Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 497, in check_model 2025-12-04T10:05:23.4161291Z correct = ref_model(*ref_inputs, **ref_kwargs) 2025-12-04T10:05:23.4161647Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 5994, in fn 2025-12-04T10:05:23.4161807Z eigenvals, eigenvecs = torch.linalg.eig(x) 2025-12-04T10:05:23.4162564Z RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T10:05:23.4163379Z Exception raised from apply_magma_eig at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebra.cpp:2024 (most recent call first): 2025-12-04T10:05:23.4163502Z C++ CapturedTraceback: 2025-12-04T10:05:23.4164823Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T10:05:23.4165318Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T10:05:23.4165713Z #6 c10::detail::torchCheckFail(char const*, char const*, unsigned int, char const*) from ??:0 2025-12-04T10:05:23.4166523Z #7 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool)::{lambda()#1}::operator()() const::{lambda()#1}::operator()() const [clone .constprop.0] from BatchLinearAlgebra.cpp:0 2025-12-04T10:05:23.4166939Z #8 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T10:05:23.4167374Z #9 at::native::linalg_eig_out_info(at::Tensor const&, at::Tensor&, at::Tensor&, at::Tensor&, bool) from BatchLinearAlgebra.cpp:0 2025-12-04T10:05:23.4167656Z #10 at::native::linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T10:05:23.4168264Z #11 at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.4170664Z #12 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&, at::Tensor&, at::Tensor&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&, at::Tensor&, at::Tensor&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.4170943Z #13 at::_ops::linalg_eig_out::call(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T10:05:23.4171108Z #14 at::native::linalg_eig(at::Tensor const&) from ??:0 2025-12-04T10:05:23.4173137Z #15 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA__linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.4173430Z #16 at::_ops::linalg_eig::redispatch(c10::DispatchKeySet, at::Tensor const&) from ??:0 2025-12-04T10:05:23.4173983Z #17 torch::autograd::VariableType::(anonymous namespace)::linalg_eig(c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T10:05:23.4176242Z #18 c10::impl::wrap_kernel_functor_unboxed_ (c10::DispatchKeySet, at::Tensor const&), &torch::autograd::VariableType::(anonymous namespace)::linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (c10::DispatchKeySet, at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T10:05:23.4176437Z #19 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T10:05:23.4176969Z #20 torch::autograd::THPVariable_linalg_eig(_object*, _object*, _object*) from python_linalg_functions.cpp:0 2025-12-04T10:05:23.4177311Z #21 cfunction_call from /usr/local/src/conda/python-3.10.14/Objects/methodobject.c:543 2025-12-04T10:05:23.4177615Z #22 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4178030Z #23 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4178419Z #24 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4178680Z #25 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4179065Z #26 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4179531Z #27 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4179906Z #28 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4180209Z #29 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T10:05:23.4180467Z #30 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4180836Z #31 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4181252Z #32 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4181619Z #33 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4181884Z #34 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4182257Z #35 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4182665Z #36 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4183042Z #37 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4183298Z #38 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4183675Z #39 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4184080Z #40 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4184449Z #41 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4184867Z #42 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4185237Z #43 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4185649Z #44 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4186024Z #45 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4186426Z #46 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4186806Z #47 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4187099Z #48 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T10:05:23.4187355Z #49 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4187735Z #50 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4188086Z #51 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4188407Z #52 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4188832Z #53 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4189135Z #54 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4189553Z #55 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4189922Z #56 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4190341Z #57 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4190706Z #58 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4190962Z #59 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4191408Z #60 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4191816Z #61 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4192195Z #62 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4192600Z #63 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4192968Z #64 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4193328Z #65 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4193627Z #66 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4193919Z #67 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4194200Z #68 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T10:05:23.4194454Z #69 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4194839Z #70 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4195244Z #71 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4195610Z #72 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4196178Z #73 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4196549Z #74 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4196818Z #75 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4197185Z #76 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4197595Z #77 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4197980Z #78 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4198384Z #79 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4198765Z #80 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4199021Z #81 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4199390Z #82 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4199804Z #83 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4200172Z #84 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4200578Z #85 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4201050Z #86 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4201402Z #87 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4201716Z #88 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4202008Z #89 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4202309Z #90 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4202725Z #91 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4203097Z #92 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4203453Z #93 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4203821Z #94 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4204222Z #95 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4204595Z #96 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4204991Z #97 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4205352Z #98 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4205706Z #99 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4206020Z #100 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4206335Z #101 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4206643Z #102 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4207064Z #103 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4207455Z #104 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4207871Z #105 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4208267Z #106 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4208680Z #107 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4209091Z #108 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4209381Z #109 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4209759Z #110 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4210196Z #111 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4210575Z #112 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4210984Z #113 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4211373Z #114 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4211731Z #115 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4212047Z #116 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4212359Z #117 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4212669Z #118 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4213146Z #119 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4213529Z #120 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4213940Z #121 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4214330Z #122 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4214741Z #123 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4215130Z #124 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4215543Z #125 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4215974Z #126 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4216404Z #127 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4216781Z #128 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4217142Z #129 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T10:05:23.4217454Z #130 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T10:05:23.4217725Z #131 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T10:05:23.4218019Z #132 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T10:05:23.4218373Z #133 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T10:05:23.4218700Z #134 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T10:05:23.4219005Z #135 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T10:05:23.4219280Z #136 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T10:05:23.4219557Z #137 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T10:05:23.4219753Z #138 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T10:05:23.4219855Z #139 _start from ??:0 2025-12-04T10:05:23.4219989Z #140 from ??:0 2025-12-04T10:05:23.4219995Z 2025-12-04T10:05:23.4220000Z 2025-12-04T10:05:23.4220222Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.4220773Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_linalg_eig_stride_consistency_cuda 2025-12-04T10:05:23.4220779Z 2025-12-04T10:05:23.4221050Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.4221272Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4222414Z /var/lib/jenkins/workspace/test/inductor/test_torchinductor.py:5994: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T10:05:23.4222562Z eigenvals, eigenvecs = torch.linalg.eig(x) 2025-12-04T10:05:23.4222714Z =================================== FAILURES =================================== 2025-12-04T10:05:23.4222963Z _______________ GPUTests.test_linalg_eig_stride_consistency_cuda _______________ 2025-12-04T10:05:23.4223087Z Traceback (most recent call last): 2025-12-04T10:05:23.4223639Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 5999, in test_linalg_eig_stride_consistency 2025-12-04T10:05:23.4223735Z self.common( 2025-12-04T10:05:23.4224018Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T10:05:23.4224139Z return func(*args, **kwds) 2025-12-04T10:05:23.4224630Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T10:05:23.4224742Z check_model( 2025-12-04T10:05:23.4225137Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 497, in check_model 2025-12-04T10:05:23.4225287Z correct = ref_model(*ref_inputs, **ref_kwargs) 2025-12-04T10:05:23.4225658Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 5994, in fn 2025-12-04T10:05:23.4225803Z eigenvals, eigenvecs = torch.linalg.eig(x) 2025-12-04T10:05:23.4226565Z RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T10:05:23.4227334Z Exception raised from apply_magma_eig at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebra.cpp:2024 (most recent call first): 2025-12-04T10:05:23.4227503Z C++ CapturedTraceback: 2025-12-04T10:05:23.4228830Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T10:05:23.4229319Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T10:05:23.4229661Z #6 c10::detail::torchCheckFail(char const*, char const*, unsigned int, char const*) from ??:0 2025-12-04T10:05:23.4230471Z #7 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool)::{lambda()#1}::operator()() const::{lambda()#1}::operator()() const [clone .constprop.0] from BatchLinearAlgebra.cpp:0 2025-12-04T10:05:23.4230873Z #8 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T10:05:23.4231321Z #9 at::native::linalg_eig_out_info(at::Tensor const&, at::Tensor&, at::Tensor&, at::Tensor&, bool) from BatchLinearAlgebra.cpp:0 2025-12-04T10:05:23.4231597Z #10 at::native::linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T10:05:23.4232215Z #11 at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.4234614Z #12 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&, at::Tensor&, at::Tensor&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&, at::Tensor&, at::Tensor&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.4234909Z #13 at::_ops::linalg_eig_out::call(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T10:05:23.4235074Z #14 at::native::linalg_eig(at::Tensor const&) from ??:0 2025-12-04T10:05:23.4237103Z #15 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA__linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.4237399Z #16 at::_ops::linalg_eig::redispatch(c10::DispatchKeySet, at::Tensor const&) from ??:0 2025-12-04T10:05:23.4237948Z #17 torch::autograd::VariableType::(anonymous namespace)::linalg_eig(c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T10:05:23.4240279Z #18 c10::impl::wrap_kernel_functor_unboxed_ (c10::DispatchKeySet, at::Tensor const&), &torch::autograd::VariableType::(anonymous namespace)::linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (c10::DispatchKeySet, at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T10:05:23.4240459Z #19 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T10:05:23.4240879Z #20 torch::autograd::THPVariable_linalg_eig(_object*, _object*, _object*) from python_linalg_functions.cpp:0 2025-12-04T10:05:23.4241252Z #21 cfunction_call from /usr/local/src/conda/python-3.10.14/Objects/methodobject.c:543 2025-12-04T10:05:23.4241557Z #22 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4241979Z #23 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4242350Z #24 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4242623Z #25 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4242994Z #26 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4243399Z #27 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4243781Z #28 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4244078Z #29 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T10:05:23.4244343Z #30 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4244718Z #31 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4245120Z #32 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4245502Z #33 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4245761Z #34 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4246139Z #35 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4246543Z #36 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4246909Z #37 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4247182Z #38 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4247554Z #39 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4247957Z #40 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4248335Z #41 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4248740Z #42 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4249116Z #43 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4249521Z #44 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4249888Z #45 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4250304Z #46 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4250743Z #47 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4251046Z #48 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T10:05:23.4251301Z #49 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4251670Z #50 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4252025Z #51 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4252329Z #52 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4252629Z #53 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4252928Z #54 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4253391Z #55 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4253772Z #56 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4254177Z #57 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4254545Z #58 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4254811Z #59 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4255200Z #60 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4255618Z #61 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4255982Z #62 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4256392Z #63 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4256775Z #64 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4257201Z #65 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4257517Z #66 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4257808Z #67 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4258075Z #68 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T10:05:23.4258348Z #69 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4258717Z #70 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4259139Z #71 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4259504Z #72 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4259914Z #73 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4260294Z #74 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4260553Z #75 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4260921Z #76 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4261337Z #77 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4261705Z #78 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4262123Z #79 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4262489Z #80 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4262808Z #81 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4263187Z #82 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4263588Z #83 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4263967Z #84 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4264371Z #85 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4264738Z #86 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4265097Z #87 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4265456Z #88 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4265760Z #89 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4266058Z #90 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4266459Z #91 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4266842Z #92 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4267100Z #93 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4267463Z #94 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4267880Z #95 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4268251Z #96 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4268663Z #97 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4269031Z #98 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4269377Z #99 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4269696Z #100 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4269994Z #101 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4270305Z #102 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4270719Z #103 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4271101Z #104 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4271523Z #105 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4271908Z #106 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4272322Z #107 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4272707Z #108 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4272973Z #109 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4273358Z #110 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4273769Z #111 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4274149Z #112 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4274571Z #113 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4275003Z #114 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4275368Z #115 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4275678Z #116 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4275976Z #117 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4276289Z #118 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4276699Z #119 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4277088Z #120 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4277555Z #121 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4277936Z #122 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4278358Z #123 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4278741Z #124 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4279165Z #125 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4279543Z #126 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4279956Z #127 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4280354Z #128 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4280650Z #129 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T10:05:23.4280962Z #130 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T10:05:23.4281244Z #131 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T10:05:23.4281529Z #132 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T10:05:23.4281895Z #133 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T10:05:23.4282221Z #134 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T10:05:23.4282511Z #135 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T10:05:23.4282798Z #136 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T10:05:23.4283069Z #137 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T10:05:23.4283280Z #138 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T10:05:23.4283382Z #139 _start from ??:0 2025-12-04T10:05:23.4283508Z #140 from ??:0 2025-12-04T10:05:23.4283514Z 2025-12-04T10:05:23.4283519Z 2025-12-04T10:05:23.4283751Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.4284295Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_linalg_eig_stride_consistency_cuda 2025-12-04T10:05:23.4284301Z 2025-12-04T10:05:23.4284569Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.4284805Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4285928Z /var/lib/jenkins/workspace/test/inductor/test_torchinductor.py:5994: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T10:05:23.4286092Z eigenvals, eigenvecs = torch.linalg.eig(x) 2025-12-04T10:05:23.4286923Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-54ad0f704656a459.xml - 2025-12-04T10:05:23.4287099Z =========================== short test summary info ============================ 2025-12-04T10:05:23.4288365Z FAILED [0.0049s] inductor/test_torchinductor.py::GPUTests::test_linalg_eig_stride_consistency_cuda - RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T10:05:23.4289132Z Exception raised from apply_magma_eig at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebra.cpp:2024 (most recent call first): 2025-12-04T10:05:23.4289257Z C++ CapturedTraceback: 2025-12-04T10:05:23.4290639Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T10:05:23.4291138Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T10:05:23.4291476Z #6 c10::detail::torchCheckFail(char const*, char const*, unsigned int, char const*) from ??:0 2025-12-04T10:05:23.4292284Z #7 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool)::{lambda()#1}::operator()() const::{lambda()#1}::operator()() const [clone .constprop.0] from BatchLinearAlgebra.cpp:0 2025-12-04T10:05:23.4292698Z #8 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T10:05:23.4293138Z #9 at::native::linalg_eig_out_info(at::Tensor const&, at::Tensor&, at::Tensor&, at::Tensor&, bool) from BatchLinearAlgebra.cpp:0 2025-12-04T10:05:23.4293430Z #10 at::native::linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T10:05:23.4294038Z #11 at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.4296601Z #12 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&, at::Tensor&, at::Tensor&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&, at::Tensor&, at::Tensor&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.4296967Z #13 at::_ops::linalg_eig_out::call(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T10:05:23.4297141Z #14 at::native::linalg_eig(at::Tensor const&) from ??:0 2025-12-04T10:05:23.4299177Z #15 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA__linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from RegisterCUDA_0.cpp:0 2025-12-04T10:05:23.4299469Z #16 at::_ops::linalg_eig::redispatch(c10::DispatchKeySet, at::Tensor const&) from ??:0 2025-12-04T10:05:23.4300027Z #17 torch::autograd::VariableType::(anonymous namespace)::linalg_eig(c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T10:05:23.4302427Z #18 c10::impl::wrap_kernel_functor_unboxed_ (c10::DispatchKeySet, at::Tensor const&), &torch::autograd::VariableType::(anonymous namespace)::linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (c10::DispatchKeySet, at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T10:05:23.4302621Z #19 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T10:05:23.4303026Z #20 torch::autograd::THPVariable_linalg_eig(_object*, _object*, _object*) from python_linalg_functions.cpp:0 2025-12-04T10:05:23.4303360Z #21 cfunction_call from /usr/local/src/conda/python-3.10.14/Objects/methodobject.c:543 2025-12-04T10:05:23.4303739Z #22 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4304149Z #23 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4304536Z #24 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4304795Z #25 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4305179Z #26 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4305581Z #27 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4305956Z #28 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4306262Z #29 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T10:05:23.4306520Z #30 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4306896Z #31 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4307314Z #32 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4307683Z #33 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4307950Z #34 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4308318Z #35 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4308722Z #36 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4309105Z #37 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4309362Z #38 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4309747Z #39 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4310153Z #40 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4310527Z #41 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4310946Z #42 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4311314Z #43 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4311729Z #44 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4312096Z #45 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4312497Z #46 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4312883Z #47 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4313174Z #48 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T10:05:23.4313487Z #49 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4313870Z #50 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4314217Z #51 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4314531Z #52 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4314824Z #53 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4315125Z #54 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4315541Z #55 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4315973Z #56 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4316395Z #57 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4316761Z #58 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4317020Z #59 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4317402Z #60 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4317808Z #61 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4318190Z #62 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4318595Z #63 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4318968Z #64 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4319331Z #65 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4319633Z #66 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4319926Z #67 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4320205Z #68 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T10:05:23.4320464Z #69 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4320846Z #70 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4321249Z #71 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4321626Z #72 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4322040Z #73 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4322412Z #74 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4322680Z #75 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4323052Z #76 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4323455Z #77 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4323838Z #78 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4324241Z #79 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4324619Z #80 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4324932Z #81 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4325356Z #82 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4325771Z #83 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4326140Z #84 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4326542Z #85 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4326923Z #86 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4327271Z #87 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4327589Z #88 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4327938Z #89 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4328243Z #90 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4328661Z #91 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4329031Z #92 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4329302Z #93 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4329668Z #94 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4330070Z #95 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4330451Z #96 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4330858Z #97 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4331239Z #98 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4331590Z #99 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4331902Z #100 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4332212Z #101 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4332518Z #102 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4332933Z #103 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4333327Z #104 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4333737Z #105 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4334134Z #106 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4334554Z #107 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4334934Z #108 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4335211Z #109 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T10:05:23.4335590Z #110 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4336011Z #111 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4336390Z #112 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4336801Z #113 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4337270Z #114 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4337701Z #115 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T10:05:23.4338026Z #116 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T10:05:23.4338322Z #117 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T10:05:23.4338627Z #118 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T10:05:23.4339054Z #119 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T10:05:23.4339432Z #120 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4339848Z #121 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4340310Z #122 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4340724Z #123 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4341116Z #124 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4341527Z #125 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4341906Z #126 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4342334Z #127 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T10:05:23.4342713Z #128 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T10:05:23.4343020Z #129 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T10:05:23.4343336Z #130 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T10:05:23.4343609Z #131 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T10:05:23.4343909Z #132 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T10:05:23.4344265Z #133 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T10:05:23.4344592Z #134 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T10:05:23.4344893Z #135 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T10:05:23.4345165Z #136 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T10:05:23.4345444Z #137 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T10:05:23.4345652Z #138 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T10:05:23.4345758Z #139 _start from ??:0 2025-12-04T10:05:23.4345892Z #140 from ??:0 2025-12-04T10:05:23.4345898Z 2025-12-04T10:05:23.4345904Z 2025-12-04T10:05:23.4346126Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.4346679Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_linalg_eig_stride_consistency_cuda 2025-12-04T10:05:23.4346685Z 2025-12-04T10:05:23.4346951Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.4347129Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T10:05:23.4347343Z ================= 1 failed, 517 deselected, 2 rerun in 16.26s ================== 2025-12-04T10:05:23.4347440Z Got exit code 1 2025-12-04T10:05:23.4347905Z FAILED CONSISTENTLY: test/inductor/test_torchinductor.py::GPUTests::test_linalg_eig_stride_consistency_cuda 2025-12-04T10:05:23.4348330Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T10:05:23.4348778Z W1204 09:59:49.039000 19501 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:05:23.4349446Z Test results will be stored in test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-fb514581ec21877e.xml 2025-12-04T10:05:23.4349612Z ============================= test session starts ============================== 2025-12-04T10:05:23.4350018Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T10:05:23.4350136Z cachedir: .pytest_cache 2025-12-04T10:05:23.4350657Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T10:05:23.4350793Z rootdir: /var/lib/jenkins/workspace 2025-12-04T10:05:23.4350900Z configfile: pytest.ini 2025-12-04T10:05:23.4351450Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T10:05:23.4351752Z collecting ... collected 981 items / 285 deselected / 696 selected 2025-12-04T10:05:23.4351904Z stepcurrent: skipping 285 already run items. 2025-12-04T10:05:23.4352019Z Running 233 items in this shard 2025-12-04T10:05:23.4352037Z 2025-12-04T10:05:23.4352806Z inductor/test_torchinductor.py::GPUTests::test_linear1_cuda W1204 09:59:54.916000 19501 site-packages/torch/_inductor/utils.py:1703] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:05:23.4352912Z PASSED [4.6786s] [ 0%] 2025-12-04T10:05:23.4353247Z inductor/test_torchinductor.py::GPUTests::test_linear2_cuda PASSED [1.5250s] [ 0%] 2025-12-04T10:05:23.4353616Z inductor/test_torchinductor.py::GPUTests::test_linear_mixed_dtype_cuda PASSED [0.2941s] [ 1%] 2025-12-04T10:05:23.4353940Z inductor/test_torchinductor.py::GPUTests::test_linspace1_cuda PASSED [0.6378s] [ 1%] 2025-12-04T10:05:23.4354275Z inductor/test_torchinductor.py::GPUTests::test_linspace2_cuda PASSED [0.5731s] [ 2%] 2025-12-04T10:05:23.4354769Z inductor/test_torchinductor.py::GPUTests::test_linspace4_cuda SKIPPED [0.0003s] (requires multiple cuda devices) [ 2%] 2025-12-04T10:05:23.4355153Z inductor/test_torchinductor.py::GPUTests::test_lite_mode_fallback_cuda PASSED [0.1174s] [ 3%] 2025-12-04T10:05:23.4355705Z inductor/test_torchinductor.py::GPUTests::test_lite_regional_compile_flex_attention_cuda ('RERUN', {'yellow': True}) [1.0667s] [ 3%] 2025-12-04T10:05:23.4356247Z inductor/test_torchinductor.py::GPUTests::test_lite_regional_compile_flex_attention_cuda ('RERUN', {'yellow': True}) [0.9849s] [ 3%] 2025-12-04T10:05:23.4356715Z inductor/test_torchinductor.py::GPUTests::test_lite_regional_compile_flex_attention_cuda FAILED [1.0054s] [ 3%] 2025-12-04T10:05:23.4356721Z 2025-12-04T10:05:23.4356863Z ==================================== RERUNS ==================================== 2025-12-04T10:05:23.4357149Z ___________ GPUTests.test_lite_regional_compile_flex_attention_cuda ____________ 2025-12-04T10:05:23.4357280Z Traceback (most recent call last): 2025-12-04T10:05:23.4357682Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T10:05:23.4357805Z return value(self) 2025-12-04T10:05:23.4358379Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in test_lite_regional_compile_flex_attention 2025-12-04T10:05:23.4358544Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:05:23.4359078Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2430, in run_fw_bw_and_get_code 2025-12-04T10:05:23.4359221Z return run_and_get_code(run_with_backward) 2025-12-04T10:05:23.4359712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2409, in run_and_get_code 2025-12-04T10:05:23.4359825Z result = fn(*args, **kwargs) 2025-12-04T10:05:23.4360310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2426, in run_with_backward 2025-12-04T10:05:23.4360424Z result = fn() 2025-12-04T10:05:23.4360879Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in 2025-12-04T10:05:23.4361060Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:05:23.4361542Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 936, in compile_wrapper 2025-12-04T10:05:23.4361782Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-12-04T10:05:23.4362001Z torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:05:23.4362285Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:05:23.4362684Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:05:23.4363246Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:05:23.4363309Z 2025-12-04T10:05:23.4363422Z Developer debug context: 2025-12-04T10:05:23.4363427Z 2025-12-04T10:05:23.4363981Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:05:23.4363986Z 2025-12-04T10:05:23.4364201Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.4364782Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_lite_regional_compile_flex_attention_cuda 2025-12-04T10:05:23.4364787Z 2025-12-04T10:05:23.4365055Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.4365275Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.4367696Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:05:23.4367820Z stats [('calls_captured', 6)] 2025-12-04T10:05:23.4368058Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('not_ok', 2)] 2025-12-04T10:05:23.4368182Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:05:23.4368280Z graph_break [] 2025-12-04T10:05:23.4368512Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4369248Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4369368Z warnings.warn( 2025-12-04T10:05:23.4370087Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4370189Z warnings.warn( 2025-12-04T10:05:23.4371464Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/attention/flex_attention.py:1624: UserWarning: flex_attention called without torch.compile() - this will use an unfused implementation that materializes the full scores matrix instead of generating a fused kernel. 2025-12-04T10:05:23.4371470Z 2025-12-04T10:05:23.4371633Z SOLUTION: Use torch.compile(flex_attention)(...) 2025-12-04T10:05:23.4371638Z 2025-12-04T10:05:23.4371832Z If you want to debug your score_mod/mask_mod, you can set: 2025-12-04T10:05:23.4372130Z torch.nn.attention.flex_attention._FLEX_ATTENTION_DISABLE_COMPILE_DEBUG = True 2025-12-04T10:05:23.4372135Z 2025-12-04T10:05:23.4372694Z This will allow you to use print statements or breakpoints. Note: This doesn't work with the backwards pass and may produce incorrect results. 2025-12-04T10:05:23.4372808Z _warn_once( 2025-12-04T10:05:23.4373139Z ___________ GPUTests.test_lite_regional_compile_flex_attention_cuda ____________ 2025-12-04T10:05:23.4373275Z Traceback (most recent call last): 2025-12-04T10:05:23.4373678Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T10:05:23.4373843Z return value(self) 2025-12-04T10:05:23.4374420Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in test_lite_regional_compile_flex_attention 2025-12-04T10:05:23.4374584Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:05:23.4375096Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2430, in run_fw_bw_and_get_code 2025-12-04T10:05:23.4375251Z return run_and_get_code(run_with_backward) 2025-12-04T10:05:23.4375777Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2409, in run_and_get_code 2025-12-04T10:05:23.4375902Z result = fn(*args, **kwargs) 2025-12-04T10:05:23.4376383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2426, in run_with_backward 2025-12-04T10:05:23.4376479Z result = fn() 2025-12-04T10:05:23.4376951Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in 2025-12-04T10:05:23.4377120Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:05:23.4377600Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 936, in compile_wrapper 2025-12-04T10:05:23.4377852Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-12-04T10:05:23.4378062Z torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:05:23.4378360Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:05:23.4378760Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:05:23.4379322Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:05:23.4379328Z 2025-12-04T10:05:23.4379451Z Developer debug context: 2025-12-04T10:05:23.4379456Z 2025-12-04T10:05:23.4379977Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:05:23.4379982Z 2025-12-04T10:05:23.4380210Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.4380774Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_lite_regional_compile_flex_attention_cuda 2025-12-04T10:05:23.4380780Z 2025-12-04T10:05:23.4381047Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.4381280Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.4383697Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:05:23.4383827Z stats [('calls_captured', 6)] 2025-12-04T10:05:23.4384051Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('not_ok', 2)] 2025-12-04T10:05:23.4384183Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:05:23.4384281Z graph_break [] 2025-12-04T10:05:23.4384496Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4385269Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4385506Z warnings.warn( 2025-12-04T10:05:23.4386232Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4386348Z warnings.warn( 2025-12-04T10:05:23.4387623Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/attention/flex_attention.py:1624: UserWarning: flex_attention called without torch.compile() - this will use an unfused implementation that materializes the full scores matrix instead of generating a fused kernel. 2025-12-04T10:05:23.4387629Z 2025-12-04T10:05:23.4387810Z SOLUTION: Use torch.compile(flex_attention)(...) 2025-12-04T10:05:23.4387815Z 2025-12-04T10:05:23.4388002Z If you want to debug your score_mod/mask_mod, you can set: 2025-12-04T10:05:23.4388399Z torch.nn.attention.flex_attention._FLEX_ATTENTION_DISABLE_COMPILE_DEBUG = True 2025-12-04T10:05:23.4388404Z 2025-12-04T10:05:23.4388971Z This will allow you to use print statements or breakpoints. Note: This doesn't work with the backwards pass and may produce incorrect results. 2025-12-04T10:05:23.4389068Z _warn_once( 2025-12-04T10:05:23.4389298Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.4391695Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:05:23.4391832Z stats [('calls_captured', 6)] 2025-12-04T10:05:23.4392055Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('not_ok', 2)] 2025-12-04T10:05:23.4392179Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:05:23.4392296Z graph_break [] 2025-12-04T10:05:23.4392517Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4393259Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4393360Z warnings.warn( 2025-12-04T10:05:23.4394080Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4394190Z warnings.warn( 2025-12-04T10:05:23.4394336Z =================================== FAILURES =================================== 2025-12-04T10:05:23.4394609Z ___________ GPUTests.test_lite_regional_compile_flex_attention_cuda ____________ 2025-12-04T10:05:23.4394740Z Traceback (most recent call last): 2025-12-04T10:05:23.4395151Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T10:05:23.4395269Z return value(self) 2025-12-04T10:05:23.4395840Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in test_lite_regional_compile_flex_attention 2025-12-04T10:05:23.4396150Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:05:23.4396676Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2430, in run_fw_bw_and_get_code 2025-12-04T10:05:23.4396816Z return run_and_get_code(run_with_backward) 2025-12-04T10:05:23.4397289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2409, in run_and_get_code 2025-12-04T10:05:23.4397415Z result = fn(*args, **kwargs) 2025-12-04T10:05:23.4397898Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2426, in run_with_backward 2025-12-04T10:05:23.4398010Z result = fn() 2025-12-04T10:05:23.4398496Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in 2025-12-04T10:05:23.4398662Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:05:23.4399159Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 936, in compile_wrapper 2025-12-04T10:05:23.4399399Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-12-04T10:05:23.4399607Z torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:05:23.4399903Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:05:23.4400300Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:05:23.4400872Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:05:23.4400953Z 2025-12-04T10:05:23.4401070Z Developer debug context: 2025-12-04T10:05:23.4401075Z 2025-12-04T10:05:23.4401614Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:05:23.4401619Z 2025-12-04T10:05:23.4401838Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.4402411Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_lite_regional_compile_flex_attention_cuda 2025-12-04T10:05:23.4402417Z 2025-12-04T10:05:23.4402700Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.4402917Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.4405343Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:05:23.4405464Z stats [('calls_captured', 6)] 2025-12-04T10:05:23.4405684Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('not_ok', 2)] 2025-12-04T10:05:23.4405817Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:05:23.4405914Z graph_break [] 2025-12-04T10:05:23.4406143Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4406876Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4406982Z warnings.warn( 2025-12-04T10:05:23.4407726Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4407826Z warnings.warn( 2025-12-04T10:05:23.4409119Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/attention/flex_attention.py:1624: UserWarning: flex_attention called without torch.compile() - this will use an unfused implementation that materializes the full scores matrix instead of generating a fused kernel. 2025-12-04T10:05:23.4409125Z 2025-12-04T10:05:23.4409290Z SOLUTION: Use torch.compile(flex_attention)(...) 2025-12-04T10:05:23.4409295Z 2025-12-04T10:05:23.4409477Z If you want to debug your score_mod/mask_mod, you can set: 2025-12-04T10:05:23.4409789Z torch.nn.attention.flex_attention._FLEX_ATTENTION_DISABLE_COMPILE_DEBUG = True 2025-12-04T10:05:23.4409794Z 2025-12-04T10:05:23.4410399Z This will allow you to use print statements or breakpoints. Note: This doesn't work with the backwards pass and may produce incorrect results. 2025-12-04T10:05:23.4410504Z _warn_once( 2025-12-04T10:05:23.4410778Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.4413179Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:05:23.4413306Z stats [('calls_captured', 6)] 2025-12-04T10:05:23.4413529Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('not_ok', 2)] 2025-12-04T10:05:23.4413724Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:05:23.4413819Z graph_break [] 2025-12-04T10:05:23.4414037Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4414781Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4414879Z warnings.warn( 2025-12-04T10:05:23.4415608Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4415705Z warnings.warn( 2025-12-04T10:05:23.4415921Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.4418464Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:05:23.4418588Z stats [('calls_captured', 6)] 2025-12-04T10:05:23.4418822Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('not_ok', 2)] 2025-12-04T10:05:23.4418945Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:05:23.4419040Z graph_break [] 2025-12-04T10:05:23.4419270Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4419994Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4420106Z warnings.warn( 2025-12-04T10:05:23.4420823Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4420923Z warnings.warn( 2025-12-04T10:05:23.4421709Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-fb514581ec21877e.xml - 2025-12-04T10:05:23.4421883Z =========================== short test summary info ============================ 2025-12-04T10:05:23.4422607Z FAILED [1.0054s] inductor/test_torchinductor.py::GPUTests::test_lite_regional_compile_flex_attention_cuda - torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:05:23.4422891Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:05:23.4423290Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:05:23.4423852Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:05:23.4423862Z 2025-12-04T10:05:23.4423976Z Developer debug context: 2025-12-04T10:05:23.4423981Z 2025-12-04T10:05:23.4424582Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:05:23.4424588Z 2025-12-04T10:05:23.4424807Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.4425385Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_lite_regional_compile_flex_attention_cuda 2025-12-04T10:05:23.4425390Z 2025-12-04T10:05:23.4425669Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.4425852Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T10:05:23.4426108Z ======= 1 failed, 6 passed, 1 skipped, 285 deselected, 2 rerun in 10.98s ======= 2025-12-04T10:05:23.4426290Z Got exit code 1 2025-12-04T10:05:23.4426400Z Retrying single test... 2025-12-04T10:05:23.4426866Z W1204 10:00:13.122000 19801 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:05:23.4427462Z Test results will be stored in test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-1a3279675b79d710.xml 2025-12-04T10:05:23.4427628Z ============================= test session starts ============================== 2025-12-04T10:05:23.4427994Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T10:05:23.4428107Z cachedir: .pytest_cache 2025-12-04T10:05:23.4428641Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T10:05:23.4428767Z rootdir: /var/lib/jenkins/workspace 2025-12-04T10:05:23.4428876Z configfile: pytest.ini 2025-12-04T10:05:23.4429435Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T10:05:23.4429671Z collecting ... collected 981 items / 517 deselected / 464 selected 2025-12-04T10:05:23.4430348Z stepcurrent: skipping 292 already run items. Running only test/inductor/test_torchinductor.py::GPUTests::test_lite_regional_compile_flex_attention_cuda 2025-12-04T10:05:23.4430465Z Running 1 items in this shard 2025-12-04T10:05:23.4430471Z 2025-12-04T10:05:23.4431026Z inductor/test_torchinductor.py::GPUTests::test_lite_regional_compile_flex_attention_cuda ('RERUN', {'yellow': True}) [3.7760s] [100%] 2025-12-04T10:05:23.4431592Z inductor/test_torchinductor.py::GPUTests::test_lite_regional_compile_flex_attention_cuda ('RERUN', {'yellow': True}) [0.9814s] [100%] 2025-12-04T10:05:23.4432049Z inductor/test_torchinductor.py::GPUTests::test_lite_regional_compile_flex_attention_cuda FAILED [0.9910s] [100%] 2025-12-04T10:05:23.4432055Z 2025-12-04T10:05:23.4432211Z ==================================== RERUNS ==================================== 2025-12-04T10:05:23.4432486Z ___________ GPUTests.test_lite_regional_compile_flex_attention_cuda ____________ 2025-12-04T10:05:23.4432611Z Traceback (most recent call last): 2025-12-04T10:05:23.4433034Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T10:05:23.4433140Z return value(self) 2025-12-04T10:05:23.4433710Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in test_lite_regional_compile_flex_attention 2025-12-04T10:05:23.4433891Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:05:23.4434409Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2430, in run_fw_bw_and_get_code 2025-12-04T10:05:23.4434565Z return run_and_get_code(run_with_backward) 2025-12-04T10:05:23.4435038Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2409, in run_and_get_code 2025-12-04T10:05:23.4435155Z result = fn(*args, **kwargs) 2025-12-04T10:05:23.4435644Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2426, in run_with_backward 2025-12-04T10:05:23.4435815Z result = fn() 2025-12-04T10:05:23.4436214Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in 2025-12-04T10:05:23.4436387Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:05:23.4436865Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 936, in compile_wrapper 2025-12-04T10:05:23.4437121Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-12-04T10:05:23.4437325Z torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:05:23.4437610Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:05:23.4438071Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:05:23.4438678Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:05:23.4438683Z 2025-12-04T10:05:23.4438816Z Developer debug context: 2025-12-04T10:05:23.4438821Z 2025-12-04T10:05:23.4439341Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:05:23.4439346Z 2025-12-04T10:05:23.4439564Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.4440145Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_lite_regional_compile_flex_attention_cuda 2025-12-04T10:05:23.4440150Z 2025-12-04T10:05:23.4440417Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.4440649Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.4443061Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:05:23.4443191Z stats [('calls_captured', 6)] 2025-12-04T10:05:23.4443415Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('not_ok', 2)] 2025-12-04T10:05:23.4443538Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:05:23.4443648Z graph_break [] 2025-12-04T10:05:23.4443865Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4444596Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4444712Z warnings.warn( 2025-12-04T10:05:23.4445436Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4445578Z warnings.warn( 2025-12-04T10:05:23.4446850Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/attention/flex_attention.py:1624: UserWarning: flex_attention called without torch.compile() - this will use an unfused implementation that materializes the full scores matrix instead of generating a fused kernel. 2025-12-04T10:05:23.4446855Z 2025-12-04T10:05:23.4447031Z SOLUTION: Use torch.compile(flex_attention)(...) 2025-12-04T10:05:23.4447036Z 2025-12-04T10:05:23.4447215Z If you want to debug your score_mod/mask_mod, you can set: 2025-12-04T10:05:23.4447517Z torch.nn.attention.flex_attention._FLEX_ATTENTION_DISABLE_COMPILE_DEBUG = True 2025-12-04T10:05:23.4447527Z 2025-12-04T10:05:23.4448102Z This will allow you to use print statements or breakpoints. Note: This doesn't work with the backwards pass and may produce incorrect results. 2025-12-04T10:05:23.4448253Z _warn_once( 2025-12-04T10:05:23.4448529Z ___________ GPUTests.test_lite_regional_compile_flex_attention_cuda ____________ 2025-12-04T10:05:23.4448662Z Traceback (most recent call last): 2025-12-04T10:05:23.4449068Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T10:05:23.4449183Z return value(self) 2025-12-04T10:05:23.4449757Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in test_lite_regional_compile_flex_attention 2025-12-04T10:05:23.4449922Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:05:23.4450443Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2430, in run_fw_bw_and_get_code 2025-12-04T10:05:23.4450642Z return run_and_get_code(run_with_backward) 2025-12-04T10:05:23.4451128Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2409, in run_and_get_code 2025-12-04T10:05:23.4451244Z result = fn(*args, **kwargs) 2025-12-04T10:05:23.4451719Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2426, in run_with_backward 2025-12-04T10:05:23.4451830Z result = fn() 2025-12-04T10:05:23.4452227Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in 2025-12-04T10:05:23.4452388Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:05:23.4452878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 936, in compile_wrapper 2025-12-04T10:05:23.4453118Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-12-04T10:05:23.4453332Z torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:05:23.4453620Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:05:23.4454018Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:05:23.4454590Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:05:23.4454595Z 2025-12-04T10:05:23.4454708Z Developer debug context: 2025-12-04T10:05:23.4454713Z 2025-12-04T10:05:23.4455242Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:05:23.4455247Z 2025-12-04T10:05:23.4455462Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.4456027Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_lite_regional_compile_flex_attention_cuda 2025-12-04T10:05:23.4456032Z 2025-12-04T10:05:23.4456316Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.4456538Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.4459042Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:05:23.4459164Z stats [('calls_captured', 6)] 2025-12-04T10:05:23.4459402Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('not_ok', 2)] 2025-12-04T10:05:23.4459528Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:05:23.4459631Z graph_break [] 2025-12-04T10:05:23.4459865Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4460661Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4460767Z warnings.warn( 2025-12-04T10:05:23.4461509Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4461612Z warnings.warn( 2025-12-04T10:05:23.4462891Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/attention/flex_attention.py:1624: UserWarning: flex_attention called without torch.compile() - this will use an unfused implementation that materializes the full scores matrix instead of generating a fused kernel. 2025-12-04T10:05:23.4462898Z 2025-12-04T10:05:23.4463062Z SOLUTION: Use torch.compile(flex_attention)(...) 2025-12-04T10:05:23.4463123Z 2025-12-04T10:05:23.4463307Z If you want to debug your score_mod/mask_mod, you can set: 2025-12-04T10:05:23.4463617Z torch.nn.attention.flex_attention._FLEX_ATTENTION_DISABLE_COMPILE_DEBUG = True 2025-12-04T10:05:23.4463622Z 2025-12-04T10:05:23.4464189Z This will allow you to use print statements or breakpoints. Note: This doesn't work with the backwards pass and may produce incorrect results. 2025-12-04T10:05:23.4464297Z _warn_once( 2025-12-04T10:05:23.4464516Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.4466936Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:05:23.4467057Z stats [('calls_captured', 6)] 2025-12-04T10:05:23.4467284Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('not_ok', 2)] 2025-12-04T10:05:23.4467417Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:05:23.4467515Z graph_break [] 2025-12-04T10:05:23.4467730Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4468470Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4468570Z warnings.warn( 2025-12-04T10:05:23.4469303Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4469437Z warnings.warn( 2025-12-04T10:05:23.4469588Z =================================== FAILURES =================================== 2025-12-04T10:05:23.4469870Z ___________ GPUTests.test_lite_regional_compile_flex_attention_cuda ____________ 2025-12-04T10:05:23.4469991Z Traceback (most recent call last): 2025-12-04T10:05:23.4470410Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T10:05:23.4470515Z return value(self) 2025-12-04T10:05:23.4471131Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in test_lite_regional_compile_flex_attention 2025-12-04T10:05:23.4471309Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:05:23.4471817Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2430, in run_fw_bw_and_get_code 2025-12-04T10:05:23.4471959Z return run_and_get_code(run_with_backward) 2025-12-04T10:05:23.4472446Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2409, in run_and_get_code 2025-12-04T10:05:23.4472566Z result = fn(*args, **kwargs) 2025-12-04T10:05:23.4473057Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2426, in run_with_backward 2025-12-04T10:05:23.4473214Z result = fn() 2025-12-04T10:05:23.4473612Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in 2025-12-04T10:05:23.4473787Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:05:23.4474266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 936, in compile_wrapper 2025-12-04T10:05:23.4474506Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-12-04T10:05:23.4474724Z torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:05:23.4475009Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:05:23.4475417Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:05:23.4476036Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:05:23.4476043Z 2025-12-04T10:05:23.4476161Z Developer debug context: 2025-12-04T10:05:23.4476166Z 2025-12-04T10:05:23.4476698Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:05:23.4476703Z 2025-12-04T10:05:23.4476920Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.4477498Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_lite_regional_compile_flex_attention_cuda 2025-12-04T10:05:23.4477503Z 2025-12-04T10:05:23.4477770Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.4477987Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.4480406Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:05:23.4480525Z stats [('calls_captured', 6)] 2025-12-04T10:05:23.4480760Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('not_ok', 2)] 2025-12-04T10:05:23.4480885Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:05:23.4480996Z graph_break [] 2025-12-04T10:05:23.4481212Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4481950Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4482068Z warnings.warn( 2025-12-04T10:05:23.4482794Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4482894Z warnings.warn( 2025-12-04T10:05:23.4484179Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/attention/flex_attention.py:1624: UserWarning: flex_attention called without torch.compile() - this will use an unfused implementation that materializes the full scores matrix instead of generating a fused kernel. 2025-12-04T10:05:23.4484185Z 2025-12-04T10:05:23.4484351Z SOLUTION: Use torch.compile(flex_attention)(...) 2025-12-04T10:05:23.4484356Z 2025-12-04T10:05:23.4484547Z If you want to debug your score_mod/mask_mod, you can set: 2025-12-04T10:05:23.4484844Z torch.nn.attention.flex_attention._FLEX_ATTENTION_DISABLE_COMPILE_DEBUG = True 2025-12-04T10:05:23.4484853Z 2025-12-04T10:05:23.4485426Z This will allow you to use print statements or breakpoints. Note: This doesn't work with the backwards pass and may produce incorrect results. 2025-12-04T10:05:23.4485600Z _warn_once( 2025-12-04T10:05:23.4485821Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.4488243Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:05:23.4488362Z stats [('calls_captured', 6)] 2025-12-04T10:05:23.4488653Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('not_ok', 2)] 2025-12-04T10:05:23.4488775Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:05:23.4488874Z graph_break [] 2025-12-04T10:05:23.4489106Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4489842Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4489957Z warnings.warn( 2025-12-04T10:05:23.4490678Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4490778Z warnings.warn( 2025-12-04T10:05:23.4491009Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.4493406Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:05:23.4493541Z stats [('calls_captured', 6)] 2025-12-04T10:05:23.4493761Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('not_ok', 2)] 2025-12-04T10:05:23.4493881Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:05:23.4493990Z graph_break [] 2025-12-04T10:05:23.4494202Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4494924Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4495040Z warnings.warn( 2025-12-04T10:05:23.4495770Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4495882Z warnings.warn( 2025-12-04T10:05:23.4496840Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-1a3279675b79d710.xml - 2025-12-04T10:05:23.4497071Z =========================== short test summary info ============================ 2025-12-04T10:05:23.4497798Z FAILED [0.9910s] inductor/test_torchinductor.py::GPUTests::test_lite_regional_compile_flex_attention_cuda - torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:05:23.4498089Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:05:23.4498499Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:05:23.4499056Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:05:23.4499062Z 2025-12-04T10:05:23.4499277Z Developer debug context: 2025-12-04T10:05:23.4499283Z 2025-12-04T10:05:23.4499824Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:05:23.4499830Z 2025-12-04T10:05:23.4500049Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.4500636Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_lite_regional_compile_flex_attention_cuda 2025-12-04T10:05:23.4500641Z 2025-12-04T10:05:23.4500912Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.4501111Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T10:05:23.4501389Z ================== 1 failed, 517 deselected, 2 rerun in 5.84s ================== 2025-12-04T10:05:23.4501488Z Got exit code 1 2025-12-04T10:05:23.4501616Z Retrying single test... 2025-12-04T10:05:23.4502066Z W1204 10:00:32.340000 20000 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:05:23.4502664Z Test results will be stored in test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-bfb3d1c8c3468bae.xml 2025-12-04T10:05:23.4502844Z ============================= test session starts ============================== 2025-12-04T10:05:23.4503194Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T10:05:23.4503319Z cachedir: .pytest_cache 2025-12-04T10:05:23.4503848Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T10:05:23.4503973Z rootdir: /var/lib/jenkins/workspace 2025-12-04T10:05:23.4504103Z configfile: pytest.ini 2025-12-04T10:05:23.4504646Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T10:05:23.4504881Z collecting ... collected 981 items / 517 deselected / 464 selected 2025-12-04T10:05:23.4505562Z stepcurrent: skipping 292 already run items. Running only test/inductor/test_torchinductor.py::GPUTests::test_lite_regional_compile_flex_attention_cuda 2025-12-04T10:05:23.4505684Z Running 1 items in this shard 2025-12-04T10:05:23.4505689Z 2025-12-04T10:05:23.4506253Z inductor/test_torchinductor.py::GPUTests::test_lite_regional_compile_flex_attention_cuda ('RERUN', {'yellow': True}) [3.7901s] [100%] 2025-12-04T10:05:23.4506802Z inductor/test_torchinductor.py::GPUTests::test_lite_regional_compile_flex_attention_cuda ('RERUN', {'yellow': True}) [0.9856s] [100%] 2025-12-04T10:05:23.4507259Z inductor/test_torchinductor.py::GPUTests::test_lite_regional_compile_flex_attention_cuda FAILED [0.9801s] [100%] 2025-12-04T10:05:23.4507282Z 2025-12-04T10:05:23.4507425Z ==================================== RERUNS ==================================== 2025-12-04T10:05:23.4507699Z ___________ GPUTests.test_lite_regional_compile_flex_attention_cuda ____________ 2025-12-04T10:05:23.4507841Z Traceback (most recent call last): 2025-12-04T10:05:23.4508244Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T10:05:23.4508351Z return value(self) 2025-12-04T10:05:23.4508931Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in test_lite_regional_compile_flex_attention 2025-12-04T10:05:23.4509098Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:05:23.4509653Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2430, in run_fw_bw_and_get_code 2025-12-04T10:05:23.4509798Z return run_and_get_code(run_with_backward) 2025-12-04T10:05:23.4510272Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2409, in run_and_get_code 2025-12-04T10:05:23.4510404Z result = fn(*args, **kwargs) 2025-12-04T10:05:23.4510951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2426, in run_with_backward 2025-12-04T10:05:23.4511051Z result = fn() 2025-12-04T10:05:23.4511461Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in 2025-12-04T10:05:23.4511624Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:05:23.4512116Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 936, in compile_wrapper 2025-12-04T10:05:23.4512355Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-12-04T10:05:23.4512565Z torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:05:23.4512863Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:05:23.4513322Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:05:23.4513886Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:05:23.4513903Z 2025-12-04T10:05:23.4514019Z Developer debug context: 2025-12-04T10:05:23.4514024Z 2025-12-04T10:05:23.4514541Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:05:23.4514546Z 2025-12-04T10:05:23.4514774Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.4515339Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_lite_regional_compile_flex_attention_cuda 2025-12-04T10:05:23.4515344Z 2025-12-04T10:05:23.4515623Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.4515847Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.4518271Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:05:23.4518402Z stats [('calls_captured', 6)] 2025-12-04T10:05:23.4518627Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('not_ok', 2)] 2025-12-04T10:05:23.4518761Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:05:23.4518861Z graph_break [] 2025-12-04T10:05:23.4519078Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4519828Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4519934Z warnings.warn( 2025-12-04T10:05:23.4520672Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4520772Z warnings.warn( 2025-12-04T10:05:23.4522046Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/attention/flex_attention.py:1624: UserWarning: flex_attention called without torch.compile() - this will use an unfused implementation that materializes the full scores matrix instead of generating a fused kernel. 2025-12-04T10:05:23.4522052Z 2025-12-04T10:05:23.4522226Z SOLUTION: Use torch.compile(flex_attention)(...) 2025-12-04T10:05:23.4522231Z 2025-12-04T10:05:23.4522413Z If you want to debug your score_mod/mask_mod, you can set: 2025-12-04T10:05:23.4522726Z torch.nn.attention.flex_attention._FLEX_ATTENTION_DISABLE_COMPILE_DEBUG = True 2025-12-04T10:05:23.4522731Z 2025-12-04T10:05:23.4523349Z This will allow you to use print statements or breakpoints. Note: This doesn't work with the backwards pass and may produce incorrect results. 2025-12-04T10:05:23.4536098Z _warn_once( 2025-12-04T10:05:23.4536465Z ___________ GPUTests.test_lite_regional_compile_flex_attention_cuda ____________ 2025-12-04T10:05:23.4536609Z Traceback (most recent call last): 2025-12-04T10:05:23.4537121Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T10:05:23.4537247Z return value(self) 2025-12-04T10:05:23.4537822Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in test_lite_regional_compile_flex_attention 2025-12-04T10:05:23.4537993Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:05:23.4538532Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2430, in run_fw_bw_and_get_code 2025-12-04T10:05:23.4538855Z return run_and_get_code(run_with_backward) 2025-12-04T10:05:23.4539337Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2409, in run_and_get_code 2025-12-04T10:05:23.4539463Z result = fn(*args, **kwargs) 2025-12-04T10:05:23.4539941Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2426, in run_with_backward 2025-12-04T10:05:23.4540052Z result = fn() 2025-12-04T10:05:23.4540446Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in 2025-12-04T10:05:23.4540608Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:05:23.4541099Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 936, in compile_wrapper 2025-12-04T10:05:23.4541336Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-12-04T10:05:23.4541559Z torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:05:23.4541842Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:05:23.4542245Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:05:23.4542819Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:05:23.4542828Z 2025-12-04T10:05:23.4542942Z Developer debug context: 2025-12-04T10:05:23.4542948Z 2025-12-04T10:05:23.4543487Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:05:23.4543493Z 2025-12-04T10:05:23.4543711Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.4544282Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_lite_regional_compile_flex_attention_cuda 2025-12-04T10:05:23.4544293Z 2025-12-04T10:05:23.4544572Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.4544799Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.4547214Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:05:23.4547331Z stats [('calls_captured', 6)] 2025-12-04T10:05:23.4547564Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('not_ok', 2)] 2025-12-04T10:05:23.4547690Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:05:23.4547790Z graph_break [] 2025-12-04T10:05:23.4548023Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4548831Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4548937Z warnings.warn( 2025-12-04T10:05:23.4549675Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4549774Z warnings.warn( 2025-12-04T10:05:23.4551068Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/attention/flex_attention.py:1624: UserWarning: flex_attention called without torch.compile() - this will use an unfused implementation that materializes the full scores matrix instead of generating a fused kernel. 2025-12-04T10:05:23.4551134Z 2025-12-04T10:05:23.4551300Z SOLUTION: Use torch.compile(flex_attention)(...) 2025-12-04T10:05:23.4551306Z 2025-12-04T10:05:23.4551486Z If you want to debug your score_mod/mask_mod, you can set: 2025-12-04T10:05:23.4551804Z torch.nn.attention.flex_attention._FLEX_ATTENTION_DISABLE_COMPILE_DEBUG = True 2025-12-04T10:05:23.4551809Z 2025-12-04T10:05:23.4552367Z This will allow you to use print statements or breakpoints. Note: This doesn't work with the backwards pass and may produce incorrect results. 2025-12-04T10:05:23.4552477Z _warn_once( 2025-12-04T10:05:23.4552696Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.4555101Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:05:23.4555238Z stats [('calls_captured', 6)] 2025-12-04T10:05:23.4555461Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('not_ok', 2)] 2025-12-04T10:05:23.4555593Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:05:23.4555692Z graph_break [] 2025-12-04T10:05:23.4555909Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4556653Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4556752Z warnings.warn( 2025-12-04T10:05:23.4557478Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4557581Z warnings.warn( 2025-12-04T10:05:23.4557722Z =================================== FAILURES =================================== 2025-12-04T10:05:23.4558007Z ___________ GPUTests.test_lite_regional_compile_flex_attention_cuda ____________ 2025-12-04T10:05:23.4558130Z Traceback (most recent call last): 2025-12-04T10:05:23.4558531Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T10:05:23.4558647Z return value(self) 2025-12-04T10:05:23.4559221Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in test_lite_regional_compile_flex_attention 2025-12-04T10:05:23.4559399Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:05:23.4559904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2430, in run_fw_bw_and_get_code 2025-12-04T10:05:23.4560046Z return run_and_get_code(run_with_backward) 2025-12-04T10:05:23.4560562Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2409, in run_and_get_code 2025-12-04T10:05:23.4560674Z result = fn(*args, **kwargs) 2025-12-04T10:05:23.4561220Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2426, in run_with_backward 2025-12-04T10:05:23.4561325Z result = fn() 2025-12-04T10:05:23.4561763Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in 2025-12-04T10:05:23.4561936Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:05:23.4562420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 936, in compile_wrapper 2025-12-04T10:05:23.4562658Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-12-04T10:05:23.4562879Z torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:05:23.4563161Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:05:23.4563632Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:05:23.4564195Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:05:23.4564202Z 2025-12-04T10:05:23.4564316Z Developer debug context: 2025-12-04T10:05:23.4564321Z 2025-12-04T10:05:23.4564851Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:05:23.4564857Z 2025-12-04T10:05:23.4565072Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.4565658Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_lite_regional_compile_flex_attention_cuda 2025-12-04T10:05:23.4565664Z 2025-12-04T10:05:23.4565932Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.4566156Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.4568579Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:05:23.4568699Z stats [('calls_captured', 6)] 2025-12-04T10:05:23.4568935Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('not_ok', 2)] 2025-12-04T10:05:23.4569059Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:05:23.4569167Z graph_break [] 2025-12-04T10:05:23.4569385Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4570126Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4570247Z warnings.warn( 2025-12-04T10:05:23.4570966Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4571064Z warnings.warn( 2025-12-04T10:05:23.4572338Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/attention/flex_attention.py:1624: UserWarning: flex_attention called without torch.compile() - this will use an unfused implementation that materializes the full scores matrix instead of generating a fused kernel. 2025-12-04T10:05:23.4572344Z 2025-12-04T10:05:23.4572508Z SOLUTION: Use torch.compile(flex_attention)(...) 2025-12-04T10:05:23.4572514Z 2025-12-04T10:05:23.4572708Z If you want to debug your score_mod/mask_mod, you can set: 2025-12-04T10:05:23.4573014Z torch.nn.attention.flex_attention._FLEX_ATTENTION_DISABLE_COMPILE_DEBUG = True 2025-12-04T10:05:23.4573018Z 2025-12-04T10:05:23.4573639Z This will allow you to use print statements or breakpoints. Note: This doesn't work with the backwards pass and may produce incorrect results. 2025-12-04T10:05:23.4573737Z _warn_once( 2025-12-04T10:05:23.4573955Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.4576365Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:05:23.4576539Z stats [('calls_captured', 6)] 2025-12-04T10:05:23.4576773Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('not_ok', 2)] 2025-12-04T10:05:23.4576984Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:05:23.4577083Z graph_break [] 2025-12-04T10:05:23.4577313Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4578039Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4578138Z warnings.warn( 2025-12-04T10:05:23.4578868Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4578968Z warnings.warn( 2025-12-04T10:05:23.4579196Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:05:23.4581604Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:05:23.4581734Z stats [('calls_captured', 6)] 2025-12-04T10:05:23.4581955Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('not_ok', 2)] 2025-12-04T10:05:23.4582075Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:05:23.4582187Z graph_break [] 2025-12-04T10:05:23.4582404Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:05:23.4583130Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4583246Z warnings.warn( 2025-12-04T10:05:23.4583967Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:05:23.4584078Z warnings.warn( 2025-12-04T10:05:23.4584851Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-bfb3d1c8c3468bae.xml - 2025-12-04T10:05:23.4585026Z =========================== short test summary info ============================ 2025-12-04T10:05:23.4585753Z FAILED [0.9801s] inductor/test_torchinductor.py::GPUTests::test_lite_regional_compile_flex_attention_cuda - torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:05:23.4586040Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:05:23.4586456Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:05:23.4587076Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:05:23.4587082Z 2025-12-04T10:05:23.4587193Z Developer debug context: 2025-12-04T10:05:23.4587199Z 2025-12-04T10:05:23.4587733Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:05:23.4587738Z 2025-12-04T10:05:23.4587956Z To execute this test, run the following from the base repo dir: 2025-12-04T10:05:23.4588652Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor.py GPUTests.test_lite_regional_compile_flex_attention_cuda 2025-12-04T10:05:23.4588658Z 2025-12-04T10:05:23.4589025Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:05:23.4589424Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T10:05:23.4589636Z ================== 1 failed, 517 deselected, 2 rerun in 5.84s ================== 2025-12-04T10:05:23.4589733Z Got exit code 1 2025-12-04T10:05:23.4590245Z FAILED CONSISTENTLY: test/inductor/test_torchinductor.py::GPUTests::test_lite_regional_compile_flex_attention_cuda 2025-12-04T10:05:23.4590661Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T10:05:23.4591109Z W1204 10:00:51.467000 20199 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:05:23.4591715Z Test results will be stored in test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-aa152592c1ebda5c.xml 2025-12-04T10:05:23.4591877Z ============================= test session starts ============================== 2025-12-04T10:05:23.4592243Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T10:05:23.4592355Z cachedir: .pytest_cache 2025-12-04T10:05:23.4592881Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T10:05:23.4593022Z rootdir: /var/lib/jenkins/workspace 2025-12-04T10:05:23.4593130Z configfile: pytest.ini 2025-12-04T10:05:23.4593675Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T10:05:23.4593914Z collecting ... collected 981 items / 293 deselected / 688 selected 2025-12-04T10:05:23.4594063Z stepcurrent: skipping 293 already run items. 2025-12-04T10:05:23.4594191Z Running 225 items in this shard 2025-12-04T10:05:23.4594197Z 2025-12-04T10:05:23.4594664Z inductor/test_torchinductor.py::GPUTests::test_lite_regional_compile_repeated_blocks_cuda PASSED [2.9070s] [ 0%] 2025-12-04T10:05:23.4595126Z inductor/test_torchinductor.py::GPUTests::test_lite_triton_kernel_wrapper_functional_cuda PASSED [0.5548s] [ 0%] 2025-12-04T10:05:23.4595463Z inductor/test_torchinductor.py::GPUTests::test_log_fp64_cuda PASSED [0.6294s] [ 1%] 2025-12-04T10:05:23.4595801Z inductor/test_torchinductor.py::GPUTests::test_log_softmax_cuda PASSED [1.8407s] [ 1%] 2025-12-04T10:05:23.4596432Z inductor/test_torchinductor.py::GPUTests::test_logaddexp_cuda SKIPPED [0.0032s] (Not implemented for CUDA) [ 2%] 2025-12-04T10:05:23.4596776Z inductor/test_torchinductor.py::GPUTests::test_logcumsumexp_cuda PASSED [13.3507s] [ 2%] 2025-12-04T10:05:23.4597162Z inductor/test_torchinductor.py::GPUTests::test_logcumsumexp_zero_dim_cuda PASSED [0.7059s] [ 3%] 2025-12-04T10:05:23.4597621Z inductor/test_torchinductor.py::GPUTests::test_low_memory_max_pool_dilation_2_dim_2_cuda PASSED [2.1680s] [ 3%] 2025-12-04T10:05:23.4598060Z inductor/test_torchinductor.py::GPUTests::test_low_memory_max_pool_dilation_2_dim_3_cuda PASSED [4.5125s] [ 4%] 2025-12-04T10:05:23.4598422Z inductor/test_torchinductor.py::GPUTests::test_masked_scatter_cuda PASSED [1.0670s] [ 4%] 2025-12-04T10:05:23.4599335Z inductor/test_torchinductor.py::GPUTests::test_matmul_layer_norm_cuda W1204 10:01:22.116000 20199 site-packages/torch/_inductor/utils.py:1703] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:05:23.4599442Z PASSED [1.6539s] [ 4%] 2025-12-04T10:05:23.4599771Z inductor/test_torchinductor.py::GPUTests::test_max_min_cuda PASSED [0.5139s] [ 5%] 2025-12-04T10:05:23.4600102Z inductor/test_torchinductor.py::GPUTests::test_max_pool2d3_cuda PASSED [2.4507s] [ 5%] 2025-12-04T10:05:23.4600429Z inductor/test_torchinductor.py::GPUTests::test_max_pool2d5_cuda PASSED [1.6155s] [ 6%] 2025-12-04T10:05:23.4600879Z inductor/test_torchinductor.py::GPUTests::test_max_pool2d_with_indices_backward2_cuda PASSED [4.3880s] [ 6%] 2025-12-04T10:05:23.4601314Z inductor/test_torchinductor.py::GPUTests::test_max_pool2d_with_indices_backward3_cuda PASSED [2.0652s] [ 7%] 2025-12-04T10:05:23.4601761Z inductor/test_torchinductor.py::GPUTests::test_max_pool2d_with_indices_backward4_cuda PASSED [8.9490s] [ 7%] 2025-12-04T10:05:23.4602277Z inductor/test_torchinductor.py::GPUTests::test_max_pool2d_with_indices_backward6_cuda PASSED [0.2636s] [ 8%] 2025-12-04T10:05:23.4602660Z inductor/test_torchinductor.py::GPUTests::test_min_max_reduction_nan_cuda PASSED [0.3871s] [ 8%] 2025-12-04T10:05:23.4603017Z inductor/test_torchinductor.py::GPUTests::test_mix_device_index_cuda PASSED [0.5393s] [ 8%] 2025-12-04T10:05:23.4603422Z inductor/test_torchinductor.py::GPUTests::test_mixed_mm2_cuda SKIPPED [0.0031s] (Requires sm80) [ 9%] 2025-12-04T10:05:23.4603831Z inductor/test_torchinductor.py::GPUTests::test_mixed_mm3_cuda SKIPPED [0.0029s] (Requires sm80) [ 9%] 2025-12-04T10:05:23.4604220Z inductor/test_torchinductor.py::GPUTests::test_mixed_mm_cuda SKIPPED [0.0028s] (Requires sm80) [ 10%] 2025-12-04T10:05:23.4604575Z inductor/test_torchinductor.py::GPUTests::test_mm_mixed_dtype_cuda PASSED [0.1199s] [ 10%] 2025-12-04T10:05:23.4604886Z inductor/test_torchinductor.py::GPUTests::test_mm_views_cuda PASSED [0.1345s] [ 11%] 2025-12-04T10:05:23.4605233Z inductor/test_torchinductor.py::GPUTests::test_mul_index_expr_cuda PASSED [0.2697s] [ 11%] 2025-12-04T10:05:23.4605618Z inductor/test_torchinductor.py::GPUTests::test_mul_softmax_symfloat_cuda PASSED [2.9852s] [ 12%] 2025-12-04T10:05:23.4606138Z inductor/test_torchinductor.py::GPUTests::test_multi_gpu_device_cuda SKIPPED [0.0003s] (requires multiple cuda devices) [ 12%] 2025-12-04T10:05:23.4606725Z inductor/test_torchinductor.py::GPUTests::test_multi_gpu_recompile_on_index_cuda SKIPPED [0.0002s] (requires multiple cuda devices) [ 12%] 2025-12-04T10:05:23.4607068Z inductor/test_torchinductor.py::GPUTests::test_multilayer_var_cuda PASSED [1.3786s] [ 13%] 2025-12-04T10:05:23.4607438Z inductor/test_torchinductor.py::GPUTests::test_multilayer_var_lowp_cuda PASSED [1.2582s] [ 13%] 2025-12-04T10:05:23.4607880Z inductor/test_torchinductor.py::GPUTests::test_mutable_custom_op_fixed_layout2_cuda PASSED [0.4215s] [ 14%] 2025-12-04T10:05:23.4608319Z inductor/test_torchinductor.py::GPUTests::test_mutable_custom_op_fixed_layout_cuda PASSED [0.7328s] [ 14%] 2025-12-04T10:05:23.4608791Z inductor/test_torchinductor.py::GPUTests::test_nan_sort_stable_False_descending_False_cuda PASSED [0.7642s] [ 15%] 2025-12-04T10:05:23.4609239Z inductor/test_torchinductor.py::GPUTests::test_nan_sort_stable_False_descending_True_cuda PASSED [0.7559s] [ 15%] 2025-12-04T10:05:23.4609564Z inductor/test_torchinductor.py::GPUTests::test_nan_to_num_cuda PASSED [0.4424s] [ 16%] 2025-12-04T10:05:23.4609898Z inductor/test_torchinductor.py::GPUTests::test_neg_index_cuda PASSED [1.5705s] [ 16%] 2025-12-04T10:05:23.4610238Z inductor/test_torchinductor.py::GPUTests::test_neg_max_uint8_cuda PASSED [0.2583s] [ 16%] 2025-12-04T10:05:23.4610593Z inductor/test_torchinductor.py::GPUTests::test_nll_loss_backward_cuda PASSED [0.6813s] [ 17%] 2025-12-04T10:05:23.4610957Z inductor/test_torchinductor.py::GPUTests::test_nll_loss_forward_cuda PASSED [0.8967s] [ 17%] 2025-12-04T10:05:23.4611378Z inductor/test_torchinductor.py::GPUTests::test_no_mega_fusion_during_lowering_cuda PASSED [0.8309s] [ 18%] 2025-12-04T10:05:23.4611739Z inductor/test_torchinductor.py::GPUTests::test_no_op_reduction_cuda PASSED [0.6980s] [ 18%] 2025-12-04T10:05:23.4612239Z inductor/test_torchinductor.py::GPUTests::test_no_specization_over_symbolic_value_cuda PASSED [0.7633s] [ 19%] 2025-12-04T10:05:23.4612656Z inductor/test_torchinductor.py::GPUTests::test_nonzero_unbacked_refinement_cuda PASSED [1.0922s] [ 19%] 2025-12-04T10:05:23.4613053Z inductor/test_torchinductor.py::GPUTests::test_norm_constant_overflow_cuda PASSED [0.6619s] [ 20%] 2025-12-04T10:05:23.4613392Z inductor/test_torchinductor.py::GPUTests::test_output_strides_cuda PASSED [0.4353s] [ 20%] 2025-12-04T10:05:23.4613805Z inductor/test_torchinductor.py::GPUTests::test_pattern_matcher_multi_user_cuda PASSED [1.9268s] [ 20%] 2025-12-04T10:05:23.4614198Z inductor/test_torchinductor.py::GPUTests::test_pattern_matcher_unbacked_cuda PASSED [0.3704s] [ 21%] 2025-12-04T10:05:23.4614573Z inductor/test_torchinductor.py::GPUTests::test_permute1_cuda PASSED [0.4186s] [ 21%] 2025-12-04T10:05:23.4614900Z inductor/test_torchinductor.py::GPUTests::test_permute2_cuda PASSED [0.2488s] [ 22%] 2025-12-04T10:05:23.4615273Z inductor/test_torchinductor.py::GPUTests::test_pointwise_bessel_j0_cuda PASSED [0.9207s] [ 22%] 2025-12-04T10:05:23.4615652Z inductor/test_torchinductor.py::GPUTests::test_pointwise_bessel_j1_cuda PASSED [0.3612s] [ 23%] 2025-12-04T10:05:23.4616112Z inductor/test_torchinductor.py::GPUTests::test_pointwise_chebyshev_polynomial_u_cuda PASSED [1.2525s] [ 23%] 2025-12-04T10:05:23.4616548Z inductor/test_torchinductor.py::GPUTests::test_pointwise_chebyshev_polynomial_v_cuda PASSED [1.0140s] [ 24%] 2025-12-04T10:05:23.4617070Z inductor/test_torchinductor.py::GPUTests::test_pointwise_chebyshev_polynomial_w_cuda PASSED [0.9960s] [ 24%] 2025-12-04T10:05:23.4617436Z inductor/test_torchinductor.py::GPUTests::test_pointwise_digamma_cuda PASSED [1.0532s] [ 24%] 2025-12-04T10:05:23.4617784Z inductor/test_torchinductor.py::GPUTests::test_pointwise_entr_cuda PASSED [0.7264s] [ 25%] 2025-12-04T10:05:23.4618139Z inductor/test_torchinductor.py::GPUTests::test_pointwise_erfc_cuda PASSED [0.4365s] [ 25%] 2025-12-04T10:05:23.4618489Z inductor/test_torchinductor.py::GPUTests::test_pointwise_expit_cuda PASSED [0.3272s] [ 26%] 2025-12-04T10:05:23.4618847Z inductor/test_torchinductor.py::GPUTests::test_pointwise_expm1_cuda PASSED [0.4151s] [ 26%] 2025-12-04T10:05:23.4619215Z inductor/test_torchinductor.py::GPUTests::test_pointwise_gammainc_cuda PASSED [0.1227s] [ 27%] 2025-12-04T10:05:23.4619583Z inductor/test_torchinductor.py::GPUTests::test_pointwise_gammaincc_cuda PASSED [0.1218s] [ 27%] 2025-12-04T10:05:23.4620032Z inductor/test_torchinductor.py::GPUTests::test_pointwise_hermite_polynomial_he_cuda PASSED [0.2660s] [ 28%] 2025-12-04T10:05:23.4620365Z inductor/test_torchinductor.py::GPUTests::test_pointwise_i0_cuda PASSED [0.9370s] [ 28%] 2025-12-04T10:05:23.4620713Z inductor/test_torchinductor.py::GPUTests::test_pointwise_i0e_cuda PASSED [0.6474s] [ 28%] 2025-12-04T10:05:23.4621054Z inductor/test_torchinductor.py::GPUTests::test_pointwise_i1e_cuda PASSED [0.2717s] [ 29%] 2025-12-04T10:05:23.4621403Z inductor/test_torchinductor.py::GPUTests::test_pointwise_logit_cuda PASSED [0.5243s] [ 29%] 2025-12-04T10:05:23.4621831Z inductor/test_torchinductor.py::GPUTests::test_pointwise_modified_bessel_k0_cuda PASSED [0.6182s] [ 30%] 2025-12-04T10:05:23.4622173Z inductor/test_torchinductor.py::GPUTests::test_pointwise_ndtr_cuda PASSED [0.4511s] [ 30%] 2025-12-04T10:05:23.4622556Z inductor/test_torchinductor.py::GPUTests::test_pointwise_polygamma_cuda PASSED [1.1334s] [ 31%] 2025-12-04T10:05:23.4623005Z inductor/test_torchinductor.py::GPUTests::test_pointwise_scaled_modified_bessel_k1_cuda PASSED [0.5621s] [ 31%] 2025-12-04T10:05:23.4623485Z inductor/test_torchinductor.py::GPUTests::test_pointwise_shifted_chebyshev_polynomial_t_cuda PASSED [0.4738s] [ 32%] 2025-12-04T10:05:23.4623976Z inductor/test_torchinductor.py::GPUTests::test_pointwise_shifted_chebyshev_polynomial_u_cuda PASSED [1.2809s] [ 32%] 2025-12-04T10:05:23.4624325Z inductor/test_torchinductor.py::GPUTests::test_pointwise_sinc_cuda PASSED [1.0356s] [ 32%] 2025-12-04T10:05:23.4624753Z inductor/test_torchinductor.py::GPUTests::test_pointwise_xlog1py_cuda PASSED [0.5123s] [ 33%] 2025-12-04T10:05:23.4625113Z inductor/test_torchinductor.py::GPUTests::test_pointwise_xlogy_cuda PASSED [0.5157s] [ 33%] 2025-12-04T10:05:23.4625419Z inductor/test_torchinductor.py::GPUTests::test_polar_cuda PASSED [0.2709s] [ 34%] 2025-12-04T10:05:23.4625732Z inductor/test_torchinductor.py::GPUTests::test_pow1_cuda PASSED [1.1344s] [ 34%] 2025-12-04T10:05:23.4626029Z inductor/test_torchinductor.py::GPUTests::test_pow2_cuda PASSED [0.5210s] [ 35%] 2025-12-04T10:05:23.4626326Z inductor/test_torchinductor.py::GPUTests::test_pow3_cuda PASSED [0.2126s] [ 35%] 2025-12-04T10:05:23.4626644Z inductor/test_torchinductor.py::GPUTests::test_pow_int_cuda PASSED [1.6186s] [ 36%] 2025-12-04T10:05:23.4627036Z inductor/test_torchinductor.py::GPUTests::test_pow_symfloat_cuda PASSED [0.4522s] [ 36%] 2025-12-04T10:05:23.4627470Z inductor/test_torchinductor.py::GPUTests::test_prepare_softmax_with_fast_math_cuda PASSED [0.1396s] [ 36%] 2025-12-04T10:05:23.4627880Z inductor/test_torchinductor.py::GPUTests::test_profiler_mark_wrapper_call_cuda PASSED [0.1999s] [ 37%] 2025-12-04T10:05:23.4628191Z inductor/test_torchinductor.py::GPUTests::test_randint_cuda PASSED [0.3556s] [ 37%] 2025-12-04T10:05:23.4628577Z inductor/test_torchinductor.py::GPUTests::test_randint_distribution_cuda PASSED [0.3378s] [ 38%] 2025-12-04T10:05:23.4628933Z inductor/test_torchinductor.py::GPUTests::test_randint_int64_mod_cuda PASSED [0.2918s] [ 38%] 2025-12-04T10:05:23.4629290Z inductor/test_torchinductor.py::GPUTests::test_randn_generator_cuda PASSED [0.5363s] [ 39%] 2025-12-04T10:05:23.4629642Z inductor/test_torchinductor.py::GPUTests::test_randn_like_empty_cuda PASSED [0.2521s] [ 39%] 2025-12-04T10:05:23.4629969Z inductor/test_torchinductor.py::GPUTests::test_reduction1_cuda PASSED [0.5193s] [ 40%] 2025-12-04T10:05:23.4630307Z inductor/test_torchinductor.py::GPUTests::test_reduction2_cuda PASSED [0.4908s] [ 40%] 2025-12-04T10:05:23.4630626Z inductor/test_torchinductor.py::GPUTests::test_reduction3_cuda PASSED [0.4796s] [ 40%] 2025-12-04T10:05:23.4630943Z inductor/test_torchinductor.py::GPUTests::test_reduction5_cuda PASSED [0.4442s] [ 41%] 2025-12-04T10:05:23.4631340Z inductor/test_torchinductor.py::GPUTests::test_reduction_config_limit_cuda PASSED [0.0034s] [ 41%] 2025-12-04T10:05:23.4631740Z inductor/test_torchinductor.py::GPUTests::test_reflection_pad2d_backward_cuda PASSED [7.5233s] [ 42%] 2025-12-04T10:05:23.4632137Z inductor/test_torchinductor.py::GPUTests::test_reinterpret_dtypeview_cuda PASSED [0.3377s] [ 42%] 2025-12-04T10:05:23.4632437Z inductor/test_torchinductor.py::GPUTests::test_relu_cuda PASSED [0.4355s] [ 43%] 2025-12-04T10:05:23.4632772Z inductor/test_torchinductor.py::GPUTests::test_remove_no_ops_cuda PASSED [1.3231s] [ 43%] 2025-12-04T10:05:23.4633146Z inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_cuda PASSED [0.3668s] [ 44%] 2025-12-04T10:05:23.4634941Z inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda SKIPPED [0.0009s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/151378 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 44%] 2025-12-04T10:05:23.4636711Z inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda SKIPPED [0.0008s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/151511 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 44%] 2025-12-04T10:05:23.4638529Z inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda SKIPPED [0.0007s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/151541 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 45%] 2025-12-04T10:05:23.4638917Z inductor/test_torchinductor.py::GPUTests::test_repeat_interleave_2_cuda PASSED [0.2871s] [ 45%] 2025-12-04T10:05:23.4639401Z inductor/test_torchinductor.py::GPUTests::test_repeat_interleave_Tensor_decomp_int32_nd_1_cuda PASSED [0.2875s] [ 46%] 2025-12-04T10:05:23.4639852Z inductor/test_torchinductor.py::GPUTests::test_replication_pad_errors_with_bool_cuda PASSED [1.0157s] [ 46%] 2025-12-04T10:05:23.4640242Z inductor/test_torchinductor.py::GPUTests::test_require_stride_expanded_cuda PASSED [1.5864s] [ 47%] 2025-12-04T10:05:23.4640564Z inductor/test_torchinductor.py::GPUTests::test_resize_as_cuda PASSED [12.4452s] [ 47%] 2025-12-04T10:05:23.4640940Z inductor/test_torchinductor.py::GPUTests::test_resize_cuda PASSED [4.5387s] [ 48%] 2025-12-04T10:05:23.4641348Z inductor/test_torchinductor.py::GPUTests::test_reuse_buffers_with_aliasing_cuda PASSED [1.2905s] [ 48%] 2025-12-04T10:05:23.4641669Z inductor/test_torchinductor.py::GPUTests::test_roll_cuda PASSED [0.7125s] [ 48%] 2025-12-04T10:05:23.4641974Z inductor/test_torchinductor.py::GPUTests::test_rsqrt_cuda PASSED [0.4660s] [ 49%] 2025-12-04T10:05:23.4642311Z inductor/test_torchinductor.py::GPUTests::test_scalar_output_cuda PASSED [1.5252s] [ 49%] 2025-12-04T10:05:23.4642946Z inductor/test_torchinductor.py::GPUTests::test_scaled_dot_product_attention_cuda SKIPPED [0.0034s] (Can't run flash attention on this platform) [ 50%] 2025-12-04T10:05:23.4643261Z inductor/test_torchinductor.py::GPUTests::test_scatter1_cuda PASSED [0.4656s] [ 50%] 2025-12-04T10:05:23.4643676Z inductor/test_torchinductor.py::GPUTests::test_scatter2_cuda SKIPPED [0.0034s] (unstable on sm86) [ 51%] 2025-12-04T10:05:23.4644010Z inductor/test_torchinductor.py::GPUTests::test_scatter3_cuda PASSED [0.4690s] [ 51%] 2025-12-04T10:05:23.4644325Z inductor/test_torchinductor.py::GPUTests::test_scatter4_cuda PASSED [0.7886s] [ 52%] 2025-12-04T10:05:23.4644644Z inductor/test_torchinductor.py::GPUTests::test_scatter5_cuda PASSED [0.9444s] [ 52%] 2025-12-04T10:05:23.4645124Z inductor/test_torchinductor.py::GPUTests::test_scatter_add1_cuda SKIPPED [0.0003s] (Flaky test, needs debugging) [ 52%] 2025-12-04T10:05:23.4645454Z inductor/test_torchinductor.py::GPUTests::test_scatter_add3_cuda PASSED [0.7695s] [ 53%] 2025-12-04T10:05:23.4645815Z inductor/test_torchinductor.py::GPUTests::test_scatter_reduce1_cuda PASSED [0.4736s] [ 53%] 2025-12-04T10:05:23.4646508Z inductor/test_torchinductor.py::GPUTests::test_sdpa_prefer_nd_tiling_True_use_block_ptr_False_cuda SKIPPED [0.0003s] (Does not support SDPA or pre-SM80 hardware) [ 54%] 2025-12-04T10:05:23.4646930Z inductor/test_torchinductor.py::GPUTests::test_sdpa_unaligned_mask_freezing_cuda PASSED [0.6241s] [ 54%] 2025-12-04T10:05:23.4647269Z inductor/test_torchinductor.py::GPUTests::test_searchsorted_cuda PASSED [8.2569s] [ 55%] 2025-12-04T10:05:23.4647614Z inductor/test_torchinductor.py::GPUTests::test_select_scatter_cuda PASSED [0.5207s] [ 55%] 2025-12-04T10:05:23.4647959Z inductor/test_torchinductor.py::GPUTests::test_sgn_extremal_cuda PASSED [0.3514s] [ 56%] 2025-12-04T10:05:23.4648297Z inductor/test_torchinductor.py::GPUTests::test_shape_padding_cuda PASSED [2.6976s] [ 56%] 2025-12-04T10:05:23.4648695Z inductor/test_torchinductor.py::GPUTests::test_should_pad_bench_for_bmm_cuda PASSED [0.0040s] [ 56%] 2025-12-04T10:05:23.4649008Z inductor/test_torchinductor.py::GPUTests::test_sigmoid_cuda PASSED [0.4037s] [ 57%] 2025-12-04T10:05:23.4649320Z inductor/test_torchinductor.py::GPUTests::test_signbit_cuda PASSED [0.4926s] [ 57%] 2025-12-04T10:05:23.4649633Z inductor/test_torchinductor.py::GPUTests::test_silu_cuda PASSED [0.3520s] [ 58%] 2025-12-04T10:05:23.4649937Z inductor/test_torchinductor.py::GPUTests::test_sin_cuda PASSED [0.5726s] [ 58%] 2025-12-04T10:05:23.4650264Z inductor/test_torchinductor.py::GPUTests::test_single_elem_cuda PASSED [0.3045s] [ 59%] 2025-12-04T10:05:23.4650635Z inductor/test_torchinductor.py::GPUTests::test_slice1_cuda PASSED [0.5493s] [ 59%] 2025-12-04T10:05:23.4650941Z inductor/test_torchinductor.py::GPUTests::test_slice3_cuda PASSED [0.5756s] [ 60%] 2025-12-04T10:05:23.4651252Z inductor/test_torchinductor.py::GPUTests::test_slice4_cuda PASSED [0.2599s] [ 60%] 2025-12-04T10:05:23.4651596Z inductor/test_torchinductor.py::GPUTests::test_slice_mutation1_cuda PASSED [0.5590s] [ 60%] 2025-12-04T10:05:23.4651943Z inductor/test_torchinductor.py::GPUTests::test_slice_mutation2_cuda PASSED [0.2976s] [ 61%] 2025-12-04T10:05:23.4652293Z inductor/test_torchinductor.py::GPUTests::test_slice_mutation3_cuda PASSED [0.1869s] [ 61%] 2025-12-04T10:05:23.4652631Z inductor/test_torchinductor.py::GPUTests::test_slice_scatter2_cuda PASSED [0.3655s] [ 62%] 2025-12-04T10:05:23.4653036Z inductor/test_torchinductor.py::GPUTests::test_slice_scatter4_cuda PASSED [0.4177s] [ 62%] 2025-12-04T10:05:23.4653374Z inductor/test_torchinductor.py::GPUTests::test_slice_scatter5_cuda PASSED [0.4903s] [ 63%] 2025-12-04T10:05:23.4653807Z inductor/test_torchinductor.py::GPUTests::test_slice_scatter_dtype_consistency_cuda PASSED [0.6345s] [ 63%] 2025-12-04T10:05:23.4654204Z inductor/test_torchinductor.py::GPUTests::test_slice_scatter_reinplace_cuda PASSED [0.7973s] [ 64%] 2025-12-04T10:05:23.4654577Z inductor/test_torchinductor.py::GPUTests::test_softmax_backward_data_cuda PASSED [0.7851s] [ 64%] 2025-12-04T10:05:23.4654974Z inductor/test_torchinductor.py::GPUTests::test_softmax_one_kernel_loop_cuda PASSED [0.3454s] [ 64%] 2025-12-04T10:05:23.4655295Z inductor/test_torchinductor.py::GPUTests::test_sort_bool_cuda PASSED [15.3624s] [ 65%] 2025-12-04T10:05:23.4655639Z inductor/test_torchinductor.py::GPUTests::test_sort_transpose_cuda PASSED [27.5862s] [ 65%] 2025-12-04T10:05:23.4656016Z inductor/test_torchinductor.py::GPUTests::test_special_polygamma_cuda PASSED [4.7143s] [ 66%] 2025-12-04T10:05:23.4656321Z inductor/test_torchinductor.py::GPUTests::test_split_cuda PASSED [0.6036s] [ 66%] 2025-12-04T10:05:23.4656782Z inductor/test_torchinductor.py::GPUTests::test_split_cumprod_low_prec_cuda SKIPPED [0.0033s] (Requires sm80) [ 67%] 2025-12-04T10:05:23.4657195Z inductor/test_torchinductor.py::GPUTests::test_split_cumsum_cuda PASSED [4.4421s] [ 67%] 2025-12-04T10:05:23.4657556Z inductor/test_torchinductor.py::GPUTests::test_split_cumsum_index_cuda PASSED [1.7508s] [ 68%] 2025-12-04T10:05:23.4658021Z inductor/test_torchinductor.py::GPUTests::test_split_cumsum_low_prec_cuda SKIPPED [0.0033s] (Requires sm80) [ 68%] 2025-12-04T10:05:23.4658436Z inductor/test_torchinductor.py::GPUTests::test_split_reduction_dynamic_shape_cuda PASSED [0.4753s] [ 68%] 2025-12-04T10:05:23.4658798Z inductor/test_torchinductor.py::GPUTests::test_split_with_integer_cuda PASSED [0.8448s] [ 69%] 2025-12-04T10:05:23.4659126Z inductor/test_torchinductor.py::GPUTests::test_squeeze2_cuda PASSED [0.4266s] [ 69%] 2025-12-04T10:05:23.4659472Z inductor/test_torchinductor.py::GPUTests::test_squeeze_varargs_cuda PASSED [0.3287s] [ 70%] 2025-12-04T10:05:23.4659786Z inductor/test_torchinductor.py::GPUTests::test_std_cuda PASSED [1.7158s] [ 70%] 2025-12-04T10:05:23.4660307Z inductor/test_torchinductor.py::GPUTests::test_stride_preservation_with_stride_modifying_fx_pass_cuda PASSED [0.1676s] [ 71%] 2025-12-04T10:05:23.4660607Z inductor/test_torchinductor.py::GPUTests::test_sum1_cuda PASSED [0.5587s] [ 71%] 2025-12-04T10:05:23.4660915Z inductor/test_torchinductor.py::GPUTests::test_sum2_cuda PASSED [2.1141s] [ 72%] 2025-12-04T10:05:23.4661209Z inductor/test_torchinductor.py::GPUTests::test_sum3_cuda PASSED [0.6461s] [ 72%] 2025-12-04T10:05:23.4661506Z inductor/test_torchinductor.py::GPUTests::test_sum5_cuda PASSED [1.2603s] [ 72%] 2025-12-04T10:05:23.4661824Z inductor/test_torchinductor.py::GPUTests::test_tensor2_cuda PASSED [0.3477s] [ 73%] 2025-12-04T10:05:23.4662191Z inductor/test_torchinductor.py::GPUTests::test_tensor_index_slice_cuda PASSED [0.7920s] [ 73%] 2025-12-04T10:05:23.4662733Z inductor/test_torchinductor.py::GPUTests::test_tmp_not_defined_issue1_use_block_ptr_True_cuda PASSED [1.3244s] [ 74%] 2025-12-04T10:05:23.4663108Z inductor/test_torchinductor.py::GPUTests::test_tmp_not_defined_issue2_cuda PASSED [0.7401s] [ 74%] 2025-12-04T10:05:23.4663464Z inductor/test_torchinductor.py::GPUTests::test_to_device_constant_cuda PASSED [0.4391s] [ 75%] 2025-12-04T10:05:23.4663790Z inductor/test_torchinductor.py::GPUTests::test_to_device_cuda PASSED [0.2925s] [ 75%] 2025-12-04T10:05:23.4664101Z inductor/test_torchinductor.py::GPUTests::test_to_dtype_cuda PASSED [0.4161s] [ 76%] 2025-12-04T10:05:23.4664456Z inductor/test_torchinductor.py::GPUTests::test_to_memory_format_cuda PASSED [0.7974s] [ 76%] 2025-12-04T10:05:23.4664836Z inductor/test_torchinductor.py::GPUTests::test_transposed_propagates_cuda PASSED [0.1800s] [ 76%] 2025-12-04T10:05:23.4665385Z inductor/test_torchinductor.py::GPUTests::test_triton_argmin_argmax_transpose_logical_index_cuda PASSED [3.9667s] [ 77%] 2025-12-04T10:05:23.4665698Z inductor/test_torchinductor.py::GPUTests::test_triu_cuda PASSED [0.5829s] [ 77%] 2025-12-04T10:05:23.4666046Z inductor/test_torchinductor.py::GPUTests::test_uint4x2_mixed_mm_cuda PASSED [0.7390s] [ 78%] 2025-12-04T10:05:23.4666459Z inductor/test_torchinductor.py::GPUTests::test_unbacked_floordiv_simplify_cuda PASSED [0.8827s] [ 78%] 2025-12-04T10:05:23.4666896Z inductor/test_torchinductor.py::GPUTests::test_unbacked_floordiv_simplify_errors_cuda PASSED [0.0230s] [ 79%] 2025-12-04T10:05:23.4667204Z inductor/test_torchinductor.py::GPUTests::test_unbind_cuda PASSED [0.2897s] [ 79%] 2025-12-04T10:05:23.4667621Z inductor/test_torchinductor.py::GPUTests::test_unfold_zero_dimension_tensor_cuda PASSED [0.1106s] [ 80%] 2025-12-04T10:05:23.4668014Z inductor/test_torchinductor.py::GPUTests::test_unsigned_constant_tensors_cuda PASSED [0.9558s] [ 80%] 2025-12-04T10:05:23.4668393Z inductor/test_torchinductor.py::GPUTests::test_unspec_inputs_float32_cuda PASSED [0.5033s] [ 80%] 2025-12-04T10:05:23.4668777Z inductor/test_torchinductor.py::GPUTests::test_unspec_inputs_float64_cuda PASSED [0.7093s] [ 81%] 2025-12-04T10:05:23.4669144Z inductor/test_torchinductor.py::GPUTests::test_unspec_inputs_int32_cuda PASSED [0.4488s] [ 81%] 2025-12-04T10:05:23.4669518Z inductor/test_torchinductor.py::GPUTests::test_unspec_inputs_uint8_cuda PASSED [0.4519s] [ 82%] 2025-12-04T10:05:23.4669894Z inductor/test_torchinductor.py::GPUTests::test_upsample_bilinear2d_a_cuda PASSED [2.2250s] [ 82%] 2025-12-04T10:05:23.4670270Z inductor/test_torchinductor.py::GPUTests::test_upsample_bilinear2d_b_cuda PASSED [1.2275s] [ 83%] 2025-12-04T10:05:23.4670640Z inductor/test_torchinductor.py::GPUTests::test_upsample_nearest1d_cuda PASSED [1.3413s] [ 83%] 2025-12-04T10:05:23.4671000Z inductor/test_torchinductor.py::GPUTests::test_upsample_nearest2d_cuda PASSED [1.2938s] [ 84%] 2025-12-04T10:05:23.4671373Z inductor/test_torchinductor.py::GPUTests::test_upsample_nearest3d_cuda PASSED [1.7441s] [ 84%] 2025-12-04T10:05:23.4671711Z inductor/test_torchinductor.py::GPUTests::test_var_correction_cuda PASSED [1.2009s] [ 84%] 2025-12-04T10:05:23.4672125Z inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_False_cuda PASSED [0.7220s] [ 85%] 2025-12-04T10:05:23.4672536Z inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda PASSED [0.7021s] [ 85%] 2025-12-04T10:05:23.4672909Z inductor/test_torchinductor.py::GPUTests::test_vectorized_ops_masked_cuda PASSED [0.5343s] [ 86%] 2025-12-04T10:05:23.4673349Z inductor/test_torchinductor.py::GPUTests::test_vectorized_ops_masked_var_novec_cuda PASSED [0.4902s] [ 86%] 2025-12-04T10:05:23.4673696Z inductor/test_torchinductor.py::GPUTests::test_view_as_complex_cuda PASSED [0.2534s] [ 87%] 2025-12-04T10:05:23.4674023Z inductor/test_torchinductor.py::GPUTests::test_view_detach_cuda PASSED [0.2273s] [ 87%] 2025-12-04T10:05:23.4674379Z inductor/test_torchinductor.py::GPUTests::test_view_on_aliased_cuda PASSED [1.2777s] [ 88%] 2025-12-04T10:05:23.4674902Z inductor/test_torchinductor.py::GPUTests::test_view_uint8_through_differing_bitwidths_cuda PASSED [1.1252s] [ 88%] 2025-12-04T10:05:23.4675267Z inductor/test_torchinductor.py::GPUTests::test_views4_cuda PASSED [1.3849s] [ 88%] 2025-12-04T10:05:23.4675580Z inductor/test_torchinductor.py::GPUTests::test_views5_cuda PASSED [0.2057s] [ 89%] 2025-12-04T10:05:23.4675884Z inductor/test_torchinductor.py::GPUTests::test_views6_cuda PASSED [0.3764s] [ 89%] 2025-12-04T10:05:23.4676192Z inductor/test_torchinductor.py::GPUTests::test_views7_cuda PASSED [0.3966s] [ 90%] 2025-12-04T10:05:23.4676536Z inductor/test_torchinductor.py::GPUTests::test_weight_norm_bwd_cuda PASSED [1.1900s] [ 90%] 2025-12-04T10:05:23.4676909Z inductor/test_torchinductor.py::GPUTests::test_xblock_divides_xnumel_cuda PASSED [1.1076s] [ 91%] 2025-12-04T10:05:23.4677280Z inductor/test_torchinductor.py::GPUTests::test_zero_dim_reductions_cuda PASSED [0.2931s] [ 91%] 2025-12-04T10:05:23.4677706Z inductor/test_torchinductor.py::GPUTests::test_zero_element_mutation_cuda PASSED [0.2234s] [ 92%] 2025-12-04T10:05:23.4678139Z inductor/test_torchinductor.py::TritonCodeGenTests::test_cant_optimize_compute PASSED [0.7508s] [ 92%] 2025-12-04T10:05:23.4678657Z inductor/test_torchinductor.py::TritonCodeGenTests::test_ctr_not_moved_to_cuda_when_used_in_index_put PASSED [0.7763s] [ 92%] 2025-12-04T10:05:23.4679078Z inductor/test_torchinductor.py::TritonCodeGenTests::test_donated_buffer_inplace PASSED [1.0678s] [ 93%] 2025-12-04T10:05:23.4679591Z inductor/test_torchinductor.py::TritonCodeGenTests::test_evict_last_non_coalesced_loads_block_ptr PASSED [3.8914s] [ 93%] 2025-12-04T10:05:23.4680092Z inductor/test_torchinductor.py::TritonCodeGenTests::test_graph_partition_default_device_context PASSED [0.2592s] [ 94%] 2025-12-04T10:05:23.4680542Z inductor/test_torchinductor.py::TritonCodeGenTests::test_grouped_mm SKIPPED [0.0037s] (Requires sm90) [ 94%] 2025-12-04T10:05:23.4681189Z inductor/test_torchinductor.py::TritonCodeGenTests::test_has_constant_mask_block_multiple_False_ynumel_exceed_ygrid_size_False PASSED [0.2160s] [ 95%] 2025-12-04T10:05:23.4681834Z inductor/test_torchinductor.py::TritonCodeGenTests::test_has_constant_mask_block_multiple_True_ynumel_exceed_ygrid_size_False PASSED [0.2008s] [ 95%] 2025-12-04T10:05:23.4682470Z inductor/test_torchinductor.py::TritonCodeGenTests::test_has_constant_mask_block_multiple_True_ynumel_exceed_ygrid_size_True PASSED [0.2003s] [ 96%] 2025-12-04T10:05:23.4682962Z inductor/test_torchinductor.py::TritonCodeGenTests::test_inductor_detach_view_backend_inductor PASSED [0.2257s] [ 96%] 2025-12-04T10:05:23.4683388Z inductor/test_torchinductor.py::TritonCodeGenTests::test_inductor_sequence_nr PASSED [2.1993s] [ 96%] 2025-12-04T10:05:23.4683874Z inductor/test_torchinductor.py::TritonCodeGenTests::test_not_materialize_pointwise_reduction PASSED [0.4829s] [ 97%] 2025-12-04T10:05:23.4684248Z inductor/test_torchinductor.py::TritonCodeGenTests::test_numpy_on_gpu PASSED [0.2915s] [ 97%] 2025-12-04T10:05:23.4684654Z inductor/test_torchinductor.py::TritonCodeGenTests::test_optimize_compute PASSED [0.5357s] [ 98%] 2025-12-04T10:05:23.4685091Z inductor/test_torchinductor.py::TritonCodeGenTests::test_optimize_indexing_assert PASSED [1.3702s] [ 98%] 2025-12-04T10:05:23.4685608Z inductor/test_torchinductor.py::TritonCodeGenTests::test_optimize_indexing_dtype_with_constraint PASSED [0.7327s] [ 99%] 2025-12-04T10:05:23.4685930Z inductor/test_torchinductor.py::RNNTest::test_rnn_compile_safe PASSED [0.4197s] [ 99%] 2025-12-04T10:05:23.4686300Z inductor/test_torchinductor.py::NanCheckerTest::test_nan_checker_pass PASSED [0.3150s] [100%] 2025-12-04T10:05:23.4686306Z 2025-12-04T10:05:23.4687098Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-aa152592c1ebda5c.xml - 2025-12-04T10:05:23.4687326Z ========= 209 passed, 16 skipped, 293 deselected in 266.78s (0:04:26) ========== 2025-12-04T10:05:23.4688756Z The following tests failed consistently: ['test/inductor/test_torchinductor.py::GPUTests::test_isinf_cuda', 'test/inductor/test_torchinductor.py::GPUTests::test_linalg_eig_stride_consistency_cuda', 'test/inductor/test_torchinductor.py::GPUTests::test_lite_regional_compile_flex_attention_cuda'] 2025-12-04T10:05:23.4688764Z 2025-12-04T10:05:23.4689343Z FINISHED PRINTING LOG FILE of inductor/test_torchinductor 2/2 (test/test-reports/inductor.test_torchinductor_2.2_198ec367be5c71a3_.log) 2025-12-04T10:05:23.4689349Z 2025-12-04T10:05:23.4689726Z Finished inductor/test_torchinductor 2/2 ... [2025-12-04 10:05:23.246300][3106.936682766], took 10.58min 2025-12-04T10:05:23.4690550Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-6ce11cbc918a017c.xml 2025-12-04T10:05:23.4691370Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-c850c8e21fb6d04c.xml 2025-12-04T10:05:23.4692282Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-17e608243d035d18.xml 2025-12-04T10:05:23.4693100Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-7d6e32f4c213ad66.xml 2025-12-04T10:05:23.4693932Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-e7bcb2cd86408aa6.xml 2025-12-04T10:05:23.4902241Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-54ad0f704656a459.xml 2025-12-04T10:05:23.5268122Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-fb514581ec21877e.xml 2025-12-04T10:05:23.5608281Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-1a3279675b79d710.xml 2025-12-04T10:05:23.5925964Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-bfb3d1c8c3468bae.xml 2025-12-04T10:05:23.6230868Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-aa152592c1ebda5c.xml 2025-12-04T10:05:23.9185069Z Uploading logs for 57119749282 to S3 2025-12-04T10:05:23.9615131Z Uploading artifacts took 0.30 seconds 2025-12-04T10:05:23.9615561Z inductor/test_torchinductor 2/2 failed! 2025-12-04T10:05:23.9620466Z Running inductor/test_torchinductor_dynamic_shapes 4/5 ... [2025-12-04 10:05:23.961878][3107.652272224] 2025-12-04T10:05:23.9621161Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T10:05:23.9624852Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_dynamic_shapes.py', '--shard-id=4', '--num-shards=5', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 10:05:23.962252] 2025-12-04T10:15:32.9600903Z 2025-12-04T10:15:32.9602168Z PRINTING LOG FILE of inductor/test_torchinductor_dynamic_shapes 4/5 (test/test-reports/inductor.test_torchinductor_dynamic_shapes_4.5_2869e792aa617cfa_.log) 2025-12-04T10:15:32.9603745Z W1204 10:05:33.011000 26102 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:15:32.9605294Z Test results will be stored in test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-535feae09e5a3ebd.xml 2025-12-04T10:15:32.9606473Z ============================= test session starts ============================== 2025-12-04T10:15:32.9607503Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T10:15:32.9608477Z cachedir: .pytest_cache 2025-12-04T10:15:32.9609922Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T10:15:32.9610826Z rootdir: /var/lib/jenkins/workspace 2025-12-04T10:15:32.9611414Z configfile: pytest.ini 2025-12-04T10:15:32.9612554Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T10:15:32.9613856Z collecting ... collected 1815 items 2025-12-04T10:15:32.9614456Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T10:15:32.9837565Z Running 363 items in this shard: test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test__dyn_quant_pack_4bit_weight_fp32_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test__unsafe_masked_index_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_adaptive_avg_pool2d2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_adaptive_avg_pool2d_low_prec_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_adaptive_max_pool2d2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_add_complex7_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_add_complex_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_add_const_float_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_add_inplace_permuted_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_adding_tensor_offsets_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_addmv_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_allow_reuse_disable_if_exceed_peak_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_aoti_eager_support_out_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_arange3_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_arange5_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_argmax_argmin2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_assert_alignment_op_name_fail_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_assert_size_stride_op_name_fail_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_avg_pool2d_backward2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_avg_pool2d_backward3_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_avg_pool2d_backward4_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_batch_norm_2d_2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bitwise2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bucketize_computed_offsets_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bucketize_int_int32_int64_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bucketize_int_int32_uint8_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bucketize_int_int8_int8_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bucketize_int_int8_uint8_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bucketize_int_uint8_int64_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bucketize_int_uint8_uint8_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bucketize_nd_tiling_True_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_builtins_round_float_ndigits_zero_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cat_empty_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cat_empty_index_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cat_unbacked_2d_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_check_stack_no_cycles_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_complex_fallback_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_config_option_dont_assume_alignment_cudagraphs_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_config_option_dont_assume_alignment_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_consecutive_split_cumsum_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_constant_pad_1d_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_constant_pad_2d_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_constant_pad_fill_dtype_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_constant_pad_nd_inplace_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_conv2d_channels_last_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_conv_with_as_strided_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cpu_scalar_with_cpu_tensor_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cpu_scalar_with_gpu_tensor_cpp_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cumprod_zero_dim_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cumsum_no_mask_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_custom_op_fixed_layout_channels_last_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_custom_scan_op_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_custom_scan_op_multi_input_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_data_type_propogation_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_deterministic_codegen_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_device_assert_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_div4_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_div_precision_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_div_presicion_accuracy_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_div_prim_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtype_mismatch_issue_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_float16_float32_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_float16_int16_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_float16_int32_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_float16_uint8_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_float32_int8_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_int16_int16_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_int32_int64_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_int32_uint8_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_int8_int16_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_int8_uint8_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_uint8_float64_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_uint8_int16_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_uint8_int32_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_uint8_int64_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_expand_as_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_expanded_reduction_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_fill1_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_fill2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_float32_to_int32_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_fuse_large_params_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_fuse_tiled_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_gather1_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_graph_partition_arange1_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_graph_partition_misaligned_input_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_index3_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_index_propagation_abs_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_index_propagation_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_index_put1_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_index_put_deterministic_fallback_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_index_put_fallback1_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_inf_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_inner_reduction_detection_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_inplace_add_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_inplace_mixed_dtype_ops_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_large_grid_use_block_ptr_False_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_large_grid_use_block_ptr_True_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_large_offset_pointwise_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_like_rands3_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_lite_triton_kernel_wrapper_functional_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_logcumsumexp_zero_dim_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_long_tensor_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_masked_fill_promotion_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_max_pool2d2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_max_pool2d_with_indices_backward2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_max_pool2d_with_indices_backward6_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_mixed_mm2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_mixed_mm3_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_mm_views_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_mul_softmax_symfloat_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_multilayer_prime_size_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_multilayer_var_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_mutations_loop_fusion_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_needs_contiguous_strides_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_new_empty_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_no_mega_fusion_during_lowering_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_nonzero_unbacked_refinement_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pad_single_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_permute1_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_bessel_y1_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_chebyshev_polynomial_u_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_erf_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_log1p_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_log_ndtr_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_modified_bessel_i1_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_modified_bessel_k0_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_modified_bessel_k1_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_ndtr_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_scaled_modified_bessel_k0_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_scaled_modified_bessel_k1_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_sinc_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_spherical_bessel_j0_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_polar_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pow2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_prepare_softmax_with_fast_math_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_randint_int64_mod_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_remove_no_ops_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_remove_noop_copy_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_repeat_interleave_2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_replication_pad_errors_with_bool_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_resize_as_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_round_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_rsqrt_dynamic_shapes_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_scatter1_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_scatter2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_scatter3_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_sdpa_prefer_nd_tiling_False_use_block_ptr_False_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_sdpa_prefer_nd_tiling_True_use_block_ptr_True_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_sdpa_unaligned_mask_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_sgn_extremal_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_shape_prop_torch_ones_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_sign_dtype_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_simplify_loops_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_single_elem_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_slice_scatter_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_softmax_one_kernel_loop_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_softmax_one_kernel_persist_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_sqrt_dynamic_shapes_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_squeeze2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_squeeze_varargs_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_std_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_sum5_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_sum_dtype_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_tanh_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_transpose_add_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_transposed_propagates_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_triu_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_uint_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_unbacked_floordiv_simplify_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_unsqueeze_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_upsample_bicubic2d_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_vectorized_ops_masked_var_novec_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_views3_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test__dyn_quant_matmul_4bit_fp32_input_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test__dyn_quant_pack_4bit_weight_bf16_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_abs_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_adaptive_avg_pool1d_argmax_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_adaptive_avg_pool_with_output_size_0_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_adaptive_pool_errors_with_long_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_add_complex5_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_add_complex9_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_add_complex_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_add_complex_strided_fallback_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_add_const_int_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_add_inplace_permuted_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_alexnet_prefix_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_aliased_buffer_reuse_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_allow_reuse_active_if_under_peak_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_allow_reuse_disable_if_exceed_peak_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_aoti_eager_support_str_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_aoti_eager_with_persistent_cache_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_argmax_argmin_with_duplicates_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_argmax_to_float_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_assert_alignment_op_name_pass_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_avg_pool2d2_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_avg_pool2d5_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_avg_pool2d_backward2_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_avg_pool3d_backward_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_bitwise3_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_bucketize_int_int16_int32_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_bucketize_int_int16_int8_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_bucketize_int_int32_int64_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_bucketize_int_uint8_int32_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_buffer_batch_norm_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_buffer_copied_in_graph_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_buffer_copied_in_graph_with_different_shapes_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_builtins_round_float_ndigits_neg_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cat_empty_index_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cat_extern_kernel_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cat_uint8_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cat_unbacked_empty_1d_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cauchy_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_check_stack_no_cycles_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_compar_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_config_option_dont_assume_alignment_cudagraphs_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_consecutive_split_cumprod_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_constant_pad_1d_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_constant_pad_float64_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_conv2d_backward_channels_last_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_conv3d_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_conv_functional_bn_fuse_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_conv_inference_heuristics_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_convolution1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cos_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cummin_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cumprod_zero_dim_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cumsum_no_mask_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cumsum_zero_dim_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_custom_op_fixed_layout_channels_last_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_custom_scan_op_multi_input_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dense_mask_index_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_deterministic_codegen_on_graph_break_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_diagonal_copy_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dist_bf16_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_div7_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_div_zero_dim_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dropout_trivial_1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtype_mismatch_issue_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_float32_int32_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_float32_int64_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_float32_uint8_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_float64_float32_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_float64_int16_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_float64_uint8_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_int16_float16_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_int32_float32_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_int64_float16_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_uint8_float16_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_uint8_int32_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_uint8_int64_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_uint8_uint8_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_embedding_bag_byte_unpack_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_embedding_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_erfc_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_erfinv_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_fallback_mutable_op_with_return_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_fmod_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_fmod_zero_dim_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_functionalize_rng_wrappers_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_gather1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_generate_rand_fp8_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_graph_partition_arange1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_graph_partition_pad_dynamic_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_hardswish_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_index_propagation_remainder_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_index_put2_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_index_put_deterministic_fallback_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_indirect_load_broadcast_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_inductor_multiple_specializations_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_inductor_triton_bucketize_respects_masking_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_kernel_names_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_l1_loss_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_large_grid_use_block_ptr_True_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_layer_norm_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_like_channels_last_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_like_rands3_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_like_rands_sliced_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_linear1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_linspace3_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_list_clearing_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_lite_regional_compile_flex_attention_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_lite_regional_compile_repeated_blocks_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_low_memory_max_pool_dilation_1_dim_2_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_masked_scatter_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_max_pool2d1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_max_pool2d6_dilation_1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_max_pool2d_with_indices_backward4_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_multi_device_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_multi_gpu_recompile_on_index_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_nan_sort_stable_True_descending_False_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_new_empty_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_nll_loss_backward_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_norm_constant_overflow_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_output_strides_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pad_view_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_chebyshev_polynomial_t_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_chebyshev_polynomial_w_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_erfcx_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_expit_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_expm1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_gammainc_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_hermite_polynomial_he_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_i0e_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_modified_bessel_k0_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_ndtri_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_round_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_spherical_bessel_j0_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pow2_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_randn_generator_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_reduction4_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_remove_noop_slice_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_remove_noop_view_default_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_repeat_interleave_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_reuse_buffers_with_aliasing_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_rsqrt_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_scaled_dot_product_attention_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_scatter1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_scatter6_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_scatter_add3_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_scatter_bf16_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_sdpa_prefer_nd_tiling_True_use_block_ptr_False_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_shape_padding_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_sigmoid_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_sin_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_single_elem_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_single_elem_indirect_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_sizehint_issue1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_slice1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_slice3_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_slice_mutation3_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_slice_scatter2_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_sort_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_sort_transpose_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_special_polygamma_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_split_cumprod_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_split_cumsum_low_prec_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_split_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_split_with_unbacked_symints_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_strided_inputs_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_sum4_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_tensor_index_slice_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_tmp_not_defined_issue2_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_tmp_not_defined_issue3_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_transposed_propagates_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_triu_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_uint_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_unbacked_floordiv_simplify_errors_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_unspec_inputs_int32_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_unspec_inputs_int64_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_unspec_inputs_int8_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_vdd_clamp_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_vectorized_ops_masked_var_novec_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_constant_fold_uniform_value_dynamic_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_item_zeros_nobreak_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_math_ops_op10_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_math_ops_op1_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_math_ops_op8_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_math_ops_op9_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_sym_stride_lowering_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_sym_sum_unbacked_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_unbacked_cat_backwards_save_data_dependent_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_unbacked_reduction_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_unspecialized_float_dynamic_cuda 2025-12-04T10:15:33.0058066Z 2025-12-04T10:15:33.0058970Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test__dyn_quant_pack_4bit_weight_fp32_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [7.6670s] [ 0%] 2025-12-04T10:15:33.0060766Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test__unsafe_masked_index_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8737s] [ 0%] 2025-12-04T10:15:33.0062495Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_adaptive_avg_pool2d2_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.2033s] [ 0%] 2025-12-04T10:15:33.0064292Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_adaptive_avg_pool2d_low_prec_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [2.0094s] [ 1%] 2025-12-04T10:15:33.0066368Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_adaptive_max_pool2d2_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.1782s] [ 1%] 2025-12-04T10:15:33.0068045Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_add_complex7_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7796s] [ 1%] 2025-12-04T10:15:33.0069691Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_add_complex_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8932s] [ 1%] 2025-12-04T10:15:33.0071343Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_add_const_float_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7552s] [ 2%] 2025-12-04T10:15:33.0073176Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_add_inplace_permuted_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.1451s] [ 2%] 2025-12-04T10:15:33.0074898Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_adding_tensor_offsets_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.2726s] [ 2%] 2025-12-04T10:15:33.0076552Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_addmv_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7326s] [ 3%] 2025-12-04T10:15:33.0078268Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_allow_reuse_disable_if_exceed_peak_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.2134s] [ 3%] 2025-12-04T10:15:33.0080412Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_aoti_eager_support_out_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py W1204 10:05:52.005000 26102 site-packages/torch/_export/__init__.py:71] +============================+ 2025-12-04T10:15:33.0082065Z W1204 10:05:52.005000 26102 site-packages/torch/_export/__init__.py:72] | !!! WARNING !!! | 2025-12-04T10:15:33.0082925Z W1204 10:05:52.005000 26102 site-packages/torch/_export/__init__.py:73] +============================+ 2025-12-04T10:15:33.0084631Z W1204 10:05:52.005000 26102 site-packages/torch/_export/__init__.py:74] torch._export.aot_compile()/torch._export.aot_load() is being deprecated, please switch to directly calling torch._inductor.aoti_compile_and_package(torch.export.export())/torch._inductor.aoti_load_package() instead. 2025-12-04T10:15:33.0086101Z PASSED [6.6021s] [ 3%] 2025-12-04T10:15:33.0087036Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_arange3_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7396s] [ 3%] 2025-12-04T10:15:33.0088647Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_arange5_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.6130s] [ 4%] 2025-12-04T10:15:33.0090273Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_argmax_argmin2_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.1603s] [ 4%] 2025-12-04T10:15:33.0092009Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_assert_alignment_op_name_fail_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.0033s] [ 4%] 2025-12-04T10:15:33.0093809Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_assert_size_stride_op_name_fail_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.0029s] [ 4%] 2025-12-04T10:15:33.0095578Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_avg_pool2d_backward2_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [32.8493s] [ 5%] 2025-12-04T10:15:33.0097573Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_avg_pool2d_backward3_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [2.8392s] [ 5%] 2025-12-04T10:15:33.0099395Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_avg_pool2d_backward4_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.3174s] [ 5%] 2025-12-04T10:15:33.0101173Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_batch_norm_2d_2_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py SKIPPED [0.0032s] (requires cuda) [ 6%] 2025-12-04T10:15:33.0102901Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bitwise2_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.0124s] [ 6%] 2025-12-04T10:15:33.0104610Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bucketize_computed_offsets_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7741s] [ 6%] 2025-12-04T10:15:33.0106403Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bucketize_int_int32_int64_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.2776s] [ 6%] 2025-12-04T10:15:33.0108243Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bucketize_int_int32_uint8_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.2606s] [ 7%] 2025-12-04T10:15:33.0110000Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bucketize_int_int8_int8_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.5796s] [ 7%] 2025-12-04T10:15:33.0111743Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bucketize_int_int8_uint8_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.2477s] [ 7%] 2025-12-04T10:15:33.0113506Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bucketize_int_uint8_int64_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.2928s] [ 7%] 2025-12-04T10:15:33.0115286Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bucketize_int_uint8_uint8_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.5804s] [ 8%] 2025-12-04T10:15:33.0117040Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bucketize_nd_tiling_True_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.3751s] [ 8%] 2025-12-04T10:15:33.0118847Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_builtins_round_float_ndigits_zero_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7791s] [ 8%] 2025-12-04T10:15:33.0120575Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cat_empty_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.8531s] [ 9%] 2025-12-04T10:15:33.0122211Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cat_empty_index_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7885s] [ 9%] 2025-12-04T10:15:33.0123870Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cat_unbacked_2d_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.0081s] [ 9%] 2025-12-04T10:15:33.0125553Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_check_stack_no_cycles_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8127s] [ 9%] 2025-12-04T10:15:33.0127256Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_complex_fallback_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.0319s] [ 10%] 2025-12-04T10:15:33.0129102Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_config_option_dont_assume_alignment_cudagraphs_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.9031s] [ 10%] 2025-12-04T10:15:33.0131037Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_config_option_dont_assume_alignment_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.9752s] [ 10%] 2025-12-04T10:15:33.0132860Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_consecutive_split_cumsum_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.2769s] [ 11%] 2025-12-04T10:15:33.0134638Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_constant_pad_1d_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.1727s] [ 11%] 2025-12-04T10:15:33.0136303Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_constant_pad_2d_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.2342s] [ 11%] 2025-12-04T10:15:33.0138074Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_constant_pad_fill_dtype_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.0421s] [ 11%] 2025-12-04T10:15:33.0139831Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_constant_pad_nd_inplace_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7365s] [ 12%] 2025-12-04T10:15:33.0141637Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_conv2d_channels_last_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.8915s] [ 12%] 2025-12-04T10:15:33.0143343Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_conv_with_as_strided_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [4.5323s] [ 12%] 2025-12-04T10:15:33.0145083Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cpu_scalar_with_cpu_tensor_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7484s] [ 12%] 2025-12-04T10:15:33.0147255Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cpu_scalar_with_gpu_tensor_cpp_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py W1204 10:07:09.456000 26102 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:15:33.0148721Z PASSED [6.2930s] [ 13%] 2025-12-04T10:15:33.0149672Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cumprod_zero_dim_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.6836s] [ 13%] 2025-12-04T10:15:33.0151342Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cumsum_no_mask_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.2698s] [ 13%] 2025-12-04T10:15:33.0153100Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_custom_op_fixed_layout_channels_last_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.6472s] [ 14%] 2025-12-04T10:15:33.0155065Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_custom_scan_op_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py SKIPPED [0.0032s] (associative_scan only supported on GPU) [ 14%] 2025-12-04T10:15:33.0157178Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_custom_scan_op_multi_input_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py SKIPPED [0.0029s] (associative_scan only supported on GPU) [ 14%] 2025-12-04T10:15:33.0159130Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_data_type_propogation_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.7145s] [ 14%] 2025-12-04T10:15:33.0160883Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_deterministic_codegen_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [10.3781s] [ 15%] 2025-12-04T10:15:33.0162591Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_device_assert_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8752s] [ 15%] 2025-12-04T10:15:33.0164201Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_div4_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.0504s] [ 15%] 2025-12-04T10:15:33.0165813Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_div_precision_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [2.8281s] [ 15%] 2025-12-04T10:15:33.0167585Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_div_presicion_accuracy_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7803s] [ 16%] 2025-12-04T10:15:33.0169270Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_div_prim_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.5016s] [ 16%] 2025-12-04T10:15:33.0170927Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtype_mismatch_issue_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.3947s] [ 16%] 2025-12-04T10:15:33.0172690Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_float16_float32_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.8910s] [ 17%] 2025-12-04T10:15:33.0174451Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_float16_int16_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.8226s] [ 17%] 2025-12-04T10:15:33.0176283Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_float16_int32_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.8617s] [ 17%] 2025-12-04T10:15:33.0178109Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_float16_uint8_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.9521s] [ 17%] 2025-12-04T10:15:33.0179859Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_float32_int8_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.0042s] [ 18%] 2025-12-04T10:15:33.0181596Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_int16_int16_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.7691s] [ 18%] 2025-12-04T10:15:33.0183322Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_int32_int64_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.8730s] [ 18%] 2025-12-04T10:15:33.0185039Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_int32_uint8_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.0042s] [ 19%] 2025-12-04T10:15:33.0186767Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_int8_int16_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.8997s] [ 19%] 2025-12-04T10:15:33.0188510Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_int8_uint8_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [2.1833s] [ 19%] 2025-12-04T10:15:33.0190255Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_uint8_float64_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.9006s] [ 19%] 2025-12-04T10:15:33.0192007Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_uint8_int16_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.9076s] [ 20%] 2025-12-04T10:15:33.0193739Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_uint8_int32_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.9060s] [ 20%] 2025-12-04T10:15:33.0195473Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_uint8_int64_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.8862s] [ 20%] 2025-12-04T10:15:33.0197351Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_expand_as_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.0417s] [ 20%] 2025-12-04T10:15:33.0199027Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_expanded_reduction_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.9928s] [ 21%] 2025-12-04T10:15:33.0200659Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_fill1_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7874s] [ 21%] 2025-12-04T10:15:33.0202365Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_fill2_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7860s] [ 21%] 2025-12-04T10:15:33.0203999Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_float32_to_int32_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8124s] [ 22%] 2025-12-04T10:15:33.0205969Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_fuse_large_params_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py SKIPPED [0.0005s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 22%] 2025-12-04T10:15:33.0207917Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_fuse_tiled_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8770s] [ 22%] 2025-12-04T10:15:33.0209525Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_gather1_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.9547s] [ 22%] 2025-12-04T10:15:33.0211393Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_graph_partition_arange1_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.5515s] [ 23%] 2025-12-04T10:15:33.0213199Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_graph_partition_misaligned_input_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.8960s] [ 23%] 2025-12-04T10:15:33.0214927Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_index3_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.2782s] [ 23%] 2025-12-04T10:15:33.0216593Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_index_propagation_abs_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.9151s] [ 23%] 2025-12-04T10:15:33.0218368Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_index_propagation_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7359s] [ 24%] 2025-12-04T10:15:33.0220043Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_index_put1_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [2.6207s] [ 24%] 2025-12-04T10:15:33.0221783Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_index_put_deterministic_fallback_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7599s] [ 24%] 2025-12-04T10:15:33.0223575Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_index_put_fallback1_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8959s] [ 25%] 2025-12-04T10:15:33.0225211Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_inf_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7655s] [ 25%] 2025-12-04T10:15:33.0226982Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_inner_reduction_detection_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py SKIPPED [0.0033s] (Skip for CPU device) [ 25%] 2025-12-04T10:15:33.0228816Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_inplace_add_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7751s] [ 25%] 2025-12-04T10:15:33.0230509Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_inplace_mixed_dtype_ops_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.9071s] [ 26%] 2025-12-04T10:15:33.0232279Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_large_grid_use_block_ptr_False_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.3668s] [ 26%] 2025-12-04T10:15:33.0234250Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_large_grid_use_block_ptr_True_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py SKIPPED [0.0033s] (triton backend is required for cpu) [ 26%] 2025-12-04T10:15:33.0236183Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_large_offset_pointwise_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.4855s] [ 26%] 2025-12-04T10:15:33.0237940Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_like_rands3_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.3674s] [ 27%] 2025-12-04T10:15:33.0239793Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_lite_triton_kernel_wrapper_functional_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py SKIPPED [0.0033s] (requires GPU) [ 27%] 2025-12-04T10:15:33.0241705Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_logcumsumexp_zero_dim_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.6914s] [ 27%] 2025-12-04T10:15:33.0243391Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_long_tensor_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7881s] [ 28%] 2025-12-04T10:15:33.0245136Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_masked_fill_promotion_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.6587s] [ 28%] 2025-12-04T10:15:33.0246822Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_max_pool2d2_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [3.0699s] [ 28%] 2025-12-04T10:15:33.0248561Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_max_pool2d_with_indices_backward2_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [8.1995s] [ 28%] 2025-12-04T10:15:33.0250409Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_max_pool2d_with_indices_backward6_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.2987s] [ 29%] 2025-12-04T10:15:33.0252140Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_mixed_mm2_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.9039s] [ 29%] 2025-12-04T10:15:33.0253741Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_mixed_mm3_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8101s] [ 29%] 2025-12-04T10:15:33.0255343Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_mm_views_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.1746s] [ 30%] 2025-12-04T10:15:33.0257064Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_mul_softmax_symfloat_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [2.5801s] [ 30%] 2025-12-04T10:15:33.0258799Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_multilayer_prime_size_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8648s] [ 30%] 2025-12-04T10:15:33.0260493Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_multilayer_var_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [2.3153s] [ 30%] 2025-12-04T10:15:33.0262212Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_mutations_loop_fusion_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7633s] [ 31%] 2025-12-04T10:15:33.0263973Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_needs_contiguous_strides_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8662s] [ 31%] 2025-12-04T10:15:33.0265664Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_new_empty_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.1017s] [ 31%] 2025-12-04T10:15:33.0267377Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_no_mega_fusion_during_lowering_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.8061s] [ 31%] 2025-12-04T10:15:33.0269177Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_nonzero_unbacked_refinement_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.9485s] [ 32%] 2025-12-04T10:15:33.0270969Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pad_single_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7652s] [ 32%] 2025-12-04T10:15:33.0272596Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_permute1_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.1546s] [ 32%] 2025-12-04T10:15:33.0274260Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_bessel_y1_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8614s] [ 33%] 2025-12-04T10:15:33.0276048Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_chebyshev_polynomial_u_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7863s] [ 33%] 2025-12-04T10:15:33.0277794Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_erf_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.9087s] [ 33%] 2025-12-04T10:15:33.0279532Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_log1p_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.9001s] [ 33%] 2025-12-04T10:15:33.0281226Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_log_ndtr_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.2536s] [ 34%] 2025-12-04T10:15:33.0282976Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_modified_bessel_i1_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8270s] [ 34%] 2025-12-04T10:15:33.0284780Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_modified_bessel_k0_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8248s] [ 34%] 2025-12-04T10:15:33.0286568Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_modified_bessel_k1_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7993s] [ 34%] 2025-12-04T10:15:33.0288310Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_ndtr_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.0944s] [ 35%] 2025-12-04T10:15:33.0290078Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_scaled_modified_bessel_k0_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8197s] [ 35%] 2025-12-04T10:15:33.0291942Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_scaled_modified_bessel_k1_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8224s] [ 35%] 2025-12-04T10:15:33.0293718Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_sinc_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.0250s] [ 36%] 2025-12-04T10:15:33.0295453Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_spherical_bessel_j0_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7513s] [ 36%] 2025-12-04T10:15:33.0297386Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_polar_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.1424s] [ 36%] 2025-12-04T10:15:33.0299057Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pow2_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.9560s] [ 36%] 2025-12-04T10:15:33.0300751Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_prepare_softmax_with_fast_math_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.2274s] [ 37%] 2025-12-04T10:15:33.0302507Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_randint_int64_mod_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8977s] [ 37%] 2025-12-04T10:15:33.0304158Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_remove_no_ops_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [4.7927s] [ 37%] 2025-12-04T10:15:33.0305953Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_remove_noop_copy_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.8535s] [ 38%] 2025-12-04T10:15:33.0307661Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_repeat_interleave_2_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8301s] [ 38%] 2025-12-04T10:15:33.0310150Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_replication_pad_errors_with_bool_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py E1204 10:09:16.237000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/0] failed while attempting to run meta for aten.replication_pad1d.default 2025-12-04T10:15:33.0312225Z E1204 10:09:16.237000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/0] Traceback (most recent call last): 2025-12-04T10:15:33.0313832Z E1204 10:09:16.237000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_subclasses/fake_tensor.py", line 2823, in _dispatch_impl 2025-12-04T10:15:33.0315290Z E1204 10:09:16.237000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/0] r = func(*args, **kwargs) 2025-12-04T10:15:33.0316597Z E1204 10:09:16.237000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_ops.py", line 836, in __call__ 2025-12-04T10:15:33.0317938Z E1204 10:09:16.237000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/0] return self._op(*args, **kwargs) 2025-12-04T10:15:33.0319332Z E1204 10:09:16.237000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_prims_common/wrappers.py", line 315, in _fn 2025-12-04T10:15:33.0320734Z E1204 10:09:16.237000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/0] result = fn(*args, **kwargs) 2025-12-04T10:15:33.0322230Z E1204 10:09:16.237000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_meta_registrations.py", line 1867, in meta_replication_pad1d 2025-12-04T10:15:33.0323663Z E1204 10:09:16.237000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/0] torch._check( 2025-12-04T10:15:33.0324936Z E1204 10:09:16.237000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1734, in _check 2025-12-04T10:15:33.0326485Z E1204 10:09:16.237000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/0] _check_with(RuntimeError, cond, message) # pyrefly: ignore [bad-argument-type] 2025-12-04T10:15:33.0328068Z E1204 10:09:16.237000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/0] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1716, in _check_with 2025-12-04T10:15:33.0329468Z E1204 10:09:16.237000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/0] raise error_type(message_evaluated) 2025-12-04T10:15:33.0330688Z E1204 10:09:16.237000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/0] RuntimeError: "replication_pad1d" not implemented for 'torch.bool' 2025-12-04T10:15:33.0332049Z E1204 10:09:16.360000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/1] failed while attempting to run meta for aten.replication_pad2d.default 2025-12-04T10:15:33.0333269Z E1204 10:09:16.360000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/1] Traceback (most recent call last): 2025-12-04T10:15:33.0334760Z E1204 10:09:16.360000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/1] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_subclasses/fake_tensor.py", line 2823, in _dispatch_impl 2025-12-04T10:15:33.0336213Z E1204 10:09:16.360000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/1] r = func(*args, **kwargs) 2025-12-04T10:15:33.0337697Z E1204 10:09:16.360000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/1] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_ops.py", line 836, in __call__ 2025-12-04T10:15:33.0339029Z E1204 10:09:16.360000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/1] return self._op(*args, **kwargs) 2025-12-04T10:15:33.0340444Z E1204 10:09:16.360000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/1] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_prims_common/wrappers.py", line 315, in _fn 2025-12-04T10:15:33.0341846Z E1204 10:09:16.360000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/1] result = fn(*args, **kwargs) 2025-12-04T10:15:33.0343347Z E1204 10:09:16.360000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/1] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_meta_registrations.py", line 1978, in meta_replication_pad2d 2025-12-04T10:15:33.0344845Z E1204 10:09:16.360000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/1] torch._check( 2025-12-04T10:15:33.0346102Z E1204 10:09:16.360000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/1] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1734, in _check 2025-12-04T10:15:33.0347665Z E1204 10:09:16.360000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/1] _check_with(RuntimeError, cond, message) # pyrefly: ignore [bad-argument-type] 2025-12-04T10:15:33.0349249Z E1204 10:09:16.360000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/1] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1716, in _check_with 2025-12-04T10:15:33.0350639Z E1204 10:09:16.360000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/1] raise error_type(message_evaluated) 2025-12-04T10:15:33.0351854Z E1204 10:09:16.360000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/1] RuntimeError: "replication_pad2d" not implemented for 'torch.bool' 2025-12-04T10:15:33.0353233Z E1204 10:09:16.481000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/2] failed while attempting to run meta for aten.replication_pad3d.default 2025-12-04T10:15:33.0354467Z E1204 10:09:16.481000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/2] Traceback (most recent call last): 2025-12-04T10:15:33.0355964Z E1204 10:09:16.481000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/2] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_subclasses/fake_tensor.py", line 2823, in _dispatch_impl 2025-12-04T10:15:33.0357425Z E1204 10:09:16.481000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/2] r = func(*args, **kwargs) 2025-12-04T10:15:33.0358720Z E1204 10:09:16.481000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/2] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_ops.py", line 836, in __call__ 2025-12-04T10:15:33.0360071Z E1204 10:09:16.481000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/2] return self._op(*args, **kwargs) 2025-12-04T10:15:33.0361487Z E1204 10:09:16.481000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/2] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_prims_common/wrappers.py", line 315, in _fn 2025-12-04T10:15:33.0362887Z E1204 10:09:16.481000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/2] result = fn(*args, **kwargs) 2025-12-04T10:15:33.0364360Z E1204 10:09:16.481000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/2] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_meta_registrations.py", line 2104, in meta_replication_pad3d 2025-12-04T10:15:33.0365781Z E1204 10:09:16.481000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/2] torch._check( 2025-12-04T10:15:33.0367050Z E1204 10:09:16.481000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/2] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1734, in _check 2025-12-04T10:15:33.0368679Z E1204 10:09:16.481000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/2] _check_with(RuntimeError, cond, message) # pyrefly: ignore [bad-argument-type] 2025-12-04T10:15:33.0370266Z E1204 10:09:16.481000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/2] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/__init__.py", line 1716, in _check_with 2025-12-04T10:15:33.0371640Z E1204 10:09:16.481000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/2] raise error_type(message_evaluated) 2025-12-04T10:15:33.0372864Z E1204 10:09:16.481000 26102 site-packages/torch/_subclasses/fake_tensor.py:2827] [1/2] RuntimeError: "replication_pad3d" not implemented for 'torch.bool' 2025-12-04T10:15:33.0373777Z PASSED [0.9608s] [ 38%] 2025-12-04T10:15:33.0374726Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_resize_as_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [21.7332s] [ 38%] 2025-12-04T10:15:33.0376324Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_round_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8524s] [ 39%] 2025-12-04T10:15:33.0378038Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_rsqrt_dynamic_shapes_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.6798s] [ 39%] 2025-12-04T10:15:33.0379712Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_scatter1_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8053s] [ 39%] 2025-12-04T10:15:33.0382476Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_scatter2_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py [W1204 10:09:41.262701973 TensorAdvancedIndexing.cpp:226] Warning: The reduce argument of torch.scatter with Tensor src is deprecated and will be removed in a future PyTorch release. Use torch.scatter_reduce instead for more reduction options. (function operator()) 2025-12-04T10:15:33.0384593Z PASSED [0.7870s] [ 39%] 2025-12-04T10:15:33.0385510Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_scatter3_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8910s] [ 40%] 2025-12-04T10:15:33.0387514Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_sdpa_prefer_nd_tiling_False_use_block_ptr_False_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py SKIPPED [0.0003s] (Does not support SDPA or pre-SM80 hardware) [ 40%] 2025-12-04T10:15:33.0389879Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_sdpa_prefer_nd_tiling_True_use_block_ptr_True_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (Does not support SDPA or pre-SM80 hardware) [ 40%] 2025-12-04T10:15:33.0391914Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_sdpa_unaligned_mask_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8428s] [ 41%] 2025-12-04T10:15:33.0393593Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_sgn_extremal_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7447s] [ 41%] 2025-12-04T10:15:33.0395272Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_shape_prop_torch_ones_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.2338s] [ 41%] 2025-12-04T10:15:33.0397106Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_sign_dtype_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.0549s] [ 41%] 2025-12-04T10:15:33.0398756Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_simplify_loops_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.1795s] [ 42%] 2025-12-04T10:15:33.0400419Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_single_elem_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.6789s] [ 42%] 2025-12-04T10:15:33.0402189Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_slice_scatter_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.6155s] [ 42%] 2025-12-04T10:15:33.0403881Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_softmax_one_kernel_loop_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.1385s] [ 42%] 2025-12-04T10:15:33.0405644Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_softmax_one_kernel_persist_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.1028s] [ 43%] 2025-12-04T10:15:33.0407394Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_sqrt_dynamic_shapes_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7851s] [ 43%] 2025-12-04T10:15:33.0409143Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_squeeze2_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.9548s] [ 43%] 2025-12-04T10:15:33.0410802Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_squeeze_varargs_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.7341s] [ 44%] 2025-12-04T10:15:33.0412408Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_std_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [2.0461s] [ 44%] 2025-12-04T10:15:33.0413986Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_sum5_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.9788s] [ 44%] 2025-12-04T10:15:33.0415579Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_sum_dtype_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.9487s] [ 44%] 2025-12-04T10:15:33.0417232Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_tanh_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.0097s] [ 45%] 2025-12-04T10:15:33.0418840Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_transpose_add_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.6069s] [ 45%] 2025-12-04T10:15:33.0420559Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_transposed_propagates_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8789s] [ 45%] 2025-12-04T10:15:33.0422229Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_triu_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.6740s] [ 46%] 2025-12-04T10:15:33.0423803Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_uint_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.6954s] [ 46%] 2025-12-04T10:15:33.0425494Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_unbacked_floordiv_simplify_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.9131s] [ 46%] 2025-12-04T10:15:33.0427207Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_unsqueeze_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [1.0793s] [ 46%] 2025-12-04T10:15:33.0428880Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_upsample_bicubic2d_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [8.3396s] [ 47%] 2025-12-04T10:15:33.0430646Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_vectorized_ops_masked_var_novec_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.8707s] [ 47%] 2025-12-04T10:15:33.0432360Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_views3_dynamic_shapes_cpu <- test/inductor/test_torchinductor.py PASSED [0.9568s] [ 47%] 2025-12-04T10:15:33.0434305Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test__dyn_quant_matmul_4bit_fp32_input_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0032s] (No _dyn_quant_matmul_4bit implementation on CUDA) [ 47%] 2025-12-04T10:15:33.0436718Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test__dyn_quant_pack_4bit_weight_bf16_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0030s] (No _dyn_quant_pack_4bit_weight implementation on CUDA) [ 48%] 2025-12-04T10:15:33.0438670Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_abs_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.7808s] [ 48%] 2025-12-04T10:15:33.0440348Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_adaptive_avg_pool1d_argmax_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.2977s] [ 48%] 2025-12-04T10:15:33.0442181Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_adaptive_avg_pool_with_output_size_0_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.4722s] [ 49%] 2025-12-04T10:15:33.0444104Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_adaptive_pool_errors_with_long_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.5952s] [ 49%] 2025-12-04T10:15:33.0445832Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_add_complex5_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.8756s] [ 49%] 2025-12-04T10:15:33.0447488Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_add_complex9_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.5718s] [ 49%] 2025-12-04T10:15:33.0449142Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_add_complex_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.6740s] [ 50%] 2025-12-04T10:15:33.0450862Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_add_complex_strided_fallback_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.1602s] [ 50%] 2025-12-04T10:15:33.0452601Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_add_const_int_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.3796s] [ 50%] 2025-12-04T10:15:33.0454287Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_add_inplace_permuted_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [2.1642s] [ 50%] 2025-12-04T10:15:33.0456460Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_alexnet_prefix_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py W1204 10:10:29.459000 26102 site-packages/torch/_inductor/utils.py:1703] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:15:33.0458493Z W1204 10:10:30.035000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0459785Z W1204 10:10:32.986000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0460602Z PASSED [7.7289s] [ 51%] 2025-12-04T10:15:33.0461582Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_aliased_buffer_reuse_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.6342s] [ 51%] 2025-12-04T10:15:33.0463369Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_allow_reuse_active_if_under_peak_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [2.0329s] [ 51%] 2025-12-04T10:15:33.0465205Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_allow_reuse_disable_if_exceed_peak_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.6318s] [ 52%] 2025-12-04T10:15:33.0467084Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_aoti_eager_support_str_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0034s] (Requires sm80) [ 52%] 2025-12-04T10:15:33.0469127Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_aoti_eager_with_persistent_cache_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0031s] (Requires sm80) [ 52%] 2025-12-04T10:15:33.0471055Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_argmax_argmin_with_duplicates_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.1917s] [ 52%] 2025-12-04T10:15:33.0472807Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_argmax_to_float_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.8328s] [ 53%] 2025-12-04T10:15:33.0474561Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_assert_alignment_op_name_pass_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.0034s] [ 53%] 2025-12-04T10:15:33.0476292Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_avg_pool2d2_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [2.3827s] [ 53%] 2025-12-04T10:15:33.0477991Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_avg_pool2d5_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.9370s] [ 53%] 2025-12-04T10:15:33.0479687Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_avg_pool2d_backward2_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [17.5536s] [ 54%] 2025-12-04T10:15:33.0481423Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_avg_pool3d_backward_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [4.1454s] [ 54%] 2025-12-04T10:15:33.0483100Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_bitwise3_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.6466s] [ 54%] 2025-12-04T10:15:33.0484803Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_bucketize_int_int16_int32_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.6698s] [ 55%] 2025-12-04T10:15:33.0486567Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_bucketize_int_int16_int8_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.7300s] [ 55%] 2025-12-04T10:15:33.0488335Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_bucketize_int_int32_int64_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.7022s] [ 55%] 2025-12-04T10:15:33.0490110Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_bucketize_int_uint8_int32_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.5630s] [ 55%] 2025-12-04T10:15:33.0491848Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_buffer_batch_norm_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [3.0906s] [ 56%] 2025-12-04T10:15:33.0493596Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_buffer_copied_in_graph_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.2332s] [ 56%] 2025-12-04T10:15:33.0495448Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_buffer_copied_in_graph_with_different_shapes_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.6799s] [ 56%] 2025-12-04T10:15:33.0497615Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_builtins_round_float_ndigits_neg_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.2490s] [ 57%] 2025-12-04T10:15:33.0499387Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cat_empty_index_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.5717s] [ 57%] 2025-12-04T10:15:33.0501088Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cat_extern_kernel_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.6296s] [ 57%] 2025-12-04T10:15:33.0502754Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cat_uint8_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.5606s] [ 57%] 2025-12-04T10:15:33.0504541Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cat_unbacked_empty_1d_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.3874s] [ 58%] 2025-12-04T10:15:33.0506218Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cauchy_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.6179s] [ 58%] 2025-12-04T10:15:33.0507884Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_check_stack_no_cycles_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.2432s] [ 58%] 2025-12-04T10:15:33.0509555Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_compar_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.3034s] [ 58%] 2025-12-04T10:15:33.0511356Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_config_option_dont_assume_alignment_cudagraphs_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.9520s] [ 59%] 2025-12-04T10:15:33.0513332Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_consecutive_split_cumprod_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.6585s] [ 59%] 2025-12-04T10:15:33.0515080Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_constant_pad_1d_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.2618s] [ 59%] 2025-12-04T10:15:33.0516793Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_constant_pad_float64_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.5978s] [ 60%] 2025-12-04T10:15:33.0518578Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_conv2d_backward_channels_last_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.8243s] [ 60%] 2025-12-04T10:15:33.0520881Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_conv3d_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py W1204 10:11:26.416000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[1, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0522881Z W1204 10:11:28.820000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[1, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0523710Z PASSED [3.9746s] [ 60%] 2025-12-04T10:15:33.0524881Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_conv_functional_bn_fuse_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0033s] (only support cpu conv bn test) [ 60%] 2025-12-04T10:15:33.0527385Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_conv_inference_heuristics_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py W1204 10:11:30.194000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0529506Z W1204 10:11:31.323000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0530314Z XFAIL [2.1498s] [ 61%] 2025-12-04T10:15:33.0531826Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_convolution1_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py W1204 10:11:32.547000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0533863Z W1204 10:11:32.551000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0535139Z W1204 10:11:32.553000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0536413Z W1204 10:11:32.556000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0537837Z W1204 10:11:32.558000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0539120Z W1204 10:11:32.560000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0540400Z W1204 10:11:32.625000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0541666Z W1204 10:11:33.657000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0542948Z W1204 10:11:34.907000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0544293Z W1204 10:11:34.911000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0545569Z W1204 10:11:34.913000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0546839Z W1204 10:11:34.916000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0548109Z W1204 10:11:34.918000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0549389Z W1204 10:11:34.920000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0550661Z W1204 10:11:34.971000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0551939Z W1204 10:11:35.409000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0552744Z PASSED [4.5430s] [ 61%] 2025-12-04T10:15:33.0553652Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cos_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.1469s] [ 61%] 2025-12-04T10:15:33.0555248Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cummin_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [2.6559s] [ 61%] 2025-12-04T10:15:33.0556898Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cumprod_zero_dim_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.3106s] [ 62%] 2025-12-04T10:15:33.0558556Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cumsum_no_mask_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.9019s] [ 62%] 2025-12-04T10:15:33.0560242Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cumsum_zero_dim_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.3111s] [ 62%] 2025-12-04T10:15:33.0562019Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_custom_op_fixed_layout_channels_last_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.2474s] [ 63%] 2025-12-04T10:15:33.0563852Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_custom_scan_op_multi_input_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.1695s] [ 63%] 2025-12-04T10:15:33.0565586Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dense_mask_index_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.7638s] [ 63%] 2025-12-04T10:15:33.0567370Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_deterministic_codegen_on_graph_break_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [2.5766s] [ 63%] 2025-12-04T10:15:33.0569262Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_diagonal_copy_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.1452s] [ 64%] 2025-12-04T10:15:33.0570998Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dist_bf16_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0033s] (Requires sm80) [ 64%] 2025-12-04T10:15:33.0572688Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_div7_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.7457s] [ 64%] 2025-12-04T10:15:33.0574301Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_div_zero_dim_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [2.3720s] [ 65%] 2025-12-04T10:15:33.0575964Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dropout_trivial_1_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.2684s] [ 65%] 2025-12-04T10:15:33.0577830Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtype_mismatch_issue_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [3.1293s] [ 65%] 2025-12-04T10:15:33.0579790Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_float32_int32_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0034s] (uses bfloat16 which requires SM >= 80) [ 65%] 2025-12-04T10:15:33.0581942Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_float32_int64_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0031s] (uses bfloat16 which requires SM >= 80) [ 66%] 2025-12-04T10:15:33.0584087Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_float32_uint8_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 66%] 2025-12-04T10:15:33.0586232Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_float64_float32_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 66%] 2025-12-04T10:15:33.0588406Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_float64_int16_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 66%] 2025-12-04T10:15:33.0590557Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_float64_uint8_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 67%] 2025-12-04T10:15:33.0592690Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_int16_float16_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 67%] 2025-12-04T10:15:33.0594838Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_int32_float32_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 67%] 2025-12-04T10:15:33.0597193Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_int64_float16_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 68%] 2025-12-04T10:15:33.0599325Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_uint8_float16_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0028s] (uses bfloat16 which requires SM >= 80) [ 68%] 2025-12-04T10:15:33.0601456Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_uint8_int32_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0031s] (uses bfloat16 which requires SM >= 80) [ 68%] 2025-12-04T10:15:33.0603591Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_uint8_int64_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 68%] 2025-12-04T10:15:33.0605870Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_uint8_uint8_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0029s] (uses bfloat16 which requires SM >= 80) [ 69%] 2025-12-04T10:15:33.0608038Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_embedding_bag_byte_unpack_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0029s] (No cuda implementation (it returns empty)) [ 69%] 2025-12-04T10:15:33.0609968Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_embedding_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.9345s] [ 69%] 2025-12-04T10:15:33.0611575Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_erfc_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.5443s] [ 69%] 2025-12-04T10:15:33.0613254Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_erfinv_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.5118s] [ 70%] 2025-12-04T10:15:33.0614979Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_fallback_mutable_op_with_return_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.0524s] [ 70%] 2025-12-04T10:15:33.0616691Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_fmod_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.8778s] [ 70%] 2025-12-04T10:15:33.0618357Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_fmod_zero_dim_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.9969s] [ 71%] 2025-12-04T10:15:33.0620094Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_functionalize_rng_wrappers_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.6832s] [ 71%] 2025-12-04T10:15:33.0621831Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_gather1_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.0701s] [ 71%] 2025-12-04T10:15:33.0623497Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_generate_rand_fp8_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.0034s] [ 71%] 2025-12-04T10:15:33.0625234Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_graph_partition_arange1_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.6103s] [ 72%] 2025-12-04T10:15:33.0627594Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_graph_partition_pad_dynamic_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py W1204 10:12:03.898000 26102 site-packages/torch/utils/_sympy/interp.py:179] [0/0] failed while executing pow_by_natural([VR[2, int_oo], VR[-1, -1]]) 2025-12-04T10:15:33.0629244Z PASSED [3.2075s] [ 72%] 2025-12-04T10:15:33.0630186Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_hardswish_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.7433s] [ 72%] 2025-12-04T10:15:33.0631922Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_index_propagation_remainder_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.6864s] [ 73%] 2025-12-04T10:15:33.0633657Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_index_put2_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.0501s] [ 73%] 2025-12-04T10:15:33.0635398Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_index_put_deterministic_fallback_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.3200s] [ 73%] 2025-12-04T10:15:33.0637227Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_indirect_load_broadcast_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [2.0454s] [ 73%] 2025-12-04T10:15:33.0639465Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_inductor_multiple_specializations_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0003s] (Skipping triton backend only since not big GPU (not enough SM)) [ 74%] 2025-12-04T10:15:33.0641714Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_inductor_triton_bucketize_respects_masking_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.4419s] [ 74%] 2025-12-04T10:15:33.0643535Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_kernel_names_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.1898s] [ 74%] 2025-12-04T10:15:33.0645150Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_l1_loss_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.7020s] [ 74%] 2025-12-04T10:15:33.0647662Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_large_grid_use_block_ptr_True_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py W1204 10:12:13.027000 26102 site-packages/torch/fx/experimental/symbolic_shapes.py:7441] [0/0] failed during evaluate_expr(2 < ps1, hint=None, size_oblivious=True, forcing_spec=False 2025-12-04T10:15:33.0649474Z PASSED [1.6252s] [ 75%] 2025-12-04T10:15:33.0651466Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_layer_norm_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py W1204 10:12:14.253000 26102 site-packages/torch/_inductor/debug.py:518] [0/0_1] model__525_inference_495 debug trace: /var/lib/jenkins/workspace/test/torch_compile_debug/run_2025_12_04_10_12_13_865809-pid_26102/torchinductor/model__525_inference_495.0 2025-12-04T10:15:33.0653495Z PASSED [0.8139s] [ 75%] 2025-12-04T10:15:33.0654462Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_like_channels_last_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.6801s] [ 75%] 2025-12-04T10:15:33.0656146Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_like_rands3_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.0911s] [ 76%] 2025-12-04T10:15:33.0657891Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_like_rands_sliced_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.5356s] [ 76%] 2025-12-04T10:15:33.0659553Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_linear1_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.8685s] [ 76%] 2025-12-04T10:15:33.0661312Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_linspace3_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.1660s] [ 76%] 2025-12-04T10:15:33.0662960Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_list_clearing_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [9.3616s] [ 77%] 2025-12-04T10:15:33.0664863Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_lite_regional_compile_flex_attention_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py ('RERUN', {'yellow': True}) [1.9967s] [ 77%] 2025-12-04T10:15:33.0666961Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_lite_regional_compile_flex_attention_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py ('RERUN', {'yellow': True}) [1.7710s] [ 77%] 2025-12-04T10:15:33.0668954Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_lite_regional_compile_flex_attention_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py FAILED [1.7728s] [ 77%] 2025-12-04T10:15:33.0669973Z 2025-12-04T10:15:33.0670135Z ==================================== RERUNS ==================================== 2025-12-04T10:15:33.0670782Z _ DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda _ 2025-12-04T10:15:33.0671419Z Traceback (most recent call last): 2025-12-04T10:15:33.0672241Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in test_lite_regional_compile_flex_attention 2025-12-04T10:15:33.0673218Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:15:33.0674035Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2430, in run_fw_bw_and_get_code 2025-12-04T10:15:33.0674832Z return run_and_get_code(run_with_backward) 2025-12-04T10:15:33.0675585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2409, in run_and_get_code 2025-12-04T10:15:33.0676304Z result = fn(*args, **kwargs) 2025-12-04T10:15:33.0677014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2426, in run_with_backward 2025-12-04T10:15:33.0677738Z result = fn() 2025-12-04T10:15:33.0678296Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in 2025-12-04T10:15:33.0679080Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:15:33.0679870Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 936, in compile_wrapper 2025-12-04T10:15:33.0680739Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-12-04T10:15:33.0681316Z torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:15:33.0681943Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:15:33.0682770Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:15:33.0683869Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:15:33.0684564Z 2025-12-04T10:15:33.0684679Z Developer debug context: 2025-12-04T10:15:33.0684897Z 2025-12-04T10:15:33.0685444Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:15:33.0686110Z 2025-12-04T10:15:33.0686331Z To execute this test, run the following from the base repo dir: 2025-12-04T10:15:33.0687496Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor_dynamic_shapes.py DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda 2025-12-04T10:15:33.0688432Z 2025-12-04T10:15:33.0688700Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:15:33.0689334Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:15:33.0706987Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:15:33.0709881Z stats [('calls_captured', 12), ('unique_graphs', 1)] 2025-12-04T10:15:33.0710414Z aot_autograd [('total', 3), ('autograd_cache_miss', 3), ('not_ok', 2), ('ok', 1)] 2025-12-04T10:15:33.0710927Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:15:33.0711279Z graph_break [] 2025-12-04T10:15:33.0711663Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:15:33.0712758Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0713740Z warnings.warn( 2025-12-04T10:15:33.0714632Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0715605Z warnings.warn( 2025-12-04T10:15:33.0717312Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/attention/flex_attention.py:1624: UserWarning: flex_attention called without torch.compile() - this will use an unfused implementation that materializes the full scores matrix instead of generating a fused kernel. 2025-12-04T10:15:33.0718728Z 2025-12-04T10:15:33.0718889Z SOLUTION: Use torch.compile(flex_attention)(...) 2025-12-04T10:15:33.0719197Z 2025-12-04T10:15:33.0719381Z If you want to debug your score_mod/mask_mod, you can set: 2025-12-04T10:15:33.0720002Z torch.nn.attention.flex_attention._FLEX_ATTENTION_DISABLE_COMPILE_DEBUG = True 2025-12-04T10:15:33.0720442Z 2025-12-04T10:15:33.0721001Z This will allow you to use print statements or breakpoints. Note: This doesn't work with the backwards pass and may produce incorrect results. 2025-12-04T10:15:33.0721904Z _warn_once( 2025-12-04T10:15:33.0722418Z _ DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda _ 2025-12-04T10:15:33.0723082Z Traceback (most recent call last): 2025-12-04T10:15:33.0723903Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in test_lite_regional_compile_flex_attention 2025-12-04T10:15:33.0724785Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:15:33.0725601Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2430, in run_fw_bw_and_get_code 2025-12-04T10:15:33.0726380Z return run_and_get_code(run_with_backward) 2025-12-04T10:15:33.0727130Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2409, in run_and_get_code 2025-12-04T10:15:33.0727864Z result = fn(*args, **kwargs) 2025-12-04T10:15:33.0728568Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2426, in run_with_backward 2025-12-04T10:15:33.0729278Z result = fn() 2025-12-04T10:15:33.0729844Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in 2025-12-04T10:15:33.0730553Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:15:33.0731327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 936, in compile_wrapper 2025-12-04T10:15:33.0732188Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-12-04T10:15:33.0732777Z torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:15:33.0733404Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:15:33.0734217Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:15:33.0735322Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:15:33.0736032Z 2025-12-04T10:15:33.0736153Z Developer debug context: 2025-12-04T10:15:33.0736360Z 2025-12-04T10:15:33.0737004Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:15:33.0737673Z 2025-12-04T10:15:33.0737892Z To execute this test, run the following from the base repo dir: 2025-12-04T10:15:33.0739056Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor_dynamic_shapes.py DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda 2025-12-04T10:15:33.0740010Z 2025-12-04T10:15:33.0740277Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:15:33.0740920Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:15:33.0743761Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:15:33.0746436Z stats [('calls_captured', 12), ('unique_graphs', 1)] 2025-12-04T10:15:33.0746993Z aot_autograd [('total', 3), ('autograd_cache_miss', 3), ('not_ok', 2), ('ok', 1)] 2025-12-04T10:15:33.0747524Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:15:33.0747874Z graph_break [] 2025-12-04T10:15:33.0748241Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:15:33.0749341Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0750382Z warnings.warn( 2025-12-04T10:15:33.0751266Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0752234Z warnings.warn( 2025-12-04T10:15:33.0753680Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/attention/flex_attention.py:1624: UserWarning: flex_attention called without torch.compile() - this will use an unfused implementation that materializes the full scores matrix instead of generating a fused kernel. 2025-12-04T10:15:33.0755079Z 2025-12-04T10:15:33.0755254Z SOLUTION: Use torch.compile(flex_attention)(...) 2025-12-04T10:15:33.0755548Z 2025-12-04T10:15:33.0755743Z If you want to debug your score_mod/mask_mod, you can set: 2025-12-04T10:15:33.0756353Z torch.nn.attention.flex_attention._FLEX_ATTENTION_DISABLE_COMPILE_DEBUG = True 2025-12-04T10:15:33.0756803Z 2025-12-04T10:15:33.0757368Z This will allow you to use print statements or breakpoints. Note: This doesn't work with the backwards pass and may produce incorrect results. 2025-12-04T10:15:33.0758169Z _warn_once( 2025-12-04T10:15:33.0758528Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:15:33.0761426Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:15:33.0764112Z stats [('calls_captured', 12), ('unique_graphs', 1)] 2025-12-04T10:15:33.0764665Z aot_autograd [('total', 3), ('autograd_cache_miss', 3), ('not_ok', 2), ('ok', 1)] 2025-12-04T10:15:33.0765194Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:15:33.0765535Z graph_break [] 2025-12-04T10:15:33.0765914Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:15:33.0767022Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0767996Z warnings.warn( 2025-12-04T10:15:33.0768877Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0769846Z warnings.warn( 2025-12-04T10:15:33.0770158Z =================================== FAILURES =================================== 2025-12-04T10:15:33.0770801Z _ DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda _ 2025-12-04T10:15:33.0771423Z Traceback (most recent call last): 2025-12-04T10:15:33.0772241Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in test_lite_regional_compile_flex_attention 2025-12-04T10:15:33.0773123Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:15:33.0774027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2430, in run_fw_bw_and_get_code 2025-12-04T10:15:33.0774831Z return run_and_get_code(run_with_backward) 2025-12-04T10:15:33.0775586Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2409, in run_and_get_code 2025-12-04T10:15:33.0776310Z result = fn(*args, **kwargs) 2025-12-04T10:15:33.0777081Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2426, in run_with_backward 2025-12-04T10:15:33.0777804Z result = fn() 2025-12-04T10:15:33.0778377Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in 2025-12-04T10:15:33.0779074Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:15:33.0779932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 936, in compile_wrapper 2025-12-04T10:15:33.0780808Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-12-04T10:15:33.0781388Z torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:15:33.0782090Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:15:33.0783002Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:15:33.0784106Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:15:33.0784795Z 2025-12-04T10:15:33.0784912Z Developer debug context: 2025-12-04T10:15:33.0785130Z 2025-12-04T10:15:33.0785657Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:15:33.0786329Z 2025-12-04T10:15:33.0786549Z To execute this test, run the following from the base repo dir: 2025-12-04T10:15:33.0787721Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor_dynamic_shapes.py DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda 2025-12-04T10:15:33.0788650Z 2025-12-04T10:15:33.0788933Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:15:33.0789555Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:15:33.0792324Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:15:33.0795024Z stats [('calls_captured', 12), ('unique_graphs', 1)] 2025-12-04T10:15:33.0795579Z aot_autograd [('total', 3), ('autograd_cache_miss', 3), ('not_ok', 2), ('ok', 1)] 2025-12-04T10:15:33.0796293Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:15:33.0796648Z graph_break [] 2025-12-04T10:15:33.0797030Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:15:33.0798137Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0799104Z warnings.warn( 2025-12-04T10:15:33.0799998Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0800975Z warnings.warn( 2025-12-04T10:15:33.0802548Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/attention/flex_attention.py:1624: UserWarning: flex_attention called without torch.compile() - this will use an unfused implementation that materializes the full scores matrix instead of generating a fused kernel. 2025-12-04T10:15:33.0803966Z 2025-12-04T10:15:33.0804128Z SOLUTION: Use torch.compile(flex_attention)(...) 2025-12-04T10:15:33.0804437Z 2025-12-04T10:15:33.0804620Z If you want to debug your score_mod/mask_mod, you can set: 2025-12-04T10:15:33.0805236Z torch.nn.attention.flex_attention._FLEX_ATTENTION_DISABLE_COMPILE_DEBUG = True 2025-12-04T10:15:33.0805684Z 2025-12-04T10:15:33.0806252Z This will allow you to use print statements or breakpoints. Note: This doesn't work with the backwards pass and may produce incorrect results. 2025-12-04T10:15:33.0807050Z _warn_once( 2025-12-04T10:15:33.0807407Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:15:33.0810275Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:15:33.0812981Z stats [('calls_captured', 12), ('unique_graphs', 1)] 2025-12-04T10:15:33.0813530Z aot_autograd [('total', 3), ('autograd_cache_miss', 3), ('not_ok', 2), ('ok', 1)] 2025-12-04T10:15:33.0814040Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:15:33.0814419Z graph_break [] 2025-12-04T10:15:33.0814788Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:15:33.0815888Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0816862Z warnings.warn( 2025-12-04T10:15:33.0817832Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0818800Z warnings.warn( 2025-12-04T10:15:33.0819172Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:15:33.0821963Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:15:33.0824657Z stats [('calls_captured', 12), ('unique_graphs', 1)] 2025-12-04T10:15:33.0825211Z aot_autograd [('total', 3), ('autograd_cache_miss', 3), ('not_ok', 2), ('ok', 1)] 2025-12-04T10:15:33.0825731Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:15:33.0826061Z graph_break [] 2025-12-04T10:15:33.0826429Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:15:33.0827522Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0828492Z warnings.warn( 2025-12-04T10:15:33.0829364Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0830328Z warnings.warn( 2025-12-04T10:15:33.0831423Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-535feae09e5a3ebd.xml - 2025-12-04T10:15:33.0832732Z =========================== short test summary info ============================ 2025-12-04T10:15:33.0834020Z FAILED [1.7728s] inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_lite_regional_compile_flex_attention_dynamic_shapes_cuda - torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:15:33.0835386Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:15:33.0836235Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:15:33.0837338Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:15:33.0838033Z 2025-12-04T10:15:33.0838151Z Developer debug context: 2025-12-04T10:15:33.0838439Z 2025-12-04T10:15:33.0838965Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:15:33.0839618Z 2025-12-04T10:15:33.0839851Z To execute this test, run the following from the base repo dir: 2025-12-04T10:15:33.0841011Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor_dynamic_shapes.py DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda 2025-12-04T10:15:33.0841943Z 2025-12-04T10:15:33.0842211Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:15:33.0842803Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T10:15:33.0843375Z == 1 failed, 249 passed, 30 skipped, 1 xfailed, 2 rerun in 418.63s (0:06:58) === 2025-12-04T10:15:33.0843870Z Got exit code 1 2025-12-04T10:15:33.0844135Z Retrying single test... 2025-12-04T10:15:33.0844776Z W1204 10:12:46.423000 34522 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:15:33.0846120Z Test results will be stored in test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-b527a05b10cde953.xml 2025-12-04T10:15:33.0847169Z ============================= test session starts ============================== 2025-12-04T10:15:33.0847836Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T10:15:33.0848442Z cachedir: .pytest_cache 2025-12-04T10:15:33.0849163Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T10:15:33.0849940Z rootdir: /var/lib/jenkins/workspace 2025-12-04T10:15:33.0850293Z configfile: pytest.ini 2025-12-04T10:15:33.0851037Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T10:15:33.0851946Z collecting ... collected 1815 items / 362 deselected / 1453 selected 2025-12-04T10:15:33.0853236Z stepcurrent: skipping 280 already run items. Running only test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_lite_regional_compile_flex_attention_dynamic_shapes_cuda 2025-12-04T10:15:33.0854381Z Running 1 items in this shard 2025-12-04T10:15:33.0854591Z 2025-12-04T10:15:33.0855586Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_lite_regional_compile_flex_attention_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py ('RERUN', {'yellow': True}) [4.5939s] [100%] 2025-12-04T10:15:33.0857772Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_lite_regional_compile_flex_attention_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py ('RERUN', {'yellow': True}) [1.9647s] [100%] 2025-12-04T10:15:33.0859766Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_lite_regional_compile_flex_attention_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py FAILED [1.7391s] [100%] 2025-12-04T10:15:33.0860807Z 2025-12-04T10:15:33.0860951Z ==================================== RERUNS ==================================== 2025-12-04T10:15:33.0861704Z _ DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda _ 2025-12-04T10:15:33.0862339Z Traceback (most recent call last): 2025-12-04T10:15:33.0863149Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in test_lite_regional_compile_flex_attention 2025-12-04T10:15:33.0864048Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:15:33.0864871Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2430, in run_fw_bw_and_get_code 2025-12-04T10:15:33.0865674Z return run_and_get_code(run_with_backward) 2025-12-04T10:15:33.0866420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2409, in run_and_get_code 2025-12-04T10:15:33.0867211Z result = fn(*args, **kwargs) 2025-12-04T10:15:33.0867918Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2426, in run_with_backward 2025-12-04T10:15:33.0868635Z result = fn() 2025-12-04T10:15:33.0869204Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in 2025-12-04T10:15:33.0869913Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:15:33.0870705Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 936, in compile_wrapper 2025-12-04T10:15:33.0871559Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-12-04T10:15:33.0872153Z torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:15:33.0872786Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:15:33.0873606Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:15:33.0874712Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:15:33.0875417Z 2025-12-04T10:15:33.0875537Z Developer debug context: 2025-12-04T10:15:33.0875742Z 2025-12-04T10:15:33.0876284Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:15:33.0876936Z 2025-12-04T10:15:33.0877165Z To execute this test, run the following from the base repo dir: 2025-12-04T10:15:33.0878306Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor_dynamic_shapes.py DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda 2025-12-04T10:15:33.0879258Z 2025-12-04T10:15:33.0879526Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:15:33.0880161Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:15:33.0882939Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:15:33.0885627Z stats [('calls_captured', 12), ('unique_graphs', 1)] 2025-12-04T10:15:33.0886163Z aot_autograd [('total', 3), ('autograd_cache_miss', 3), ('not_ok', 2), ('ok', 1)] 2025-12-04T10:15:33.0886688Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:15:33.0887030Z graph_break [] 2025-12-04T10:15:33.0887390Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:15:33.0888494Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0889571Z warnings.warn( 2025-12-04T10:15:33.0890459Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0891409Z warnings.warn( 2025-12-04T10:15:33.0892847Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/attention/flex_attention.py:1624: UserWarning: flex_attention called without torch.compile() - this will use an unfused implementation that materializes the full scores matrix instead of generating a fused kernel. 2025-12-04T10:15:33.0894258Z 2025-12-04T10:15:33.0894421Z SOLUTION: Use torch.compile(flex_attention)(...) 2025-12-04T10:15:33.0894713Z 2025-12-04T10:15:33.0894907Z If you want to debug your score_mod/mask_mod, you can set: 2025-12-04T10:15:33.0895572Z torch.nn.attention.flex_attention._FLEX_ATTENTION_DISABLE_COMPILE_DEBUG = True 2025-12-04T10:15:33.0896180Z 2025-12-04T10:15:33.0896747Z This will allow you to use print statements or breakpoints. Note: This doesn't work with the backwards pass and may produce incorrect results. 2025-12-04T10:15:33.0897598Z _warn_once( 2025-12-04T10:15:33.0898117Z _ DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda _ 2025-12-04T10:15:33.0898734Z Traceback (most recent call last): 2025-12-04T10:15:33.0899542Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in test_lite_regional_compile_flex_attention 2025-12-04T10:15:33.0900425Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:15:33.0901229Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2430, in run_fw_bw_and_get_code 2025-12-04T10:15:33.0902030Z return run_and_get_code(run_with_backward) 2025-12-04T10:15:33.0902781Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2409, in run_and_get_code 2025-12-04T10:15:33.0903508Z result = fn(*args, **kwargs) 2025-12-04T10:15:33.0904207Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2426, in run_with_backward 2025-12-04T10:15:33.0904923Z result = fn() 2025-12-04T10:15:33.0905490Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in 2025-12-04T10:15:33.0906184Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:15:33.0906967Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 936, in compile_wrapper 2025-12-04T10:15:33.0907833Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-12-04T10:15:33.0908418Z torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:15:33.0909034Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:15:33.0909867Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:15:33.0910970Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:15:33.0911659Z 2025-12-04T10:15:33.0911791Z Developer debug context: 2025-12-04T10:15:33.0911996Z 2025-12-04T10:15:33.0912516Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:15:33.0913184Z 2025-12-04T10:15:33.0913400Z To execute this test, run the following from the base repo dir: 2025-12-04T10:15:33.0914558Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor_dynamic_shapes.py DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda 2025-12-04T10:15:33.0915491Z 2025-12-04T10:15:33.0915770Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:15:33.0916389Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:15:33.0919299Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:15:33.0922004Z stats [('calls_captured', 12), ('unique_graphs', 1)] 2025-12-04T10:15:33.0922553Z aot_autograd [('total', 3), ('autograd_cache_miss', 3), ('not_ok', 2), ('ok', 1)] 2025-12-04T10:15:33.0923079Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:15:33.0923501Z graph_break [] 2025-12-04T10:15:33.0923877Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:15:33.0924978Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0925946Z warnings.warn( 2025-12-04T10:15:33.0926837Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0927808Z warnings.warn( 2025-12-04T10:15:33.0929247Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/attention/flex_attention.py:1624: UserWarning: flex_attention called without torch.compile() - this will use an unfused implementation that materializes the full scores matrix instead of generating a fused kernel. 2025-12-04T10:15:33.0930650Z 2025-12-04T10:15:33.0930821Z SOLUTION: Use torch.compile(flex_attention)(...) 2025-12-04T10:15:33.0931118Z 2025-12-04T10:15:33.0931299Z If you want to debug your score_mod/mask_mod, you can set: 2025-12-04T10:15:33.0931910Z torch.nn.attention.flex_attention._FLEX_ATTENTION_DISABLE_COMPILE_DEBUG = True 2025-12-04T10:15:33.0932349Z 2025-12-04T10:15:33.0932923Z This will allow you to use print statements or breakpoints. Note: This doesn't work with the backwards pass and may produce incorrect results. 2025-12-04T10:15:33.0933705Z _warn_once( 2025-12-04T10:15:33.0934073Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:15:33.0936849Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:15:33.0939617Z stats [('calls_captured', 12), ('unique_graphs', 1)] 2025-12-04T10:15:33.0940170Z aot_autograd [('total', 3), ('autograd_cache_miss', 3), ('not_ok', 2), ('ok', 1)] 2025-12-04T10:15:33.0940683Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:15:33.0941031Z graph_break [] 2025-12-04T10:15:33.0941409Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:15:33.0942497Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0943470Z warnings.warn( 2025-12-04T10:15:33.0944356Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0945324Z warnings.warn( 2025-12-04T10:15:33.0945621Z =================================== FAILURES =================================== 2025-12-04T10:15:33.0946272Z _ DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda _ 2025-12-04T10:15:33.0946979Z Traceback (most recent call last): 2025-12-04T10:15:33.0947798Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in test_lite_regional_compile_flex_attention 2025-12-04T10:15:33.0948665Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:15:33.0949484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2430, in run_fw_bw_and_get_code 2025-12-04T10:15:33.0950282Z return run_and_get_code(run_with_backward) 2025-12-04T10:15:33.0951025Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2409, in run_and_get_code 2025-12-04T10:15:33.0951754Z result = fn(*args, **kwargs) 2025-12-04T10:15:33.0952520Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2426, in run_with_backward 2025-12-04T10:15:33.0953247Z result = fn() 2025-12-04T10:15:33.0953805Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in 2025-12-04T10:15:33.0954521Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:15:33.0955314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 936, in compile_wrapper 2025-12-04T10:15:33.0956169Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-12-04T10:15:33.0956761Z torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:15:33.0957393Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:15:33.0958222Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:15:33.0959304Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:15:33.0960003Z 2025-12-04T10:15:33.0960120Z Developer debug context: 2025-12-04T10:15:33.0960342Z 2025-12-04T10:15:33.0960869Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:15:33.0961526Z 2025-12-04T10:15:33.0961753Z To execute this test, run the following from the base repo dir: 2025-12-04T10:15:33.0962904Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor_dynamic_shapes.py DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda 2025-12-04T10:15:33.0963852Z 2025-12-04T10:15:33.0964120Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:15:33.0964751Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:15:33.0967656Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:15:33.0970359Z stats [('calls_captured', 12), ('unique_graphs', 1)] 2025-12-04T10:15:33.0970902Z aot_autograd [('total', 3), ('autograd_cache_miss', 3), ('not_ok', 2), ('ok', 1)] 2025-12-04T10:15:33.0971424Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:15:33.0971769Z graph_break [] 2025-12-04T10:15:33.0972138Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:15:33.0973231Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0974205Z warnings.warn( 2025-12-04T10:15:33.0975175Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0976142Z warnings.warn( 2025-12-04T10:15:33.0977638Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/attention/flex_attention.py:1624: UserWarning: flex_attention called without torch.compile() - this will use an unfused implementation that materializes the full scores matrix instead of generating a fused kernel. 2025-12-04T10:15:33.0979052Z 2025-12-04T10:15:33.0979215Z SOLUTION: Use torch.compile(flex_attention)(...) 2025-12-04T10:15:33.0979518Z 2025-12-04T10:15:33.0979701Z If you want to debug your score_mod/mask_mod, you can set: 2025-12-04T10:15:33.0980324Z torch.nn.attention.flex_attention._FLEX_ATTENTION_DISABLE_COMPILE_DEBUG = True 2025-12-04T10:15:33.0980855Z 2025-12-04T10:15:33.0981415Z This will allow you to use print statements or breakpoints. Note: This doesn't work with the backwards pass and may produce incorrect results. 2025-12-04T10:15:33.0982212Z _warn_once( 2025-12-04T10:15:33.0982590Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:15:33.0985377Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:15:33.0988079Z stats [('calls_captured', 12), ('unique_graphs', 1)] 2025-12-04T10:15:33.0988623Z aot_autograd [('total', 3), ('autograd_cache_miss', 3), ('not_ok', 2), ('ok', 1)] 2025-12-04T10:15:33.0989146Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:15:33.0989492Z graph_break [] 2025-12-04T10:15:33.0989858Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:15:33.0990953Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0991925Z warnings.warn( 2025-12-04T10:15:33.0992813Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.0993766Z warnings.warn( 2025-12-04T10:15:33.0994148Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:15:33.0997080Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:15:33.0999779Z stats [('calls_captured', 12), ('unique_graphs', 1)] 2025-12-04T10:15:33.1000323Z aot_autograd [('total', 3), ('autograd_cache_miss', 3), ('not_ok', 2), ('ok', 1)] 2025-12-04T10:15:33.1000849Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:15:33.1001203Z graph_break [] 2025-12-04T10:15:33.1001580Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:15:33.1002664Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.1003645Z warnings.warn( 2025-12-04T10:15:33.1004658Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.1005617Z warnings.warn( 2025-12-04T10:15:33.1006712Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-b527a05b10cde953.xml - 2025-12-04T10:15:33.1007947Z =========================== short test summary info ============================ 2025-12-04T10:15:33.1009219Z FAILED [1.7391s] inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_lite_regional_compile_flex_attention_dynamic_shapes_cuda - torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:15:33.1010584Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:15:33.1011489Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:15:33.1012589Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:15:33.1013278Z 2025-12-04T10:15:33.1013408Z Developer debug context: 2025-12-04T10:15:33.1013615Z 2025-12-04T10:15:33.1014154Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:15:33.1014809Z 2025-12-04T10:15:33.1015025Z To execute this test, run the following from the base repo dir: 2025-12-04T10:15:33.1016183Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor_dynamic_shapes.py DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda 2025-12-04T10:15:33.1017205Z 2025-12-04T10:15:33.1017471Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:15:33.1018069Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T10:15:33.1018581Z ================== 1 failed, 362 deselected, 2 rerun in 8.58s ================== 2025-12-04T10:15:33.1019034Z Got exit code 1 2025-12-04T10:15:33.1019305Z Retrying single test... 2025-12-04T10:15:33.1019925Z W1204 10:13:08.097000 34721 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:15:33.1021253Z Test results will be stored in test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-0142110b0a563d3e.xml 2025-12-04T10:15:33.1022296Z ============================= test session starts ============================== 2025-12-04T10:15:33.1022956Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T10:15:33.1023548Z cachedir: .pytest_cache 2025-12-04T10:15:33.1024260Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T10:15:33.1025053Z rootdir: /var/lib/jenkins/workspace 2025-12-04T10:15:33.1025398Z configfile: pytest.ini 2025-12-04T10:15:33.1026131Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T10:15:33.1027050Z collecting ... collected 1815 items / 362 deselected / 1453 selected 2025-12-04T10:15:33.1028317Z stepcurrent: skipping 280 already run items. Running only test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_lite_regional_compile_flex_attention_dynamic_shapes_cuda 2025-12-04T10:15:33.1029446Z Running 1 items in this shard 2025-12-04T10:15:33.1029668Z 2025-12-04T10:15:33.1030648Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_lite_regional_compile_flex_attention_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py ('RERUN', {'yellow': True}) [4.6435s] [100%] 2025-12-04T10:15:33.1032754Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_lite_regional_compile_flex_attention_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py ('RERUN', {'yellow': True}) [1.9838s] [100%] 2025-12-04T10:15:33.1034836Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_lite_regional_compile_flex_attention_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py FAILED [1.7391s] [100%] 2025-12-04T10:15:33.1035848Z 2025-12-04T10:15:33.1036004Z ==================================== RERUNS ==================================== 2025-12-04T10:15:33.1036635Z _ DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda _ 2025-12-04T10:15:33.1037272Z Traceback (most recent call last): 2025-12-04T10:15:33.1038091Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in test_lite_regional_compile_flex_attention 2025-12-04T10:15:33.1038975Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:15:33.1039859Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2430, in run_fw_bw_and_get_code 2025-12-04T10:15:33.1040657Z return run_and_get_code(run_with_backward) 2025-12-04T10:15:33.1041420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2409, in run_and_get_code 2025-12-04T10:15:33.1042142Z result = fn(*args, **kwargs) 2025-12-04T10:15:33.1042848Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2426, in run_with_backward 2025-12-04T10:15:33.1043571Z result = fn() 2025-12-04T10:15:33.1044139Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in 2025-12-04T10:15:33.1044835Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:15:33.1045623Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 936, in compile_wrapper 2025-12-04T10:15:33.1046501Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-12-04T10:15:33.1047081Z torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:15:33.1047723Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:15:33.1048555Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:15:33.1049655Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:15:33.1050345Z 2025-12-04T10:15:33.1050464Z Developer debug context: 2025-12-04T10:15:33.1050680Z 2025-12-04T10:15:33.1051200Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:15:33.1051866Z 2025-12-04T10:15:33.1052084Z To execute this test, run the following from the base repo dir: 2025-12-04T10:15:33.1053254Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor_dynamic_shapes.py DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda 2025-12-04T10:15:33.1054187Z 2025-12-04T10:15:33.1054470Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:15:33.1055090Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:15:33.1057932Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:15:33.1060626Z stats [('calls_captured', 12), ('unique_graphs', 1)] 2025-12-04T10:15:33.1061176Z aot_autograd [('total', 3), ('autograd_cache_miss', 3), ('not_ok', 2), ('ok', 1)] 2025-12-04T10:15:33.1061686Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:15:33.1062113Z graph_break [] 2025-12-04T10:15:33.1062489Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:15:33.1063587Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.1064544Z warnings.warn( 2025-12-04T10:15:33.1065432Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.1066396Z warnings.warn( 2025-12-04T10:15:33.1067843Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/attention/flex_attention.py:1624: UserWarning: flex_attention called without torch.compile() - this will use an unfused implementation that materializes the full scores matrix instead of generating a fused kernel. 2025-12-04T10:15:33.1069306Z 2025-12-04T10:15:33.1069466Z SOLUTION: Use torch.compile(flex_attention)(...) 2025-12-04T10:15:33.1069774Z 2025-12-04T10:15:33.1069955Z If you want to debug your score_mod/mask_mod, you can set: 2025-12-04T10:15:33.1070575Z torch.nn.attention.flex_attention._FLEX_ATTENTION_DISABLE_COMPILE_DEBUG = True 2025-12-04T10:15:33.1071015Z 2025-12-04T10:15:33.1071590Z This will allow you to use print statements or breakpoints. Note: This doesn't work with the backwards pass and may produce incorrect results. 2025-12-04T10:15:33.1072371Z _warn_once( 2025-12-04T10:15:33.1072877Z _ DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda _ 2025-12-04T10:15:33.1073505Z Traceback (most recent call last): 2025-12-04T10:15:33.1074310Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in test_lite_regional_compile_flex_attention 2025-12-04T10:15:33.1075199Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:15:33.1076018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2430, in run_fw_bw_and_get_code 2025-12-04T10:15:33.1076813Z return run_and_get_code(run_with_backward) 2025-12-04T10:15:33.1077552Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2409, in run_and_get_code 2025-12-04T10:15:33.1078283Z result = fn(*args, **kwargs) 2025-12-04T10:15:33.1078987Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2426, in run_with_backward 2025-12-04T10:15:33.1079706Z result = fn() 2025-12-04T10:15:33.1080261Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in 2025-12-04T10:15:33.1080972Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:15:33.1081758Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 936, in compile_wrapper 2025-12-04T10:15:33.1082617Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-12-04T10:15:33.1083206Z torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:15:33.1083830Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:15:33.1084657Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:15:33.1085737Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:15:33.1086437Z 2025-12-04T10:15:33.1086553Z Developer debug context: 2025-12-04T10:15:33.1086757Z 2025-12-04T10:15:33.1087298Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:15:33.1087951Z 2025-12-04T10:15:33.1088184Z To execute this test, run the following from the base repo dir: 2025-12-04T10:15:33.1089402Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor_dynamic_shapes.py DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda 2025-12-04T10:15:33.1090359Z 2025-12-04T10:15:33.1090627Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:15:33.1091255Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:15:33.1094017Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:15:33.1097089Z stats [('calls_captured', 12), ('unique_graphs', 1)] 2025-12-04T10:15:33.1097631Z aot_autograd [('total', 3), ('autograd_cache_miss', 3), ('not_ok', 2), ('ok', 1)] 2025-12-04T10:15:33.1098160Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:15:33.1098506Z graph_break [] 2025-12-04T10:15:33.1098870Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:15:33.1099976Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.1100952Z warnings.warn( 2025-12-04T10:15:33.1101838Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.1102792Z warnings.warn( 2025-12-04T10:15:33.1104228Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/attention/flex_attention.py:1624: UserWarning: flex_attention called without torch.compile() - this will use an unfused implementation that materializes the full scores matrix instead of generating a fused kernel. 2025-12-04T10:15:33.1105640Z 2025-12-04T10:15:33.1105801Z SOLUTION: Use torch.compile(flex_attention)(...) 2025-12-04T10:15:33.1106091Z 2025-12-04T10:15:33.1106284Z If you want to debug your score_mod/mask_mod, you can set: 2025-12-04T10:15:33.1106885Z torch.nn.attention.flex_attention._FLEX_ATTENTION_DISABLE_COMPILE_DEBUG = True 2025-12-04T10:15:33.1107331Z 2025-12-04T10:15:33.1107892Z This will allow you to use print statements or breakpoints. Note: This doesn't work with the backwards pass and may produce incorrect results. 2025-12-04T10:15:33.1108683Z _warn_once( 2025-12-04T10:15:33.1109054Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:15:33.1111818Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:15:33.1114510Z stats [('calls_captured', 12), ('unique_graphs', 1)] 2025-12-04T10:15:33.1115063Z aot_autograd [('total', 3), ('autograd_cache_miss', 3), ('not_ok', 2), ('ok', 1)] 2025-12-04T10:15:33.1115592Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:15:33.1115936Z graph_break [] 2025-12-04T10:15:33.1116299Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:15:33.1117399Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.1118381Z warnings.warn( 2025-12-04T10:15:33.1119384Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.1120354Z warnings.warn( 2025-12-04T10:15:33.1120664Z =================================== FAILURES =================================== 2025-12-04T10:15:33.1121321Z _ DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda _ 2025-12-04T10:15:33.1121933Z Traceback (most recent call last): 2025-12-04T10:15:33.1122748Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in test_lite_regional_compile_flex_attention 2025-12-04T10:15:33.1123627Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:15:33.1124430Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2430, in run_fw_bw_and_get_code 2025-12-04T10:15:33.1125319Z return run_and_get_code(run_with_backward) 2025-12-04T10:15:33.1126074Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2409, in run_and_get_code 2025-12-04T10:15:33.1126815Z result = fn(*args, **kwargs) 2025-12-04T10:15:33.1127505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/utils.py", line 2426, in run_with_backward 2025-12-04T10:15:33.1128232Z result = fn() 2025-12-04T10:15:33.1128804Z File "/var/lib/jenkins/workspace/test/inductor/test_torchinductor.py", line 13781, in 2025-12-04T10:15:33.1129498Z _, codes = run_fw_bw_and_get_code(lambda: opt_fn(x)) 2025-12-04T10:15:33.1130290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 936, in compile_wrapper 2025-12-04T10:15:33.1131159Z raise e.with_traceback(None) from e.__cause__ # User compiler error 2025-12-04T10:15:33.1131752Z torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:15:33.1132373Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:15:33.1133208Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:15:33.1134306Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:15:33.1134995Z 2025-12-04T10:15:33.1135129Z Developer debug context: 2025-12-04T10:15:33.1135336Z 2025-12-04T10:15:33.1135861Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:15:33.1136531Z 2025-12-04T10:15:33.1136748Z To execute this test, run the following from the base repo dir: 2025-12-04T10:15:33.1137970Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor_dynamic_shapes.py DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda 2025-12-04T10:15:33.1138942Z 2025-12-04T10:15:33.1139223Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:15:33.1139848Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:15:33.1142629Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:15:33.1145322Z stats [('calls_captured', 12), ('unique_graphs', 1)] 2025-12-04T10:15:33.1145873Z aot_autograd [('total', 3), ('autograd_cache_miss', 3), ('not_ok', 2), ('ok', 1)] 2025-12-04T10:15:33.1146402Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:15:33.1146733Z graph_break [] 2025-12-04T10:15:33.1147102Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:15:33.1148275Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.1149236Z warnings.warn( 2025-12-04T10:15:33.1150122Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.1151087Z warnings.warn( 2025-12-04T10:15:33.1152534Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/nn/attention/flex_attention.py:1624: UserWarning: flex_attention called without torch.compile() - this will use an unfused implementation that materializes the full scores matrix instead of generating a fused kernel. 2025-12-04T10:15:33.1154001Z 2025-12-04T10:15:33.1154173Z SOLUTION: Use torch.compile(flex_attention)(...) 2025-12-04T10:15:33.1154463Z 2025-12-04T10:15:33.1154645Z If you want to debug your score_mod/mask_mod, you can set: 2025-12-04T10:15:33.1155267Z torch.nn.attention.flex_attention._FLEX_ATTENTION_DISABLE_COMPILE_DEBUG = True 2025-12-04T10:15:33.1155705Z 2025-12-04T10:15:33.1156282Z This will allow you to use print statements or breakpoints. Note: This doesn't work with the backwards pass and may produce incorrect results. 2025-12-04T10:15:33.1157065Z _warn_once( 2025-12-04T10:15:33.1157436Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:15:33.1160212Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:15:33.1162927Z stats [('calls_captured', 12), ('unique_graphs', 1)] 2025-12-04T10:15:33.1163480Z aot_autograd [('total', 3), ('autograd_cache_miss', 3), ('not_ok', 2), ('ok', 1)] 2025-12-04T10:15:33.1164124Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:15:33.1164474Z graph_break [] 2025-12-04T10:15:33.1164847Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:15:33.1165935Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.1166912Z warnings.warn( 2025-12-04T10:15:33.1167799Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.1168775Z warnings.warn( 2025-12-04T10:15:33.1169142Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T10:15:33.1171907Z unimplemented [('Attempt to trace generator\n Explanation: Generators cannot be compiled directly with `torch.compile`.\n Hint: Call a generator from inside of a non-generator Python function and compile that function instead.\n Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround.\n\n Developer debug context: \n\n For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html', 1)] 2025-12-04T10:15:33.1174604Z stats [('calls_captured', 12), ('unique_graphs', 1)] 2025-12-04T10:15:33.1175153Z aot_autograd [('total', 3), ('autograd_cache_miss', 3), ('not_ok', 2), ('ok', 1)] 2025-12-04T10:15:33.1175681Z inductor [('fxgraph_cache_miss', 2)] 2025-12-04T10:15:33.1176012Z graph_break [] 2025-12-04T10:15:33.1176383Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T10:15:33.1177720Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.1178689Z warnings.warn( 2025-12-04T10:15:33.1179581Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T10:15:33.1180551Z warnings.warn( 2025-12-04T10:15:33.1181644Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-0142110b0a563d3e.xml - 2025-12-04T10:15:33.1182862Z =========================== short test summary info ============================ 2025-12-04T10:15:33.1184204Z FAILED [1.7391s] inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_lite_regional_compile_flex_attention_dynamic_shapes_cuda - torch._dynamo.exc.Unsupported: Attempt to trace generator 2025-12-04T10:15:33.1185573Z Explanation: Generators cannot be compiled directly with `torch.compile`. 2025-12-04T10:15:33.1186402Z Hint: Call a generator from inside of a non-generator Python function and compile that function instead. 2025-12-04T10:15:33.1187480Z Hint: This graph break is fundamental - it is unlikely that Dynamo will ever be able to trace through your code. Consider finding a workaround. 2025-12-04T10:15:33.1188180Z 2025-12-04T10:15:33.1188298Z Developer debug context: 2025-12-04T10:15:33.1188516Z 2025-12-04T10:15:33.1189046Z For more details about this graph break, please visit: https://meta-pytorch.github.io/compile-graph-break-site/gb/gb0003.html 2025-12-04T10:15:33.1189702Z 2025-12-04T10:15:33.1189928Z To execute this test, run the following from the base repo dir: 2025-12-04T10:15:33.1191084Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_torchinductor_dynamic_shapes.py DynamicShapesGPUTests.test_lite_regional_compile_flex_attention_dynamic_shapes_cuda 2025-12-04T10:15:33.1192039Z 2025-12-04T10:15:33.1192311Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:15:33.1192902Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T10:15:33.1193422Z ================== 1 failed, 362 deselected, 2 rerun in 8.66s ================== 2025-12-04T10:15:33.1193856Z Got exit code 1 2025-12-04T10:15:33.1194737Z FAILED CONSISTENTLY: test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_lite_regional_compile_flex_attention_dynamic_shapes_cuda 2025-12-04T10:15:33.1196242Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T10:15:33.1197245Z W1204 10:13:29.660000 34920 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T10:15:33.1198566Z Test results will be stored in test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-92fbe8f0c3defdbe.xml 2025-12-04T10:15:33.1199617Z ============================= test session starts ============================== 2025-12-04T10:15:33.1200276Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T10:15:33.1200879Z cachedir: .pytest_cache 2025-12-04T10:15:33.1201583Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T10:15:33.1202373Z rootdir: /var/lib/jenkins/workspace 2025-12-04T10:15:33.1202732Z configfile: pytest.ini 2025-12-04T10:15:33.1203458Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T10:15:33.1204378Z collecting ... collected 1815 items / 281 deselected / 1534 selected 2025-12-04T10:15:33.1204901Z stepcurrent: skipping 281 already run items. 2025-12-04T10:15:33.1205299Z Running 82 items in this shard 2025-12-04T10:15:33.1205510Z 2025-12-04T10:15:33.1206556Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_lite_regional_compile_repeated_blocks_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [2.8998s] [ 1%] 2025-12-04T10:15:33.1208444Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_low_memory_max_pool_dilation_1_dim_2_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [4.4181s] [ 2%] 2025-12-04T10:15:33.1210208Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_masked_scatter_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.3940s] [ 3%] 2025-12-04T10:15:33.1211874Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_max_pool2d1_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [4.5949s] [ 4%] 2025-12-04T10:15:33.1213665Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_max_pool2d6_dilation_1_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [5.0405s] [ 6%] 2025-12-04T10:15:33.1215465Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_max_pool2d_with_indices_backward4_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [25.1310s] [ 7%] 2025-12-04T10:15:33.1217697Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_multi_device_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py W1204 10:14:15.711000 34920 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:15:33.1219404Z W1204 10:14:15.715000 34920 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:15:33.1220350Z W1204 10:14:15.719000 34920 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T10:15:33.1220986Z PASSED [1.4879s] [ 8%] 2025-12-04T10:15:33.1222185Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_multi_gpu_recompile_on_index_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0003s] (requires multiple cuda devices) [ 9%] 2025-12-04T10:15:33.1224209Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_nan_sort_stable_True_descending_False_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.1435s] [ 10%] 2025-12-04T10:15:33.1225969Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_new_empty_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.2355s] [ 12%] 2025-12-04T10:15:33.1227637Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_nll_loss_backward_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.5679s] [ 13%] 2025-12-04T10:15:33.1229358Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_norm_constant_overflow_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [2.2056s] [ 14%] 2025-12-04T10:15:33.1231103Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_output_strides_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.7590s] [ 15%] 2025-12-04T10:15:33.1232754Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pad_view_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [5.5147s] [ 17%] 2025-12-04T10:15:33.1234504Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_chebyshev_polynomial_t_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.5131s] [ 18%] 2025-12-04T10:15:33.1236386Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_chebyshev_polynomial_w_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.1404s] [ 19%] 2025-12-04T10:15:33.1238157Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_erfcx_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [5.7459s] [ 20%] 2025-12-04T10:15:33.1239954Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_expit_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.3378s] [ 21%] 2025-12-04T10:15:33.1241651Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_expm1_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.4469s] [ 23%] 2025-12-04T10:15:33.1243366Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_gammainc_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.1447s] [ 24%] 2025-12-04T10:15:33.1245154Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_hermite_polynomial_he_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.1243s] [ 25%] 2025-12-04T10:15:33.1246985Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_i0e_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.2420s] [ 26%] 2025-12-04T10:15:33.1248743Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_modified_bessel_k0_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.1225s] [ 28%] 2025-12-04T10:15:33.1250506Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_ndtri_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.4376s] [ 29%] 2025-12-04T10:15:33.1252205Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_round_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.4660s] [ 30%] 2025-12-04T10:15:33.1253953Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_spherical_bessel_j0_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.3862s] [ 31%] 2025-12-04T10:15:33.1255668Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pow2_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.8506s] [ 32%] 2025-12-04T10:15:33.1257383Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_randn_generator_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.5975s] [ 34%] 2025-12-04T10:15:33.1259054Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_reduction4_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.9840s] [ 35%] 2025-12-04T10:15:33.1260725Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_remove_noop_slice_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.3436s] [ 36%] 2025-12-04T10:15:33.1262450Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_remove_noop_view_default_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.3618s] [ 37%] 2025-12-04T10:15:33.1264192Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_repeat_interleave_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [2.3998s] [ 39%] 2025-12-04T10:15:33.1265957Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_reuse_buffers_with_aliasing_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.5012s] [ 40%] 2025-12-04T10:15:33.1267670Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_rsqrt_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.4968s] [ 41%] 2025-12-04T10:15:33.1269599Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_scaled_dot_product_attention_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0033s] (Can't run flash attention on this platform) [ 42%] 2025-12-04T10:15:33.1271521Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_scatter1_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.5491s] [ 43%] 2025-12-04T10:15:33.1273150Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_scatter6_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.4264s] [ 45%] 2025-12-04T10:15:33.1274870Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_scatter_add3_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.0821s] [ 46%] 2025-12-04T10:15:33.1276722Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_scatter_bf16_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0033s] (uses bfloat16 which requires SM >= 80) [ 47%] 2025-12-04T10:15:33.1278926Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_sdpa_prefer_nd_tiling_True_use_block_ptr_False_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0002s] (Does not support SDPA or pre-SM80 hardware) [ 48%] 2025-12-04T10:15:33.1280938Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_shape_padding_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [4.7141s] [ 50%] 2025-12-04T10:15:33.1282662Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_sigmoid_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.5257s] [ 51%] 2025-12-04T10:15:33.1284259Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_sin_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.9480s] [ 52%] 2025-12-04T10:15:33.1285864Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_single_elem_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.2991s] [ 53%] 2025-12-04T10:15:33.1287548Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_single_elem_indirect_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.3289s] [ 54%] 2025-12-04T10:15:33.1289254Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_sizehint_issue1_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [3.5346s] [ 56%] 2025-12-04T10:15:33.1290906Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_slice1_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.5403s] [ 57%] 2025-12-04T10:15:33.1292506Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_slice3_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.8910s] [ 58%] 2025-12-04T10:15:33.1294144Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_slice_mutation3_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.3432s] [ 59%] 2025-12-04T10:15:33.1295823Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_slice_scatter2_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.6967s] [ 60%] 2025-12-04T10:15:33.1297638Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_sort_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.5532s] [ 62%] 2025-12-04T10:15:33.1299289Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_sort_transpose_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.6235s] [ 63%] 2025-12-04T10:15:33.1300995Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_special_polygamma_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.6691s] [ 64%] 2025-12-04T10:15:33.1302700Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_split_cumprod_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.7455s] [ 65%] 2025-12-04T10:15:33.1304479Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_split_cumsum_low_prec_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py SKIPPED [0.0034s] (Requires sm80) [ 67%] 2025-12-04T10:15:33.1306414Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_split_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.4007s] [ 68%] 2025-12-04T10:15:33.1308114Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_split_with_unbacked_symints_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.7972s] [ 69%] 2025-12-04T10:15:33.1309993Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_strided_inputs_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.3725s] [ 70%] 2025-12-04T10:15:33.1311630Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_sum4_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.9549s] [ 71%] 2025-12-04T10:15:33.1313262Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_tensor_index_slice_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [3.1050s] [ 73%] 2025-12-04T10:15:33.1314994Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_tmp_not_defined_issue2_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [2.5549s] [ 74%] 2025-12-04T10:15:33.1316818Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_tmp_not_defined_issue3_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.8441s] [ 75%] 2025-12-04T10:15:33.1318583Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_transposed_propagates_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.5117s] [ 76%] 2025-12-04T10:15:33.1320257Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_triu_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [1.4738s] [ 78%] 2025-12-04T10:15:33.1321829Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_uint_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.2328s] [ 79%] 2025-12-04T10:15:33.1323558Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_unbacked_floordiv_simplify_errors_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.0221s] [ 80%] 2025-12-04T10:15:33.1325362Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_unspec_inputs_int32_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.5630s] [ 81%] 2025-12-04T10:15:33.1327086Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_unspec_inputs_int64_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.4945s] [ 82%] 2025-12-04T10:15:33.1328806Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_unspec_inputs_int8_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.4935s] [ 84%] 2025-12-04T10:15:33.1330452Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_vdd_clamp_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.6870s] [ 85%] 2025-12-04T10:15:33.1332185Z inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_vectorized_ops_masked_var_novec_dynamic_shapes_cuda <- test/inductor/test_torchinductor.py PASSED [0.5906s] [ 86%] 2025-12-04T10:15:33.1333798Z inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_constant_fold_uniform_value_dynamic_cuda PASSED [1.5422s] [ 87%] 2025-12-04T10:15:33.1335098Z inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_item_zeros_nobreak_cuda PASSED [0.1916s] [ 89%] 2025-12-04T10:15:33.1336277Z inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_math_ops_op10_cuda PASSED [0.2318s] [ 90%] 2025-12-04T10:15:33.1337517Z inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_math_ops_op1_cuda PASSED [0.2546s] [ 91%] 2025-12-04T10:15:33.1338688Z inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_math_ops_op8_cuda PASSED [0.2270s] [ 92%] 2025-12-04T10:15:33.1339847Z inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_math_ops_op9_cuda PASSED [0.2258s] [ 93%] 2025-12-04T10:15:33.1341025Z inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_sym_stride_lowering_cuda PASSED [0.2875s] [ 95%] 2025-12-04T10:15:33.1342248Z inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_sym_sum_unbacked_cuda PASSED [2.2009s] [ 96%] 2025-12-04T10:15:33.1343671Z inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_unbacked_cat_backwards_save_data_dependent_cuda PASSED [1.9727s] [ 97%] 2025-12-04T10:15:33.1345012Z inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_unbacked_reduction_cuda PASSED [0.4634s] [ 98%] 2025-12-04T10:15:33.1346263Z inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_unspecialized_float_dynamic_cuda PASSED [0.7873s] [100%] 2025-12-04T10:15:33.1346998Z 2025-12-04T10:15:33.1347928Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-92fbe8f0c3defdbe.xml - 2025-12-04T10:15:33.1349298Z ========== 77 passed, 5 skipped, 281 deselected in 118.42s (0:01:58) =========== 2025-12-04T10:15:33.1350502Z The following tests failed consistently: ['test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_lite_regional_compile_flex_attention_dynamic_shapes_cuda'] 2025-12-04T10:15:33.1351461Z 2025-12-04T10:15:33.1352199Z FINISHED PRINTING LOG FILE of inductor/test_torchinductor_dynamic_shapes 4/5 (test/test-reports/inductor.test_torchinductor_dynamic_shapes_4.5_2869e792aa617cfa_.log) 2025-12-04T10:15:33.1353051Z 2025-12-04T10:15:33.1353494Z Finished inductor/test_torchinductor_dynamic_shapes 4/5 ... [2025-12-04 10:15:32.961356][3716.65174301], took 10.15min 2025-12-04T10:15:33.1355053Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-535feae09e5a3ebd.xml 2025-12-04T10:15:33.1357152Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-b527a05b10cde953.xml 2025-12-04T10:15:33.1359252Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-0142110b0a563d3e.xml 2025-12-04T10:15:33.1361365Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-92fbe8f0c3defdbe.xml 2025-12-04T10:15:33.4855586Z Uploading logs for 57119749282 to S3 2025-12-04T10:15:33.5307679Z Uploading artifacts took 0.38 seconds 2025-12-04T10:15:33.5308204Z inductor/test_torchinductor_dynamic_shapes 4/5 failed! 2025-12-04T10:15:33.5312833Z Running inductor/test_torchinductor_codegen_dynamic_shapes 4/4 ... [2025-12-04 10:15:33.531096][3717.221488347] 2025-12-04T10:15:33.5313545Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T10:15:33.5317606Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_codegen_dynamic_shapes.py', '--shard-id=4', '--num-shards=4', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 10:15:33.531523] 2025-12-04T10:24:45.5588650Z 2025-12-04T10:24:45.5589987Z inductor/test_torchinductor_codegen_dynamic_shapes 4/4 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_codegen_dynamic_shapes_4.4_0fcd01f998dd3903_.log 2025-12-04T10:24:45.5902898Z Running 437 items in this shard: test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_adaptive_avg_pool2d2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_adaptive_avg_pool2d_low_prec_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_adaptive_avg_pool_errors_with_long_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_adaptive_max_pool2d1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_adaptive_max_pool2d3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_add_complex5_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_add_complex_strided_fallback_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_add_const_int_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_addmm_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_aoti_eager_dtype_device_layout_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_aoti_eager_override_registration_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_aoti_eager_support_str_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_argmax_argmin2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_argmax_argmin_with_nan_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_argmax_to_float_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_avg_pool2d6_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_avg_pool2d7_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_batch_norm_2d_2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_bitwise_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_bool_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_bucketize_add_autotune_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_bucketize_broadcast_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_bucketize_int_int16_int32_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_bucketize_int_int16_int64_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_bucketize_int_int16_int8_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_bucketize_int_int32_int64_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_bucketize_int_int32_uint8_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_bucketize_int_int64_int64_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_bucketize_int_int8_int16_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_bucketize_int_uint8_int64_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_bucketize_nd_tiling_False_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_buffer_copied_in_graph_with_different_shapes_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_builtins_round_float_ndigits_pos_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_cat_single_empty_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_clamp_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_clone_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_complex_fallback_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_computed_buffer_inlining_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_config_option_dont_assume_alignment_cudagraphs_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_constant_pad_2d_strides_nonpositive_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_constant_pad_3d_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_conv2d_backward_channels_last_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_conv_shape_check_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_convolution5_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_copy_with_scalar_src_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_cpu_scalar_with_cpu_tensor_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_cpu_scalar_with_gpu_tensor_dynamic_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_cummin_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_cumsum_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_cumsum_inf_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_cumsum_zero_dim_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_custom_op_2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_custom_op_3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_custom_op_fixed_layout_channels_last_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_custom_op_unbacked_symints_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_custom_scan_op_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_custom_scan_op_multi_input_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dense_mask_index_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dist_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_div5_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_div8_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_div_precision_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_div_prim_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_div_softmax_symfloat_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dropout_deterministic_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_float16_float32_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_float16_float64_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_float16_int8_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_float32_float32_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_int16_float16_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_int16_int64_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_int16_int8_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_int16_uint8_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_int32_int32_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_int32_int8_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_int64_float64_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_int64_int16_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_int64_int8_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_uint8_float16_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_uint8_int16_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_uint8_int32_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_uint8_int64_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_embedding_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_embedding_sparse_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_empty1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_emulate_precision_triton_fp_fusion_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_erfc_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_expand_as_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_expm1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_fallback_mutable_op_no_mutated_tensors_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_fft_real_input_real_output_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_fill2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_flip_cat_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_float16_to_int16_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_float_index_expression_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_float_index_expression_type_promotion_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_float_repr_dynamic_shapes_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_fmod_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_full_boolean_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_full_like_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_full_truncation_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_fuse_large_params_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_fusing_write_into_disjoint_read_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_gather3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_generate_rand_fp8_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_graph_partition_argmax_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_graph_partition_constant_tensor1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_graph_partition_pad_dynamic_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_graph_partition_refcount_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_graph_partition_scalar_inputs_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_index1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_index_propagation_device_assert_masked_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_index_propagation_nested_indirect_indexing_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_index_put1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_index_put_deterministic_fallback_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_index_remainder_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_index_tensor_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_inductor_assert_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_inductor_layout_optimization_input_mutations_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_inf_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_inner_reduction_detection_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_input_mutation3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_input_mutation4_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_insignificant_strides_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_isinf_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_kernel_names_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_kwargs_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_large_block_sizes_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_large_strided_reduction_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_large_tensor_reduction_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_like_rands2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_like_rands3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_linalg_eig_stride_consistency_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_linear_mixed_dtype_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_linspace2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_lite_dynamic_shape_assertion_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_lite_regional_compile_repeated_blocks_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_lite_triton_kernel_wrapper_functional_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_log_softmax_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_low_memory_max_pool_dilation_1_dim_2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_low_memory_max_pool_dilation_2_dim_2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_low_memory_max_pool_dilation_2_dim_3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_max_pool2d3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_max_pool2d6_dilation_1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_max_pool2d8_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_max_pool2d_with_indices_backward3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_max_pool2d_with_indices_backward6_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_mm_views_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_mul_softmax_symfloat_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_multi_device_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_multilayer_any_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_mutations_loop_fusion_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_nan_sort_stable_True_descending_False_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_needs_contiguous_strides_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_neg_index_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_new_ones_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_nll_loss_backward_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_norm_constant_overflow_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_pad_single_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_pad_view_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_pointwise_bessel_y0_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_pointwise_chebyshev_polynomial_u_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_pointwise_chebyshev_polynomial_w_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_pointwise_hermite_polynomial_h_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_pointwise_i0_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_pointwise_i0e_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_pointwise_i1e_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_pointwise_logit_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_pointwise_sinc_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_pointwise_spherical_bessel_j0_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_pointwise_zeta_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_randint_distribution_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_randn_with_dtype_and_device_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_reduction3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_reduction4_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_reduction5_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_relu_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_remove_no_ops_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_remove_noop_slice1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_repeat_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_repeat_interleave_Tensor_decomp_int64_nd_2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_resize_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_roi_align_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_scatter3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_scatter_add3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_scatter_reduce3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_sdpa_unaligned_mask_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_single_elem_indirect_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_size_asserts_for_multi_output_fallback_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_slice_mutation1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_slice_mutation2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_softmax_one_kernel_loop_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_squeeze1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_squeeze2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_tan_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_tanh_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_tensor1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_tensor2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_tensor_index_put_slice_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_tmp_not_defined_issue3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_transpose_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_transposed_propagates_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_unspec_inputs_uint8_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_unsqueeze_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_upsample_bicubic2d_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_upsample_nearest1d_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_var_mean_tile_reduction_False_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_vectorized_ops_masked_var_novec_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_view_as_complex_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_view_detach_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_views2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_views6_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_weight_norm_bwd_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_where_with_logical_op_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test__unsafe_masked_index_put_accumulate_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_add_complex10_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_add_complex3_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_add_complex4_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_add_complex7_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_add_complex9_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_aoti_eager_cache_hit_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_aoti_eager_dtype_device_layout_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_aoti_eager_support_out_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_arange1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_arange3_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_arange4_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_as_strided_on_views_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_avg_pool2d1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_avg_pool2d4_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_avg_pool2d_backward4_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_avg_pool2d_backward_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_avg_pool3d_backward2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bfloat16_to_int16_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bitwise_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bmm1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bmm2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bucketize_int_int16_int32_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bucketize_int_int16_int8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bucketize_int_int64_int64_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bucketize_int_int8_int32_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bucketize_int_int8_int8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bucketize_int_uint8_int8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bucketize_int_uint8_uint8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_buffer_use_after_remove_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_builtins_round_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_builtins_round_int_ndigits_zero_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_cat_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_cat_extern_kernel_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_cat_single_empty_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_cat_unbacked_2d_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_cat_unbacked_legacy_empty_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_check_stack_no_cycles_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_complex_fallback_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_concat_add_inplace_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_config_option_dont_assume_alignment_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_consecutive_split_cumsum_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_conv1d_with_permute_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_conv2d_channels_last_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_conv3d_channels_last_use_block_ptr_False_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_conv3d_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_conv_shape_check_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_convolution1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_convolution2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_convolution4_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_cpu_scalar_with_cpu_tensor_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_custom_op_fixed_layout_sequential_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dense_mask_index_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_deterministic_codegen_on_graph_break_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_deterministic_codegen_with_suffix_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_diagonal_copy_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dist_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_div2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_div6_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_div7_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_div8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_div_by_zero_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_div_presicion_accuracy_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_div_prim_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dropout_trivial_0_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dropout_trivial_1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_float16_float16_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_float16_int32_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_float16_int64_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_float16_int8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_float32_float16_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_float32_float32_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_float64_uint8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_int16_int64_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_int16_int8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_int32_int32_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_int32_int8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_int64_float64_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_int64_int32_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_int8_float64_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_int8_int8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_uint8_float32_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_uint8_int64_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_uint8_int8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_elu_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_embedding_sparse_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_empty1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_erfinv_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_fallback_mutable_op_list_tensor_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_fft_real_input_real_output_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_fill1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_float_index_expression_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_fmin_fmax_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_fmod_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_fractional_max_pool2d2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_fractional_max_pool2d4_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_fusing_write_into_disjoint_read_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_gather2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_gather3_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_gelu_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_generate_rand_fp8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_gpu_scalar_with_cpu_tensor_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_gpu_scalar_with_gpu_tensor_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_graph_partition_argmax_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_graph_partition_no_inputs_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_graph_partition_unbacked_symint_as_output_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_grid_sampler_2d_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_hardswish_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_index_propagation_remainder_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_index_put_deterministic_fallback_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_index_tensor_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_inductor_multiple_specializations_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_inductor_triton_bucketize_respects_masking_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_inf_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_inner_fn_str_and_stride_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_inplace_add_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_l1_loss_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_large_broadcast_reduction_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_large_grid_use_block_ptr_True_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_large_strided_reduction_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_lgamma_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_like_rands2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_linear_mixed_dtype_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_linspace2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_lite_dynamic_shape_assertion_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_lite_triton_kernel_wrapper_functional_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_logaddexp_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_low_memory_max_pool_dilation_2_dim_3_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_mark_dynamic_with_hint_override_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_masked_scatter_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_matmul_layer_norm_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_max_pool2d3_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_max_pool2d_with_indices_backward3_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_max_pool2d_with_indices_backward_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_mixed_mm3_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_mm_mixed_dtype_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_multi_threading_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_mutable_custom_op_fixed_layout2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_nan_sort_stable_True_descending_True_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_needs_contiguous_strides_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_new_empty_strided_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_nll_loss_backward_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_no_specization_over_symbolic_value_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pattern_matcher_multi_user_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_permute1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_bessel_j0_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_bessel_y0_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_erfc_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_erfinv_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_expm1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_gammaln_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_hermite_polynomial_h_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_i1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_i1e_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_log_ndtr_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_logit_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_multigammaln_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_scaled_modified_bessel_k0_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_shifted_chebyshev_polynomial_t_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_shifted_chebyshev_polynomial_w_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_xlog1py_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_zeta_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_polar_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pow1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pow2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pow_by_natural_log2_dynamic_shapes_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_prepare_softmax_with_fast_math_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_rand_like_deterministic_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_randn_with_dtype_and_device_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_reduction2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_remove_no_ops_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_repeat_as_strided_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_repeat_interleave_Tensor_decomp_int32_nd_1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_repeat_interleave_decomposition_has_clamp_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_resize_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_reuse_buffers_with_aliasing_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_scalar_cpu_tensor_arg_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_scalar_input_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_scaled_dot_product_efficient_attention_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_scatter2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_scatter3_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_scatter_add2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_scatter_add3_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_sdpa_prefer_nd_tiling_True_use_block_ptr_False_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_sgn_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_sgn_extremal_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_shape_prop_torch_ones_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_silu_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_single_elem_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_single_elem_indirect_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_slice2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_slice_mutation2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_softmax_backward_data_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_sort_bool_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_sort_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_sort_stable_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_split_cumprod_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_split_failed_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_split_with_integer_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_squeeze2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_sum1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_sum4_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_tensor_index_slice_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_tmp_not_defined_issue2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_tmp_not_defined_issue3_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_torch_device_split_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_transpose_add_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_transpose_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_triu_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_uint4x2_mixed_mm_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_unbacked_float_item_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_unbacked_floordiv_simplify_errors_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_unspec_inputs_int16_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_unspec_inputs_uint8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_upsample_nearest2d_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_var_correction_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_var_mean_tile_reduction_False_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_view_on_aliased_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_view_uint8_through_differing_bitwidths_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_views4_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_weight_norm_bwd_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_where_with_logical_op_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_zero_dim_reductions_dynamic_shapes_cuda 2025-12-04T10:24:45.6210179Z 2025-12-04T10:24:45.6210695Z Finished inductor/test_torchinductor_codegen_dynamic_shapes 4/4 ... [2025-12-04 10:24:45.559925][4269.25031588], took 9.20min 2025-12-04T10:24:45.6212404Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_codegen_dynamic_shapes/inductor.test_torchinductor_codegen_dynamic_shapes-3b673269684e2f76.xml 2025-12-04T10:24:45.6571207Z Running inductor/test_torchinductor_opinfo 5/17 ... [2025-12-04 10:24:45.656822][4269.347214502] 2025-12-04T10:24:45.6571856Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T10:24:45.6574984Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '--shard-id=5', '--num-shards=17', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 10:24:45.657243] 2025-12-04T10:35:56.5112668Z 2025-12-04T10:35:56.5116086Z inductor/test_torchinductor_opinfo 5/17 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_5.17_b8d59ea0e8d50306_.log 2025-12-04T10:35:56.5262333Z Running 242 items in this shard: test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_T_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___radd___cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rmul___cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive__chunk_cat_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive__chunk_cat_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive__segment_reduce_offsets_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive__unsafe_masked_index_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive__unsafe_masked_index_put_accumulate_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_abs_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_acos_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_acosh_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_addbmm_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_addbmm_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_addr_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_any_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_argmax_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_argmin_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_argsort_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_as_strided_partial_views_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_baddbmm_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bitwise_right_shift_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bool_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_broadcast_tensors_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_byte_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_byte_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cartesian_prod_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cartesian_prod_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cat_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cholesky_inverse_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_clamp_max_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_combinations_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_conj_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_constant_pad_nd_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_contiguous_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cosh_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_count_nonzero_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cummax_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cummax_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cummin_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cummin_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_diagonal_copy_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_diagonal_scatter_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_diff_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_digamma_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_div_floor_rounding_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_double_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_empty_like_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_empty_permuted_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_eq_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_equal_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_equal_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_erf_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_exp2_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_expand_copy_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_expand_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_expm1_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_fft_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_fft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_hfft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ifft_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ihfft2_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ihfft2_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ihfft2_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_irfft2_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_irfft2_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_flatten_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fliplr_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_flipud_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_floor_divide_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_floor_divide_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fmin_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fmin_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_full_like_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_geometric_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_grid_sampler_3d_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_gt_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_half_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_i0_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_index_fill_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_index_fill_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_index_reduce_prod_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isclose_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isposinf_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isreal_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_item_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_item_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_jiterator_2inputs_2outputs_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_jiterator_4inputs_with_extra_args_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_jiterator_4inputs_with_extra_args_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_jiterator_binary_return_by_ref_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_kron_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_ldexp_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_lgamma_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_lgamma_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_cholesky_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_cross_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_det_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_diagonal_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_householder_product_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_matrix_norm_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_matrix_rank_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_matrix_rank_hermitian_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_norm_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_solve_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_vander_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_vecdot_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log1p_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log1p_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log_softmax_with_dtype_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log_softmax_with_dtype_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logaddexp2_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logical_and_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logical_not_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logical_or_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logical_or_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logical_xor_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logit_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mH_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mT_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_argmax_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_fill_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_logaddexp_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_normalize_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_scatter_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_softmin_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_softmin_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_std_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_sum_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_sum_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_max_reduction_no_dim_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_median_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_median_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_meshgrid_list_of_tensors_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_meshgrid_variadic_tensors_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_min_reduction_no_dim_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_min_reduction_with_dim_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mode_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mul_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nansum_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_native_dropout_backward_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_ne_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_neg_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_new_empty_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_max_pool3d_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_cosine_embedding_loss_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_glu_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_group_norm_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_hinge_embedding_loss_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bicubic_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_linear_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_max_pool1d_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_max_unpool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_max_unpool3d_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_mse_loss_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_multi_head_attention_forward_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_multilabel_margin_loss_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_multilabel_soft_margin_loss_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pad_circular_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pad_reflect_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pad_replicate_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pixel_unshuffle_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_rms_norm_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_smooth_l1_loss_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_softmin_with_dtype_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_softsign_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_tanhshrink_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_triplet_margin_loss_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_unfold_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_norm_nuc_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_normal_number_mean_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_outer_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_permute_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_permute_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_polygamma_polygamma_n_0_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_polygamma_polygamma_n_3_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_positive_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_randint_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_renorm_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_resize__cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_roll_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rot90_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rot90_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_round_decimals_3_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsqrt_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_scatter_reduce_prod_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_select_scatter_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sign_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_signbit_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_signbit_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_signbit_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sinc_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sinh_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sinh_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_slice_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_softmax_with_dtype_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_softmax_with_dtype_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_softmax_with_dtype_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_airy_ai_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_airy_ai_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_bessel_j1_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_bessel_y0_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_chebyshev_polynomial_u_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_hermite_polynomial_h_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_hermite_polynomial_he_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_i0e_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_i0e_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_i1_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_i1e_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_modified_bessel_k0_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_modified_bessel_k1_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_ndtr_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_ndtr_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_polygamma_special_polygamma_n_0_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_t_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_u_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_w_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_split_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_split_list_args_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_split_with_sizes_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_squeeze_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stack_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sum_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_t_copy_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_to_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_to_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_trace_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_transpose_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_trapezoid_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_tril_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_triu_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_triu_indices_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_true_divide_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unfold_copy_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unravel_index_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_var_mean_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_vsplit_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_vsplit_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_zero__cuda_int64 2025-12-04T10:35:56.5403989Z 2025-12-04T10:35:56.5404412Z Finished inductor/test_torchinductor_opinfo 5/17 ... [2025-12-04 10:35:56.511541][4940.201931999], took 11.18min 2025-12-04T10:35:56.5405892Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-a02012f2395cf6ef.xml 2025-12-04T10:35:56.9113408Z Uploading artifacts took 0.31 seconds 2025-12-04T10:35:56.9117444Z Running inductor/test_torchinductor_opinfo 10/17 ... [2025-12-04 10:35:56.911556][4940.601948865] 2025-12-04T10:35:56.9118149Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T10:35:56.9121949Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '--shard-id=10', '--num-shards=17', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 10:35:56.911958] 2025-12-04T10:49:14.3899211Z 2025-12-04T10:49:14.3902386Z inductor/test_torchinductor_opinfo 10/17 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_10.17_664b3f49cadd1793_.log 2025-12-04T10:49:14.4034400Z Running 225 items in this shard: test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rdiv___cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rxor___cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive__native_batch_norm_legit_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_acos_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_acosh_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_addmm_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_all_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_allclose_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_amin_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_angle_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_arange_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_arange_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_as_strided_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_as_strided_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_as_strided_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_as_strided_partial_views_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_as_strided_scatter_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_as_strided_scatter_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_atan2_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bfloat16_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bfloat16_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_block_diag_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bool_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bool_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_broadcast_tensors_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_broadcast_to_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bucketize_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_chalf_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_chalf_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_clamp_min_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_combinations_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_combinations_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_conj_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_conj_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_conj_physical_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_conj_physical_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_constant_pad_nd_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_copysign_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_copysign_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_corrcoef_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_count_nonzero_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_diag_embed_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_diagflat_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_diagflat_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_diff_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_diff_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_dsplit_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_empty_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_empty_permuted_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_empty_strided_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_empty_strided_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_equal_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_erfc_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_erfinv_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_erfinv_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_erfinv_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_expand_copy_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_expand_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_expm1_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_fftn_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_fftn_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_fftn_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_fftshift_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_hfftn_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ihfft2_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ihfftn_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ihfftn_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_rfft2_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_flatten_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_flatten_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_flip_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_float_power_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_float_power_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fmod_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_gather_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_gcd_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_ge_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_geometric_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_gt_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_histc_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_hypot_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_i0_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_i0_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_igamma_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_index_put_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_index_put_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_index_reduce_amin_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_index_reduce_mean_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isfinite_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isinf_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isreal_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_jiterator_binary_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_jiterator_binary_return_by_ref_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_kron_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_le_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_cholesky_ex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_inv_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_lu_factor_ex_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_matrix_rank_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_multi_dot_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_norm_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_norm_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_pinv_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_qr_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_solve_ex_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_svdvals_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_vector_norm_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log_normal_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log_softmax_with_dtype_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logspace_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logspace_tensor_overload_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_lt_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mH_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mH_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_argmax_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_argmin_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_cumsum_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_fill_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_logsumexp_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_prod_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matrix_exp_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_max_reduction_with_dim_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_max_reduction_with_dim_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_median_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_meshgrid_list_of_tensors_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_meshgrid_variadic_tensors_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_movedim_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_msort_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mvlgamma_mvlgamma_p_1_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mvlgamma_mvlgamma_p_5_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_narrow_copy_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_narrow_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_ne_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_new_full_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_max_pool3d_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_cosine_embedding_loss_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_dropout2d_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_embedding_bag_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_feature_alpha_dropout_with_train_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_hinge_embedding_loss_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_leaky_relu_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_leaky_relu_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_max_pool1d_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_max_unpool3d_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pad_circular_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pixel_shuffle_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pixel_shuffle_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_relu_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_tanhshrink_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_tanhshrink_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_upsample_nearest_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nonzero_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_polygamma_polygamma_n_0_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_polygamma_polygamma_n_4_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_pow_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_pow_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_prod_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_put_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rad2deg_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rad2deg_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_randn_like_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_reciprocal_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_repeat_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_reshape_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_resize_as__cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_resize_as__cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_resolve_neg_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rot90_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_round_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_scatter_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_scatter_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_scatter_reduce_prod_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_scatter_reduce_sum_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_select_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_select_scatter_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_select_scatter_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_select_scatter_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_signal_windows_cosine_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_signbit_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sin_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_slice_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_slice_scatter_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_slice_scatter_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_softmax_with_dtype_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_bessel_j1_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_hermite_polynomial_he_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_i1e_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_laguerre_polynomial_l_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_modified_bessel_k1_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_polygamma_special_polygamma_n_0_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_v_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_xlog1py_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_split_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_split_list_args_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_split_with_sizes_copy_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_split_with_sizes_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sqrt_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_squeeze_copy_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_squeeze_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_std_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_std_unbiased_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sum_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sum_to_size_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sum_to_size_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_to_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_trapezoid_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_tril_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_tril_indices_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_triu_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_trunc_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unfold_copy_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unique_consecutive_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unique_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unsqueeze_copy_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_var_mean_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_view_as_complex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_vsplit_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_where_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_xlogy_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_zeros_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_zeros_like_cuda_float16 2025-12-04T10:49:14.4164411Z 2025-12-04T10:49:14.4164845Z Finished inductor/test_torchinductor_opinfo 10/17 ... [2025-12-04 10:49:14.389833][5738.080224586], took 13.29min 2025-12-04T10:49:14.4166341Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-96bec2e922d5155c.xml 2025-12-04T10:49:14.4899442Z Running inductor/test_torchinductor_opinfo 15/17 ... [2025-12-04 10:49:14.489569][5738.179958852] 2025-12-04T10:49:14.4900075Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T10:49:14.4903272Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '--shard-id=15', '--num-shards=17', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 10:49:14.490035] 2025-12-04T10:57:31.5804554Z 2025-12-04T10:57:31.5805881Z inductor/test_torchinductor_opinfo 15/17 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_15.17_68b5d6354fd5686b_.log 2025-12-04T10:57:31.5930418Z Running 210 items in this shard: test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_H_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___radd___cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rand___cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rdiv___cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rsub___cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive__chunk_cat_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive__segment_reduce_lengths_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive__unsafe_masked_index_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive__upsample_bilinear2d_aa_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_abs_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_acosh_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_addcmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_alias_copy_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_alias_copy_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_all_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_all_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_all_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_amax_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_aminmax_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_arange_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_arange_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_argsort_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_as_strided_copy_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_asin_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_asin_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_asinh_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_atan_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_atleast_2d_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bfloat16_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_block_diag_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_block_diag_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_byte_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_ceil_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_chunk_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_clamp_max_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_clone_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_clone_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cross_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cumulative_trapezoid_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_diff_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_div_no_rounding_mode_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_div_no_rounding_mode_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_equal_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_equal_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_exp2_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_exp_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_expand_copy_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_exponential_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_fft2_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_fft_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_fft_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_fftn_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_hfft_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ifft2_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ifft_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ifftshift_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ifftshift_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_irfft2_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_float_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_floor_divide_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fmax_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fmin_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_full_like_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_gcd_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_geometric_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_half_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_hash_tensor_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_heaviside_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_heaviside_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_hsplit_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_index_put_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_index_reduce_mean_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isfinite_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isneginf_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isposinf_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_kron_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_kthvalue_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_kthvalue_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_kthvalue_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_le_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_diagonal_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_eigvals_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_ldl_factor_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_ldl_solve_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_lstsq_grad_oriented_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_slogdet_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_solve_triangular_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_tensorsolve_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log10_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log2_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_argmin_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_cumsum_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_select_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_softmax_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_max_reduction_no_dim_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_max_reduction_no_dim_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_max_reduction_with_dim_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_meshgrid_list_of_tensors_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_meshgrid_variadic_tensors_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_min_binary_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_min_reduction_no_dim_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_min_reduction_no_dim_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_min_reduction_with_dim_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mul_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mvlgamma_mvlgamma_p_3_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nan_to_num_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nanquantile_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_narrow_copy_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_narrow_copy_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_ne_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_neg_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_new_empty_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_new_full_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_new_ones_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_new_zeros_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_new_zeros_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool3d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_avg_pool2d_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_binary_cross_entropy_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_dropout_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_glu_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_hardtanh_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_huber_loss_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_huber_loss_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_trilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_layer_norm_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_logsigmoid_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_max_pool3d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_multi_head_attention_forward_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_normalize_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pad_constant_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pad_replicate_negative_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pixel_shuffle_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pixel_shuffle_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_rms_norm_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_rrelu_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_soft_margin_loss_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_softmin_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_softsign_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_threshold_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_triplet_margin_loss_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nonzero_static_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_norm_fro_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_pca_lowrank_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_polygamma_polygamma_n_1_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_pow_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rand_like_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_randint_like_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_randn_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_ravel_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_real_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_real_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_resize_as__cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_resolve_conj_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_round_decimals_neg_3_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_scalar_tensor_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_scatter_add_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_scatter_add_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_scatter_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_scatter_reduce_mean_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_select_scatter_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sgn_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_short_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sigmoid_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sign_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sin_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sinc_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sinh_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_softmax_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_softmax_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_softmax_with_dtype_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_airy_ai_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_chebyshev_polynomial_t_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_chebyshev_polynomial_u_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_hermite_polynomial_he_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_laguerre_polynomial_l_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_ndtr_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_ndtr_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_polygamma_special_polygamma_n_0_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_scaled_modified_bessel_k1_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_u_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_v_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_split_list_args_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_split_with_sizes_copy_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_std_mean_unbiased_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sum_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_t_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_tensor_split_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_tile_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_to_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_to_sparse_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_trapz_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_trapz_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_trapz_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_triu_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_true_divide_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unique_consecutive_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unsafe_split_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unsafe_split_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unsafe_split_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unsqueeze_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_vdot_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_view_as_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_view_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_vstack_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_where_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_zeros_like_cuda_float64 2025-12-04T10:57:31.6053170Z 2025-12-04T10:57:31.6053593Z Finished inductor/test_torchinductor_opinfo 15/17 ... [2025-12-04 10:57:31.580648][6235.271038734], took 8.28min 2025-12-04T10:57:31.6055197Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-ef6350a9f8bb75e3.xml 2025-12-04T10:57:31.9921872Z Uploading artifacts took 0.31 seconds 2025-12-04T10:57:31.9925870Z Running dynamo/test_logging 1/1 ... [2025-12-04 10:57:31.992404][6235.682796116] 2025-12-04T10:57:31.9926422Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T10:57:31.9930664Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_logging.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 10:57:31.992837] 2025-12-04T10:58:15.4227686Z 2025-12-04T10:58:15.4228727Z dynamo/test_logging 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_logging_1.1_65e572a46b19d5f7_.log 2025-12-04T10:58:15.4246232Z Running 51 items in this shard: test/dynamo/test_logging.py::LoggingTests::test_all, test/dynamo/test_logging.py::LoggingTests::test_aot, test/dynamo/test_logging.py::LoggingTests::test_aot_graphs, test/dynamo/test_logging.py::LoggingTests::test_aot_joint_graph, test/dynamo/test_logging.py::LoggingTests::test_autotuning, test/dynamo/test_logging.py::LoggingTests::test_bytecode, test/dynamo/test_logging.py::LoggingTests::test_cudagraph_static_inputs, test/dynamo/test_logging.py::LoggingTests::test_cudagraphs, test/dynamo/test_logging.py::LoggingTests::test_custom_format, test/dynamo/test_logging.py::LoggingTests::test_custom_format_exc, test/dynamo/test_logging.py::LoggingTests::test_ddp_graphs, test/dynamo/test_logging.py::LoggingTests::test_default_logging, test/dynamo/test_logging.py::LoggingTests::test_distributed_rank_logging, test/dynamo/test_logging.py::LoggingTests::test_dump_compile_times, test/dynamo/test_logging.py::LoggingTests::test_dynamo_debug, test/dynamo/test_logging.py::LoggingTests::test_dynamo_debug_default_off_artifacts, test/dynamo/test_logging.py::LoggingTests::test_dynamo_error, test/dynamo/test_logging.py::LoggingTests::test_dynamo_info, test/dynamo/test_logging.py::LoggingTests::test_fusion, test/dynamo/test_logging.py::LoggingTests::test_graph_breaks, test/dynamo/test_logging.py::LoggingTests::test_graph_region_expansion, test/dynamo/test_logging.py::LoggingTests::test_guards_polyfill_sloc, test/dynamo/test_logging.py::LoggingTests::test_guards_recompiles, test/dynamo/test_logging.py::LoggingTests::test_guards_sloc, test/dynamo/test_logging.py::LoggingTests::test_guards_sloc_vr, test/dynamo/test_logging.py::LoggingTests::test_hierarchical_compile, test/dynamo/test_logging.py::LoggingTests::test_inductor_debug, test/dynamo/test_logging.py::LoggingTests::test_inductor_error, test/dynamo/test_logging.py::LoggingTests::test_inductor_info, test/dynamo/test_logging.py::LoggingTests::test_invalid_artifact_flag, test/dynamo/test_logging.py::LoggingTests::test_invalid_artifact_flag_error_msg, test/dynamo/test_logging.py::LoggingTests::test_kernel_code, test/dynamo/test_logging.py::LoggingTests::test_log_traced_frames, test/dynamo/test_logging.py::LoggingTests::test_logs_out, test/dynamo/test_logging.py::LoggingTests::test_multiline_format, test/dynamo/test_logging.py::LoggingTests::test_open_registration, test/dynamo/test_logging.py::LoggingTests::test_open_registration_python_api, test/dynamo/test_logging.py::LoggingTests::test_open_registration_with_registered_parent, test/dynamo/test_logging.py::LoggingTests::test_optimizer_non_static_param, test/dynamo/test_logging.py::LoggingTests::test_output_code, test/dynamo/test_logging.py::LoggingTests::test_recompiles, test/dynamo/test_logging.py::LoggingTests::test_schedule, test/dynamo/test_logging.py::LoggingTests::test_trace_call, test/dynamo/test_logging.py::LoggingTests::test_trace_call_graph_break, test/dynamo/test_logging.py::LoggingTests::test_trace_call_inline_call, test/dynamo/test_logging.py::LoggingTests::test_trace_call_prefix, test/dynamo/test_logging.py::LoggingTests::test_trace_source_cond, test/dynamo/test_logging.py::LoggingTests::test_trace_source_funcname, test/dynamo/test_logging.py::LoggingTests::test_trace_source_if_stmt, test/dynamo/test_logging.py::LoggingTests::test_trace_source_nested, test/dynamo/test_logging.py::LoggingTests::test_trace_source_simple 2025-12-04T10:58:15.4262814Z 2025-12-04T10:58:15.4263132Z Finished dynamo/test_logging 1/1 ... [2025-12-04 10:58:15.422608][6279.11300239], took 0.72min 2025-12-04T10:58:15.4268195Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_logging/dynamo.test_logging-77311c728804d279.xml 2025-12-04T10:58:15.5498567Z Running dynamo/test_repros 1/1 ... [2025-12-04 10:58:15.549537][6279.239927817] 2025-12-04T10:58:15.5499124Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T10:58:15.5502384Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_repros.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 10:58:15.549999] 2025-12-04T11:01:01.1542137Z 2025-12-04T11:01:01.1543441Z dynamo/test_repros 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_repros_1.1_3353f5f590b0aac4_.log 2025-12-04T11:01:01.1665102Z Running 351 items in this shard: test/dynamo/test_repros.py::LRUCacheWarningTests::test_lru_cache_warning_issued_during_tracing, test/dynamo/test_repros.py::ReproTests::test_312_local_cell_overlap, test/dynamo/test_repros.py::ReproTests::test_Size, test/dynamo/test_repros.py::ReproTests::test_abc_setattr, test/dynamo/test_repros.py::ReproTests::test_add_complex_conj, test/dynamo/test_repros.py::ReproTests::test_add_sub_alpha_out, test/dynamo/test_repros.py::ReproTests::test_addr_alpha_beta_out, test/dynamo/test_repros.py::ReproTests::test_amp_foreach_fake_impl, test/dynamo/test_repros.py::ReproTests::test_aot_autograd_runtime_wrapper_prologue_profiled, test/dynamo/test_repros.py::ReproTests::test_as_strided_on_base_with_mutation_works, test/dynamo/test_repros.py::ReproTests::test_as_strided_on_existing_view_banned, test/dynamo/test_repros.py::ReproTests::test_attached_attribute_in_dir, test/dynamo/test_repros.py::ReproTests::test_autograd_function_graph_break, test/dynamo/test_repros.py::ReproTests::test_avoid_dupe_specialization, test/dynamo/test_repros.py::ReproTests::test_batch_encoding_clone_inputs, test/dynamo/test_repros.py::ReproTests::test_batch_norm_act, test/dynamo/test_repros.py::ReproTests::test_batchnorm_e2e, test/dynamo/test_repros.py::ReproTests::test_bigbird_unsqueeze_inplace, test/dynamo/test_repros.py::ReproTests::test_bitwise_op_guard, test/dynamo/test_repros.py::ReproTests::test_bitwise_print_precedence, test/dynamo/test_repros.py::ReproTests::test_boxes_len, test/dynamo/test_repros.py::ReproTests::test_build_map_unpack_with_call, test/dynamo/test_repros.py::ReproTests::test_c_defined_metaclass, test/dynamo/test_repros.py::ReproTests::test_cells_unsupported_step_exception, test/dynamo/test_repros.py::ReproTests::test_changing_stride, test/dynamo/test_repros.py::ReproTests::test_chunk_reformer_ff, test/dynamo/test_repros.py::ReproTests::test_class_member, test/dynamo/test_repros.py::ReproTests::test_classmethod_with_slots, test/dynamo/test_repros.py::ReproTests::test_clone_not_memory_dense, test/dynamo/test_repros.py::ReproTests::test_compilation_metrics_on_error, test/dynamo/test_repros.py::ReproTests::test_compile_complex_conj, test/dynamo/test_repros.py::ReproTests::test_compile_copy__int_overload, test/dynamo/test_repros.py::ReproTests::test_compiled_module_truthiness, test/dynamo/test_repros.py::ReproTests::test_const_dict_keyerror, test/dynamo/test_repros.py::ReproTests::test_contains_range_constprop, test/dynamo/test_repros.py::ReproTests::test_convert_boxes_to_pooler_format, test/dynamo/test_repros.py::ReproTests::test_copy_weird_strides, test/dynamo/test_repros.py::ReproTests::test_create_rand_mask_from_inputs, test/dynamo/test_repros.py::ReproTests::test_dalle2_maybe, test/dynamo/test_repros.py::ReproTests::test_data_attr_mutation_after_saved_for_bw, test/dynamo/test_repros.py::ReproTests::test_dataclass_in_module, test/dynamo/test_repros.py::ReproTests::test_dataclass_init_with_default_factory_with_inputs, test/dynamo/test_repros.py::ReproTests::test_ddp_checkpoint, test/dynamo/test_repros.py::ReproTests::test_dedup_global, test/dynamo/test_repros.py::ReproTests::test_deferred_runtime_asserts, test/dynamo/test_repros.py::ReproTests::test_delattr, test/dynamo/test_repros.py::ReproTests::test_delattr_raises, test/dynamo/test_repros.py::ReproTests::test_delattr_return, test/dynamo/test_repros.py::ReproTests::test_delete_local_error, test/dynamo/test_repros.py::ReproTests::test_deleted_compile_wrapper_segfault, test/dynamo/test_repros.py::ReproTests::test_delsubscr, test/dynamo/test_repros.py::ReproTests::test_delsubscr_raises, test/dynamo/test_repros.py::ReproTests::test_detectron2_instances_cat, test/dynamo/test_repros.py::ReproTests::test_disabling_unpack_hooks_within_compiled_region, test/dynamo/test_repros.py::ReproTests::test_distributions_subclass, test/dynamo/test_repros.py::ReproTests::test_do_paste_mask, test/dynamo/test_repros.py::ReproTests::test_dont_aggressively_write_assert, test/dynamo/test_repros.py::ReproTests::test_dont_dce_rand, test/dynamo/test_repros.py::ReproTests::test_dropout_inline, test/dynamo/test_repros.py::ReproTests::test_dynamic_shape_disable_duck_size, test/dynamo/test_repros.py::ReproTests::test_dynamic_shapes_double_not_equal, test/dynamo/test_repros.py::ReproTests::test_dynamic_shapes_float_guard, test/dynamo/test_repros.py::ReproTests::test_dynamic_shapes_implicit_guard, test/dynamo/test_repros.py::ReproTests::test_dynamic_shapes_right_side, test/dynamo/test_repros.py::ReproTests::test_dynamo_default_lru_cache_behavior, test/dynamo/test_repros.py::ReproTests::test_dynamo_disable_lru_cache_behavior, test/dynamo/test_repros.py::ReproTests::test_dynamo_set_recursion_limit, test/dynamo/test_repros.py::ReproTests::test_dynamo_set_recursion_limit_usage, test/dynamo/test_repros.py::ReproTests::test_ellipsis, test/dynamo/test_repros.py::ReproTests::test_embedding_backward_broadcasting_decomp, test/dynamo/test_repros.py::ReproTests::test_empty_graph_nested_calls_fullgraph_False, test/dynamo/test_repros.py::ReproTests::test_empty_graph_nested_calls_fullgraph_True, test/dynamo/test_repros.py::ReproTests::test_empty_list_contains_with_jump, test/dynamo/test_repros.py::ReproTests::test_empty_out_dynamic, test/dynamo/test_repros.py::ReproTests::test_enum, test/dynamo/test_repros.py::ReproTests::test_ephemeral_module, test/dynamo/test_repros.py::ReproTests::test_error_return_without_exception_set, test/dynamo/test_repros.py::ReproTests::test_exception_in_dynamo_handling, test/dynamo/test_repros.py::ReproTests::test_exec_import, test/dynamo/test_repros.py::ReproTests::test_exec_wildcard_import, test/dynamo/test_repros.py::ReproTests::test_export_vs_dynamo_for_multiheadattention, test/dynamo/test_repros.py::ReproTests::test_flip_bad_accuracy, test/dynamo/test_repros.py::ReproTests::test_for_loop_graph_break, test/dynamo/test_repros.py::ReproTests::test_for_loop_graph_break_before, test/dynamo/test_repros.py::ReproTests::test_foreach_decomp_arg_names, test/dynamo/test_repros.py::ReproTests::test_fsdp_set_input_mutation_applied_when_input_gets_no_gradients, test/dynamo/test_repros.py::ReproTests::test_function_in_skipfiles, test/dynamo/test_repros.py::ReproTests::test_functools_wraps, test/dynamo/test_repros.py::ReproTests::test_gan_repro_trying_to_backward_through_the_graph_a_second_time, test/dynamo/test_repros.py::ReproTests::test_generator_dealloc, test/dynamo/test_repros.py::ReproTests::test_get_parameter_dtype, test/dynamo/test_repros.py::ReproTests::test_get_type_hints, test/dynamo/test_repros.py::ReproTests::test_global_fn_mutation, test/dynamo/test_repros.py::ReproTests::test_grad, test/dynamo/test_repros.py::ReproTests::test_grad_mode_carrying_correct_state_after_graph_break, test/dynamo/test_repros.py::ReproTests::test_grad_references_cleared, test/dynamo/test_repros.py::ReproTests::test_graph_break_on_jit_isinstance, test/dynamo/test_repros.py::ReproTests::test_graph_break_on_jit_isinstance_pep585, test/dynamo/test_repros.py::ReproTests::test_graph_break_unsupported_fake, test/dynamo/test_repros.py::ReproTests::test_guard_default_device, test/dynamo/test_repros.py::ReproTests::test_guard_fail_nested_tuple, test/dynamo/test_repros.py::ReproTests::test_guard_fail_tensor_bool, test/dynamo/test_repros.py::ReproTests::test_guard_ordering_shape_fail, test/dynamo/test_repros.py::ReproTests::test_guard_same_frame_fail_message, test/dynamo/test_repros.py::ReproTests::test_guard_with_tuple_mutation, test/dynamo/test_repros.py::ReproTests::test_hasattr_builtin, test/dynamo/test_repros.py::ReproTests::test_hf_bigbird_unsqueeze, test/dynamo/test_repros.py::ReproTests::test_hf_classinstantier, test/dynamo/test_repros.py::ReproTests::test_hf_gelu_inline, test/dynamo/test_repros.py::ReproTests::test_hf_model_output, test/dynamo/test_repros.py::ReproTests::test_hf_t5_forward, test/dynamo/test_repros.py::ReproTests::test_hf_xsoftmax_inference, test/dynamo/test_repros.py::ReproTests::test_hf_xsoftmax_training, test/dynamo/test_repros.py::ReproTests::test_iadd_graph_break, test/dynamo/test_repros.py::ReproTests::test_incompatible_configs, test/dynamo/test_repros.py::ReproTests::test_indexing_with_list, test/dynamo/test_repros.py::ReproTests::test_inductor_dynamic_shapes_broadcasting, test/dynamo/test_repros.py::ReproTests::test_inductor_no_recursionerror_on_for_loops, test/dynamo/test_repros.py::ReproTests::test_inductor_rng_default_dtype, test/dynamo/test_repros.py::ReproTests::test_inference_mode_dynamic_shapes, test/dynamo/test_repros.py::ReproTests::test_inlining_cornercase, test/dynamo/test_repros.py::ReproTests::test_inplace_unsqueeze_input, test/dynamo/test_repros.py::ReproTests::test_int_format, test/dynamo/test_repros.py::ReproTests::test_intermediate_leaf_requires_grad, test/dynamo/test_repros.py::ReproTests::test_invalid_seq_unpack, test/dynamo/test_repros.py::ReproTests::test_is_make_fx_tracing, test/dynamo/test_repros.py::ReproTests::test_is_symbolic_tracing, test/dynamo/test_repros.py::ReproTests::test_isinstance_dtype, test/dynamo/test_repros.py::ReproTests::test_isinstance_storage, test/dynamo/test_repros.py::ReproTests::test_issue111522, test/dynamo/test_repros.py::ReproTests::test_issue111918, test/dynamo/test_repros.py::ReproTests::test_issue114171, test/dynamo/test_repros.py::ReproTests::test_issue126128, test/dynamo/test_repros.py::ReproTests::test_issue134451, test/dynamo/test_repros.py::ReproTests::test_issue1466_size_aot_autograd, test/dynamo/test_repros.py::ReproTests::test_issue164247_backend_eager, test/dynamo/test_repros.py::ReproTests::test_issue164247_backend_inductor, test/dynamo/test_repros.py::ReproTests::test_issue175, test/dynamo/test_repros.py::ReproTests::test_jit_script_defaults, test/dynamo/test_repros.py::ReproTests::test_jit_trace_errors, test/dynamo/test_repros.py::ReproTests::test_kwargs_out_list_variable, test/dynamo/test_repros.py::ReproTests::test_list_aliasing, test/dynamo/test_repros.py::ReproTests::test_list_index, test/dynamo/test_repros.py::ReproTests::test_list_index_not_found, test/dynamo/test_repros.py::ReproTests::test_list_index_tensor_unsupported, test/dynamo/test_repros.py::ReproTests::test_list_reverse, test/dynamo/test_repros.py::ReproTests::test_list_self_reference, test/dynamo/test_repros.py::ReproTests::test_listcomp, test/dynamo/test_repros.py::ReproTests::test_longformer_chunk, test/dynamo/test_repros.py::ReproTests::test_longtensor_list, test/dynamo/test_repros.py::ReproTests::test_lru_cache_tracing, test/dynamo/test_repros.py::ReproTests::test_maml_item_capture, test/dynamo/test_repros.py::ReproTests::test_maml_no_item_capture, test/dynamo/test_repros.py::ReproTests::test_many_overlapping_inputs_does_not_explode_guards, test/dynamo/test_repros.py::ReproTests::test_many_views_with_mutation, test/dynamo/test_repros.py::ReproTests::test_map_with_multiple_args, test/dynamo/test_repros.py::ReproTests::test_maybe_multiply_symint, test/dynamo/test_repros.py::ReproTests::test_mem_leak_guards, test/dynamo/test_repros.py::ReproTests::test_merge_criteria_processor_list1, test/dynamo/test_repros.py::ReproTests::test_merge_criteria_processor_list2, test/dynamo/test_repros.py::ReproTests::test_method_overriding, test/dynamo/test_repros.py::ReproTests::test_module_in_skipfiles, test/dynamo/test_repros.py::ReproTests::test_modules, test/dynamo/test_repros.py::ReproTests::test_multi_dot_import, test/dynamo/test_repros.py::ReproTests::test_multi_import, test/dynamo/test_repros.py::ReproTests::test_named_buffers, test/dynamo/test_repros.py::ReproTests::test_nanmean_out, test/dynamo/test_repros.py::ReproTests::test_negative_floor_div_solve, test/dynamo/test_repros.py::ReproTests::test_negative_shape_guard, test/dynamo/test_repros.py::ReproTests::test_nested_while_loop_graph_break, test/dynamo/test_repros.py::ReproTests::test_nn_module_callable, test/dynamo/test_repros.py::ReproTests::test_nn_module_property_closure, test/dynamo/test_repros.py::ReproTests::test_nn_module_stack_bc, test/dynamo/test_repros.py::ReproTests::test_nn_param_freevar_codegen, test/dynamo/test_repros.py::ReproTests::test_nn_parameter, test/dynamo/test_repros.py::ReproTests::test_nn_parameter_ctor_graph_breaks, test/dynamo/test_repros.py::ReproTests::test_nn_parametrize, test/dynamo/test_repros.py::ReproTests::test_no_grad_inline, test/dynamo/test_repros.py::ReproTests::test_no_tracing_into_eval_frame, test/dynamo/test_repros.py::ReproTests::test_no_tracing_into_eval_frame_ctx_manager, test/dynamo/test_repros.py::ReproTests::test_nonconst_issubclass, test/dynamo/test_repros.py::ReproTests::test_not_rewrite_assert_for_other_errors, test/dynamo/test_repros.py::ReproTests::test_nullcontext1, test/dynamo/test_repros.py::ReproTests::test_nullcontext2, test/dynamo/test_repros.py::ReproTests::test_numpy_not_ndarray_recompiles, test/dynamo/test_repros.py::ReproTests::test_numpy_tobytes_no_error, test/dynamo/test_repros.py::ReproTests::test_odict_get_item_index_name, test/dynamo/test_repros.py::ReproTests::test_omegaconf_dictconfig, test/dynamo/test_repros.py::ReproTests::test_omegaconf_listconfig_contains, test/dynamo/test_repros.py::ReproTests::test_omegaconf_listconfig_iter, test/dynamo/test_repros.py::ReproTests::test_ones_out_dynamic, test/dynamo/test_repros.py::ReproTests::test_optim_state_references_cleared, test/dynamo/test_repros.py::ReproTests::test_optimized_deepcopy, test/dynamo/test_repros.py::ReproTests::test_optimized_module_patched_init, test/dynamo/test_repros.py::ReproTests::test_optimized_module_training, test/dynamo/test_repros.py::ReproTests::test_os_fspath, test/dynamo/test_repros.py::ReproTests::test_out_nested_cell_shape_change, test/dynamo/test_repros.py::ReproTests::test_out_nested_cell_tuple_shape_change, test/dynamo/test_repros.py::ReproTests::test_out_none, test/dynamo/test_repros.py::ReproTests::test_out_overload_non_contiguous, test/dynamo/test_repros.py::ReproTests::test_out_root_cell_shape_change, test/dynamo/test_repros.py::ReproTests::test_out_root_cell_tuple_shape_change, test/dynamo/test_repros.py::ReproTests::test_output_aliases_intermediate, test/dynamo/test_repros.py::ReproTests::test_overlapping_inputs_with_dynamic_shapes_error, test/dynamo/test_repros.py::ReproTests::test_overwriting_params, test/dynamo/test_repros.py::ReproTests::test_partially_initialized_module_property, test/dynamo/test_repros.py::ReproTests::test_partitioner_activation_memory_budget_with_unbacked_symints, test/dynamo/test_repros.py::ReproTests::test_partitioner_cse_respects_mutation_boundaries, test/dynamo/test_repros.py::ReproTests::test_pointless_graph_removal, test/dynamo/test_repros.py::ReproTests::test_preserve_stride_with_clone, test/dynamo/test_repros.py::ReproTests::test_primtorch, test/dynamo/test_repros.py::ReproTests::test_primtorch_no_graph_break, test/dynamo/test_repros.py::ReproTests::test_randint_out_dynamic, test/dynamo/test_repros.py::ReproTests::test_recursive_map, test/dynamo/test_repros.py::ReproTests::test_reformer_eval, test/dynamo/test_repros.py::ReproTests::test_reformer_min_chunk_len, test/dynamo/test_repros.py::ReproTests::test_reformer_sorting, test/dynamo/test_repros.py::ReproTests::test_reformer_train, test/dynamo/test_repros.py::ReproTests::test_reinplacing, test/dynamo/test_repros.py::ReproTests::test_relative_import, test/dynamo/test_repros.py::ReproTests::test_relative_import_no_modulename, test/dynamo/test_repros.py::ReproTests::test_requires_grad_guards_with_grad_mode1, test/dynamo/test_repros.py::ReproTests::test_requires_grad_guards_with_grad_mode2, test/dynamo/test_repros.py::ReproTests::test_restricted_list_subclass1, test/dynamo/test_repros.py::ReproTests::test_restricted_list_subclass2, test/dynamo/test_repros.py::ReproTests::test_restricted_list_subclass3, test/dynamo/test_repros.py::ReproTests::test_return_value_duplication_mixed_grad, test/dynamo/test_repros.py::ReproTests::test_return_value_duplication_scalar, test/dynamo/test_repros.py::ReproTests::test_return_value_duplication_tensor, test/dynamo/test_repros.py::ReproTests::test_return_weakref, test/dynamo/test_repros.py::ReproTests::test_rewrite_assert_dont_change_bytecode, test/dynamo/test_repros.py::ReproTests::test_rewrite_assert_noop, test/dynamo/test_repros.py::ReproTests::test_rewrite_assert_with_msg, test/dynamo/test_repros.py::ReproTests::test_rewrite_assert_with_non_string_msg, test/dynamo/test_repros.py::ReproTests::test_rewrite_assert_without_msg, test/dynamo/test_repros.py::ReproTests::test_rng_state, test/dynamo/test_repros.py::ReproTests::test_seq_append_list, test/dynamo/test_repros.py::ReproTests::test_setattr_requires_grad_graph_breaks, test/dynamo/test_repros.py::ReproTests::test_setitem_boolean_mask_diff, test/dynamo/test_repros.py::ReproTests::test_setitem_tensor_prop, test/dynamo/test_repros.py::ReproTests::test_setitem_tuple_boolean_mask_diff, test/dynamo/test_repros.py::ReproTests::test_sigmoid_out, test/dynamo/test_repros.py::ReproTests::test_sigmoid_out2, test/dynamo/test_repros.py::ReproTests::test_size_typematch, test/dynamo/test_repros.py::ReproTests::test_slice_into_list_mutable, test/dynamo/test_repros.py::ReproTests::test_slicing_dynamic_shape, test/dynamo/test_repros.py::ReproTests::test_slicing_dynamic_shape_setitem, test/dynamo/test_repros.py::ReproTests::test_sort_out, test/dynamo/test_repros.py::ReproTests::test_sort_out2, test/dynamo/test_repros.py::ReproTests::test_specialized_stride, test/dynamo/test_repros.py::ReproTests::test_split_with_sizes_aot_autograd, test/dynamo/test_repros.py::ReproTests::test_staticmethod_allow_in_graph, test/dynamo/test_repros.py::ReproTests::test_stk_sdd_is_transposed, test/dynamo/test_repros.py::ReproTests::test_stop_iteration_reconstruct, test/dynamo/test_repros.py::ReproTests::test_str_isalnum, test/dynamo/test_repros.py::ReproTests::test_string_format, test/dynamo/test_repros.py::ReproTests::test_subclass_graph_output_repro, test/dynamo/test_repros.py::ReproTests::test_super_classmethod, test/dynamo/test_repros.py::ReproTests::test_super_classmethod_inheritance, test/dynamo/test_repros.py::ReproTests::test_super_diamond, test/dynamo/test_repros.py::ReproTests::test_super_in_staticmethod, test/dynamo/test_repros.py::ReproTests::test_super_staticmethod, test/dynamo/test_repros.py::ReproTests::test_swin_base_tensor_attr, test/dynamo/test_repros.py::ReproTests::test_symint_bitwise, test/dynamo/test_repros.py::ReproTests::test_symnode_is_not_op, test/dynamo/test_repros.py::ReproTests::test_symnode_is_op, test/dynamo/test_repros.py::ReproTests::test_sys_monitoring, test/dynamo/test_repros.py::ReproTests::test_tensor_data_kwarg, test/dynamo/test_repros.py::ReproTests::test_tensor_isinstance_tuple, test/dynamo/test_repros.py::ReproTests::test_tensor_item, test/dynamo/test_repros.py::ReproTests::test_tensor_random, test/dynamo/test_repros.py::ReproTests::test_tensor_set_data_backend_aot_eager_func_name_func1, test/dynamo/test_repros.py::ReproTests::test_tensor_set_data_backend_aot_eager_func_name_func2, test/dynamo/test_repros.py::ReproTests::test_tensor_set_data_backend_aot_eager_func_name_func3, test/dynamo/test_repros.py::ReproTests::test_tensor_set_data_backend_eager_func_name_func1, test/dynamo/test_repros.py::ReproTests::test_tensor_set_data_backend_eager_func_name_func2, test/dynamo/test_repros.py::ReproTests::test_tensor_set_data_backend_eager_func_name_func3, test/dynamo/test_repros.py::ReproTests::test_tensor_set_data_backend_inductor_func_name_func1, test/dynamo/test_repros.py::ReproTests::test_tensor_set_data_backend_inductor_func_name_func2, test/dynamo/test_repros.py::ReproTests::test_tensor_set_data_backend_inductor_func_name_func3, test/dynamo/test_repros.py::ReproTests::test_tensor_set_data_mismatched_dtype, test/dynamo/test_repros.py::ReproTests::test_tensor_split, test/dynamo/test_repros.py::ReproTests::test_tensor_split_within_device_cm, test/dynamo/test_repros.py::ReproTests::test_tensor_uniform, test/dynamo/test_repros.py::ReproTests::test_threading_local, test/dynamo/test_repros.py::ReproTests::test_tokenization, test/dynamo/test_repros.py::ReproTests::test_torch_compile_in_compile_frame, test/dynamo/test_repros.py::ReproTests::test_torch_ops_aten, test/dynamo/test_repros.py::ReproTests::test_torch_tensor_ops, test/dynamo/test_repros.py::ReproTests::test_torch_tensor_ops_no_graph_break, test/dynamo/test_repros.py::ReproTests::test_torch_variable_type, test/dynamo/test_repros.py::ReproTests::test_torchname, test/dynamo/test_repros.py::ReproTests::test_trace_functional_tensor_with, test/dynamo/test_repros.py::ReproTests::test_tuple_enum_as_key_dict, test/dynamo/test_repros.py::ReproTests::test_typed_dict, test/dynamo/test_repros.py::ReproTests::test_typed_dict_total, test/dynamo/test_repros.py::ReproTests::test_udf_classes_reconstruction, test/dynamo/test_repros.py::ReproTests::test_unbacked_arange_in_bounds, test/dynamo/test_repros.py::ReproTests::test_unbind_copy_out, test/dynamo/test_repros.py::ReproTests::test_unpack_hooks_can_be_disabled, test/dynamo/test_repros.py::ReproTests::test_unpack_hooks_dont_run_during_tracing, test/dynamo/test_repros.py::ReproTests::test_unspecialized_nn_module_with_torch_variable_attribute, test/dynamo/test_repros.py::ReproTests::test_unsqueeze_mul_strides, test/dynamo/test_repros.py::ReproTests::test_user_ctor_ctx_manager, test/dynamo/test_repros.py::ReproTests::test_user_ctor_ctx_manager_custom_init, test/dynamo/test_repros.py::ReproTests::test_user_ctor_ctx_manager_custom_init_graph_break, test/dynamo/test_repros.py::ReproTests::test_user_defined_iter, test/dynamo/test_repros.py::ReproTests::test_user_defined_object_callable, test/dynamo/test_repros.py::ReproTests::test_validate_model_kwargs, test/dynamo/test_repros.py::ReproTests::test_vc_bumped_in_inference_graph, test/dynamo/test_repros.py::ReproTests::test_vdd_duplicate_error, test/dynamo/test_repros.py::ReproTests::test_view_dtype_overload, test/dynamo/test_repros.py::ReproTests::test_weakref, test/dynamo/test_repros.py::ReproTests::test_weakref_callback, test/dynamo/test_repros.py::ReproTests::test_weakref_construction, test/dynamo/test_repros.py::ReproTests::test_weakref_del, test/dynamo/test_repros.py::ReproTests::test_weakref_proxy, test/dynamo/test_repros.py::ReproTests::test_weakref_reconstruct, test/dynamo/test_repros.py::ReproTests::test_while_loop_graph_break, test/dynamo/test_repros.py::ReproTests::test_while_loop_graph_break_inside_call_function, test/dynamo/test_repros.py::ReproTests::test_with_on_graph_break_inst, test/dynamo/test_repros.py::ReproTests::test_with_on_graph_break_nested, test/dynamo/test_repros.py::ReproTests::test_zeros_out_dynamic, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_cuda_sync_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_current_accelerator_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_data_dependent_error_log_no_print_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_deepcopy_constant_tensor_in_aot_bwd_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_filter_safe_grad_warning_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_filter_user_warnings_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_filter_warnings_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_flash_attn_backward_mixed_strides_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_getattr_return_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_guard_default_device_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_megablocks_moe_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_memleak_when_graph_input_has_tensor_attr_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_module_attribute_error_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_named_tuple_vt_clone_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_norm_dtype_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_partial_export_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_partitioner_saves_weights_for_bw_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_pytree_get_node_type_not_traced_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_pytree_get_node_type_with_namedtuple_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_pytree_tree_is_leaf_not_traced_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_pytree_tree_is_leaf_with_namedtuple_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_sdpa_dynamic_shapes_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_sub_alpha_scalar_repro_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_tensor_size_hasattr_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_torch_cuda_is_initialized_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_truthiness_of_symints_no_recompiles_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_udf_class_source_cuda, test/dynamo/test_repros.py::ReproTestsDeviceCUDA::test_zero_dim_param_mixed_device_grad_cuda 2025-12-04T11:01:01.1783668Z 2025-12-04T11:01:01.1783996Z Finished dynamo/test_repros 1/1 ... [2025-12-04 11:01:01.154573][6444.844958886], took 2.76min 2025-12-04T11:01:01.1785175Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_repros/dynamo.test_repros-9b6b927930117494.xml 2025-12-04T11:01:01.2660083Z Running inductor/test_cuda_select_algorithm 5/5 ... [2025-12-04 11:01:01.265730][6444.95612132] 2025-12-04T11:01:01.2660728Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T11:01:01.2664030Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_cuda_select_algorithm.py', '--shard-id=5', '--num-shards=5', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 11:01:01.266153] 2025-12-04T11:17:05.9629799Z 2025-12-04T11:17:05.9631274Z PRINTING LOG FILE of inductor/test_cuda_select_algorithm 5/5 (test/test-reports/inductor.test_cuda_select_algorithm_5.5_387c3291169f376b_.log) 2025-12-04T11:17:05.9633387Z W1204 11:01:10.199000 91548 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:05.9635489Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-97143c683457795e.xml 2025-12-04T11:17:05.9637208Z ============================= test session starts ============================== 2025-12-04T11:17:05.9638539Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:05.9639558Z cachedir: .pytest_cache 2025-12-04T11:17:05.9640739Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:05.9642135Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:05.9642749Z configfile: pytest.ini 2025-12-04T11:17:05.9643918Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:05.9645136Z collecting ... collected 58 items 2025-12-04T11:17:05.9645783Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T11:17:05.9661064Z Running 11 items in this shard: test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16, test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16, test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16, test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16, test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16, test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16, test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16, test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16, test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16, test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16, test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 2025-12-04T11:17:05.9677174Z 2025-12-04T11:17:05.9678662Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 ('RERUN', {'yellow': True}) [3.8574s] [ 9%] 2025-12-04T11:17:05.9681714Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 ('RERUN', {'yellow': True}) [0.4314s] [ 9%] 2025-12-04T11:17:05.9684869Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 FAILED [0.4298s] [ 9%] 2025-12-04T11:17:05.9686504Z 2025-12-04T11:17:05.9686701Z ==================================== RERUNS ==================================== 2025-12-04T11:17:05.9687952Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:05.9689198Z Traceback (most recent call last): 2025-12-04T11:17:05.9690355Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:05.9691807Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:05.9693258Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:05.9694592Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:05.9695930Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:05.9697561Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:05.9698345Z AssertionError: Scalars are not equal! 2025-12-04T11:17:05.9698735Z 2025-12-04T11:17:05.9698848Z Expected 1 but got 2. 2025-12-04T11:17:05.9699283Z Absolute difference: 1 2025-12-04T11:17:05.9699726Z Relative difference: 1.0 2025-12-04T11:17:05.9699987Z 2025-12-04T11:17:05.9700323Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:05.9702113Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:05.9704061Z 2025-12-04T11:17:05.9704494Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:05.9705675Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:05.9706571Z stats [('calls_captured', 6)] 2025-12-04T11:17:05.9708209Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:05.9709487Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:05.9709959Z graph_break [] 2025-12-04T11:17:05.9710325Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:05.9711507Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9712494Z warnings.warn( 2025-12-04T11:17:05.9713383Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9714402Z warnings.warn( 2025-12-04T11:17:05.9715079Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:05.9715860Z Traceback (most recent call last): 2025-12-04T11:17:05.9716620Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:05.9717493Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:05.9718329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:05.9719097Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:05.9719929Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:05.9720823Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:05.9721301Z AssertionError: Scalars are not equal! 2025-12-04T11:17:05.9721554Z 2025-12-04T11:17:05.9721675Z Expected 1 but got 2. 2025-12-04T11:17:05.9721953Z Absolute difference: 1 2025-12-04T11:17:05.9722247Z Relative difference: 1.0 2025-12-04T11:17:05.9722437Z 2025-12-04T11:17:05.9722662Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:05.9723915Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:05.9724975Z 2025-12-04T11:17:05.9725241Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:05.9725876Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:05.9726355Z stats [('calls_captured', 6)] 2025-12-04T11:17:05.9727467Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:05.9728733Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:05.9729206Z graph_break [] 2025-12-04T11:17:05.9729584Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:05.9730676Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9731658Z warnings.warn( 2025-12-04T11:17:05.9732620Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9733591Z warnings.warn( 2025-12-04T11:17:05.9733964Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:05.9734441Z stats [('calls_captured', 6)] 2025-12-04T11:17:05.9734887Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:05.9736138Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:05.9737363Z graph_break [] 2025-12-04T11:17:05.9737739Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:05.9738875Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9739831Z warnings.warn( 2025-12-04T11:17:05.9740722Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9741730Z warnings.warn( 2025-12-04T11:17:05.9742043Z =================================== FAILURES =================================== 2025-12-04T11:17:05.9742841Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:05.9743622Z Traceback (most recent call last): 2025-12-04T11:17:05.9744373Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:05.9745244Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:05.9746135Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:05.9747387Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:05.9748237Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:05.9749114Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:05.9749613Z AssertionError: Scalars are not equal! 2025-12-04T11:17:05.9749864Z 2025-12-04T11:17:05.9749986Z Expected 1 but got 2. 2025-12-04T11:17:05.9750268Z Absolute difference: 1 2025-12-04T11:17:05.9750570Z Relative difference: 1.0 2025-12-04T11:17:05.9750776Z 2025-12-04T11:17:05.9750990Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:05.9752262Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:05.9753303Z 2025-12-04T11:17:05.9753571Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:05.9754212Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:05.9754705Z stats [('calls_captured', 6)] 2025-12-04T11:17:05.9755826Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:05.9757080Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:05.9757559Z graph_break [] 2025-12-04T11:17:05.9757939Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:05.9759044Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9760009Z warnings.warn( 2025-12-04T11:17:05.9760985Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9761963Z warnings.warn( 2025-12-04T11:17:05.9762338Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:05.9762821Z stats [('calls_captured', 6)] 2025-12-04T11:17:05.9763270Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:05.9764542Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:05.9765668Z graph_break [] 2025-12-04T11:17:05.9766044Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:05.9767198Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9768166Z warnings.warn( 2025-12-04T11:17:05.9769044Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9770045Z warnings.warn( 2025-12-04T11:17:05.9770424Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:05.9770888Z stats [('calls_captured', 6)] 2025-12-04T11:17:05.9771338Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:05.9772598Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:05.9773734Z graph_break [] 2025-12-04T11:17:05.9774094Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:05.9775196Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9776173Z warnings.warn( 2025-12-04T11:17:05.9777150Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9778112Z warnings.warn( 2025-12-04T11:17:05.9779118Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-97143c683457795e.xml - 2025-12-04T11:17:05.9780263Z =========================== short test summary info ============================ 2025-12-04T11:17:05.9781713Z FAILED [0.4298s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:05.9782782Z 2025-12-04T11:17:05.9782890Z Expected 1 but got 2. 2025-12-04T11:17:05.9783193Z Absolute difference: 1 2025-12-04T11:17:05.9783494Z Relative difference: 1.0 2025-12-04T11:17:05.9783687Z 2025-12-04T11:17:05.9783902Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:05.9785169Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:05.9786216Z 2025-12-04T11:17:05.9786482Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:05.9787073Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:05.9787548Z ========================== 1 failed, 2 rerun in 4.75s ========================== 2025-12-04T11:17:05.9787969Z Got exit code 1 2025-12-04T11:17:05.9788239Z Retrying single test... 2025-12-04T11:17:05.9788972Z W1204 11:01:29.894000 91717 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:05.9790227Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-c6ae8965f2e7a602.xml 2025-12-04T11:17:05.9791205Z ============================= test session starts ============================== 2025-12-04T11:17:05.9791873Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:05.9792464Z cachedir: .pytest_cache 2025-12-04T11:17:05.9793803Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:05.9794614Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:05.9795033Z configfile: pytest.ini 2025-12-04T11:17:05.9795755Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:05.9796849Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:05.9798202Z stepcurrent: skipping 0 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:05.9799521Z Running 1 items in this shard 2025-12-04T11:17:05.9799734Z 2025-12-04T11:17:05.9801013Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 [W1204 11:01:35.050852181 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9802438Z 2025-12-04T11:17:05.9802960Z [W1204 11:01:50.011207547 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9803632Z 2025-12-04T11:17:05.9804152Z [W1204 11:01:50.011467165 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9804814Z 2025-12-04T11:17:05.9805327Z [W1204 11:01:50.018726946 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9805979Z 2025-12-04T11:17:05.9806508Z [W1204 11:01:50.019408131 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9807154Z 2025-12-04T11:17:05.9807670Z [W1204 11:01:50.019598446 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9808335Z 2025-12-04T11:17:05.9808846Z [W1204 11:01:50.026474647 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9809512Z 2025-12-04T11:17:05.9810028Z [W1204 11:01:50.027234522 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9810690Z 2025-12-04T11:17:05.9811201Z [W1204 11:01:50.027423812 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9811843Z 2025-12-04T11:17:05.9812366Z [W1204 11:01:50.160028588 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9813016Z 2025-12-04T11:17:05.9813540Z [W1204 11:01:50.161701476 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9814188Z 2025-12-04T11:17:05.9814702Z [W1204 11:01:50.161908453 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9815360Z 2025-12-04T11:17:05.9815957Z [W1204 11:01:50.165795640 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9816619Z 2025-12-04T11:17:05.9817219Z [W1204 11:01:50.166444590 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9817867Z 2025-12-04T11:17:05.9818389Z [W1204 11:01:50.166638603 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9819037Z 2025-12-04T11:17:05.9819562Z [W1204 11:01:50.172655807 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9820208Z 2025-12-04T11:17:05.9820722Z [W1204 11:01:50.173284374 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9821435Z 2025-12-04T11:17:05.9821944Z [W1204 11:01:50.173477454 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9822608Z 2025-12-04T11:17:05.9822743Z ('RERUN', {'yellow': True}) [18.8376s] [100%] 2025-12-04T11:17:05.9824342Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 [W1204 11:01:50.566983216 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9825744Z 2025-12-04T11:17:05.9826272Z [W1204 11:01:50.567726173 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9826919Z 2025-12-04T11:17:05.9827433Z [W1204 11:01:50.567928020 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9828097Z 2025-12-04T11:17:05.9828609Z [W1204 11:01:50.571912046 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9829273Z 2025-12-04T11:17:05.9829784Z [W1204 11:01:50.572540789 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9830435Z 2025-12-04T11:17:05.9830960Z [W1204 11:01:50.572729074 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9831606Z 2025-12-04T11:17:05.9832132Z [W1204 11:01:50.578739381 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9832778Z 2025-12-04T11:17:05.9833291Z [W1204 11:01:50.579344449 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9833959Z 2025-12-04T11:17:05.9834472Z [W1204 11:01:50.579533131 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9835128Z 2025-12-04T11:17:05.9835643Z [W1204 11:01:50.666630962 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9836291Z 2025-12-04T11:17:05.9836815Z [W1204 11:01:50.667372439 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9837459Z 2025-12-04T11:17:05.9837981Z [W1204 11:01:50.667582130 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9838626Z 2025-12-04T11:17:05.9839139Z [W1204 11:01:50.671496257 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9839802Z 2025-12-04T11:17:05.9840319Z [W1204 11:01:50.672149757 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9840983Z 2025-12-04T11:17:05.9841556Z [W1204 11:01:50.672349417 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9842225Z 2025-12-04T11:17:05.9842737Z [W1204 11:01:50.678306317 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9843392Z 2025-12-04T11:17:05.9843921Z [W1204 11:01:50.679094079 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9844572Z 2025-12-04T11:17:05.9845099Z [W1204 11:01:50.679289609 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9845785Z 2025-12-04T11:17:05.9845920Z ('RERUN', {'yellow': True}) [0.4672s] [100%] 2025-12-04T11:17:05.9847465Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 [W1204 11:01:51.014425035 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9848903Z 2025-12-04T11:17:05.9849418Z [W1204 11:01:51.015144503 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9850073Z 2025-12-04T11:17:05.9850605Z [W1204 11:01:51.015351817 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9851252Z 2025-12-04T11:17:05.9851779Z [W1204 11:01:51.019315667 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9852433Z 2025-12-04T11:17:05.9852946Z [W1204 11:01:51.019929877 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9853611Z 2025-12-04T11:17:05.9854131Z [W1204 11:01:51.020187623 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9854798Z 2025-12-04T11:17:05.9855311Z [W1204 11:01:51.026189123 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9855965Z 2025-12-04T11:17:05.9856493Z [W1204 11:01:51.026794150 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9857215Z 2025-12-04T11:17:05.9857743Z [W1204 11:01:51.026983296 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9858395Z 2025-12-04T11:17:05.9858910Z [W1204 11:01:51.114740769 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9859650Z 2025-12-04T11:17:05.9860167Z [W1204 11:01:51.115493482 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9860833Z 2025-12-04T11:17:05.9861348Z [W1204 11:01:51.115701239 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9861995Z 2025-12-04T11:17:05.9862522Z [W1204 11:01:51.119646748 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9863170Z 2025-12-04T11:17:05.9863696Z [W1204 11:01:51.120305365 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9864343Z 2025-12-04T11:17:05.9864860Z [W1204 11:01:51.120508150 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9865518Z 2025-12-04T11:17:05.9866115Z [W1204 11:01:51.126455881 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9866777Z 2025-12-04T11:17:05.9867290Z [W1204 11:01:51.127234986 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9867935Z 2025-12-04T11:17:05.9868459Z [W1204 11:01:51.127431873 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9869102Z 2025-12-04T11:17:05.9869218Z FAILED [0.4459s] [100%] 2025-12-04T11:17:05.9869396Z 2025-12-04T11:17:05.9869539Z ==================================== RERUNS ==================================== 2025-12-04T11:17:05.9870346Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:05.9871155Z Traceback (most recent call last): 2025-12-04T11:17:05.9871916Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:05.9872787Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:05.9873657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:05.9874431Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:05.9875259Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:05.9876148Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:05.9876623Z AssertionError: Scalars are not equal! 2025-12-04T11:17:05.9876874Z 2025-12-04T11:17:05.9876997Z Expected 1 but got 2. 2025-12-04T11:17:05.9877279Z Absolute difference: 1 2025-12-04T11:17:05.9877576Z Relative difference: 1.0 2025-12-04T11:17:05.9877766Z 2025-12-04T11:17:05.9877993Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:05.9879257Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:05.9880313Z 2025-12-04T11:17:05.9880579Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:05.9881211Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:05.9881691Z stats [('calls_captured', 6)] 2025-12-04T11:17:05.9882800Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:05.9884071Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:05.9884547Z graph_break [] 2025-12-04T11:17:05.9884927Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:05.9886497Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:05.9887968Z if out == self.unknown_value: 2025-12-04T11:17:05.9888924Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9889889Z warnings.warn( 2025-12-04T11:17:05.9890765Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9891733Z warnings.warn( 2025-12-04T11:17:05.9892413Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:05.9893182Z Traceback (most recent call last): 2025-12-04T11:17:05.9893982Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:05.9894866Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:05.9895695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:05.9896671Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:05.9897963Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:05.9898854Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:05.9899331Z AssertionError: Scalars are not equal! 2025-12-04T11:17:05.9899668Z 2025-12-04T11:17:05.9899774Z Expected 1 but got 2. 2025-12-04T11:17:05.9900322Z Absolute difference: 1 2025-12-04T11:17:05.9900630Z Relative difference: 1.0 2025-12-04T11:17:05.9900823Z 2025-12-04T11:17:05.9901289Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:05.9902860Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:05.9903920Z 2025-12-04T11:17:05.9904191Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:05.9904831Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:05.9905301Z stats [('calls_captured', 6)] 2025-12-04T11:17:05.9906429Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:05.9907711Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:05.9908180Z graph_break [] 2025-12-04T11:17:05.9908550Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:05.9910468Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:05.9911933Z if out == self.unknown_value: 2025-12-04T11:17:05.9912887Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9913849Z warnings.warn( 2025-12-04T11:17:05.9914735Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9915711Z warnings.warn( 2025-12-04T11:17:05.9916105Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:05.9916574Z stats [('calls_captured', 6)] 2025-12-04T11:17:05.9918424Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:05.9919828Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:05.9920976Z graph_break [] 2025-12-04T11:17:05.9921338Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:05.9922545Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9923621Z warnings.warn( 2025-12-04T11:17:05.9924608Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9925723Z warnings.warn( 2025-12-04T11:17:05.9926050Z =================================== FAILURES =================================== 2025-12-04T11:17:05.9926863Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:05.9927628Z Traceback (most recent call last): 2025-12-04T11:17:05.9928389Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:05.9929279Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:05.9930230Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:05.9931507Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:05.9932358Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:05.9933252Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:05.9933771Z AssertionError: Scalars are not equal! 2025-12-04T11:17:05.9934039Z 2025-12-04T11:17:05.9934149Z Expected 1 but got 2. 2025-12-04T11:17:05.9934446Z Absolute difference: 1 2025-12-04T11:17:05.9934734Z Relative difference: 1.0 2025-12-04T11:17:05.9934941Z 2025-12-04T11:17:05.9935157Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:05.9936426Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:05.9937537Z 2025-12-04T11:17:05.9937823Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:05.9938449Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:05.9938935Z stats [('calls_captured', 6)] 2025-12-04T11:17:05.9940066Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:05.9941329Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:05.9941787Z graph_break [] 2025-12-04T11:17:05.9942160Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:05.9943730Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:05.9945184Z if out == self.unknown_value: 2025-12-04T11:17:05.9946120Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9947107Z warnings.warn( 2025-12-04T11:17:05.9948007Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9948974Z warnings.warn( 2025-12-04T11:17:05.9949345Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:05.9949824Z stats [('calls_captured', 6)] 2025-12-04T11:17:05.9950273Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:05.9951527Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:05.9952666Z graph_break [] 2025-12-04T11:17:05.9953040Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:05.9954211Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9955174Z warnings.warn( 2025-12-04T11:17:05.9956060Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9957027Z warnings.warn( 2025-12-04T11:17:05.9957408Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:05.9957875Z stats [('calls_captured', 6)] 2025-12-04T11:17:05.9958320Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:05.9959595Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:05.9960756Z graph_break [] 2025-12-04T11:17:05.9961140Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:05.9962264Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9963241Z warnings.warn( 2025-12-04T11:17:05.9964117Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:05.9965091Z warnings.warn( 2025-12-04T11:17:05.9966096Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-c6ae8965f2e7a602.xml - 2025-12-04T11:17:05.9967258Z =========================== short test summary info ============================ 2025-12-04T11:17:05.9968504Z FAILED [0.4459s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:05.9969579Z 2025-12-04T11:17:05.9969686Z Expected 1 but got 2. 2025-12-04T11:17:05.9969978Z Absolute difference: 1 2025-12-04T11:17:05.9970264Z Relative difference: 1.0 2025-12-04T11:17:05.9970467Z 2025-12-04T11:17:05.9970679Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:05.9971938Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:05.9972974Z 2025-12-04T11:17:05.9973254Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:05.9973846Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:05.9974360Z ================== 1 failed, 10 deselected, 2 rerun in 19.78s ================== 2025-12-04T11:17:05.9974803Z Got exit code 1 2025-12-04T11:17:05.9975073Z Retrying single test... 2025-12-04T11:17:05.9975698Z W1204 11:02:02.761000 91891 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:05.9977017Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-2f238f5f180b8489.xml 2025-12-04T11:17:05.9977992Z ============================= test session starts ============================== 2025-12-04T11:17:05.9978658Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:05.9979245Z cachedir: .pytest_cache 2025-12-04T11:17:05.9979958Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:05.9980752Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:05.9981092Z configfile: pytest.ini 2025-12-04T11:17:05.9981919Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:05.9982823Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:05.9984175Z stepcurrent: skipping 0 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:05.9985400Z Running 1 items in this shard 2025-12-04T11:17:05.9985625Z 2025-12-04T11:17:05.9986899Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 [W1204 11:02:08.918369352 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9988349Z 2025-12-04T11:17:05.9988872Z [W1204 11:02:23.478854521 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9989561Z 2025-12-04T11:17:05.9990090Z [W1204 11:02:23.479117176 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9990739Z 2025-12-04T11:17:05.9991266Z [W1204 11:02:23.486601225 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9991913Z 2025-12-04T11:17:05.9992426Z [W1204 11:02:23.487328445 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9993093Z 2025-12-04T11:17:05.9993605Z [W1204 11:02:23.487516199 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9994265Z 2025-12-04T11:17:05.9994781Z [W1204 11:02:23.494524375 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9995442Z 2025-12-04T11:17:05.9995956Z [W1204 11:02:23.495320283 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9996804Z 2025-12-04T11:17:05.9997328Z [W1204 11:02:23.495505322 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9997978Z 2025-12-04T11:17:05.9998502Z [W1204 11:02:23.632253502 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:05.9999152Z 2025-12-04T11:17:05.9999666Z [W1204 11:02:23.634002624 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0000326Z 2025-12-04T11:17:06.0000847Z [W1204 11:02:23.634209876 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0001509Z 2025-12-04T11:17:06.0002022Z [W1204 11:02:23.638199600 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0002673Z 2025-12-04T11:17:06.0003199Z [W1204 11:02:23.638829647 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0003845Z 2025-12-04T11:17:06.0004368Z [W1204 11:02:23.639034775 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0005018Z 2025-12-04T11:17:06.0005528Z [W1204 11:02:23.645114362 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0006188Z 2025-12-04T11:17:06.0006702Z [W1204 11:02:23.645744926 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0007477Z 2025-12-04T11:17:06.0007994Z [W1204 11:02:23.645934699 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0008645Z 2025-12-04T11:17:06.0008795Z ('RERUN', {'yellow': True}) [19.4400s] [100%] 2025-12-04T11:17:06.0010347Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 [W1204 11:02:24.044138843 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0011750Z 2025-12-04T11:17:06.0012271Z [W1204 11:02:24.044891662 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0012984Z 2025-12-04T11:17:06.0013505Z [W1204 11:02:24.045091271 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0014175Z 2025-12-04T11:17:06.0014689Z [W1204 11:02:24.049123256 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0015385Z 2025-12-04T11:17:06.0015913Z [W1204 11:02:24.049755215 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0016560Z 2025-12-04T11:17:06.0017143Z [W1204 11:02:24.049945714 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0017788Z 2025-12-04T11:17:06.0018301Z [W1204 11:02:24.056190487 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0018961Z 2025-12-04T11:17:06.0019478Z [W1204 11:02:24.056826689 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0020143Z 2025-12-04T11:17:06.0020661Z [W1204 11:02:24.057016052 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0021311Z 2025-12-04T11:17:06.0021840Z [W1204 11:02:24.145465313 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0022493Z 2025-12-04T11:17:06.0023025Z [W1204 11:02:24.146222890 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0023673Z 2025-12-04T11:17:06.0024187Z [W1204 11:02:24.146425479 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0024852Z 2025-12-04T11:17:06.0025364Z [W1204 11:02:24.150353667 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0026028Z 2025-12-04T11:17:06.0026546Z [W1204 11:02:24.150996687 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0027211Z 2025-12-04T11:17:06.0027725Z [W1204 11:02:24.151192460 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0028372Z 2025-12-04T11:17:06.0028903Z [W1204 11:02:24.157248937 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0029551Z 2025-12-04T11:17:06.0030076Z [W1204 11:02:24.158045925 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0030722Z 2025-12-04T11:17:06.0031237Z [W1204 11:02:24.158240918 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0031894Z 2025-12-04T11:17:06.0032025Z ('RERUN', {'yellow': True}) [0.4725s] [100%] 2025-12-04T11:17:06.0033630Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 [W1204 11:02:24.492719894 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0035044Z 2025-12-04T11:17:06.0035574Z [W1204 11:02:24.493466032 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0036221Z 2025-12-04T11:17:06.0036747Z [W1204 11:02:24.493664550 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0037395Z 2025-12-04T11:17:06.0037938Z [W1204 11:02:24.497710001 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0038594Z 2025-12-04T11:17:06.0039111Z [W1204 11:02:24.498337174 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0039805Z 2025-12-04T11:17:06.0040320Z [W1204 11:02:24.498535001 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0040968Z 2025-12-04T11:17:06.0041494Z [W1204 11:02:24.504777928 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0042137Z 2025-12-04T11:17:06.0042662Z [W1204 11:02:24.505411291 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0043311Z 2025-12-04T11:17:06.0043830Z [W1204 11:02:24.505598935 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0044493Z 2025-12-04T11:17:06.0045010Z [W1204 11:02:24.594489001 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0045671Z 2025-12-04T11:17:06.0046188Z [W1204 11:02:24.595246778 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0046835Z 2025-12-04T11:17:06.0047360Z [W1204 11:02:24.595450919 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0048008Z 2025-12-04T11:17:06.0048536Z [W1204 11:02:24.599405567 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0049183Z 2025-12-04T11:17:06.0049697Z [W1204 11:02:24.600063297 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0050362Z 2025-12-04T11:17:06.0050880Z [W1204 11:02:24.600273037 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0051542Z 2025-12-04T11:17:06.0052056Z [W1204 11:02:24.606284042 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0052705Z 2025-12-04T11:17:06.0053228Z [W1204 11:02:24.607087515 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0053879Z 2025-12-04T11:17:06.0054403Z [W1204 11:02:24.607282094 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0055052Z 2025-12-04T11:17:06.0055155Z FAILED [0.4473s] [100%] 2025-12-04T11:17:06.0055341Z 2025-12-04T11:17:06.0055487Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.0056283Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.0057188Z Traceback (most recent call last): 2025-12-04T11:17:06.0057938Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0058829Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0059670Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0060436Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0061264Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0062155Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0062663Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0062914Z 2025-12-04T11:17:06.0063020Z Expected 1 but got 2. 2025-12-04T11:17:06.0063311Z Absolute difference: 1 2025-12-04T11:17:06.0063618Z Relative difference: 1.0 2025-12-04T11:17:06.0063810Z 2025-12-04T11:17:06.0064093Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0065347Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.0066401Z 2025-12-04T11:17:06.0066668Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0067297Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0067779Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0068887Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0070167Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0070640Z graph_break [] 2025-12-04T11:17:06.0071006Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0072589Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.0074046Z if out == self.unknown_value: 2025-12-04T11:17:06.0075000Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0075959Z warnings.warn( 2025-12-04T11:17:06.0076848Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0077820Z warnings.warn( 2025-12-04T11:17:06.0078498Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.0079258Z Traceback (most recent call last): 2025-12-04T11:17:06.0080006Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0080884Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0081715Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0082463Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0083300Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0084188Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0084652Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0084922Z 2025-12-04T11:17:06.0085102Z Expected 1 but got 2. 2025-12-04T11:17:06.0085396Z Absolute difference: 1 2025-12-04T11:17:06.0085682Z Relative difference: 1.0 2025-12-04T11:17:06.0085887Z 2025-12-04T11:17:06.0086098Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0087364Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.0088397Z 2025-12-04T11:17:06.0088676Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0089307Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0089813Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0090936Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0092203Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0092699Z graph_break [] 2025-12-04T11:17:06.0093072Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0094645Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.0096255Z if out == self.unknown_value: 2025-12-04T11:17:06.0097274Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0098258Z warnings.warn( 2025-12-04T11:17:06.0099153Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0100123Z warnings.warn( 2025-12-04T11:17:06.0100497Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0100979Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0101429Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0102691Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0103830Z graph_break [] 2025-12-04T11:17:06.0104208Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0105317Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0106274Z warnings.warn( 2025-12-04T11:17:06.0107172Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0108144Z warnings.warn( 2025-12-04T11:17:06.0108461Z =================================== FAILURES =================================== 2025-12-04T11:17:06.0109257Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.0110030Z Traceback (most recent call last): 2025-12-04T11:17:06.0110780Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0111648Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0112479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0113357Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0114202Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0115079Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0115555Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0115806Z 2025-12-04T11:17:06.0115924Z Expected 1 but got 2. 2025-12-04T11:17:06.0116217Z Absolute difference: 1 2025-12-04T11:17:06.0116499Z Relative difference: 1.0 2025-12-04T11:17:06.0116699Z 2025-12-04T11:17:06.0116911Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0118185Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.0119267Z 2025-12-04T11:17:06.0119533Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0120169Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0120695Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0121820Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0123072Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0123542Z graph_break [] 2025-12-04T11:17:06.0123923Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0125507Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.0126955Z if out == self.unknown_value: 2025-12-04T11:17:06.0127907Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0128890Z warnings.warn( 2025-12-04T11:17:06.0129775Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0130721Z warnings.warn( 2025-12-04T11:17:06.0131104Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0131584Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0132016Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0133283Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0134421Z graph_break [] 2025-12-04T11:17:06.0134796Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0135874Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0136841Z warnings.warn( 2025-12-04T11:17:06.0137803Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0138775Z warnings.warn( 2025-12-04T11:17:06.0139148Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0139633Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0140088Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0141411Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0142544Z graph_break [] 2025-12-04T11:17:06.0142920Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0144015Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0144964Z warnings.warn( 2025-12-04T11:17:06.0145851Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0146816Z warnings.warn( 2025-12-04T11:17:06.0147860Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-2f238f5f180b8489.xml - 2025-12-04T11:17:06.0149006Z =========================== short test summary info ============================ 2025-12-04T11:17:06.0150287Z FAILED [0.4473s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0151363Z 2025-12-04T11:17:06.0151470Z Expected 1 but got 2. 2025-12-04T11:17:06.0151766Z Absolute difference: 1 2025-12-04T11:17:06.0152052Z Relative difference: 1.0 2025-12-04T11:17:06.0152255Z 2025-12-04T11:17:06.0152469Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0153730Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.0154764Z 2025-12-04T11:17:06.0155043Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0155625Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.0156150Z ================== 1 failed, 10 deselected, 2 rerun in 20.39s ================== 2025-12-04T11:17:06.0156592Z Got exit code 1 2025-12-04T11:17:06.0157568Z FAILED CONSISTENTLY: test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.0158942Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:17:06.0159951Z W1204 11:02:36.406000 92065 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.0161197Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-df27debc7fb08425.xml 2025-12-04T11:17:06.0162163Z ============================= test session starts ============================== 2025-12-04T11:17:06.0162835Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.0163448Z cachedir: .pytest_cache 2025-12-04T11:17:06.0164166Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.0164950Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.0165305Z configfile: pytest.ini 2025-12-04T11:17:06.0166035Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.0166920Z collecting ... collected 58 items / 1 deselected / 57 selected 2025-12-04T11:17:06.0167414Z stepcurrent: skipping 1 already run items. 2025-12-04T11:17:06.0167802Z Running 10 items in this shard 2025-12-04T11:17:06.0168010Z 2025-12-04T11:17:06.0169315Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 W1204 11:02:42.004000 92065 site-packages/torch/_inductor/utils.py:1703] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.0170797Z ('RERUN', {'yellow': True}) [3.9541s] [ 10%] 2025-12-04T11:17:06.0171919Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 ('RERUN', {'yellow': True}) [0.5073s] [ 10%] 2025-12-04T11:17:06.0173716Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 FAILED [0.5164s] [ 10%] 2025-12-04T11:17:06.0174641Z 2025-12-04T11:17:06.0174800Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.0175649Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.0176415Z Traceback (most recent call last): 2025-12-04T11:17:06.0177243Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0204169Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0205115Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0205882Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0206732Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0207612Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0208080Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0208331Z 2025-12-04T11:17:06.0208440Z Expected 1 but got 0. 2025-12-04T11:17:06.0208725Z Absolute difference: 1 2025-12-04T11:17:06.0209029Z Relative difference: 1.0 2025-12-04T11:17:06.0209220Z 2025-12-04T11:17:06.0209445Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0210737Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0211790Z 2025-12-04T11:17:06.0212064Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0212707Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0213181Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0214113Z inductor [('pattern_matcher_count', 6), ('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0215191Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0215663Z graph_break [] 2025-12-04T11:17:06.0215940Z aten_mm_info [('aten.mm_136_72_1024', 2)] 2025-12-04T11:17:06.0216421Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0217625Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0218593Z warnings.warn( 2025-12-04T11:17:06.0219488Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0220460Z warnings.warn( 2025-12-04T11:17:06.0221143Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.0221912Z Traceback (most recent call last): 2025-12-04T11:17:06.0222672Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0223728Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0224559Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0225328Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0226175Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0227065Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0227533Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0227799Z 2025-12-04T11:17:06.0227909Z Expected 1 but got 0. 2025-12-04T11:17:06.0228202Z Absolute difference: 1 2025-12-04T11:17:06.0228540Z Relative difference: 1.0 2025-12-04T11:17:06.0228746Z 2025-12-04T11:17:06.0228967Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0230241Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0231335Z 2025-12-04T11:17:06.0231620Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0232245Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0232735Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0233669Z inductor [('pattern_matcher_count', 6), ('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0234756Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0235224Z graph_break [] 2025-12-04T11:17:06.0235509Z aten_mm_info [('aten.mm_136_72_1024', 2)] 2025-12-04T11:17:06.0235985Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0237098Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0238067Z warnings.warn( 2025-12-04T11:17:06.0238965Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0239930Z warnings.warn( 2025-12-04T11:17:06.0240303Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0240790Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0241245Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0242324Z inductor [('pattern_matcher_count', 6), ('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0243258Z graph_break [] 2025-12-04T11:17:06.0243542Z aten_mm_info [('aten.mm_136_72_1024', 2)] 2025-12-04T11:17:06.0244026Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0245111Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0246084Z warnings.warn( 2025-12-04T11:17:06.0246975Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0247945Z warnings.warn( 2025-12-04T11:17:06.0248250Z =================================== FAILURES =================================== 2025-12-04T11:17:06.0249067Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.0249850Z Traceback (most recent call last): 2025-12-04T11:17:06.0250652Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0251547Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0252377Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0253148Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0253980Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0254869Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0255350Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0255604Z 2025-12-04T11:17:06.0255726Z Expected 1 but got 0. 2025-12-04T11:17:06.0256043Z Absolute difference: 1 2025-12-04T11:17:06.0256342Z Relative difference: 1.0 2025-12-04T11:17:06.0256534Z 2025-12-04T11:17:06.0256764Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0258524Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0259636Z 2025-12-04T11:17:06.0259915Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0260563Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0261051Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0261970Z inductor [('pattern_matcher_count', 6), ('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0263046Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0263526Z graph_break [] 2025-12-04T11:17:06.0263814Z aten_mm_info [('aten.mm_136_72_1024', 2)] 2025-12-04T11:17:06.0264286Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0265393Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0266370Z warnings.warn( 2025-12-04T11:17:06.0267254Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0268228Z warnings.warn( 2025-12-04T11:17:06.0268614Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0269101Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0269542Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0270617Z inductor [('pattern_matcher_count', 6), ('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0271557Z graph_break [] 2025-12-04T11:17:06.0271830Z aten_mm_info [('aten.mm_136_72_1024', 2)] 2025-12-04T11:17:06.0272309Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0273404Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0274375Z warnings.warn( 2025-12-04T11:17:06.0275251Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0276220Z warnings.warn( 2025-12-04T11:17:06.0276607Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0277075Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0277527Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0278683Z inductor [('pattern_matcher_count', 6), ('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0279635Z graph_break [] 2025-12-04T11:17:06.0279909Z aten_mm_info [('aten.mm_136_72_1024', 2)] 2025-12-04T11:17:06.0280387Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0281489Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0282460Z warnings.warn( 2025-12-04T11:17:06.0283336Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0284345Z warnings.warn( 2025-12-04T11:17:06.0285361Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-df27debc7fb08425.xml - 2025-12-04T11:17:06.0286503Z =========================== short test summary info ============================ 2025-12-04T11:17:06.0287821Z FAILED [0.5164s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0288909Z 2025-12-04T11:17:06.0289022Z Expected 1 but got 0. 2025-12-04T11:17:06.0289317Z Absolute difference: 1 2025-12-04T11:17:06.0289605Z Relative difference: 1.0 2025-12-04T11:17:06.0289808Z 2025-12-04T11:17:06.0290026Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0291292Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0292330Z 2025-12-04T11:17:06.0292617Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0293201Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.0293727Z =================== 1 failed, 1 deselected, 2 rerun in 5.01s =================== 2025-12-04T11:17:06.0294172Z Got exit code 1 2025-12-04T11:17:06.0294447Z Retrying single test... 2025-12-04T11:17:06.0295075Z W1204 11:02:56.351000 92242 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.0296509Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-39f974c40a0e5d07.xml 2025-12-04T11:17:06.0297558Z ============================= test session starts ============================== 2025-12-04T11:17:06.0298215Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.0298819Z cachedir: .pytest_cache 2025-12-04T11:17:06.0299538Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.0300329Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.0300669Z configfile: pytest.ini 2025-12-04T11:17:06.0301396Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.0302304Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.0303662Z stepcurrent: skipping 1 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0304885Z Running 1 items in this shard 2025-12-04T11:17:06.0305112Z 2025-12-04T11:17:06.0306516Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 [W1204 11:03:01.539684585 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0307948Z 2025-12-04T11:17:06.0308472Z [W1204 11:03:16.561884600 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0309124Z 2025-12-04T11:17:06.0309657Z [W1204 11:03:16.562145149 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0310305Z 2025-12-04T11:17:06.0310838Z [W1204 11:03:16.570317091 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0311491Z 2025-12-04T11:17:06.0312051Z [W1204 11:03:16.571185079 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0312710Z 2025-12-04T11:17:06.0313228Z [W1204 11:03:16.571377803 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0313933Z 2025-12-04T11:17:06.0314447Z [W1204 11:03:16.578849361 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0315108Z 2025-12-04T11:17:06.0315620Z [W1204 11:03:16.579498320 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0316271Z 2025-12-04T11:17:06.0316798Z [W1204 11:03:16.579681924 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0317442Z 2025-12-04T11:17:06.0317904Z W1204 11:03:16.991000 92242 site-packages/torch/_inductor/utils.py:1703] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.0319029Z [W1204 11:03:17.773661474 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0319687Z 2025-12-04T11:17:06.0320198Z [W1204 11:03:17.775445330 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0320851Z 2025-12-04T11:17:06.0321375Z [W1204 11:03:17.775665678 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0322022Z 2025-12-04T11:17:06.0322543Z [W1204 11:03:17.780478884 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0323190Z 2025-12-04T11:17:06.0323700Z [W1204 11:03:17.781182698 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0324364Z 2025-12-04T11:17:06.0324876Z [W1204 11:03:17.781381390 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0325540Z 2025-12-04T11:17:06.0326055Z [W1204 11:03:17.788246959 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0326717Z 2025-12-04T11:17:06.0327229Z [W1204 11:03:17.788958038 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0327876Z 2025-12-04T11:17:06.0328401Z [W1204 11:03:17.789158870 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0329045Z 2025-12-04T11:17:06.0329181Z ('RERUN', {'yellow': True}) [19.0060s] [100%] 2025-12-04T11:17:06.0330727Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 [W1204 11:03:17.232186078 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0332209Z 2025-12-04T11:17:06.0332727Z [W1204 11:03:17.232902003 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0333388Z 2025-12-04T11:17:06.0333901Z [W1204 11:03:17.233107737 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0334550Z 2025-12-04T11:17:06.0335074Z [W1204 11:03:17.237957957 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0335720Z 2025-12-04T11:17:06.0336245Z [W1204 11:03:17.238581667 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0336981Z 2025-12-04T11:17:06.0337494Z [W1204 11:03:17.238773936 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0338151Z 2025-12-04T11:17:06.0338669Z [W1204 11:03:17.245429913 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0339367Z 2025-12-04T11:17:06.0339879Z [W1204 11:03:17.246047249 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0340527Z 2025-12-04T11:17:06.0341051Z [W1204 11:03:17.246235796 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0341701Z 2025-12-04T11:17:06.0342224Z [W1204 11:03:17.350472346 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0342874Z 2025-12-04T11:17:06.0343386Z [W1204 11:03:17.351229951 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0344047Z 2025-12-04T11:17:06.0344563Z [W1204 11:03:17.351438826 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0345228Z 2025-12-04T11:17:06.0345741Z [W1204 11:03:17.356056950 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0346388Z 2025-12-04T11:17:06.0346916Z [W1204 11:03:17.356704852 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0347567Z 2025-12-04T11:17:06.0348091Z [W1204 11:03:17.356902054 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0348739Z 2025-12-04T11:17:06.0349251Z [W1204 11:03:17.363583540 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0349907Z 2025-12-04T11:17:06.0350427Z [W1204 11:03:17.364231450 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0351089Z 2025-12-04T11:17:06.0351602Z [W1204 11:03:17.364426620 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0352253Z 2025-12-04T11:17:06.0352403Z ('RERUN', {'yellow': True}) [0.5350s] [100%] 2025-12-04T11:17:06.0353960Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 [W1204 11:03:18.745572399 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0355370Z 2025-12-04T11:17:06.0355890Z [W1204 11:03:18.746308074 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0356555Z 2025-12-04T11:17:06.0357127Z [W1204 11:03:18.746510074 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0357791Z 2025-12-04T11:17:06.0358305Z [W1204 11:03:18.751430383 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0358961Z 2025-12-04T11:17:06.0359473Z [W1204 11:03:18.752059364 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0360118Z 2025-12-04T11:17:06.0360642Z [W1204 11:03:18.752266028 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0361294Z 2025-12-04T11:17:06.0361850Z [W1204 11:03:18.758919928 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0362496Z 2025-12-04T11:17:06.0363010Z [W1204 11:03:18.759536406 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0363722Z 2025-12-04T11:17:06.0364237Z [W1204 11:03:18.759725278 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0364897Z 2025-12-04T11:17:06.0365410Z [W1204 11:03:18.863946434 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0366059Z 2025-12-04T11:17:06.0366592Z [W1204 11:03:18.864715445 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0367238Z 2025-12-04T11:17:06.0367769Z [W1204 11:03:18.864921759 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0368419Z 2025-12-04T11:17:06.0368931Z [W1204 11:03:18.872144667 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0369588Z 2025-12-04T11:17:06.0370098Z [W1204 11:03:18.872876655 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0370752Z 2025-12-04T11:17:06.0371265Z [W1204 11:03:18.873076052 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0371908Z 2025-12-04T11:17:06.0372428Z [W1204 11:03:18.880957075 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0373071Z 2025-12-04T11:17:06.0373588Z [W1204 11:03:18.881733950 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0374237Z 2025-12-04T11:17:06.0374752Z [W1204 11:03:18.881930525 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0375411Z 2025-12-04T11:17:06.0375517Z FAILED [0.5161s] [100%] 2025-12-04T11:17:06.0375703Z 2025-12-04T11:17:06.0375846Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.0376643Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.0377484Z Traceback (most recent call last): 2025-12-04T11:17:06.0378234Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0379113Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0379951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0380705Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0381614Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0382497Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0382960Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0383220Z 2025-12-04T11:17:06.0383328Z Expected 1 but got 0. 2025-12-04T11:17:06.0383612Z Absolute difference: 1 2025-12-04T11:17:06.0383901Z Relative difference: 1.0 2025-12-04T11:17:06.0384087Z 2025-12-04T11:17:06.0384299Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0385559Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0386649Z 2025-12-04T11:17:06.0386924Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0387549Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0388017Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0388965Z inductor [('pattern_matcher_count', 6), ('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0390023Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0390480Z graph_break [] 2025-12-04T11:17:06.0390750Z aten_mm_info [('aten.mm_136_72_1024', 2)] 2025-12-04T11:17:06.0391219Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0392787Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.0394230Z if out == self.unknown_value: 2025-12-04T11:17:06.0395182Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0396298Z warnings.warn( 2025-12-04T11:17:06.0397187Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0398141Z warnings.warn( 2025-12-04T11:17:06.0398810Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.0399580Z Traceback (most recent call last): 2025-12-04T11:17:06.0400319Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0401189Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0402009Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0402772Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0403594Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0404470Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0404937Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0405187Z 2025-12-04T11:17:06.0405301Z Expected 1 but got 0. 2025-12-04T11:17:06.0405572Z Absolute difference: 1 2025-12-04T11:17:06.0405860Z Relative difference: 1.0 2025-12-04T11:17:06.0406050Z 2025-12-04T11:17:06.0406273Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0407520Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0408565Z 2025-12-04T11:17:06.0408934Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0409567Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0410051Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0410963Z inductor [('pattern_matcher_count', 6), ('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0412024Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0412490Z graph_break [] 2025-12-04T11:17:06.0412769Z aten_mm_info [('aten.mm_136_72_1024', 2)] 2025-12-04T11:17:06.0413226Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0414847Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.0416302Z if out == self.unknown_value: 2025-12-04T11:17:06.0417391Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0418348Z warnings.warn( 2025-12-04T11:17:06.0419240Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0420200Z warnings.warn( 2025-12-04T11:17:06.0420569Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0421039Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0421481Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0422558Z inductor [('pattern_matcher_count', 6), ('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0423491Z graph_break [] 2025-12-04T11:17:06.0423770Z aten_mm_info [('aten.mm_136_72_1024', 2)] 2025-12-04T11:17:06.0424234Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0425319Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0426284Z warnings.warn( 2025-12-04T11:17:06.0427163Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0428116Z warnings.warn( 2025-12-04T11:17:06.0428414Z =================================== FAILURES =================================== 2025-12-04T11:17:06.0429220Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.0429990Z Traceback (most recent call last): 2025-12-04T11:17:06.0430729Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0431608Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0432432Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0433194Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0434019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0434897Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0435364Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0435616Z 2025-12-04T11:17:06.0435728Z Expected 1 but got 0. 2025-12-04T11:17:06.0436003Z Absolute difference: 1 2025-12-04T11:17:06.0436291Z Relative difference: 1.0 2025-12-04T11:17:06.0436478Z 2025-12-04T11:17:06.0436771Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0438032Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0439079Z 2025-12-04T11:17:06.0439348Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0439972Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0440448Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0441363Z inductor [('pattern_matcher_count', 6), ('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0442471Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0442938Z graph_break [] 2025-12-04T11:17:06.0443230Z aten_mm_info [('aten.mm_136_72_1024', 2)] 2025-12-04T11:17:06.0443698Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0445323Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.0446788Z if out == self.unknown_value: 2025-12-04T11:17:06.0447745Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0448699Z warnings.warn( 2025-12-04T11:17:06.0449593Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0450563Z warnings.warn( 2025-12-04T11:17:06.0450946Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0451432Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0451888Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0452971Z inductor [('pattern_matcher_count', 6), ('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0453902Z graph_break [] 2025-12-04T11:17:06.0454184Z aten_mm_info [('aten.mm_136_72_1024', 2)] 2025-12-04T11:17:06.0454655Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0455739Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0456711Z warnings.warn( 2025-12-04T11:17:06.0457677Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0458647Z warnings.warn( 2025-12-04T11:17:06.0459018Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0459494Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0459942Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0461000Z inductor [('pattern_matcher_count', 6), ('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0461944Z graph_break [] 2025-12-04T11:17:06.0462227Z aten_mm_info [('aten.mm_136_72_1024', 2)] 2025-12-04T11:17:06.0462703Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0463786Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0464826Z warnings.warn( 2025-12-04T11:17:06.0465715Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0466687Z warnings.warn( 2025-12-04T11:17:06.0467682Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-39f974c40a0e5d07.xml - 2025-12-04T11:17:06.0468836Z =========================== short test summary info ============================ 2025-12-04T11:17:06.0470112Z FAILED [0.5161s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0471238Z 2025-12-04T11:17:06.0471361Z Expected 1 but got 0. 2025-12-04T11:17:06.0471643Z Absolute difference: 1 2025-12-04T11:17:06.0471940Z Relative difference: 1.0 2025-12-04T11:17:06.0472136Z 2025-12-04T11:17:06.0472366Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0473653Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0474705Z 2025-12-04T11:17:06.0474974Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0475572Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.0476097Z ================== 1 failed, 10 deselected, 2 rerun in 20.09s ================== 2025-12-04T11:17:06.0476531Z Got exit code 1 2025-12-04T11:17:06.0476807Z Retrying single test... 2025-12-04T11:17:06.0477444Z W1204 11:03:29.446000 92424 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.0478709Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-4c5f8d97ccaac9c3.xml 2025-12-04T11:17:06.0479680Z ============================= test session starts ============================== 2025-12-04T11:17:06.0480345Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.0480954Z cachedir: .pytest_cache 2025-12-04T11:17:06.0481659Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.0482445Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.0482801Z configfile: pytest.ini 2025-12-04T11:17:06.0483541Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.0484431Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.0485787Z stepcurrent: skipping 1 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0487028Z Running 1 items in this shard 2025-12-04T11:17:06.0487238Z 2025-12-04T11:17:06.0488538Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 [W1204 11:03:34.663340782 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0489949Z 2025-12-04T11:17:06.0490482Z [W1204 11:03:50.802962664 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0491139Z 2025-12-04T11:17:06.0491656Z [W1204 11:03:50.803217941 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0492320Z 2025-12-04T11:17:06.0492898Z [W1204 11:03:50.811519304 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0493566Z 2025-12-04T11:17:06.0494077Z [W1204 11:03:50.812435901 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0494724Z 2025-12-04T11:17:06.0495249Z [W1204 11:03:50.812631716 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0495897Z 2025-12-04T11:17:06.0496579Z [W1204 11:03:50.820246219 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0497346Z 2025-12-04T11:17:06.0497864Z [W1204 11:03:50.820928888 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0498521Z 2025-12-04T11:17:06.0499037Z [W1204 11:03:50.821115109 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0499751Z 2025-12-04T11:17:06.0500216Z W1204 11:03:50.232000 92424 site-packages/torch/_inductor/utils.py:1703] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.0501339Z [W1204 11:03:50.011699317 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0501985Z 2025-12-04T11:17:06.0502507Z [W1204 11:03:50.013457413 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0503152Z 2025-12-04T11:17:06.0503662Z [W1204 11:03:50.013668254 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0504323Z 2025-12-04T11:17:06.0504836Z [W1204 11:03:50.018331981 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0505494Z 2025-12-04T11:17:06.0506005Z [W1204 11:03:50.018989687 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0506652Z 2025-12-04T11:17:06.0507174Z [W1204 11:03:50.019185414 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0507822Z 2025-12-04T11:17:06.0508347Z [W1204 11:03:50.025909464 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0508996Z 2025-12-04T11:17:06.0509508Z [W1204 11:03:50.026570135 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0510165Z 2025-12-04T11:17:06.0510683Z [W1204 11:03:50.026761658 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0511345Z 2025-12-04T11:17:06.0511481Z ('RERUN', {'yellow': True}) [19.1366s] [100%] 2025-12-04T11:17:06.0513041Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 [W1204 11:03:50.471399255 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0514445Z 2025-12-04T11:17:06.0514974Z [W1204 11:03:50.472158449 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0515622Z 2025-12-04T11:17:06.0516133Z [W1204 11:03:50.472362192 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0516795Z 2025-12-04T11:17:06.0517389Z [W1204 11:03:50.477222610 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0518056Z 2025-12-04T11:17:06.0518577Z [W1204 11:03:50.477867210 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0519221Z 2025-12-04T11:17:06.0519747Z [W1204 11:03:50.478060084 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0520395Z 2025-12-04T11:17:06.0520924Z [W1204 11:03:50.484844053 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0521571Z 2025-12-04T11:17:06.0522087Z [W1204 11:03:50.485498496 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0522782Z 2025-12-04T11:17:06.0523297Z [W1204 11:03:50.485689137 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0523960Z 2025-12-04T11:17:06.0524473Z [W1204 11:03:50.591761760 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0525164Z 2025-12-04T11:17:06.0525680Z [W1204 11:03:50.592537735 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0526326Z 2025-12-04T11:17:06.0526851Z [W1204 11:03:50.592742473 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0527501Z 2025-12-04T11:17:06.0528025Z [W1204 11:03:50.597354008 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0529095Z 2025-12-04T11:17:06.0529617Z [W1204 11:03:50.597992039 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0530282Z 2025-12-04T11:17:06.0530809Z [W1204 11:03:50.598189631 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0531473Z 2025-12-04T11:17:06.0531985Z [W1204 11:03:50.604869662 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0532633Z 2025-12-04T11:17:06.0533159Z [W1204 11:03:50.605504830 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0533809Z 2025-12-04T11:17:06.0534331Z [W1204 11:03:50.605699903 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0534980Z 2025-12-04T11:17:06.0535114Z ('RERUN', {'yellow': True}) [0.5394s] [100%] 2025-12-04T11:17:06.0536660Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 [W1204 11:03:51.988312770 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0538138Z 2025-12-04T11:17:06.0538653Z [W1204 11:03:51.989022853 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0539301Z 2025-12-04T11:17:06.0539826Z [W1204 11:03:51.989219244 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0540471Z 2025-12-04T11:17:06.0540996Z [W1204 11:03:51.994151454 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0541641Z 2025-12-04T11:17:06.0542159Z [W1204 11:03:51.994787421 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0543132Z 2025-12-04T11:17:06.0544075Z [W1204 11:03:51.994976393 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0544744Z 2025-12-04T11:17:06.0545261Z [W1204 11:03:51.001706610 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0545912Z 2025-12-04T11:17:06.0546437Z [W1204 11:03:51.002335134 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0547087Z 2025-12-04T11:17:06.0547613Z [W1204 11:03:51.002521553 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0548294Z 2025-12-04T11:17:06.0548807Z [W1204 11:03:51.108129637 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0549470Z 2025-12-04T11:17:06.0549987Z [W1204 11:03:51.109204486 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0550687Z 2025-12-04T11:17:06.0551201Z [W1204 11:03:51.109410627 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0551850Z 2025-12-04T11:17:06.0552370Z [W1204 11:03:51.114919871 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0553013Z 2025-12-04T11:17:06.0553541Z [W1204 11:03:51.115709095 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0554189Z 2025-12-04T11:17:06.0554705Z [W1204 11:03:51.115903444 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0555366Z 2025-12-04T11:17:06.0555882Z [W1204 11:03:51.123682772 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0556539Z 2025-12-04T11:17:06.0557056Z [W1204 11:03:51.124444887 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0557716Z 2025-12-04T11:17:06.0558226Z [W1204 11:03:51.124636898 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0558875Z 2025-12-04T11:17:06.0558989Z FAILED [0.5169s] [100%] 2025-12-04T11:17:06.0559165Z 2025-12-04T11:17:06.0559308Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.0560112Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.0560884Z Traceback (most recent call last): 2025-12-04T11:17:06.0561644Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0562516Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0563348Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0564113Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0564937Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0565819Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0566298Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0566548Z 2025-12-04T11:17:06.0566666Z Expected 1 but got 0. 2025-12-04T11:17:06.0566943Z Absolute difference: 1 2025-12-04T11:17:06.0567240Z Relative difference: 1.0 2025-12-04T11:17:06.0567426Z 2025-12-04T11:17:06.0567655Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0568991Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0570052Z 2025-12-04T11:17:06.0570323Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0570957Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0571438Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0572357Z inductor [('pattern_matcher_count', 6), ('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0573433Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0574208Z graph_break [] 2025-12-04T11:17:06.0574491Z aten_mm_info [('aten.mm_136_72_1024', 2)] 2025-12-04T11:17:06.0574968Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0576555Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.0578144Z if out == self.unknown_value: 2025-12-04T11:17:06.0579103Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0580065Z warnings.warn( 2025-12-04T11:17:06.0580958Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0581922Z warnings.warn( 2025-12-04T11:17:06.0582587Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.0583363Z Traceback (most recent call last): 2025-12-04T11:17:06.0584113Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0584990Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0585805Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0586569Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0587401Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0588285Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0588748Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0589010Z 2025-12-04T11:17:06.0589119Z Expected 1 but got 0. 2025-12-04T11:17:06.0589407Z Absolute difference: 1 2025-12-04T11:17:06.0589691Z Relative difference: 1.0 2025-12-04T11:17:06.0589895Z 2025-12-04T11:17:06.0590110Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0591379Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0592420Z 2025-12-04T11:17:06.0592698Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0593319Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0593798Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0594720Z inductor [('pattern_matcher_count', 6), ('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0595789Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0596433Z graph_break [] 2025-12-04T11:17:06.0596829Z aten_mm_info [('aten.mm_136_72_1024', 2)] 2025-12-04T11:17:06.0597308Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0598875Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.0600334Z if out == self.unknown_value: 2025-12-04T11:17:06.0601285Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0602253Z warnings.warn( 2025-12-04T11:17:06.0603171Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0604144Z warnings.warn( 2025-12-04T11:17:06.0604532Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0605053Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0605488Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0606552Z inductor [('pattern_matcher_count', 6), ('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0607491Z graph_break [] 2025-12-04T11:17:06.0607756Z aten_mm_info [('aten.mm_136_72_1024', 2)] 2025-12-04T11:17:06.0608225Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0609307Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0610269Z warnings.warn( 2025-12-04T11:17:06.0611151Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0612116Z warnings.warn( 2025-12-04T11:17:06.0612427Z =================================== FAILURES =================================== 2025-12-04T11:17:06.0613233Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.0613994Z Traceback (most recent call last): 2025-12-04T11:17:06.0614751Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0615633Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0616454Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0617292Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0618136Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0619027Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0619493Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0619761Z 2025-12-04T11:17:06.0619869Z Expected 1 but got 0. 2025-12-04T11:17:06.0620163Z Absolute difference: 1 2025-12-04T11:17:06.0620452Z Relative difference: 1.0 2025-12-04T11:17:06.0620655Z 2025-12-04T11:17:06.0620873Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0622145Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0623191Z 2025-12-04T11:17:06.0623470Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0623691Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0623876Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0624598Z inductor [('pattern_matcher_count', 6), ('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0624830Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0624943Z graph_break [] 2025-12-04T11:17:06.0625069Z aten_mm_info [('aten.mm_136_72_1024', 2)] 2025-12-04T11:17:06.0625283Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0626509Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.0626660Z if out == self.unknown_value: 2025-12-04T11:17:06.0627399Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0627534Z warnings.warn( 2025-12-04T11:17:06.0628256Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0628367Z warnings.warn( 2025-12-04T11:17:06.0628584Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0628701Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0628938Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0629639Z inductor [('pattern_matcher_count', 6), ('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0629752Z graph_break [] 2025-12-04T11:17:06.0629875Z aten_mm_info [('aten.mm_136_72_1024', 2)] 2025-12-04T11:17:06.0630094Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0630833Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0630933Z warnings.warn( 2025-12-04T11:17:06.0631663Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0631765Z warnings.warn( 2025-12-04T11:17:06.0631984Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0632108Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0632335Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0633036Z inductor [('pattern_matcher_count', 6), ('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0633148Z graph_break [] 2025-12-04T11:17:06.0633271Z aten_mm_info [('aten.mm_136_72_1024', 2)] 2025-12-04T11:17:06.0633503Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0634227Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0634326Z warnings.warn( 2025-12-04T11:17:06.0635057Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0635155Z warnings.warn( 2025-12-04T11:17:06.0635997Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-4c5f8d97ccaac9c3.xml - 2025-12-04T11:17:06.0636182Z =========================== short test summary info ============================ 2025-12-04T11:17:06.0637192Z FAILED [0.5169s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0637200Z 2025-12-04T11:17:06.0637323Z Expected 1 but got 0. 2025-12-04T11:17:06.0637430Z Absolute difference: 1 2025-12-04T11:17:06.0637540Z Relative difference: 1.0 2025-12-04T11:17:06.0637545Z 2025-12-04T11:17:06.0637776Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0638692Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0638727Z 2025-12-04T11:17:06.0639006Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0639190Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.0639391Z ================== 1 failed, 10 deselected, 2 rerun in 20.23s ================== 2025-12-04T11:17:06.0639530Z Got exit code 1 2025-12-04T11:17:06.0640356Z FAILED CONSISTENTLY: test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0640784Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:17:06.0641229Z W1204 11:04:02.810000 92606 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.0641901Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-fda94603e2b70838.xml 2025-12-04T11:17:06.0642078Z ============================= test session starts ============================== 2025-12-04T11:17:06.0642432Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.0642554Z cachedir: .pytest_cache 2025-12-04T11:17:06.0643075Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.0643197Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.0643317Z configfile: pytest.ini 2025-12-04T11:17:06.0643863Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.0644077Z collecting ... collected 58 items / 2 deselected / 56 selected 2025-12-04T11:17:06.0644230Z stepcurrent: skipping 2 already run items. 2025-12-04T11:17:06.0644342Z Running 9 items in this shard 2025-12-04T11:17:06.0644349Z 2025-12-04T11:17:06.0645249Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 ('RERUN', {'yellow': True}) [3.9356s] [ 11%] 2025-12-04T11:17:06.0646116Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 ('RERUN', {'yellow': True}) [0.4840s] [ 11%] 2025-12-04T11:17:06.0646911Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 FAILED [0.4848s] [ 11%] 2025-12-04T11:17:06.0646917Z 2025-12-04T11:17:06.0647059Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.0647570Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.0647707Z Traceback (most recent call last): 2025-12-04T11:17:06.0648217Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0648554Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0649020Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0649186Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0649735Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0649945Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0650079Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0650084Z 2025-12-04T11:17:06.0650207Z Expected 1 but got 2. 2025-12-04T11:17:06.0650315Z Absolute difference: 1 2025-12-04T11:17:06.0650469Z Relative difference: 1.0 2025-12-04T11:17:06.0650475Z 2025-12-04T11:17:06.0650691Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0651615Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.0651654Z 2025-12-04T11:17:06.0651938Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0652160Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0652285Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0652819Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.0653045Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0653153Z graph_break [] 2025-12-04T11:17:06.0653372Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0654111Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0654224Z warnings.warn( 2025-12-04T11:17:06.0654949Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0655057Z warnings.warn( 2025-12-04T11:17:06.0655570Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.0655689Z Traceback (most recent call last): 2025-12-04T11:17:06.0656214Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0656445Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0656992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0657158Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0657700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0657919Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0658051Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0658056Z 2025-12-04T11:17:06.0658162Z Expected 1 but got 2. 2025-12-04T11:17:06.0658282Z Absolute difference: 1 2025-12-04T11:17:06.0658391Z Relative difference: 1.0 2025-12-04T11:17:06.0658396Z 2025-12-04T11:17:06.0658622Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0659541Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.0659548Z 2025-12-04T11:17:06.0659820Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0660936Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0661070Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0661625Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.0661852Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0661950Z graph_break [] 2025-12-04T11:17:06.0662179Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0662913Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0663050Z warnings.warn( 2025-12-04T11:17:06.0663788Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0663892Z warnings.warn( 2025-12-04T11:17:06.0664176Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0664290Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0664518Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0665058Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.0665153Z graph_break [] 2025-12-04T11:17:06.0665370Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0666107Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0666207Z warnings.warn( 2025-12-04T11:17:06.0666943Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0667042Z warnings.warn( 2025-12-04T11:17:06.0667187Z =================================== FAILURES =================================== 2025-12-04T11:17:06.0667711Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.0667837Z Traceback (most recent call last): 2025-12-04T11:17:06.0668359Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0668590Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0669045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0669226Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0669765Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0669973Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0670117Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0670123Z 2025-12-04T11:17:06.0670227Z Expected 1 but got 2. 2025-12-04T11:17:06.0670348Z Absolute difference: 1 2025-12-04T11:17:06.0670457Z Relative difference: 1.0 2025-12-04T11:17:06.0670462Z 2025-12-04T11:17:06.0670675Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0671604Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.0671612Z 2025-12-04T11:17:06.0671878Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0672107Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0672276Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0672810Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.0673051Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0673148Z graph_break [] 2025-12-04T11:17:06.0673362Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0674103Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0674201Z warnings.warn( 2025-12-04T11:17:06.0674933Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0675062Z warnings.warn( 2025-12-04T11:17:06.0675284Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0675444Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0675672Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0676221Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.0676320Z graph_break [] 2025-12-04T11:17:06.0676537Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0677277Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0677378Z warnings.warn( 2025-12-04T11:17:06.0678108Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0678222Z warnings.warn( 2025-12-04T11:17:06.0678444Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0678575Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0678805Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0679338Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.0679451Z graph_break [] 2025-12-04T11:17:06.0679668Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0680411Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0680516Z warnings.warn( 2025-12-04T11:17:06.0681245Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0681356Z warnings.warn( 2025-12-04T11:17:06.0682205Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-fda94603e2b70838.xml - 2025-12-04T11:17:06.0682382Z =========================== short test summary info ============================ 2025-12-04T11:17:06.0683347Z FAILED [0.4848s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0683353Z 2025-12-04T11:17:06.0683462Z Expected 1 but got 2. 2025-12-04T11:17:06.0683588Z Absolute difference: 1 2025-12-04T11:17:06.0683702Z Relative difference: 1.0 2025-12-04T11:17:06.0683708Z 2025-12-04T11:17:06.0683926Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0684918Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.0684926Z 2025-12-04T11:17:06.0685197Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0685393Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.0685594Z =================== 1 failed, 2 deselected, 2 rerun in 4.94s =================== 2025-12-04T11:17:06.0685695Z Got exit code 1 2025-12-04T11:17:06.0685816Z Retrying single test... 2025-12-04T11:17:06.0686260Z W1204 11:04:23.120000 92782 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.0686971Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-ace483b73b471725.xml 2025-12-04T11:17:06.0687138Z ============================= test session starts ============================== 2025-12-04T11:17:06.0687490Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.0687660Z cachedir: .pytest_cache 2025-12-04T11:17:06.0688186Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.0688311Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.0688430Z configfile: pytest.ini 2025-12-04T11:17:06.0688984Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.0689221Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.0690222Z stepcurrent: skipping 2 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.0690341Z Running 1 items in this shard 2025-12-04T11:17:06.0690346Z 2025-12-04T11:17:06.0691646Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 [W1204 11:04:26.475594792 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0691652Z 2025-12-04T11:17:06.0692171Z [W1204 11:04:42.563328283 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0692176Z 2025-12-04T11:17:06.0692700Z [W1204 11:04:42.563590787 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0692708Z 2025-12-04T11:17:06.0693219Z [W1204 11:04:42.570857307 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0693224Z 2025-12-04T11:17:06.0693753Z [W1204 11:04:42.571510756 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0693760Z 2025-12-04T11:17:06.0694269Z [W1204 11:04:42.571699249 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0694273Z 2025-12-04T11:17:06.0694793Z [W1204 11:04:42.578477151 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0694798Z 2025-12-04T11:17:06.0695306Z [W1204 11:04:42.579099128 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0695313Z 2025-12-04T11:17:06.0695836Z [W1204 11:04:42.579283936 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0695840Z 2025-12-04T11:17:06.0696614Z [W1204 11:04:44.572818735 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0696622Z 2025-12-04T11:17:06.0697195Z [W1204 11:04:44.574458916 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0697219Z 2025-12-04T11:17:06.0697728Z [W1204 11:04:44.574670448 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0697733Z 2025-12-04T11:17:06.0698247Z [W1204 11:04:44.578507365 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0698251Z 2025-12-04T11:17:06.0698824Z [W1204 11:04:44.579126802 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0698828Z 2025-12-04T11:17:06.0699340Z [W1204 11:04:44.579321045 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0699384Z 2025-12-04T11:17:06.0699906Z [W1204 11:04:44.585302282 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0699911Z 2025-12-04T11:17:06.0700422Z [W1204 11:04:44.585917548 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0700427Z 2025-12-04T11:17:06.0700950Z [W1204 11:04:44.586112012 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0700955Z 2025-12-04T11:17:06.0701089Z ('RERUN', {'yellow': True}) [20.0183s] [100%] 2025-12-04T11:17:06.0702385Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 [W1204 11:04:45.003845327 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0702405Z 2025-12-04T11:17:06.0702918Z [W1204 11:04:45.004568077 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0702923Z 2025-12-04T11:17:06.0703431Z [W1204 11:04:45.004768053 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0703436Z 2025-12-04T11:17:06.0703961Z [W1204 11:04:45.008629117 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0703965Z 2025-12-04T11:17:06.0704480Z [W1204 11:04:45.009379812 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0704485Z 2025-12-04T11:17:06.0705012Z [W1204 11:04:45.009570244 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0705019Z 2025-12-04T11:17:06.0705527Z [W1204 11:04:45.015520804 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0705532Z 2025-12-04T11:17:06.0706055Z [W1204 11:04:45.016137214 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0706059Z 2025-12-04T11:17:06.0706570Z [W1204 11:04:45.016325140 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0706575Z 2025-12-04T11:17:06.0707097Z [W1204 11:04:45.100230961 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0707104Z 2025-12-04T11:17:06.0707674Z [W1204 11:04:45.100922429 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0707682Z 2025-12-04T11:17:06.0708193Z [W1204 11:04:45.101125147 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0708209Z 2025-12-04T11:17:06.0708718Z [W1204 11:04:45.104982697 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0708723Z 2025-12-04T11:17:06.0709233Z [W1204 11:04:45.105591660 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0709237Z 2025-12-04T11:17:06.0709759Z [W1204 11:04:45.105785950 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0709792Z 2025-12-04T11:17:06.0710306Z [W1204 11:04:45.111764726 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0710311Z 2025-12-04T11:17:06.0710859Z [W1204 11:04:45.112545715 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0710864Z 2025-12-04T11:17:06.0711370Z [W1204 11:04:45.112738204 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0711375Z 2025-12-04T11:17:06.0711518Z ('RERUN', {'yellow': True}) [0.4878s] [100%] 2025-12-04T11:17:06.0712809Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 [W1204 11:04:45.469241158 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0712817Z 2025-12-04T11:17:06.0713348Z [W1204 11:04:45.469963718 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0713354Z 2025-12-04T11:17:06.0713864Z [W1204 11:04:45.470188221 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0713869Z 2025-12-04T11:17:06.0714378Z [W1204 11:04:45.474095589 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0714383Z 2025-12-04T11:17:06.0714906Z [W1204 11:04:45.474870018 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0714910Z 2025-12-04T11:17:06.0715418Z [W1204 11:04:45.475065033 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0715424Z 2025-12-04T11:17:06.0715951Z [W1204 11:04:45.481080949 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0715955Z 2025-12-04T11:17:06.0716466Z [W1204 11:04:45.481689661 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0716473Z 2025-12-04T11:17:06.0716993Z [W1204 11:04:45.481877033 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0716998Z 2025-12-04T11:17:06.0717506Z [W1204 11:04:45.567092784 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0717511Z 2025-12-04T11:17:06.0718040Z [W1204 11:04:45.567808881 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0718047Z 2025-12-04T11:17:06.0718557Z [W1204 11:04:45.568011948 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0718618Z 2025-12-04T11:17:06.0719133Z [W1204 11:04:45.571925034 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0719153Z 2025-12-04T11:17:06.0719666Z [W1204 11:04:45.572567109 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0719671Z 2025-12-04T11:17:06.0720180Z [W1204 11:04:45.572762678 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0720184Z 2025-12-04T11:17:06.0720704Z [W1204 11:04:45.578687852 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0720736Z 2025-12-04T11:17:06.0721249Z [W1204 11:04:45.579463261 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0721253Z 2025-12-04T11:17:06.0721782Z [W1204 11:04:45.579658487 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0721816Z 2025-12-04T11:17:06.0721922Z FAILED [0.4661s] [100%] 2025-12-04T11:17:06.0721927Z 2025-12-04T11:17:06.0722083Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.0722600Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.0722724Z Traceback (most recent call last): 2025-12-04T11:17:06.0723250Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0723485Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0723947Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0724132Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0724669Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0724893Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0725024Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0725029Z 2025-12-04T11:17:06.0725134Z Expected 1 but got 2. 2025-12-04T11:17:06.0725251Z Absolute difference: 1 2025-12-04T11:17:06.0725358Z Relative difference: 1.0 2025-12-04T11:17:06.0725363Z 2025-12-04T11:17:06.0725591Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0726511Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.0726520Z 2025-12-04T11:17:06.0726792Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0727025Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0727142Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0727690Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.0727916Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0728015Z graph_break [] 2025-12-04T11:17:06.0728244Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0729465Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.0729585Z if out == self.unknown_value: 2025-12-04T11:17:06.0730383Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0730488Z warnings.warn( 2025-12-04T11:17:06.0731343Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0731486Z warnings.warn( 2025-12-04T11:17:06.0732471Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.0732611Z Traceback (most recent call last): 2025-12-04T11:17:06.0733120Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0733426Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0733888Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0734056Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0734640Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0734847Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0734979Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0735002Z 2025-12-04T11:17:06.0735109Z Expected 1 but got 2. 2025-12-04T11:17:06.0735215Z Absolute difference: 1 2025-12-04T11:17:06.0735337Z Relative difference: 1.0 2025-12-04T11:17:06.0735342Z 2025-12-04T11:17:06.0735561Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0736487Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.0736495Z 2025-12-04T11:17:06.0736781Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0737074Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0737209Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0737744Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.0737970Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0738084Z graph_break [] 2025-12-04T11:17:06.0738302Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0739528Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.0739648Z if out == self.unknown_value: 2025-12-04T11:17:06.0740379Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0740498Z warnings.warn( 2025-12-04T11:17:06.0741221Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0741323Z warnings.warn( 2025-12-04T11:17:06.0741556Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0741670Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0741913Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0742450Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.0742551Z graph_break [] 2025-12-04T11:17:06.0742841Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0743569Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0743682Z warnings.warn( 2025-12-04T11:17:06.0744405Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0744505Z warnings.warn( 2025-12-04T11:17:06.0744663Z =================================== FAILURES =================================== 2025-12-04T11:17:06.0745177Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.0745329Z Traceback (most recent call last): 2025-12-04T11:17:06.0745854Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0746092Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0746599Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0746765Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0747304Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0747522Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0747654Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0747659Z 2025-12-04T11:17:06.0747776Z Expected 1 but got 2. 2025-12-04T11:17:06.0747883Z Absolute difference: 1 2025-12-04T11:17:06.0747996Z Relative difference: 1.0 2025-12-04T11:17:06.0748000Z 2025-12-04T11:17:06.0748228Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0749156Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.0749164Z 2025-12-04T11:17:06.0749431Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0749662Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0749778Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0750326Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.0750552Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0750655Z graph_break [] 2025-12-04T11:17:06.0750887Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0752108Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.0752243Z if out == self.unknown_value: 2025-12-04T11:17:06.0752975Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0753077Z warnings.warn( 2025-12-04T11:17:06.0753819Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0753920Z warnings.warn( 2025-12-04T11:17:06.0754139Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0754270Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0754499Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0755103Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.0755206Z graph_break [] 2025-12-04T11:17:06.0755422Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0756162Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0756267Z warnings.warn( 2025-12-04T11:17:06.0756998Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0757127Z warnings.warn( 2025-12-04T11:17:06.0757341Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0757473Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0757708Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0758241Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.0758393Z graph_break [] 2025-12-04T11:17:06.0758612Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0759353Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0759454Z warnings.warn( 2025-12-04T11:17:06.0760177Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0760291Z warnings.warn( 2025-12-04T11:17:06.0761137Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-ace483b73b471725.xml - 2025-12-04T11:17:06.0761324Z =========================== short test summary info ============================ 2025-12-04T11:17:06.0762281Z FAILED [0.4661s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0762288Z 2025-12-04T11:17:06.0762391Z Expected 1 but got 2. 2025-12-04T11:17:06.0762510Z Absolute difference: 1 2025-12-04T11:17:06.0762619Z Relative difference: 1.0 2025-12-04T11:17:06.0762624Z 2025-12-04T11:17:06.0762843Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0763783Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.0763790Z 2025-12-04T11:17:06.0764063Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0764258Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.0764459Z ================== 1 failed, 10 deselected, 2 rerun in 21.01s ================== 2025-12-04T11:17:06.0764558Z Got exit code 1 2025-12-04T11:17:06.0764675Z Retrying single test... 2025-12-04T11:17:06.0765125Z W1204 11:04:57.044000 92963 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.0765796Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-841659f3603156bc.xml 2025-12-04T11:17:06.0765958Z ============================= test session starts ============================== 2025-12-04T11:17:06.0766307Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.0766428Z cachedir: .pytest_cache 2025-12-04T11:17:06.0767008Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.0767154Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.0767263Z configfile: pytest.ini 2025-12-04T11:17:06.0767808Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.0768038Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.0769039Z stepcurrent: skipping 2 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.0769187Z Running 1 items in this shard 2025-12-04T11:17:06.0769203Z 2025-12-04T11:17:06.0770496Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 [W1204 11:05:00.361397339 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0770529Z 2025-12-04T11:17:06.0771051Z [W1204 11:05:15.643786530 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0771070Z 2025-12-04T11:17:06.0771586Z [W1204 11:05:15.644049316 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0771591Z 2025-12-04T11:17:06.0772105Z [W1204 11:05:15.651464064 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0772112Z 2025-12-04T11:17:06.0772641Z [W1204 11:05:15.652170094 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0772645Z 2025-12-04T11:17:06.0773157Z [W1204 11:05:15.652364877 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0773164Z 2025-12-04T11:17:06.0773687Z [W1204 11:05:15.659147610 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0773692Z 2025-12-04T11:17:06.0774200Z [W1204 11:05:15.659795551 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0774204Z 2025-12-04T11:17:06.0774728Z [W1204 11:05:15.659981301 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0774733Z 2025-12-04T11:17:06.0775246Z [W1204 11:05:17.660031764 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0775250Z 2025-12-04T11:17:06.0775774Z [W1204 11:05:17.661764265 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0775781Z 2025-12-04T11:17:06.0776288Z [W1204 11:05:17.661978076 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0776293Z 2025-12-04T11:17:06.0776805Z [W1204 11:05:17.665927697 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0776810Z 2025-12-04T11:17:06.0777411Z [W1204 11:05:17.666583675 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0777417Z 2025-12-04T11:17:06.0777923Z [W1204 11:05:17.666779197 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0777930Z 2025-12-04T11:17:06.0778532Z [W1204 11:05:17.672884926 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0778539Z 2025-12-04T11:17:06.0779053Z [W1204 11:05:17.673533717 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0779058Z 2025-12-04T11:17:06.0779581Z [W1204 11:05:17.673726706 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0779586Z 2025-12-04T11:17:06.0779718Z ('RERUN', {'yellow': True}) [19.1910s] [100%] 2025-12-04T11:17:06.0781018Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 [W1204 11:05:18.101689502 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0781056Z 2025-12-04T11:17:06.0781573Z [W1204 11:05:18.102435182 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0781606Z 2025-12-04T11:17:06.0782119Z [W1204 11:05:18.102635276 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0782137Z 2025-12-04T11:17:06.0782646Z [W1204 11:05:18.106558783 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0782651Z 2025-12-04T11:17:06.0783162Z [W1204 11:05:18.107332002 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0783166Z 2025-12-04T11:17:06.0783687Z [W1204 11:05:18.107520875 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0783694Z 2025-12-04T11:17:06.0784208Z [W1204 11:05:18.113599401 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0784215Z 2025-12-04T11:17:06.0784740Z [W1204 11:05:18.114219187 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0784745Z 2025-12-04T11:17:06.0785250Z [W1204 11:05:18.114405623 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0785255Z 2025-12-04T11:17:06.0785777Z [W1204 11:05:18.200575394 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0785782Z 2025-12-04T11:17:06.0786291Z [W1204 11:05:18.201309727 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0786298Z 2025-12-04T11:17:06.0786823Z [W1204 11:05:18.201517424 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0786827Z 2025-12-04T11:17:06.0787339Z [W1204 11:05:18.205437200 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0787343Z 2025-12-04T11:17:06.0787851Z [W1204 11:05:18.206064832 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0787868Z 2025-12-04T11:17:06.0788377Z [W1204 11:05:18.206272626 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0788381Z 2025-12-04T11:17:06.0788887Z [W1204 11:05:18.212274187 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0788894Z 2025-12-04T11:17:06.0789417Z [W1204 11:05:18.213053745 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0789475Z 2025-12-04T11:17:06.0789989Z [W1204 11:05:18.213247453 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0789996Z 2025-12-04T11:17:06.0790141Z ('RERUN', {'yellow': True}) [0.4991s] [100%] 2025-12-04T11:17:06.0791423Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 [W1204 11:05:18.573432688 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0791428Z 2025-12-04T11:17:06.0791953Z [W1204 11:05:18.574152454 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0791986Z 2025-12-04T11:17:06.0792500Z [W1204 11:05:18.574354337 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0792509Z 2025-12-04T11:17:06.0793020Z [W1204 11:05:18.578282421 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0793070Z 2025-12-04T11:17:06.0793580Z [W1204 11:05:18.579035625 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0793585Z 2025-12-04T11:17:06.0794089Z [W1204 11:05:18.579225918 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0794094Z 2025-12-04T11:17:06.0794613Z [W1204 11:05:18.585197247 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0794621Z 2025-12-04T11:17:06.0795128Z [W1204 11:05:18.585815466 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0795133Z 2025-12-04T11:17:06.0795655Z [W1204 11:05:18.586001607 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0795662Z 2025-12-04T11:17:06.0796411Z [W1204 11:05:18.671089220 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0796419Z 2025-12-04T11:17:06.0796947Z [W1204 11:05:18.671806714 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0796951Z 2025-12-04T11:17:06.0797461Z [W1204 11:05:18.672011584 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0797469Z 2025-12-04T11:17:06.0797993Z [W1204 11:05:18.675888853 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0797998Z 2025-12-04T11:17:06.0798509Z [W1204 11:05:18.676510776 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0798517Z 2025-12-04T11:17:06.0799028Z [W1204 11:05:18.676705112 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0799047Z 2025-12-04T11:17:06.0799562Z [W1204 11:05:18.682681999 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0799567Z 2025-12-04T11:17:06.0800079Z [W1204 11:05:18.683469537 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0800084Z 2025-12-04T11:17:06.0800606Z [W1204 11:05:18.683664017 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0800611Z 2025-12-04T11:17:06.0800715Z FAILED [0.4695s] [100%] 2025-12-04T11:17:06.0800827Z 2025-12-04T11:17:06.0800987Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.0801511Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.0801636Z Traceback (most recent call last): 2025-12-04T11:17:06.0802170Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0802403Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0802887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0803090Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0803632Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0803859Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0804034Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0804039Z 2025-12-04T11:17:06.0804146Z Expected 1 but got 2. 2025-12-04T11:17:06.0804266Z Absolute difference: 1 2025-12-04T11:17:06.0804377Z Relative difference: 1.0 2025-12-04T11:17:06.0804382Z 2025-12-04T11:17:06.0804614Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0805538Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.0805544Z 2025-12-04T11:17:06.0805815Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0806050Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0806168Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0806718Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.0806948Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0807052Z graph_break [] 2025-12-04T11:17:06.0807283Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0808498Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.0808626Z if out == self.unknown_value: 2025-12-04T11:17:06.0809351Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0809455Z warnings.warn( 2025-12-04T11:17:06.0810195Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0810297Z warnings.warn( 2025-12-04T11:17:06.0810809Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.0810943Z Traceback (most recent call last): 2025-12-04T11:17:06.0811450Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0811696Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0812155Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0812323Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0812871Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0813136Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0813283Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0813289Z 2025-12-04T11:17:06.0813395Z Expected 1 but got 2. 2025-12-04T11:17:06.0813503Z Absolute difference: 1 2025-12-04T11:17:06.0813623Z Relative difference: 1.0 2025-12-04T11:17:06.0813628Z 2025-12-04T11:17:06.0813845Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0814761Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.0814778Z 2025-12-04T11:17:06.0815075Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0815297Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0815426Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0815966Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.0816227Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0816336Z graph_break [] 2025-12-04T11:17:06.0816554Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0817847Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.0817967Z if out == self.unknown_value: 2025-12-04T11:17:06.0818703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0818818Z warnings.warn( 2025-12-04T11:17:06.0819542Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0819658Z warnings.warn( 2025-12-04T11:17:06.0819872Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0819983Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0820220Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0820753Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.0820850Z graph_break [] 2025-12-04T11:17:06.0821076Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0821800Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0821911Z warnings.warn( 2025-12-04T11:17:06.0822632Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0822730Z warnings.warn( 2025-12-04T11:17:06.0822889Z =================================== FAILURES =================================== 2025-12-04T11:17:06.0823408Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.0823541Z Traceback (most recent call last): 2025-12-04T11:17:06.0824062Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0824297Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0824770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0825021Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0825564Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0825785Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0825918Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0825923Z 2025-12-04T11:17:06.0826045Z Expected 1 but got 2. 2025-12-04T11:17:06.0826154Z Absolute difference: 1 2025-12-04T11:17:06.0826265Z Relative difference: 1.0 2025-12-04T11:17:06.0826270Z 2025-12-04T11:17:06.0826504Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0827426Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.0827463Z 2025-12-04T11:17:06.0827751Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0827999Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0828114Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0828667Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.0828896Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0828996Z graph_break [] 2025-12-04T11:17:06.0829230Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0830437Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.0830572Z if out == self.unknown_value: 2025-12-04T11:17:06.0831304Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0831410Z warnings.warn( 2025-12-04T11:17:06.0832146Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0832246Z warnings.warn( 2025-12-04T11:17:06.0832483Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0832600Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0832829Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0833375Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.0833473Z graph_break [] 2025-12-04T11:17:06.0833694Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0834436Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0834539Z warnings.warn( 2025-12-04T11:17:06.0835270Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0835366Z warnings.warn( 2025-12-04T11:17:06.0835582Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0835710Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0835939Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0836470Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.0836634Z graph_break [] 2025-12-04T11:17:06.0836854Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0837596Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0837695Z warnings.warn( 2025-12-04T11:17:06.0838412Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0838521Z warnings.warn( 2025-12-04T11:17:06.0839360Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-841659f3603156bc.xml - 2025-12-04T11:17:06.0839577Z =========================== short test summary info ============================ 2025-12-04T11:17:06.0840530Z FAILED [0.4695s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0840564Z 2025-12-04T11:17:06.0840671Z Expected 1 but got 2. 2025-12-04T11:17:06.0840792Z Absolute difference: 1 2025-12-04T11:17:06.0840898Z Relative difference: 1.0 2025-12-04T11:17:06.0840903Z 2025-12-04T11:17:06.0841137Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0842052Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.0842056Z 2025-12-04T11:17:06.0842321Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0842513Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.0842713Z ================== 1 failed, 10 deselected, 2 rerun in 20.19s ================== 2025-12-04T11:17:06.0842825Z Got exit code 1 2025-12-04T11:17:06.0843660Z FAILED CONSISTENTLY: test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.0844071Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:17:06.0844527Z W1204 11:05:30.249000 93144 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.0845197Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-e4e48c14b07c6744.xml 2025-12-04T11:17:06.0845376Z ============================= test session starts ============================== 2025-12-04T11:17:06.0845725Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.0845845Z cachedir: .pytest_cache 2025-12-04T11:17:06.0846382Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.0846507Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.0846614Z configfile: pytest.ini 2025-12-04T11:17:06.0847172Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.0847387Z collecting ... collected 58 items / 3 deselected / 55 selected 2025-12-04T11:17:06.0847541Z stepcurrent: skipping 3 already run items. 2025-12-04T11:17:06.0847657Z Running 8 items in this shard 2025-12-04T11:17:06.0847662Z 2025-12-04T11:17:06.0848529Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 ('RERUN', {'yellow': True}) [4.1293s] [ 12%] 2025-12-04T11:17:06.0849459Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 ('RERUN', {'yellow': True}) [0.4528s] [ 12%] 2025-12-04T11:17:06.0850243Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 FAILED [0.4420s] [ 12%] 2025-12-04T11:17:06.0850249Z 2025-12-04T11:17:06.0850402Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.0850907Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.0851026Z Traceback (most recent call last): 2025-12-04T11:17:06.0851577Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0851808Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0852283Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0852486Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0853023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0853238Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0853369Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0853375Z 2025-12-04T11:17:06.0853490Z Expected 1 but got 2. 2025-12-04T11:17:06.0853595Z Absolute difference: 1 2025-12-04T11:17:06.0853702Z Relative difference: 1.0 2025-12-04T11:17:06.0853707Z 2025-12-04T11:17:06.0853933Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0854843Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0854849Z 2025-12-04T11:17:06.0855126Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0855341Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0855457Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0856369Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0856597Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0856693Z graph_break [] 2025-12-04T11:17:06.0856999Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0857754Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0857869Z warnings.warn( 2025-12-04T11:17:06.0858595Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0858696Z warnings.warn( 2025-12-04T11:17:06.0859218Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.0859339Z Traceback (most recent call last): 2025-12-04T11:17:06.0859867Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0860098Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0860561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0860741Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0861340Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0861550Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0861691Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0861696Z 2025-12-04T11:17:06.0861801Z Expected 1 but got 2. 2025-12-04T11:17:06.0861915Z Absolute difference: 1 2025-12-04T11:17:06.0862023Z Relative difference: 1.0 2025-12-04T11:17:06.0862029Z 2025-12-04T11:17:06.0862241Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0863155Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0863188Z 2025-12-04T11:17:06.0863456Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0863686Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0863842Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0864737Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0864972Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0865067Z graph_break [] 2025-12-04T11:17:06.0865284Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0866024Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0866124Z warnings.warn( 2025-12-04T11:17:06.0866856Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0866957Z warnings.warn( 2025-12-04T11:17:06.0867176Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0867306Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0867530Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0868441Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0868538Z graph_break [] 2025-12-04T11:17:06.0868754Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0869492Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0869594Z warnings.warn( 2025-12-04T11:17:06.0870317Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0870427Z warnings.warn( 2025-12-04T11:17:06.0870568Z =================================== FAILURES =================================== 2025-12-04T11:17:06.0871090Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.0871210Z Traceback (most recent call last): 2025-12-04T11:17:06.0871714Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0871957Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0872417Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0872591Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0873181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0873389Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0873531Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0873536Z 2025-12-04T11:17:06.0873646Z Expected 1 but got 2. 2025-12-04T11:17:06.0873749Z Absolute difference: 1 2025-12-04T11:17:06.0873867Z Relative difference: 1.0 2025-12-04T11:17:06.0873872Z 2025-12-04T11:17:06.0874086Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0875325Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0875374Z 2025-12-04T11:17:06.0875645Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0875869Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0876028Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0876922Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0877155Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0877254Z graph_break [] 2025-12-04T11:17:06.0877470Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0878213Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0878319Z warnings.warn( 2025-12-04T11:17:06.0879053Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0879156Z warnings.warn( 2025-12-04T11:17:06.0879374Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0879502Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0879732Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0880635Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0880743Z graph_break [] 2025-12-04T11:17:06.0880956Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0881692Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0881794Z warnings.warn( 2025-12-04T11:17:06.0882517Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0882631Z warnings.warn( 2025-12-04T11:17:06.0882844Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0882968Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0896542Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0898190Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0898325Z graph_break [] 2025-12-04T11:17:06.0898560Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0899473Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0899599Z warnings.warn( 2025-12-04T11:17:06.0900329Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0900432Z warnings.warn( 2025-12-04T11:17:06.0901295Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-e4e48c14b07c6744.xml - 2025-12-04T11:17:06.0901473Z =========================== short test summary info ============================ 2025-12-04T11:17:06.0902431Z FAILED [0.4420s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0902486Z 2025-12-04T11:17:06.0902603Z Expected 1 but got 2. 2025-12-04T11:17:06.0902713Z Absolute difference: 1 2025-12-04T11:17:06.0902885Z Relative difference: 1.0 2025-12-04T11:17:06.0902891Z 2025-12-04T11:17:06.0903116Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0904034Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0904039Z 2025-12-04T11:17:06.0904310Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0904495Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.0904709Z =================== 1 failed, 3 deselected, 2 rerun in 5.06s =================== 2025-12-04T11:17:06.0904810Z Got exit code 1 2025-12-04T11:17:06.0904931Z Retrying single test... 2025-12-04T11:17:06.0905386Z W1204 11:05:50.123000 93340 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.0906059Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-887ea8bf31a36206.xml 2025-12-04T11:17:06.0906238Z ============================= test session starts ============================== 2025-12-04T11:17:06.0906593Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.0906718Z cachedir: .pytest_cache 2025-12-04T11:17:06.0907240Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.0907366Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.0907493Z configfile: pytest.ini 2025-12-04T11:17:06.0908041Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.0908269Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.0909276Z stepcurrent: skipping 3 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0909393Z Running 1 items in this shard 2025-12-04T11:17:06.0909399Z 2025-12-04T11:17:06.0910690Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 [W1204 11:05:55.542956251 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0910696Z 2025-12-04T11:17:06.0911219Z [W1204 11:06:11.602871059 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0911224Z 2025-12-04T11:17:06.0911809Z [W1204 11:06:11.603184220 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0911817Z 2025-12-04T11:17:06.0912333Z [W1204 11:06:11.610594249 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0912338Z 2025-12-04T11:17:06.0912865Z [W1204 11:06:11.611295129 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0912870Z 2025-12-04T11:17:06.0913382Z [W1204 11:06:11.611490676 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0913386Z 2025-12-04T11:17:06.0913900Z [W1204 11:06:11.618491313 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0913943Z 2025-12-04T11:17:06.0914462Z [W1204 11:06:11.619122734 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0914494Z 2025-12-04T11:17:06.0915005Z [W1204 11:06:11.619330582 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0915009Z 2025-12-04T11:17:06.0915535Z [W1204 11:06:12.752182801 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0915540Z 2025-12-04T11:17:06.0916050Z [W1204 11:06:12.753878720 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0916055Z 2025-12-04T11:17:06.0916579Z [W1204 11:06:12.754085523 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0916586Z 2025-12-04T11:17:06.0917386Z [W1204 11:06:12.757979659 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0917392Z 2025-12-04T11:17:06.0917921Z [W1204 11:06:12.758617235 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0917926Z 2025-12-04T11:17:06.0918437Z [W1204 11:06:12.758817935 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0918442Z 2025-12-04T11:17:06.0918971Z [W1204 11:06:12.764854069 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0918976Z 2025-12-04T11:17:06.0919485Z [W1204 11:06:12.765481338 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0919493Z 2025-12-04T11:17:06.0920000Z [W1204 11:06:12.765675506 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0920009Z 2025-12-04T11:17:06.0920157Z ('RERUN', {'yellow': True}) [20.2055s] [100%] 2025-12-04T11:17:06.0921430Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 [W1204 11:06:12.154698154 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0921435Z 2025-12-04T11:17:06.0921961Z [W1204 11:06:12.155406492 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0921966Z 2025-12-04T11:17:06.0922473Z [W1204 11:06:12.155600645 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0922480Z 2025-12-04T11:17:06.0923105Z [W1204 11:06:12.159483676 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0923110Z 2025-12-04T11:17:06.0923628Z [W1204 11:06:12.160158891 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0923632Z 2025-12-04T11:17:06.0924156Z [W1204 11:06:12.160352039 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0924160Z 2025-12-04T11:17:06.0924670Z [W1204 11:06:12.166350257 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0924675Z 2025-12-04T11:17:06.0925183Z [W1204 11:06:12.166954956 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0925245Z 2025-12-04T11:17:06.0925754Z [W1204 11:06:12.167143236 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0925763Z 2025-12-04T11:17:06.0926273Z [W1204 11:06:12.252957578 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0926307Z 2025-12-04T11:17:06.0926829Z [W1204 11:06:12.253673541 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0926834Z 2025-12-04T11:17:06.0927343Z [W1204 11:06:12.253876931 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0927348Z 2025-12-04T11:17:06.0927870Z [W1204 11:06:12.257727729 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0927878Z 2025-12-04T11:17:06.0928386Z [W1204 11:06:12.258341506 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0928391Z 2025-12-04T11:17:06.0928919Z [W1204 11:06:12.258537991 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0928926Z 2025-12-04T11:17:06.0929436Z [W1204 11:06:12.264518760 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0929440Z 2025-12-04T11:17:06.0929959Z [W1204 11:06:12.265305724 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0929964Z 2025-12-04T11:17:06.0930474Z [W1204 11:06:12.265500730 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0930481Z 2025-12-04T11:17:06.0930615Z ('RERUN', {'yellow': True}) [0.4605s] [100%] 2025-12-04T11:17:06.0931913Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 [W1204 11:06:12.585085159 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0931920Z 2025-12-04T11:17:06.0932437Z [W1204 11:06:12.585782317 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0932441Z 2025-12-04T11:17:06.0932968Z [W1204 11:06:12.585978285 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0932973Z 2025-12-04T11:17:06.0933484Z [W1204 11:06:12.589855892 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0933491Z 2025-12-04T11:17:06.0934016Z [W1204 11:06:12.590521142 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0934020Z 2025-12-04T11:17:06.0934590Z [W1204 11:06:12.590714753 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0934598Z 2025-12-04T11:17:06.0935128Z [W1204 11:06:12.596655624 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0935132Z 2025-12-04T11:17:06.0935646Z [W1204 11:06:12.597254870 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0935650Z 2025-12-04T11:17:06.0936180Z [W1204 11:06:12.597442477 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0936184Z 2025-12-04T11:17:06.0936724Z [W1204 11:06:12.682288261 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0936729Z 2025-12-04T11:17:06.0937324Z [W1204 11:06:12.682996017 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0937362Z 2025-12-04T11:17:06.0937890Z [W1204 11:06:12.683199732 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0937894Z 2025-12-04T11:17:06.0938407Z [W1204 11:06:12.687025856 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0938412Z 2025-12-04T11:17:06.0938939Z [W1204 11:06:12.687643216 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0938943Z 2025-12-04T11:17:06.0939457Z [W1204 11:06:12.687837463 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0939464Z 2025-12-04T11:17:06.0939994Z [W1204 11:06:12.693808796 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0940000Z 2025-12-04T11:17:06.0940507Z [W1204 11:06:12.694587899 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0940512Z 2025-12-04T11:17:06.0941033Z [W1204 11:06:12.694783664 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0941038Z 2025-12-04T11:17:06.0941142Z FAILED [0.4274s] [100%] 2025-12-04T11:17:06.0941147Z 2025-12-04T11:17:06.0941294Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.0941821Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.0941950Z Traceback (most recent call last): 2025-12-04T11:17:06.0942482Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0942718Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0943190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0943367Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0943912Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0944121Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0944271Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0944276Z 2025-12-04T11:17:06.0944384Z Expected 1 but got 2. 2025-12-04T11:17:06.0944508Z Absolute difference: 1 2025-12-04T11:17:06.0944617Z Relative difference: 1.0 2025-12-04T11:17:06.0944622Z 2025-12-04T11:17:06.0944838Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0945898Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0945906Z 2025-12-04T11:17:06.0946181Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0946421Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0946539Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0947446Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0947727Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0947828Z graph_break [] 2025-12-04T11:17:06.0948062Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0949282Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.0949434Z if out == self.unknown_value: 2025-12-04T11:17:06.0950175Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0950278Z warnings.warn( 2025-12-04T11:17:06.0951012Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0951118Z warnings.warn( 2025-12-04T11:17:06.0951629Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.0951772Z Traceback (most recent call last): 2025-12-04T11:17:06.0952283Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0952521Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0952995Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0953163Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0953716Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0953927Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0954062Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0954068Z 2025-12-04T11:17:06.0954189Z Expected 1 but got 2. 2025-12-04T11:17:06.0954296Z Absolute difference: 1 2025-12-04T11:17:06.0954407Z Relative difference: 1.0 2025-12-04T11:17:06.0954412Z 2025-12-04T11:17:06.0954646Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0955558Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0955565Z 2025-12-04T11:17:06.0955850Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0956070Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0956189Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0957104Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0957336Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0957514Z graph_break [] 2025-12-04T11:17:06.0957738Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0958954Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.0959084Z if out == self.unknown_value: 2025-12-04T11:17:06.0959815Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0959930Z warnings.warn( 2025-12-04T11:17:06.0960653Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0960785Z warnings.warn( 2025-12-04T11:17:06.0961022Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0961139Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0961404Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0962312Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0962412Z graph_break [] 2025-12-04T11:17:06.0962642Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0963372Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0963476Z warnings.warn( 2025-12-04T11:17:06.0964216Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0964315Z warnings.warn( 2025-12-04T11:17:06.0964475Z =================================== FAILURES =================================== 2025-12-04T11:17:06.0964984Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.0965109Z Traceback (most recent call last): 2025-12-04T11:17:06.0965632Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.0965866Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.0966340Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.0966509Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.0967045Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.0967270Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.0967408Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0967413Z 2025-12-04T11:17:06.0967521Z Expected 1 but got 2. 2025-12-04T11:17:06.0967644Z Absolute difference: 1 2025-12-04T11:17:06.0967755Z Relative difference: 1.0 2025-12-04T11:17:06.0967760Z 2025-12-04T11:17:06.0967991Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0968900Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0968905Z 2025-12-04T11:17:06.0969181Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0969412Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0969534Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0970525Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0970761Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0970862Z graph_break [] 2025-12-04T11:17:06.0971095Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0972303Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.0972466Z if out == self.unknown_value: 2025-12-04T11:17:06.0973201Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0973306Z warnings.warn( 2025-12-04T11:17:06.0974074Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0974178Z warnings.warn( 2025-12-04T11:17:06.0974396Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0974528Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0974758Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0975664Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0975765Z graph_break [] 2025-12-04T11:17:06.0975982Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0976731Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0976835Z warnings.warn( 2025-12-04T11:17:06.0977655Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0977759Z warnings.warn( 2025-12-04T11:17:06.0977976Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.0978107Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.0978338Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.0979234Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.0979350Z graph_break [] 2025-12-04T11:17:06.0979571Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.0980318Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0980419Z warnings.warn( 2025-12-04T11:17:06.0981140Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.0981256Z warnings.warn( 2025-12-04T11:17:06.0982094Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-887ea8bf31a36206.xml - 2025-12-04T11:17:06.0982284Z =========================== short test summary info ============================ 2025-12-04T11:17:06.0983293Z FAILED [0.4274s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.0983301Z 2025-12-04T11:17:06.0983411Z Expected 1 but got 2. 2025-12-04T11:17:06.0983538Z Absolute difference: 1 2025-12-04T11:17:06.0983650Z Relative difference: 1.0 2025-12-04T11:17:06.0983655Z 2025-12-04T11:17:06.0983887Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.0984792Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0984798Z 2025-12-04T11:17:06.0985067Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.0985294Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.0985499Z ================== 1 failed, 10 deselected, 2 rerun in 21.13s ================== 2025-12-04T11:17:06.0985613Z Got exit code 1 2025-12-04T11:17:06.0985752Z Retrying single test... 2025-12-04T11:17:06.0986201Z W1204 11:06:24.137000 93541 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.0986883Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-d486c0fac95f9182.xml 2025-12-04T11:17:06.0987048Z ============================= test session starts ============================== 2025-12-04T11:17:06.0987399Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.0987525Z cachedir: .pytest_cache 2025-12-04T11:17:06.0988048Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.0988193Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.0988303Z configfile: pytest.ini 2025-12-04T11:17:06.0988855Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.0989092Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.0990083Z stepcurrent: skipping 3 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.0990213Z Running 1 items in this shard 2025-12-04T11:17:06.0990218Z 2025-12-04T11:17:06.0991503Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 [W1204 11:06:29.569013009 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0991511Z 2025-12-04T11:17:06.0992035Z [W1204 11:06:45.714036068 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0992057Z 2025-12-04T11:17:06.0992577Z [W1204 11:06:45.714291142 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0992582Z 2025-12-04T11:17:06.0993093Z [W1204 11:06:45.721600129 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0993099Z 2025-12-04T11:17:06.0993618Z [W1204 11:06:45.722284712 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0993622Z 2025-12-04T11:17:06.0994130Z [W1204 11:06:45.722474104 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0994137Z 2025-12-04T11:17:06.0994720Z [W1204 11:06:45.729473343 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0994727Z 2025-12-04T11:17:06.0995238Z [W1204 11:06:45.730104033 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0995243Z 2025-12-04T11:17:06.0995765Z [W1204 11:06:45.730292282 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0995769Z 2025-12-04T11:17:06.0996442Z [W1204 11:06:45.863151553 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0996447Z 2025-12-04T11:17:06.0996969Z [W1204 11:06:45.864841574 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0997038Z 2025-12-04T11:17:06.0997552Z [W1204 11:06:45.865050301 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0997557Z 2025-12-04T11:17:06.0998112Z [W1204 11:06:45.868976604 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0998131Z 2025-12-04T11:17:06.0998641Z [W1204 11:06:45.869617985 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0998645Z 2025-12-04T11:17:06.0999150Z [W1204 11:06:45.869812584 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0999154Z 2025-12-04T11:17:06.0999675Z [W1204 11:06:45.875860092 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.0999682Z 2025-12-04T11:17:06.1000197Z [W1204 11:06:45.876498261 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1000206Z 2025-12-04T11:17:06.1000723Z [W1204 11:06:45.876690600 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1000729Z 2025-12-04T11:17:06.1000879Z ('RERUN', {'yellow': True}) [19.2877s] [100%] 2025-12-04T11:17:06.1002158Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 [W1204 11:06:45.264534533 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1002164Z 2025-12-04T11:17:06.1002676Z [W1204 11:06:45.265265738 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1002696Z 2025-12-04T11:17:06.1003200Z [W1204 11:06:45.265464470 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1003208Z 2025-12-04T11:17:06.1003713Z [W1204 11:06:45.269442739 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1003720Z 2025-12-04T11:17:06.1004243Z [W1204 11:06:45.270102438 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1004248Z 2025-12-04T11:17:06.1004756Z [W1204 11:06:45.270294140 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1004761Z 2025-12-04T11:17:06.1005286Z [W1204 11:06:45.276368141 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1005294Z 2025-12-04T11:17:06.1005802Z [W1204 11:06:45.276973790 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1005890Z 2025-12-04T11:17:06.1006412Z [W1204 11:06:45.277161380 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1006419Z 2025-12-04T11:17:06.1006930Z [W1204 11:06:45.366682491 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1006934Z 2025-12-04T11:17:06.1007450Z [W1204 11:06:45.367444748 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1007456Z 2025-12-04T11:17:06.1007963Z [W1204 11:06:45.367654145 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1007998Z 2025-12-04T11:17:06.1008511Z [W1204 11:06:45.371715722 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1008528Z 2025-12-04T11:17:06.1009041Z [W1204 11:06:45.372385046 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1009076Z 2025-12-04T11:17:06.1009585Z [W1204 11:06:45.372586377 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1009590Z 2025-12-04T11:17:06.1010113Z [W1204 11:06:45.378636987 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1010118Z 2025-12-04T11:17:06.1010630Z [W1204 11:06:45.379436967 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1010637Z 2025-12-04T11:17:06.1011164Z [W1204 11:06:45.379630011 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1011168Z 2025-12-04T11:17:06.1011307Z ('RERUN', {'yellow': True}) [0.4647s] [100%] 2025-12-04T11:17:06.1012597Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 [W1204 11:06:46.721037584 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1012605Z 2025-12-04T11:17:06.1013117Z [W1204 11:06:46.721756368 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1013122Z 2025-12-04T11:17:06.1013646Z [W1204 11:06:46.721953850 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1013653Z 2025-12-04T11:17:06.1014163Z [W1204 11:06:46.725941930 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1014168Z 2025-12-04T11:17:06.1014684Z [W1204 11:06:46.726547936 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1014691Z 2025-12-04T11:17:06.1015213Z [W1204 11:06:46.726743701 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1015217Z 2025-12-04T11:17:06.1015725Z [W1204 11:06:46.732881092 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1015730Z 2025-12-04T11:17:06.1016250Z [W1204 11:06:46.733488933 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1016254Z 2025-12-04T11:17:06.1016767Z [W1204 11:06:46.733676289 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1016772Z 2025-12-04T11:17:06.1017470Z [W1204 11:06:46.820948437 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1017478Z 2025-12-04T11:17:06.1017989Z [W1204 11:06:46.821691905 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1017993Z 2025-12-04T11:17:06.1018517Z [W1204 11:06:46.821896841 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1018521Z 2025-12-04T11:17:06.1019029Z [W1204 11:06:46.825850280 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1019034Z 2025-12-04T11:17:06.1019547Z [W1204 11:06:46.826472610 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1019596Z 2025-12-04T11:17:06.1020114Z [W1204 11:06:46.826665769 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1020150Z 2025-12-04T11:17:06.1020655Z [W1204 11:06:46.832728292 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1020660Z 2025-12-04T11:17:06.1021176Z [W1204 11:06:46.833510414 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1021181Z 2025-12-04T11:17:06.1021690Z [W1204 11:06:46.833704751 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1021694Z 2025-12-04T11:17:06.1021807Z FAILED [0.4515s] [100%] 2025-12-04T11:17:06.1021814Z 2025-12-04T11:17:06.1021957Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.1022481Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.1022606Z Traceback (most recent call last): 2025-12-04T11:17:06.1023122Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1023369Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1023837Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1024002Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1024557Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1024762Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1024909Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1024914Z 2025-12-04T11:17:06.1025021Z Expected 1 but got 2. 2025-12-04T11:17:06.1025127Z Absolute difference: 1 2025-12-04T11:17:06.1025252Z Relative difference: 1.0 2025-12-04T11:17:06.1025257Z 2025-12-04T11:17:06.1025475Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1026383Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1026401Z 2025-12-04T11:17:06.1026674Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1026897Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1027023Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1027923Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1028223Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1028323Z graph_break [] 2025-12-04T11:17:06.1028540Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1029764Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1029882Z if out == self.unknown_value: 2025-12-04T11:17:06.1030613Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1030733Z warnings.warn( 2025-12-04T11:17:06.1031481Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1031592Z warnings.warn( 2025-12-04T11:17:06.1032105Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.1032258Z Traceback (most recent call last): 2025-12-04T11:17:06.1032778Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1033012Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1033484Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1033649Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1034189Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1034412Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1034543Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1034548Z 2025-12-04T11:17:06.1034658Z Expected 1 but got 2. 2025-12-04T11:17:06.1034777Z Absolute difference: 1 2025-12-04T11:17:06.1034891Z Relative difference: 1.0 2025-12-04T11:17:06.1034896Z 2025-12-04T11:17:06.1035121Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1036029Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1036034Z 2025-12-04T11:17:06.1036303Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1036536Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1036654Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1037571Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1037802Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1037901Z graph_break [] 2025-12-04T11:17:06.1038131Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1039348Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1039477Z if out == self.unknown_value: 2025-12-04T11:17:06.1040204Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1040310Z warnings.warn( 2025-12-04T11:17:06.1041096Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1041200Z warnings.warn( 2025-12-04T11:17:06.1041422Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1041547Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1041775Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1042684Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1042783Z graph_break [] 2025-12-04T11:17:06.1042997Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1043769Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1043874Z warnings.warn( 2025-12-04T11:17:06.1044603Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1044735Z warnings.warn( 2025-12-04T11:17:06.1044883Z =================================== FAILURES =================================== 2025-12-04T11:17:06.1045403Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.1045526Z Traceback (most recent call last): 2025-12-04T11:17:06.1046036Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1046280Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1046742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1046927Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1047761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1047975Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1048132Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1048137Z 2025-12-04T11:17:06.1048245Z Expected 1 but got 2. 2025-12-04T11:17:06.1048361Z Absolute difference: 1 2025-12-04T11:17:06.1048473Z Relative difference: 1.0 2025-12-04T11:17:06.1048478Z 2025-12-04T11:17:06.1048691Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1049613Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1049621Z 2025-12-04T11:17:06.1049894Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1050120Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1050238Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1051141Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1051377Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1051478Z graph_break [] 2025-12-04T11:17:06.1051692Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1052919Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1053122Z if out == self.unknown_value: 2025-12-04T11:17:06.1053856Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1053962Z warnings.warn( 2025-12-04T11:17:06.1054686Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1054797Z warnings.warn( 2025-12-04T11:17:06.1055018Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1055139Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1055370Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1056300Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1056411Z graph_break [] 2025-12-04T11:17:06.1056692Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1057487Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1057588Z warnings.warn( 2025-12-04T11:17:06.1058312Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1058422Z warnings.warn( 2025-12-04T11:17:06.1058638Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1058752Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1058993Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1059898Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1060003Z graph_break [] 2025-12-04T11:17:06.1060217Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1060943Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1061052Z warnings.warn( 2025-12-04T11:17:06.1061774Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1061880Z warnings.warn( 2025-12-04T11:17:06.1062723Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-d486c0fac95f9182.xml - 2025-12-04T11:17:06.1062897Z =========================== short test summary info ============================ 2025-12-04T11:17:06.1063844Z FAILED [0.4515s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1063850Z 2025-12-04T11:17:06.1063955Z Expected 1 but got 2. 2025-12-04T11:17:06.1064068Z Absolute difference: 1 2025-12-04T11:17:06.1064176Z Relative difference: 1.0 2025-12-04T11:17:06.1064181Z 2025-12-04T11:17:06.1064399Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1065309Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1065317Z 2025-12-04T11:17:06.1065584Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1065843Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.1066044Z ================== 1 failed, 10 deselected, 2 rerun in 20.24s ================== 2025-12-04T11:17:06.1066143Z Got exit code 1 2025-12-04T11:17:06.1066969Z FAILED CONSISTENTLY: test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1067384Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:17:06.1067828Z W1204 11:06:57.409000 93742 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.1068533Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-73bc989177a97c92.xml 2025-12-04T11:17:06.1068704Z ============================= test session starts ============================== 2025-12-04T11:17:06.1069063Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.1069207Z cachedir: .pytest_cache 2025-12-04T11:17:06.1069730Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.1069863Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.1069974Z configfile: pytest.ini 2025-12-04T11:17:06.1070529Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.1070747Z collecting ... collected 58 items / 4 deselected / 54 selected 2025-12-04T11:17:06.1070886Z stepcurrent: skipping 4 already run items. 2025-12-04T11:17:06.1071012Z Running 7 items in this shard 2025-12-04T11:17:06.1071017Z 2025-12-04T11:17:06.1071888Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 ('RERUN', {'yellow': True}) [3.8212s] [ 14%] 2025-12-04T11:17:06.1072760Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 ('RERUN', {'yellow': True}) [0.4082s] [ 14%] 2025-12-04T11:17:06.1073538Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 FAILED [0.4074s] [ 14%] 2025-12-04T11:17:06.1073543Z 2025-12-04T11:17:06.1073684Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.1074195Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.1074318Z Traceback (most recent call last): 2025-12-04T11:17:06.1074842Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1075072Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1075536Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1075707Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1076247Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1076457Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1076589Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1076594Z 2025-12-04T11:17:06.1076700Z Expected 1 but got 2. 2025-12-04T11:17:06.1076814Z Absolute difference: 1 2025-12-04T11:17:06.1076921Z Relative difference: 1.0 2025-12-04T11:17:06.1076926Z 2025-12-04T11:17:06.1077136Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1078104Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1078112Z 2025-12-04T11:17:06.1078384Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1078607Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1078723Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1079259Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1079494Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1079625Z graph_break [] 2025-12-04T11:17:06.1079844Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1080581Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1080714Z warnings.warn( 2025-12-04T11:17:06.1081447Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1081547Z warnings.warn( 2025-12-04T11:17:06.1082050Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.1082180Z Traceback (most recent call last): 2025-12-04T11:17:06.1082688Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1082929Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1083384Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1083552Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1084101Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1084305Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1084442Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1084448Z 2025-12-04T11:17:06.1084556Z Expected 1 but got 2. 2025-12-04T11:17:06.1084665Z Absolute difference: 1 2025-12-04T11:17:06.1084781Z Relative difference: 1.0 2025-12-04T11:17:06.1084786Z 2025-12-04T11:17:06.1084998Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1085902Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1085920Z 2025-12-04T11:17:06.1086193Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1086409Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1086532Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1087063Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1087287Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1087391Z graph_break [] 2025-12-04T11:17:06.1087603Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1088338Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1088438Z warnings.warn( 2025-12-04T11:17:06.1089215Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1089330Z warnings.warn( 2025-12-04T11:17:06.1089545Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1089656Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1089891Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1090420Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1090524Z graph_break [] 2025-12-04T11:17:06.1090735Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1091458Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1091597Z warnings.warn( 2025-12-04T11:17:06.1092318Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1092442Z warnings.warn( 2025-12-04T11:17:06.1092595Z =================================== FAILURES =================================== 2025-12-04T11:17:06.1093103Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.1093231Z Traceback (most recent call last): 2025-12-04T11:17:06.1093741Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1093966Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1094434Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1094596Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1095148Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1095356Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1095484Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1095489Z 2025-12-04T11:17:06.1095604Z Expected 1 but got 2. 2025-12-04T11:17:06.1095709Z Absolute difference: 1 2025-12-04T11:17:06.1095811Z Relative difference: 1.0 2025-12-04T11:17:06.1095816Z 2025-12-04T11:17:06.1096208Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1097177Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1097187Z 2025-12-04T11:17:06.1097465Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1097686Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1097800Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1098345Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1098568Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1098671Z graph_break [] 2025-12-04T11:17:06.1098885Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1099616Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1099726Z warnings.warn( 2025-12-04T11:17:06.1100448Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1100556Z warnings.warn( 2025-12-04T11:17:06.1100879Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1100994Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1101230Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1101760Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1101854Z graph_break [] 2025-12-04T11:17:06.1102084Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1102807Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1102970Z warnings.warn( 2025-12-04T11:17:06.1103686Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1103787Z warnings.warn( 2025-12-04T11:17:06.1104055Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1104169Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1104393Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1104934Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1105030Z graph_break [] 2025-12-04T11:17:06.1105252Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1105971Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1106068Z warnings.warn( 2025-12-04T11:17:06.1106801Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1106901Z warnings.warn( 2025-12-04T11:17:06.1107758Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-73bc989177a97c92.xml - 2025-12-04T11:17:06.1107931Z =========================== short test summary info ============================ 2025-12-04T11:17:06.1108867Z FAILED [0.4074s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1108873Z 2025-12-04T11:17:06.1108991Z Expected 1 but got 2. 2025-12-04T11:17:06.1109101Z Absolute difference: 1 2025-12-04T11:17:06.1109210Z Relative difference: 1.0 2025-12-04T11:17:06.1109228Z 2025-12-04T11:17:06.1109445Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1110357Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1110365Z 2025-12-04T11:17:06.1110647Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1110828Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.1111027Z =================== 1 failed, 4 deselected, 2 rerun in 4.67s =================== 2025-12-04T11:17:06.1111140Z Got exit code 1 2025-12-04T11:17:06.1111253Z Retrying single test... 2025-12-04T11:17:06.1111714Z W1204 11:07:17.406000 93918 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.1112381Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-9368d7516886ec65.xml 2025-12-04T11:17:06.1112607Z ============================= test session starts ============================== 2025-12-04T11:17:06.1112975Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.1113089Z cachedir: .pytest_cache 2025-12-04T11:17:06.1113628Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.1113753Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.1113864Z configfile: pytest.ini 2025-12-04T11:17:06.1114423Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.1114645Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.1115675Z stepcurrent: skipping 4 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1115804Z Running 1 items in this shard 2025-12-04T11:17:06.1115838Z 2025-12-04T11:17:06.1117129Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 [W1204 11:07:20.662896908 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1117135Z 2025-12-04T11:17:06.1117673Z [W1204 11:07:35.673008060 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1117679Z 2025-12-04T11:17:06.1118192Z [W1204 11:07:35.673267095 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1118200Z 2025-12-04T11:17:06.1118731Z [W1204 11:07:35.680574671 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1118736Z 2025-12-04T11:17:06.1119249Z [W1204 11:07:35.681260318 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1119256Z 2025-12-04T11:17:06.1119779Z [W1204 11:07:35.681448870 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1119785Z 2025-12-04T11:17:06.1120295Z [W1204 11:07:35.688322452 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1120300Z 2025-12-04T11:17:06.1120824Z [W1204 11:07:35.688954487 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1120831Z 2025-12-04T11:17:06.1121341Z [W1204 11:07:35.689139249 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1121349Z 2025-12-04T11:17:06.1121860Z [W1204 11:07:37.686607529 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1121880Z 2025-12-04T11:17:06.1122391Z [W1204 11:07:37.688256035 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1122396Z 2025-12-04T11:17:06.1122909Z [W1204 11:07:37.688460563 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1122913Z 2025-12-04T11:17:06.1123435Z [W1204 11:07:37.692363475 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1123442Z 2025-12-04T11:17:06.1123952Z [W1204 11:07:37.692999512 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1123956Z 2025-12-04T11:17:06.1124541Z [W1204 11:07:37.693192795 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1124548Z 2025-12-04T11:17:06.1125055Z [W1204 11:07:37.699200867 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1125060Z 2025-12-04T11:17:06.1125589Z [W1204 11:07:37.699807531 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1125593Z 2025-12-04T11:17:06.1126103Z [W1204 11:07:38.700019024 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1126141Z 2025-12-04T11:17:06.1126276Z ('RERUN', {'yellow': True}) [18.8456s] [100%] 2025-12-04T11:17:06.1127564Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 [W1204 11:07:38.047418756 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1127604Z 2025-12-04T11:17:06.1128114Z [W1204 11:07:38.048164256 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1128119Z 2025-12-04T11:17:06.1128645Z [W1204 11:07:38.048366544 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1128650Z 2025-12-04T11:17:06.1129160Z [W1204 11:07:38.052311353 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1129167Z 2025-12-04T11:17:06.1129689Z [W1204 11:07:38.053099005 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1129694Z 2025-12-04T11:17:06.1130207Z [W1204 11:07:38.053289152 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1130214Z 2025-12-04T11:17:06.1130736Z [W1204 11:07:38.059257124 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1130741Z 2025-12-04T11:17:06.1131249Z [W1204 11:07:38.059868071 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1131254Z 2025-12-04T11:17:06.1131773Z [W1204 11:07:38.060080500 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1131777Z 2025-12-04T11:17:06.1132289Z [W1204 11:07:38.145403525 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1132293Z 2025-12-04T11:17:06.1132809Z [W1204 11:07:38.146121484 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1132830Z 2025-12-04T11:17:06.1133340Z [W1204 11:07:38.146327074 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1133345Z 2025-12-04T11:17:06.1133855Z [W1204 11:07:38.150232015 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1133859Z 2025-12-04T11:17:06.1134381Z [W1204 11:07:38.150854172 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1134386Z 2025-12-04T11:17:06.1134896Z [W1204 11:07:38.151047988 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1134902Z 2025-12-04T11:17:06.1135552Z [W1204 11:07:38.157020066 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1135560Z 2025-12-04T11:17:06.1136070Z [W1204 11:07:38.157796456 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1136075Z 2025-12-04T11:17:06.1136597Z [W1204 11:07:38.157989762 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1136602Z 2025-12-04T11:17:06.1136734Z ('RERUN', {'yellow': True}) [0.4196s] [100%] 2025-12-04T11:17:06.1138087Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 [W1204 11:07:38.446963534 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1138144Z 2025-12-04T11:17:06.1138662Z [W1204 11:07:38.447714204 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1138698Z 2025-12-04T11:17:06.1139206Z [W1204 11:07:38.447917346 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1139211Z 2025-12-04T11:17:06.1139737Z [W1204 11:07:38.451889855 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1139741Z 2025-12-04T11:17:06.1140252Z [W1204 11:07:38.452699187 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1140256Z 2025-12-04T11:17:06.1140777Z [W1204 11:07:38.452889767 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1140784Z 2025-12-04T11:17:06.1141295Z [W1204 11:07:38.458888525 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1141302Z 2025-12-04T11:17:06.1141822Z [W1204 11:07:38.459501284 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1141827Z 2025-12-04T11:17:06.1142338Z [W1204 11:07:38.459687979 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1142344Z 2025-12-04T11:17:06.1142865Z [W1204 11:07:38.546535133 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1142870Z 2025-12-04T11:17:06.1143382Z [W1204 11:07:38.547326180 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1143389Z 2025-12-04T11:17:06.1143905Z [W1204 11:07:38.547535834 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1143922Z 2025-12-04T11:17:06.1144433Z [W1204 11:07:38.551521163 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1144438Z 2025-12-04T11:17:06.1144946Z [W1204 11:07:38.552177551 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1144951Z 2025-12-04T11:17:06.1145472Z [W1204 11:07:38.552371023 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1145477Z 2025-12-04T11:17:06.1145988Z [W1204 11:07:38.558390681 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1145995Z 2025-12-04T11:17:06.1146518Z [W1204 11:07:38.559195616 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1146600Z 2025-12-04T11:17:06.1147113Z [W1204 11:07:38.559389464 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1147119Z 2025-12-04T11:17:06.1147238Z FAILED [0.4006s] [100%] 2025-12-04T11:17:06.1147243Z 2025-12-04T11:17:06.1147390Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.1147900Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.1148037Z Traceback (most recent call last): 2025-12-04T11:17:06.1148552Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1148836Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1149301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1149472Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1150058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1150265Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1150400Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1150418Z 2025-12-04T11:17:06.1150524Z Expected 1 but got 2. 2025-12-04T11:17:06.1150632Z Absolute difference: 1 2025-12-04T11:17:06.1150754Z Relative difference: 1.0 2025-12-04T11:17:06.1150759Z 2025-12-04T11:17:06.1150976Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1151884Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1151892Z 2025-12-04T11:17:06.1152180Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1152404Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1152535Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1153073Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1153305Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1153419Z graph_break [] 2025-12-04T11:17:06.1153637Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1154862Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1154982Z if out == self.unknown_value: 2025-12-04T11:17:06.1155723Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1155840Z warnings.warn( 2025-12-04T11:17:06.1156568Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1156669Z warnings.warn( 2025-12-04T11:17:06.1157193Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.1157316Z Traceback (most recent call last): 2025-12-04T11:17:06.1157837Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1158069Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1158597Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1158780Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1159319Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1159539Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1159673Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1159679Z 2025-12-04T11:17:06.1159783Z Expected 1 but got 2. 2025-12-04T11:17:06.1159906Z Absolute difference: 1 2025-12-04T11:17:06.1160013Z Relative difference: 1.0 2025-12-04T11:17:06.1160018Z 2025-12-04T11:17:06.1160234Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1161195Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1161202Z 2025-12-04T11:17:06.1161474Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1161738Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1161856Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1162390Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1162630Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1162731Z graph_break [] 2025-12-04T11:17:06.1162964Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1164177Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1164302Z if out == self.unknown_value: 2025-12-04T11:17:06.1165045Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1165149Z warnings.warn( 2025-12-04T11:17:06.1165892Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1165993Z warnings.warn( 2025-12-04T11:17:06.1166212Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1166340Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1166574Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1167109Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1167221Z graph_break [] 2025-12-04T11:17:06.1167444Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1168186Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1168285Z warnings.warn( 2025-12-04T11:17:06.1169008Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1169118Z warnings.warn( 2025-12-04T11:17:06.1169263Z =================================== FAILURES =================================== 2025-12-04T11:17:06.1169767Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.1169905Z Traceback (most recent call last): 2025-12-04T11:17:06.1170478Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1170727Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1171188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1171353Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1171900Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1172108Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1172255Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1172260Z 2025-12-04T11:17:06.1172365Z Expected 1 but got 2. 2025-12-04T11:17:06.1172504Z Absolute difference: 1 2025-12-04T11:17:06.1172625Z Relative difference: 1.0 2025-12-04T11:17:06.1172630Z 2025-12-04T11:17:06.1172850Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1173756Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1173807Z 2025-12-04T11:17:06.1174079Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1174301Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1174432Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1174968Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1175197Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1175311Z graph_break [] 2025-12-04T11:17:06.1175529Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1176755Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1176949Z if out == self.unknown_value: 2025-12-04T11:17:06.1177681Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1177797Z warnings.warn( 2025-12-04T11:17:06.1178518Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1178631Z warnings.warn( 2025-12-04T11:17:06.1178853Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1178967Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1179207Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1179748Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1179848Z graph_break [] 2025-12-04T11:17:06.1180077Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1180808Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1180922Z warnings.warn( 2025-12-04T11:17:06.1181647Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1181750Z warnings.warn( 2025-12-04T11:17:06.1181982Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1182097Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1182399Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1182947Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1183044Z graph_break [] 2025-12-04T11:17:06.1183282Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1184010Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1184110Z warnings.warn( 2025-12-04T11:17:06.1184839Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1184973Z warnings.warn( 2025-12-04T11:17:06.1185832Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-9368d7516886ec65.xml - 2025-12-04T11:17:06.1186037Z =========================== short test summary info ============================ 2025-12-04T11:17:06.1186977Z FAILED [0.4006s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1186983Z 2025-12-04T11:17:06.1187102Z Expected 1 but got 2. 2025-12-04T11:17:06.1187209Z Absolute difference: 1 2025-12-04T11:17:06.1187338Z Relative difference: 1.0 2025-12-04T11:17:06.1187343Z 2025-12-04T11:17:06.1187562Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1188464Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1188472Z 2025-12-04T11:17:06.1188755Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1188941Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.1189152Z ================== 1 failed, 10 deselected, 2 rerun in 19.70s ================== 2025-12-04T11:17:06.1189253Z Got exit code 1 2025-12-04T11:17:06.1189361Z Retrying single test... 2025-12-04T11:17:06.1189817Z W1204 11:07:49.954000 94099 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.1190485Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-921cae3f6acd6a84.xml 2025-12-04T11:17:06.1190651Z ============================= test session starts ============================== 2025-12-04T11:17:06.1191017Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.1191129Z cachedir: .pytest_cache 2025-12-04T11:17:06.1191671Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.1191801Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.1191909Z configfile: pytest.ini 2025-12-04T11:17:06.1192468Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.1192686Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.1193681Z stepcurrent: skipping 4 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1193802Z Running 1 items in this shard 2025-12-04T11:17:06.1193807Z 2025-12-04T11:17:06.1195160Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 [W1204 11:07:53.221938437 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1195182Z 2025-12-04T11:17:06.1195708Z [W1204 11:08:08.385277327 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1195713Z 2025-12-04T11:17:06.1196440Z [W1204 11:08:08.385545526 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1196446Z 2025-12-04T11:17:06.1196974Z [W1204 11:08:08.392922794 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1197040Z 2025-12-04T11:17:06.1197548Z [W1204 11:08:08.393617869 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1197553Z 2025-12-04T11:17:06.1198075Z [W1204 11:08:08.393805669 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1198128Z 2025-12-04T11:17:06.1198635Z [W1204 11:08:08.400746730 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1198640Z 2025-12-04T11:17:06.1199162Z [W1204 11:08:08.401388216 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1199166Z 2025-12-04T11:17:06.1199673Z [W1204 11:08:08.401574732 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1199681Z 2025-12-04T11:17:06.1200187Z [W1204 11:08:10.402649211 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1200205Z 2025-12-04T11:17:06.1200720Z [W1204 11:08:10.404376337 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1200727Z 2025-12-04T11:17:06.1201239Z [W1204 11:08:10.404591725 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1201244Z 2025-12-04T11:17:06.1201769Z [W1204 11:08:10.408493181 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1201773Z 2025-12-04T11:17:06.1202279Z [W1204 11:08:10.409168562 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1202284Z 2025-12-04T11:17:06.1202806Z [W1204 11:08:10.409374685 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1202810Z 2025-12-04T11:17:06.1203322Z [W1204 11:08:10.415483761 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1203329Z 2025-12-04T11:17:06.1203856Z [W1204 11:08:10.416157161 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1203861Z 2025-12-04T11:17:06.1204363Z [W1204 11:08:10.416358556 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1204368Z 2025-12-04T11:17:06.1204515Z ('RERUN', {'yellow': True}) [19.0100s] [100%] 2025-12-04T11:17:06.1205786Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 [W1204 11:08:11.784553111 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1205794Z 2025-12-04T11:17:06.1206385Z [W1204 11:08:11.785314353 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1206404Z 2025-12-04T11:17:06.1206914Z [W1204 11:08:11.785516902 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1206919Z 2025-12-04T11:17:06.1207431Z [W1204 11:08:11.789407751 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1207436Z 2025-12-04T11:17:06.1207961Z [W1204 11:08:11.790211236 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1207966Z 2025-12-04T11:17:06.1208474Z [W1204 11:08:11.790413646 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1208510Z 2025-12-04T11:17:06.1209037Z [W1204 11:08:11.796399278 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1209073Z 2025-12-04T11:17:06.1209584Z [W1204 11:08:11.797023942 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1209588Z 2025-12-04T11:17:06.1210114Z [W1204 11:08:11.797213353 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1210119Z 2025-12-04T11:17:06.1210627Z [W1204 11:08:11.884413977 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1210632Z 2025-12-04T11:17:06.1211153Z [W1204 11:08:11.885191274 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1211160Z 2025-12-04T11:17:06.1211671Z [W1204 11:08:11.885400758 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1211676Z 2025-12-04T11:17:06.1212187Z [W1204 11:08:11.889317034 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1212192Z 2025-12-04T11:17:06.1212719Z [W1204 11:08:11.889953719 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1212724Z 2025-12-04T11:17:06.1213232Z [W1204 11:08:11.890172439 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1213237Z 2025-12-04T11:17:06.1213755Z [W1204 11:08:11.896212624 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1213762Z 2025-12-04T11:17:06.1214269Z [W1204 11:08:11.897013298 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1214278Z 2025-12-04T11:17:06.1214799Z [W1204 11:08:11.897208948 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1214806Z 2025-12-04T11:17:06.1214937Z ('RERUN', {'yellow': True}) [0.4414s] [100%] 2025-12-04T11:17:06.1216221Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 [W1204 11:08:11.190158169 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1216227Z 2025-12-04T11:17:06.1216733Z [W1204 11:08:11.190914938 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1216741Z 2025-12-04T11:17:06.1217386Z [W1204 11:08:11.191119562 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1217406Z 2025-12-04T11:17:06.1217918Z [W1204 11:08:11.195015340 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1217922Z 2025-12-04T11:17:06.1218431Z [W1204 11:08:11.195787820 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1218436Z 2025-12-04T11:17:06.1218958Z [W1204 11:08:11.195979163 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1218963Z 2025-12-04T11:17:06.1219470Z [W1204 11:08:11.201991779 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1219506Z 2025-12-04T11:17:06.1220033Z [W1204 11:08:11.202606347 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1220042Z 2025-12-04T11:17:06.1220550Z [W1204 11:08:11.202792427 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1220586Z 2025-12-04T11:17:06.1221106Z [W1204 11:08:11.288920559 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1221110Z 2025-12-04T11:17:06.1221619Z [W1204 11:08:11.289696628 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1221624Z 2025-12-04T11:17:06.1222140Z [W1204 11:08:11.289902828 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1222147Z 2025-12-04T11:17:06.1222656Z [W1204 11:08:11.293831295 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1222660Z 2025-12-04T11:17:06.1223172Z [W1204 11:08:11.294473259 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1223192Z 2025-12-04T11:17:06.1223697Z [W1204 11:08:11.294669108 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1223702Z 2025-12-04T11:17:06.1224211Z [W1204 11:08:11.300693040 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1224215Z 2025-12-04T11:17:06.1224735Z [W1204 11:08:11.301497211 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1224742Z 2025-12-04T11:17:06.1225248Z [W1204 11:08:11.301692712 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1225252Z 2025-12-04T11:17:06.1225366Z FAILED [0.4020s] [100%] 2025-12-04T11:17:06.1225375Z 2025-12-04T11:17:06.1225520Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.1226033Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.1226169Z Traceback (most recent call last): 2025-12-04T11:17:06.1226683Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1226930Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1227396Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1227564Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1228111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1228388Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1228540Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1228545Z 2025-12-04T11:17:06.1228651Z Expected 1 but got 2. 2025-12-04T11:17:06.1228757Z Absolute difference: 1 2025-12-04T11:17:06.1228878Z Relative difference: 1.0 2025-12-04T11:17:06.1228883Z 2025-12-04T11:17:06.1229098Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1230009Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1230028Z 2025-12-04T11:17:06.1230297Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1230553Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1230681Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1231220Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1231497Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1231605Z graph_break [] 2025-12-04T11:17:06.1231821Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1233046Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1233162Z if out == self.unknown_value: 2025-12-04T11:17:06.1233889Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1234001Z warnings.warn( 2025-12-04T11:17:06.1234723Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1234846Z warnings.warn( 2025-12-04T11:17:06.1235354Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.1235476Z Traceback (most recent call last): 2025-12-04T11:17:06.1235996Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1236227Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1236687Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1236864Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1237405Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1237623Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1237757Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1237763Z 2025-12-04T11:17:06.1237868Z Expected 1 but got 2. 2025-12-04T11:17:06.1237988Z Absolute difference: 1 2025-12-04T11:17:06.1238099Z Relative difference: 1.0 2025-12-04T11:17:06.1238104Z 2025-12-04T11:17:06.1238320Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1239235Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1239241Z 2025-12-04T11:17:06.1239514Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1239745Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1239862Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1240469Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1240712Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1240810Z graph_break [] 2025-12-04T11:17:06.1241036Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1242244Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1242360Z if out == self.unknown_value: 2025-12-04T11:17:06.1243129Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1243231Z warnings.warn( 2025-12-04T11:17:06.1243965Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1244105Z warnings.warn( 2025-12-04T11:17:06.1244325Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1244449Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1244679Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1245214Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1245322Z graph_break [] 2025-12-04T11:17:06.1245540Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1246286Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1246387Z warnings.warn( 2025-12-04T11:17:06.1247111Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1247223Z warnings.warn( 2025-12-04T11:17:06.1247370Z =================================== FAILURES =================================== 2025-12-04T11:17:06.1247892Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.1248014Z Traceback (most recent call last): 2025-12-04T11:17:06.1248524Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1248774Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1249232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1249403Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1249953Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1250158Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1250305Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1250311Z 2025-12-04T11:17:06.1250415Z Expected 1 but got 2. 2025-12-04T11:17:06.1250523Z Absolute difference: 1 2025-12-04T11:17:06.1250642Z Relative difference: 1.0 2025-12-04T11:17:06.1250647Z 2025-12-04T11:17:06.1250862Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1252165Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1252176Z 2025-12-04T11:17:06.1252523Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1252748Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1252876Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1253411Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1253652Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1253751Z graph_break [] 2025-12-04T11:17:06.1253968Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1255196Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1255347Z if out == self.unknown_value: 2025-12-04T11:17:06.1256084Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1256228Z warnings.warn( 2025-12-04T11:17:06.1257010Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1257126Z warnings.warn( 2025-12-04T11:17:06.1257347Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1257462Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1257707Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1258245Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1258358Z graph_break [] 2025-12-04T11:17:06.1258579Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1259306Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1259421Z warnings.warn( 2025-12-04T11:17:06.1260147Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1260247Z warnings.warn( 2025-12-04T11:17:06.1260479Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1260598Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1260843Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1261379Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1261481Z graph_break [] 2025-12-04T11:17:06.1261709Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1262438Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1262540Z warnings.warn( 2025-12-04T11:17:06.1263279Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1263380Z warnings.warn( 2025-12-04T11:17:06.1264795Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-921cae3f6acd6a84.xml - 2025-12-04T11:17:06.1264998Z =========================== short test summary info ============================ 2025-12-04T11:17:06.1266023Z FAILED [0.4020s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1266046Z 2025-12-04T11:17:06.1266157Z Expected 1 but got 2. 2025-12-04T11:17:06.1266266Z Absolute difference: 1 2025-12-04T11:17:06.1266393Z Relative difference: 1.0 2025-12-04T11:17:06.1266398Z 2025-12-04T11:17:06.1266618Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1267532Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1267538Z 2025-12-04T11:17:06.1267820Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1268035Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.1268257Z ================== 1 failed, 10 deselected, 2 rerun in 19.89s ================== 2025-12-04T11:17:06.1268357Z Got exit code 1 2025-12-04T11:17:06.1269217Z FAILED CONSISTENTLY: test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1269644Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:17:06.1270090Z W1204 11:08:23.201000 94280 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.1270762Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-8d346e15c82776ef.xml 2025-12-04T11:17:06.1270930Z ============================= test session starts ============================== 2025-12-04T11:17:06.1271284Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.1271411Z cachedir: .pytest_cache 2025-12-04T11:17:06.1271943Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.1272082Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.1272194Z configfile: pytest.ini 2025-12-04T11:17:06.1272739Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.1272967Z collecting ... collected 58 items / 5 deselected / 53 selected 2025-12-04T11:17:06.1273109Z stepcurrent: skipping 5 already run items. 2025-12-04T11:17:06.1273222Z Running 6 items in this shard 2025-12-04T11:17:06.1273227Z 2025-12-04T11:17:06.1274115Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 ('RERUN', {'yellow': True}) [4.3195s] [ 16%] 2025-12-04T11:17:06.1274993Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 ('RERUN', {'yellow': True}) [0.8817s] [ 16%] 2025-12-04T11:17:06.1275792Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 FAILED [0.8839s] [ 16%] 2025-12-04T11:17:06.1275798Z 2025-12-04T11:17:06.1275937Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.1276461Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.1276584Z Traceback (most recent call last): 2025-12-04T11:17:06.1277095Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1277340Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1277864Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1278032Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1278585Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1278791Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1278935Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1278940Z 2025-12-04T11:17:06.1279046Z Expected 1 but got 2. 2025-12-04T11:17:06.1279155Z Absolute difference: 1 2025-12-04T11:17:06.1279278Z Relative difference: 1.0 2025-12-04T11:17:06.1279283Z 2025-12-04T11:17:06.1279496Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1280486Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.1280492Z 2025-12-04T11:17:06.1280793Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1281014Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1281143Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1281677Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1281906Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1282021Z graph_break [] 2025-12-04T11:17:06.1282241Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1282994Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1283093Z warnings.warn( 2025-12-04T11:17:06.1283822Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1283940Z warnings.warn( 2025-12-04T11:17:06.1284456Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.1284594Z Traceback (most recent call last): 2025-12-04T11:17:06.1285100Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1285333Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1285803Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1285972Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1286515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1286735Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1286867Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1286873Z 2025-12-04T11:17:06.1286993Z Expected 1 but got 2. 2025-12-04T11:17:06.1287099Z Absolute difference: 1 2025-12-04T11:17:06.1287207Z Relative difference: 1.0 2025-12-04T11:17:06.1287212Z 2025-12-04T11:17:06.1287437Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1288354Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.1288361Z 2025-12-04T11:17:06.1288643Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1288863Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1289037Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1289587Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1289816Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1289915Z graph_break [] 2025-12-04T11:17:06.1290148Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1290884Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1290997Z warnings.warn( 2025-12-04T11:17:06.1291722Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1291882Z warnings.warn( 2025-12-04T11:17:06.1292116Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1292262Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1292501Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1293038Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1293134Z graph_break [] 2025-12-04T11:17:06.1293363Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1294356Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1294464Z warnings.warn( 2025-12-04T11:17:06.1295202Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1295309Z warnings.warn( 2025-12-04T11:17:06.1295465Z =================================== FAILURES =================================== 2025-12-04T11:17:06.1296152Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.1296277Z Traceback (most recent call last): 2025-12-04T11:17:06.1296803Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1297097Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1297574Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1297741Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1298282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1298507Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1298639Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1298647Z 2025-12-04T11:17:06.1298752Z Expected 1 but got 2. 2025-12-04T11:17:06.1298874Z Absolute difference: 1 2025-12-04T11:17:06.1298984Z Relative difference: 1.0 2025-12-04T11:17:06.1298989Z 2025-12-04T11:17:06.1299218Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1300132Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.1300138Z 2025-12-04T11:17:06.1300407Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1300639Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1300755Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1301432Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1301665Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1301765Z graph_break [] 2025-12-04T11:17:06.1301997Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1302732Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1302832Z warnings.warn( 2025-12-04T11:17:06.1303569Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1303710Z warnings.warn( 2025-12-04T11:17:06.1303941Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1304054Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1304282Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1304868Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1304964Z graph_break [] 2025-12-04T11:17:06.1305177Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1305918Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1306016Z warnings.warn( 2025-12-04T11:17:06.1306747Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1306848Z warnings.warn( 2025-12-04T11:17:06.1307062Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1307192Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1307420Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1307963Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1308059Z graph_break [] 2025-12-04T11:17:06.1308273Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1309005Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1309105Z warnings.warn( 2025-12-04T11:17:06.1309830Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1309943Z warnings.warn( 2025-12-04T11:17:06.1310788Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-8d346e15c82776ef.xml - 2025-12-04T11:17:06.1310974Z =========================== short test summary info ============================ 2025-12-04T11:17:06.1311927Z FAILED [0.8839s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1311933Z 2025-12-04T11:17:06.1312041Z Expected 1 but got 2. 2025-12-04T11:17:06.1312162Z Absolute difference: 1 2025-12-04T11:17:06.1312271Z Relative difference: 1.0 2025-12-04T11:17:06.1312276Z 2025-12-04T11:17:06.1312513Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1313492Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.1313500Z 2025-12-04T11:17:06.1313769Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1313964Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.1314162Z =================== 1 failed, 5 deselected, 2 rerun in 6.12s =================== 2025-12-04T11:17:06.1314274Z Got exit code 1 2025-12-04T11:17:06.1314382Z Retrying single test... 2025-12-04T11:17:06.1314827Z W1204 11:08:43.939000 94457 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.1315509Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-1b686c821c7b9c77.xml 2025-12-04T11:17:06.1315703Z ============================= test session starts ============================== 2025-12-04T11:17:06.1316055Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.1316283Z cachedir: .pytest_cache 2025-12-04T11:17:06.1316804Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.1316940Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.1317048Z configfile: pytest.ini 2025-12-04T11:17:06.1317595Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.1317828Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.1318822Z stepcurrent: skipping 5 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.1318951Z Running 1 items in this shard 2025-12-04T11:17:06.1318956Z 2025-12-04T11:17:06.1320243Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 [W1204 11:08:47.671857073 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1320251Z 2025-12-04T11:17:06.1320785Z [W1204 11:09:03.828641431 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1320790Z 2025-12-04T11:17:06.1321307Z [W1204 11:09:03.828904657 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1321312Z 2025-12-04T11:17:06.1321819Z [W1204 11:09:03.836262841 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1321849Z 2025-12-04T11:17:06.1322363Z [W1204 11:09:03.836931457 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1322371Z 2025-12-04T11:17:06.1322881Z [W1204 11:09:03.837119950 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1322886Z 2025-12-04T11:17:06.1323410Z [W1204 11:09:03.843927290 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1323414Z 2025-12-04T11:17:06.1323923Z [W1204 11:09:03.844570016 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1323927Z 2025-12-04T11:17:06.1324453Z [W1204 11:09:03.844754840 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1324461Z 2025-12-04T11:17:06.1325048Z [W1204 11:09:05.839633098 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1325053Z 2025-12-04T11:17:06.1325585Z [W1204 11:09:05.841401626 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1325590Z 2025-12-04T11:17:06.1326099Z [W1204 11:09:05.841624861 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1326104Z 2025-12-04T11:17:06.1326617Z [W1204 11:09:05.845547729 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1326635Z 2025-12-04T11:17:06.1327149Z [W1204 11:09:05.846175979 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1327184Z 2025-12-04T11:17:06.1327699Z [W1204 11:09:05.846373962 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1327709Z 2025-12-04T11:17:06.1328233Z [W1204 11:09:05.852440408 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1328266Z 2025-12-04T11:17:06.1328781Z [W1204 11:09:05.853055214 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1328785Z 2025-12-04T11:17:06.1329313Z [W1204 11:09:05.853247124 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1329318Z 2025-12-04T11:17:06.1329453Z ('RERUN', {'yellow': True}) [19.4691s] [100%] 2025-12-04T11:17:06.1330764Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 [W1204 11:09:05.673027588 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1330773Z 2025-12-04T11:17:06.1331288Z [W1204 11:09:05.673733274 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1331293Z 2025-12-04T11:17:06.1331818Z [W1204 11:09:05.673930613 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1331823Z 2025-12-04T11:17:06.1332337Z [W1204 11:09:05.677797873 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1332341Z 2025-12-04T11:17:06.1332852Z [W1204 11:09:05.678543424 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1332873Z 2025-12-04T11:17:06.1333384Z [W1204 11:09:05.678731504 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1333393Z 2025-12-04T11:17:06.1333905Z [W1204 11:09:05.684720659 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1333913Z 2025-12-04T11:17:06.1334440Z [W1204 11:09:05.685323759 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1334445Z 2025-12-04T11:17:06.1334953Z [W1204 11:09:05.685512180 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1334958Z 2025-12-04T11:17:06.1335483Z [W1204 11:09:06.769209359 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1335490Z 2025-12-04T11:17:06.1336005Z [W1204 11:09:06.769855117 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1336009Z 2025-12-04T11:17:06.1336592Z [W1204 11:09:06.770074157 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1336600Z 2025-12-04T11:17:06.1337175Z [W1204 11:09:06.773945659 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1337181Z 2025-12-04T11:17:06.1337707Z [W1204 11:09:06.774545756 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1337712Z 2025-12-04T11:17:06.1338224Z [W1204 11:09:06.774737349 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1338263Z 2025-12-04T11:17:06.1338779Z [W1204 11:09:06.780717823 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1338798Z 2025-12-04T11:17:06.1339315Z [W1204 11:09:06.781477232 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1339348Z 2025-12-04T11:17:06.1339862Z [W1204 11:09:06.781670997 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1339866Z 2025-12-04T11:17:06.1340015Z ('RERUN', {'yellow': True}) [0.8882s] [100%] 2025-12-04T11:17:06.1341307Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 [W1204 11:09:06.528402800 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1341315Z 2025-12-04T11:17:06.1341844Z [W1204 11:09:06.529146320 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1341848Z 2025-12-04T11:17:06.1342362Z [W1204 11:09:06.529345308 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1342369Z 2025-12-04T11:17:06.1342891Z [W1204 11:09:06.533271536 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1342897Z 2025-12-04T11:17:06.1343409Z [W1204 11:09:06.533880622 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1343414Z 2025-12-04T11:17:06.1343924Z [W1204 11:09:06.534069255 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1343941Z 2025-12-04T11:17:06.1344451Z [W1204 11:09:06.540100414 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1344456Z 2025-12-04T11:17:06.1344968Z [W1204 11:09:06.540704633 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1344975Z 2025-12-04T11:17:06.1345496Z [W1204 11:09:06.540889396 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1345501Z 2025-12-04T11:17:06.1346009Z [W1204 11:09:06.626323087 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1346014Z 2025-12-04T11:17:06.1346532Z [W1204 11:09:06.627059315 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1346536Z 2025-12-04T11:17:06.1347045Z [W1204 11:09:06.627265744 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1347055Z 2025-12-04T11:17:06.1347629Z [W1204 11:09:06.631191427 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1347636Z 2025-12-04T11:17:06.1348142Z [W1204 11:09:06.631817319 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1348147Z 2025-12-04T11:17:06.1348665Z [W1204 11:09:06.632011958 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1348670Z 2025-12-04T11:17:06.1349179Z [W1204 11:09:06.637996884 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1349184Z 2025-12-04T11:17:06.1349689Z [W1204 11:09:06.638771539 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1349733Z 2025-12-04T11:17:06.1350251Z [W1204 11:09:06.638966443 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1350286Z 2025-12-04T11:17:06.1350393Z FAILED [0.8566s] [100%] 2025-12-04T11:17:06.1350397Z 2025-12-04T11:17:06.1350553Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.1351071Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.1351209Z Traceback (most recent call last): 2025-12-04T11:17:06.1351721Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1351957Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1352437Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1352604Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1353150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1353374Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1353508Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1353513Z 2025-12-04T11:17:06.1353634Z Expected 1 but got 2. 2025-12-04T11:17:06.1353743Z Absolute difference: 1 2025-12-04T11:17:06.1353850Z Relative difference: 1.0 2025-12-04T11:17:06.1353855Z 2025-12-04T11:17:06.1354082Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1355004Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.1355013Z 2025-12-04T11:17:06.1355293Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1355521Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1355638Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1356189Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1356417Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1356515Z graph_break [] 2025-12-04T11:17:06.1356745Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1357968Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1358101Z if out == self.unknown_value: 2025-12-04T11:17:06.1358892Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1358997Z warnings.warn( 2025-12-04T11:17:06.1359737Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1359837Z warnings.warn( 2025-12-04T11:17:06.1360368Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.1360489Z Traceback (most recent call last): 2025-12-04T11:17:06.1360998Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1361243Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1361731Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1361895Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1362449Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1362689Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1362834Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1362839Z 2025-12-04T11:17:06.1362944Z Expected 1 but got 2. 2025-12-04T11:17:06.1363052Z Absolute difference: 1 2025-12-04T11:17:06.1363174Z Relative difference: 1.0 2025-12-04T11:17:06.1363178Z 2025-12-04T11:17:06.1363394Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1364322Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.1364330Z 2025-12-04T11:17:06.1364597Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1364820Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1364950Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1365489Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1365730Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1365830Z graph_break [] 2025-12-04T11:17:06.1366047Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1367280Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1367400Z if out == self.unknown_value: 2025-12-04T11:17:06.1368126Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1368241Z warnings.warn( 2025-12-04T11:17:06.1368962Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1369075Z warnings.warn( 2025-12-04T11:17:06.1369293Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1369406Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1369643Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1370175Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1370286Z graph_break [] 2025-12-04T11:17:06.1370501Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1371305Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1371423Z warnings.warn( 2025-12-04T11:17:06.1372144Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1372245Z warnings.warn( 2025-12-04T11:17:06.1372403Z =================================== FAILURES =================================== 2025-12-04T11:17:06.1372919Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.1373057Z Traceback (most recent call last): 2025-12-04T11:17:06.1373596Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1373831Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1374309Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1374507Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1375058Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1375264Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1375396Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1375402Z 2025-12-04T11:17:06.1375522Z Expected 1 but got 2. 2025-12-04T11:17:06.1375630Z Absolute difference: 1 2025-12-04T11:17:06.1375740Z Relative difference: 1.0 2025-12-04T11:17:06.1375745Z 2025-12-04T11:17:06.1375974Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1376963Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.1376973Z 2025-12-04T11:17:06.1377261Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1377479Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1377593Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1378140Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1378366Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1378478Z graph_break [] 2025-12-04T11:17:06.1378693Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1379904Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1380036Z if out == self.unknown_value: 2025-12-04T11:17:06.1380761Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1380874Z warnings.warn( 2025-12-04T11:17:06.1381600Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1381702Z warnings.warn( 2025-12-04T11:17:06.1381934Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1382049Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1382281Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1382899Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1383000Z graph_break [] 2025-12-04T11:17:06.1383238Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1383969Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1384068Z warnings.warn( 2025-12-04T11:17:06.1384800Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1384901Z warnings.warn( 2025-12-04T11:17:06.1385118Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1385278Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1385508Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1386057Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1386186Z graph_break [] 2025-12-04T11:17:06.1386406Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1387148Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1387248Z warnings.warn( 2025-12-04T11:17:06.1387983Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1388084Z warnings.warn( 2025-12-04T11:17:06.1388932Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-1b686c821c7b9c77.xml - 2025-12-04T11:17:06.1389124Z =========================== short test summary info ============================ 2025-12-04T11:17:06.1390070Z FAILED [0.8566s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1390078Z 2025-12-04T11:17:06.1390195Z Expected 1 but got 2. 2025-12-04T11:17:06.1390300Z Absolute difference: 1 2025-12-04T11:17:06.1390410Z Relative difference: 1.0 2025-12-04T11:17:06.1390416Z 2025-12-04T11:17:06.1390644Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1391565Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.1391573Z 2025-12-04T11:17:06.1391849Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1392033Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.1392235Z ================== 1 failed, 10 deselected, 2 rerun in 21.25s ================== 2025-12-04T11:17:06.1392345Z Got exit code 1 2025-12-04T11:17:06.1392454Z Retrying single test... 2025-12-04T11:17:06.1392902Z W1204 11:09:18.014000 94639 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.1393578Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-21dcaad0846c92ef.xml 2025-12-04T11:17:06.1393744Z ============================= test session starts ============================== 2025-12-04T11:17:06.1394107Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.1394219Z cachedir: .pytest_cache 2025-12-04T11:17:06.1394801Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.1394944Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.1395056Z configfile: pytest.ini 2025-12-04T11:17:06.1395605Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.1395842Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.1396975Z stepcurrent: skipping 5 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.1397106Z Running 1 items in this shard 2025-12-04T11:17:06.1397111Z 2025-12-04T11:17:06.1398472Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 [W1204 11:09:22.734490846 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1398517Z 2025-12-04T11:17:06.1399058Z [W1204 11:09:37.744443585 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1399063Z 2025-12-04T11:17:06.1399579Z [W1204 11:09:37.744701344 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1399584Z 2025-12-04T11:17:06.1400117Z [W1204 11:09:37.752232724 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1400122Z 2025-12-04T11:17:06.1400636Z [W1204 11:09:37.752948511 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1400643Z 2025-12-04T11:17:06.1401426Z [W1204 11:09:37.753138901 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1401432Z 2025-12-04T11:17:06.1401961Z [W1204 11:09:37.760115128 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1401966Z 2025-12-04T11:17:06.1402478Z [W1204 11:09:37.760760134 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1402482Z 2025-12-04T11:17:06.1403009Z [W1204 11:09:37.760946128 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1403014Z 2025-12-04T11:17:06.1403527Z [W1204 11:09:39.763098933 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1403534Z 2025-12-04T11:17:06.1404061Z [W1204 11:09:39.764861572 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1404070Z 2025-12-04T11:17:06.1404582Z [W1204 11:09:39.765081931 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1404589Z 2025-12-04T11:17:06.1405114Z [W1204 11:09:39.769107620 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1405119Z 2025-12-04T11:17:06.1405632Z [W1204 11:09:39.769769341 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1405637Z 2025-12-04T11:17:06.1406147Z [W1204 11:09:39.769967622 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1406167Z 2025-12-04T11:17:06.1406680Z [W1204 11:09:39.776125410 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1406685Z 2025-12-04T11:17:06.1407291Z [W1204 11:09:39.776806401 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1407298Z 2025-12-04T11:17:06.1407825Z [W1204 11:09:39.777003072 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1407830Z 2025-12-04T11:17:06.1407964Z ('RERUN', {'yellow': True}) [19.3275s] [100%] 2025-12-04T11:17:06.1409269Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 [W1204 11:09:39.621230531 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1409305Z 2025-12-04T11:17:06.1409820Z [W1204 11:09:39.621977018 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1409825Z 2025-12-04T11:17:06.1410354Z [W1204 11:09:39.622180013 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1410407Z 2025-12-04T11:17:06.1410919Z [W1204 11:09:39.626136050 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1410923Z 2025-12-04T11:17:06.1411446Z [W1204 11:09:39.626906492 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1411451Z 2025-12-04T11:17:06.1411963Z [W1204 11:09:39.627097537 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1411969Z 2025-12-04T11:17:06.1412480Z [W1204 11:09:39.633154471 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1412499Z 2025-12-04T11:17:06.1413014Z [W1204 11:09:39.633769179 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1413020Z 2025-12-04T11:17:06.1413527Z [W1204 11:09:39.633957226 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1413532Z 2025-12-04T11:17:06.1414055Z [W1204 11:09:40.719989421 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1414060Z 2025-12-04T11:17:06.1414570Z [W1204 11:09:40.720746405 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1414576Z 2025-12-04T11:17:06.1415098Z [W1204 11:09:40.720955078 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1415102Z 2025-12-04T11:17:06.1415615Z [W1204 11:09:40.724855388 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1415622Z 2025-12-04T11:17:06.1416144Z [W1204 11:09:40.725478898 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1416149Z 2025-12-04T11:17:06.1416658Z [W1204 11:09:40.725674567 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1416663Z 2025-12-04T11:17:06.1417239Z [W1204 11:09:40.731663810 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1417245Z 2025-12-04T11:17:06.1417753Z [W1204 11:09:40.732456199 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1417758Z 2025-12-04T11:17:06.1418361Z [W1204 11:09:40.732651080 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1418369Z 2025-12-04T11:17:06.1418519Z ('RERUN', {'yellow': True}) [0.9128s] [100%] 2025-12-04T11:17:06.1419814Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 [W1204 11:09:40.494591766 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1419819Z 2025-12-04T11:17:06.1420340Z [W1204 11:09:40.495353044 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1420345Z 2025-12-04T11:17:06.1420888Z [W1204 11:09:40.495556604 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1420893Z 2025-12-04T11:17:06.1421421Z [W1204 11:09:40.499459493 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1421458Z 2025-12-04T11:17:06.1421967Z [W1204 11:09:40.500115314 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1421972Z 2025-12-04T11:17:06.1422493Z [W1204 11:09:40.500312256 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1422498Z 2025-12-04T11:17:06.1423007Z [W1204 11:09:40.506302933 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1423012Z 2025-12-04T11:17:06.1423522Z [W1204 11:09:40.506917638 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1423543Z 2025-12-04T11:17:06.1424054Z [W1204 11:09:40.507105570 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1424061Z 2025-12-04T11:17:06.1424571Z [W1204 11:09:40.593623144 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1424575Z 2025-12-04T11:17:06.1425094Z [W1204 11:09:40.594377848 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1425099Z 2025-12-04T11:17:06.1425609Z [W1204 11:09:40.594586114 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1425614Z 2025-12-04T11:17:06.1426139Z [W1204 11:09:40.598523162 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1426146Z 2025-12-04T11:17:06.1426659Z [W1204 11:09:40.599156275 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1426664Z 2025-12-04T11:17:06.1427187Z [W1204 11:09:40.599351862 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1427192Z 2025-12-04T11:17:06.1427701Z [W1204 11:09:40.605404432 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1427706Z 2025-12-04T11:17:06.1428227Z [W1204 11:09:40.606219988 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1428232Z 2025-12-04T11:17:06.1428741Z [W1204 11:09:40.606415627 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1428748Z 2025-12-04T11:17:06.1428852Z FAILED [0.8729s] [100%] 2025-12-04T11:17:06.1428857Z 2025-12-04T11:17:06.1429070Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.1429589Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.1429727Z Traceback (most recent call last): 2025-12-04T11:17:06.1430239Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1430473Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1430954Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1431120Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1431673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1431913Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1432049Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1432054Z 2025-12-04T11:17:06.1432214Z Expected 1 but got 2. 2025-12-04T11:17:06.1432321Z Absolute difference: 1 2025-12-04T11:17:06.1432431Z Relative difference: 1.0 2025-12-04T11:17:06.1432436Z 2025-12-04T11:17:06.1432666Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1433588Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.1433593Z 2025-12-04T11:17:06.1433875Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1434101Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1434219Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1434771Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1435002Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1435116Z graph_break [] 2025-12-04T11:17:06.1435332Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1436543Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1436674Z if out == self.unknown_value: 2025-12-04T11:17:06.1437404Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1437523Z warnings.warn( 2025-12-04T11:17:06.1438249Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1438350Z warnings.warn( 2025-12-04T11:17:06.1438880Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.1439002Z Traceback (most recent call last): 2025-12-04T11:17:06.1439511Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1439753Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1440214Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1440391Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1440930Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1441135Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1441340Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1441347Z 2025-12-04T11:17:06.1441453Z Expected 1 but got 2. 2025-12-04T11:17:06.1441562Z Absolute difference: 1 2025-12-04T11:17:06.1441684Z Relative difference: 1.0 2025-12-04T11:17:06.1441688Z 2025-12-04T11:17:06.1441902Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1442822Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.1442827Z 2025-12-04T11:17:06.1443095Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1443346Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1443473Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1444010Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1444284Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1444383Z graph_break [] 2025-12-04T11:17:06.1444601Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1445828Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1445946Z if out == self.unknown_value: 2025-12-04T11:17:06.1446683Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1446789Z warnings.warn( 2025-12-04T11:17:06.1447516Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1447630Z warnings.warn( 2025-12-04T11:17:06.1447847Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1447961Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1448202Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1448735Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1448847Z graph_break [] 2025-12-04T11:17:06.1449063Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1449788Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1449899Z warnings.warn( 2025-12-04T11:17:06.1450625Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1450738Z warnings.warn( 2025-12-04T11:17:06.1450882Z =================================== FAILURES =================================== 2025-12-04T11:17:06.1451396Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.1451529Z Traceback (most recent call last): 2025-12-04T11:17:06.1452039Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1452270Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1452743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1452906Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1453519Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1453730Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1453861Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1453866Z 2025-12-04T11:17:06.1453983Z Expected 1 but got 2. 2025-12-04T11:17:06.1454090Z Absolute difference: 1 2025-12-04T11:17:06.1454201Z Relative difference: 1.0 2025-12-04T11:17:06.1454206Z 2025-12-04T11:17:06.1454433Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1455354Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.1455405Z 2025-12-04T11:17:06.1455689Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1455910Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1456057Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1456604Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1456831Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1457013Z graph_break [] 2025-12-04T11:17:06.1457231Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1458438Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1458573Z if out == self.unknown_value: 2025-12-04T11:17:06.1459308Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1459427Z warnings.warn( 2025-12-04T11:17:06.1460145Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1460245Z warnings.warn( 2025-12-04T11:17:06.1460474Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1460589Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1460817Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1461361Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1461462Z graph_break [] 2025-12-04T11:17:06.1461689Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1462419Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1462523Z warnings.warn( 2025-12-04T11:17:06.1463251Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1463350Z warnings.warn( 2025-12-04T11:17:06.1463576Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1463691Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1463919Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1464462Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1464561Z graph_break [] 2025-12-04T11:17:06.1464845Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1465585Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1465687Z warnings.warn( 2025-12-04T11:17:06.1466419Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1466519Z warnings.warn( 2025-12-04T11:17:06.1467363Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-21dcaad0846c92ef.xml - 2025-12-04T11:17:06.1467554Z =========================== short test summary info ============================ 2025-12-04T11:17:06.1468546Z FAILED [0.8729s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1468582Z 2025-12-04T11:17:06.1468704Z Expected 1 but got 2. 2025-12-04T11:17:06.1468813Z Absolute difference: 1 2025-12-04T11:17:06.1468923Z Relative difference: 1.0 2025-12-04T11:17:06.1468928Z 2025-12-04T11:17:06.1469164Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1470081Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.1470086Z 2025-12-04T11:17:06.1470369Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1470552Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.1470756Z ================== 1 failed, 10 deselected, 2 rerun in 21.15s ================== 2025-12-04T11:17:06.1470871Z Got exit code 1 2025-12-04T11:17:06.1471703Z FAILED CONSISTENTLY: test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.1472137Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:17:06.1472587Z W1204 11:09:52.082000 94821 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.1473253Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-633a2407c7cc9faa.xml 2025-12-04T11:17:06.1473433Z ============================= test session starts ============================== 2025-12-04T11:17:06.1473789Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.1473901Z cachedir: .pytest_cache 2025-12-04T11:17:06.1474445Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.1474574Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.1474701Z configfile: pytest.ini 2025-12-04T11:17:06.1475248Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.1475466Z collecting ... collected 58 items / 6 deselected / 52 selected 2025-12-04T11:17:06.1475627Z stepcurrent: skipping 6 already run items. 2025-12-04T11:17:06.1475745Z Running 5 items in this shard 2025-12-04T11:17:06.1475750Z 2025-12-04T11:17:06.1476631Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 ('RERUN', {'yellow': True}) [3.8650s] [ 20%] 2025-12-04T11:17:06.1477558Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 ('RERUN', {'yellow': True}) [0.4471s] [ 20%] 2025-12-04T11:17:06.1478345Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 FAILED [0.4391s] [ 20%] 2025-12-04T11:17:06.1478365Z 2025-12-04T11:17:06.1478509Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.1479015Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.1479154Z Traceback (most recent call last): 2025-12-04T11:17:06.1479665Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1479928Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1480403Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1480573Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1481162Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1481371Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1481505Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1481510Z 2025-12-04T11:17:06.1481631Z Expected 1 but got 2. 2025-12-04T11:17:06.1481739Z Absolute difference: 1 2025-12-04T11:17:06.1481848Z Relative difference: 1.0 2025-12-04T11:17:06.1481853Z 2025-12-04T11:17:06.1482081Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1482989Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1482996Z 2025-12-04T11:17:06.1483280Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1483503Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1483619Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1484533Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1484761Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1484871Z graph_break [] 2025-12-04T11:17:06.1485088Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1485829Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1485942Z warnings.warn( 2025-12-04T11:17:06.1486671Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1486785Z warnings.warn( 2025-12-04T11:17:06.1487293Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.1487415Z Traceback (most recent call last): 2025-12-04T11:17:06.1487938Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1488171Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1488628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1488806Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1489408Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1489628Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1489762Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1489767Z 2025-12-04T11:17:06.1489873Z Expected 1 but got 2. 2025-12-04T11:17:06.1489995Z Absolute difference: 1 2025-12-04T11:17:06.1490104Z Relative difference: 1.0 2025-12-04T11:17:06.1490109Z 2025-12-04T11:17:06.1490327Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1491258Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1491263Z 2025-12-04T11:17:06.1491567Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1491798Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1491916Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1492830Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1493096Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1493196Z graph_break [] 2025-12-04T11:17:06.1493425Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1494157Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1494259Z warnings.warn( 2025-12-04T11:17:06.1494997Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1495098Z warnings.warn( 2025-12-04T11:17:06.1495328Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1495445Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1495679Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1496781Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1496956Z graph_break [] 2025-12-04T11:17:06.1497178Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1497916Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1498020Z warnings.warn( 2025-12-04T11:17:06.1498758Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1498862Z warnings.warn( 2025-12-04T11:17:06.1499009Z =================================== FAILURES =================================== 2025-12-04T11:17:06.1499534Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.1499658Z Traceback (most recent call last): 2025-12-04T11:17:06.1500186Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1500419Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1500881Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1501065Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1501725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1501940Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1502092Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1502097Z 2025-12-04T11:17:06.1502207Z Expected 1 but got 2. 2025-12-04T11:17:06.1502333Z Absolute difference: 1 2025-12-04T11:17:06.1502448Z Relative difference: 1.0 2025-12-04T11:17:06.1502452Z 2025-12-04T11:17:06.1502668Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1503599Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1503715Z 2025-12-04T11:17:06.1503989Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1504221Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1504342Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1505278Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1505521Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1505622Z graph_break [] 2025-12-04T11:17:06.1505852Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1506586Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1506691Z warnings.warn( 2025-12-04T11:17:06.1507421Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1507525Z warnings.warn( 2025-12-04T11:17:06.1507743Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1507871Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1508098Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1509012Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1509112Z graph_break [] 2025-12-04T11:17:06.1509328Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1510070Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1510172Z warnings.warn( 2025-12-04T11:17:06.1510906Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1511007Z warnings.warn( 2025-12-04T11:17:06.1511222Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1511347Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1511577Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1512472Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1512582Z graph_break [] 2025-12-04T11:17:06.1512794Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1513601Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1513705Z warnings.warn( 2025-12-04T11:17:06.1514426Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1514537Z warnings.warn( 2025-12-04T11:17:06.1515376Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-633a2407c7cc9faa.xml - 2025-12-04T11:17:06.1515557Z =========================== short test summary info ============================ 2025-12-04T11:17:06.1516494Z FAILED [0.4391s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1516531Z 2025-12-04T11:17:06.1516636Z Expected 1 but got 2. 2025-12-04T11:17:06.1516756Z Absolute difference: 1 2025-12-04T11:17:06.1516867Z Relative difference: 1.0 2025-12-04T11:17:06.1516872Z 2025-12-04T11:17:06.1517134Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1518040Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1518045Z 2025-12-04T11:17:06.1518313Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1518506Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.1518703Z =================== 1 failed, 6 deselected, 2 rerun in 4.78s =================== 2025-12-04T11:17:06.1518816Z Got exit code 1 2025-12-04T11:17:06.1518924Z Retrying single test... 2025-12-04T11:17:06.1519374Z W1204 11:10:12.108000 94990 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.1520060Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-2a4f8a9a7e80b95a.xml 2025-12-04T11:17:06.1520226Z ============================= test session starts ============================== 2025-12-04T11:17:06.1520577Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.1520700Z cachedir: .pytest_cache 2025-12-04T11:17:06.1521221Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.1521362Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.1521470Z configfile: pytest.ini 2025-12-04T11:17:06.1522018Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.1522251Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.1523247Z stepcurrent: skipping 6 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1523378Z Running 1 items in this shard 2025-12-04T11:17:06.1523383Z 2025-12-04T11:17:06.1524656Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 [W1204 11:10:17.238176254 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1524662Z 2025-12-04T11:17:06.1525184Z [W1204 11:10:32.377891202 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1525205Z 2025-12-04T11:17:06.1525720Z [W1204 11:10:32.378148518 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1525786Z 2025-12-04T11:17:06.1526302Z [W1204 11:10:32.385490443 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1526310Z 2025-12-04T11:17:06.1526834Z [W1204 11:10:32.386214470 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1526838Z 2025-12-04T11:17:06.1527351Z [W1204 11:10:32.386406506 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1527356Z 2025-12-04T11:17:06.1527876Z [W1204 11:10:32.393374428 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1527911Z 2025-12-04T11:17:06.1528421Z [W1204 11:10:32.394150484 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1528425Z 2025-12-04T11:17:06.1528955Z [W1204 11:10:32.394340141 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1528990Z 2025-12-04T11:17:06.1529503Z [W1204 11:10:32.529238251 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1529507Z 2025-12-04T11:17:06.1530027Z [W1204 11:10:32.530961285 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1530032Z 2025-12-04T11:17:06.1530539Z [W1204 11:10:32.531174200 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1530546Z 2025-12-04T11:17:06.1531057Z [W1204 11:10:32.535066195 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1531074Z 2025-12-04T11:17:06.1531586Z [W1204 11:10:32.535717164 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1531593Z 2025-12-04T11:17:06.1532106Z [W1204 11:10:32.535911844 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1532112Z 2025-12-04T11:17:06.1540719Z [W1204 11:10:32.541964712 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1540736Z 2025-12-04T11:17:06.1541335Z [W1204 11:10:32.542605399 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1541341Z 2025-12-04T11:17:06.1541870Z [W1204 11:10:32.542797419 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1541876Z 2025-12-04T11:17:06.1542034Z ('RERUN', {'yellow': True}) [18.9929s] [100%] 2025-12-04T11:17:06.1543322Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 [W1204 11:10:33.930864297 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1543332Z 2025-12-04T11:17:06.1543850Z [W1204 11:10:33.931602522 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1543869Z 2025-12-04T11:17:06.1544382Z [W1204 11:10:33.931798618 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1544389Z 2025-12-04T11:17:06.1544902Z [W1204 11:10:33.935698718 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1544908Z 2025-12-04T11:17:06.1545573Z [W1204 11:10:33.936324316 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1545582Z 2025-12-04T11:17:06.1546098Z [W1204 11:10:33.936515033 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1546102Z 2025-12-04T11:17:06.1546625Z [W1204 11:10:33.942585717 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1546630Z 2025-12-04T11:17:06.1547142Z [W1204 11:10:33.943195512 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1547147Z 2025-12-04T11:17:06.1547710Z [W1204 11:10:33.943384521 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1547715Z 2025-12-04T11:17:06.1548230Z [W1204 11:10:33.029298562 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1548270Z 2025-12-04T11:17:06.1548790Z [W1204 11:10:33.030028578 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1548795Z 2025-12-04T11:17:06.1549302Z [W1204 11:10:33.030236591 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1549307Z 2025-12-04T11:17:06.1549819Z [W1204 11:10:33.034098629 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1549838Z 2025-12-04T11:17:06.1550350Z [W1204 11:10:33.034709569 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1550357Z 2025-12-04T11:17:06.1550871Z [W1204 11:10:33.034904370 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1550878Z 2025-12-04T11:17:06.1551401Z [W1204 11:10:33.040906733 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1551405Z 2025-12-04T11:17:06.1551915Z [W1204 11:10:33.041685483 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1551920Z 2025-12-04T11:17:06.1552443Z [W1204 11:10:33.041878649 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1552448Z 2025-12-04T11:17:06.1552583Z ('RERUN', {'yellow': True}) [0.4596s] [100%] 2025-12-04T11:17:06.1553875Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 [W1204 11:10:33.365514873 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1553883Z 2025-12-04T11:17:06.1554392Z [W1204 11:10:33.366236336 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1554398Z 2025-12-04T11:17:06.1554918Z [W1204 11:10:33.366432298 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1554923Z 2025-12-04T11:17:06.1555439Z [W1204 11:10:33.370480885 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1555444Z 2025-12-04T11:17:06.1555953Z [W1204 11:10:33.371091800 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1555972Z 2025-12-04T11:17:06.1556564Z [W1204 11:10:33.371280286 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1556569Z 2025-12-04T11:17:06.1557086Z [W1204 11:10:33.377292212 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1557091Z 2025-12-04T11:17:06.1557614Z [W1204 11:10:33.377894729 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1557620Z 2025-12-04T11:17:06.1558125Z [W1204 11:10:33.378081938 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1558130Z 2025-12-04T11:17:06.1558653Z [W1204 11:10:33.465128277 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1558687Z 2025-12-04T11:17:06.1559194Z [W1204 11:10:33.465870265 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1559203Z 2025-12-04T11:17:06.1559723Z [W1204 11:10:33.466077506 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1559756Z 2025-12-04T11:17:06.1560264Z [W1204 11:10:33.469944668 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1560268Z 2025-12-04T11:17:06.1560773Z [W1204 11:10:33.470597262 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1560791Z 2025-12-04T11:17:06.1561298Z [W1204 11:10:33.470805088 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1561305Z 2025-12-04T11:17:06.1561814Z [W1204 11:10:33.476815584 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1561819Z 2025-12-04T11:17:06.1562345Z [W1204 11:10:33.477602043 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1562352Z 2025-12-04T11:17:06.1562859Z [W1204 11:10:33.477794103 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1562864Z 2025-12-04T11:17:06.1562981Z FAILED [0.4348s] [100%] 2025-12-04T11:17:06.1562986Z 2025-12-04T11:17:06.1563132Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.1563659Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.1563785Z Traceback (most recent call last): 2025-12-04T11:17:06.1564298Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1564549Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1565016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1565184Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1565739Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1565947Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1566094Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1566099Z 2025-12-04T11:17:06.1566204Z Expected 1 but got 2. 2025-12-04T11:17:06.1566312Z Absolute difference: 1 2025-12-04T11:17:06.1566434Z Relative difference: 1.0 2025-12-04T11:17:06.1566442Z 2025-12-04T11:17:06.1566657Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1567631Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1567653Z 2025-12-04T11:17:06.1567926Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1568150Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1568288Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1569198Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1569440Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1569573Z graph_break [] 2025-12-04T11:17:06.1569796Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1571036Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1571186Z if out == self.unknown_value: 2025-12-04T11:17:06.1571923Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1572036Z warnings.warn( 2025-12-04T11:17:06.1572758Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1572873Z warnings.warn( 2025-12-04T11:17:06.1573383Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.1573508Z Traceback (most recent call last): 2025-12-04T11:17:06.1574036Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1574271Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1574744Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1574909Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1575448Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1575667Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1575796Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1575801Z 2025-12-04T11:17:06.1575910Z Expected 1 but got 2. 2025-12-04T11:17:06.1576031Z Absolute difference: 1 2025-12-04T11:17:06.1576142Z Relative difference: 1.0 2025-12-04T11:17:06.1576147Z 2025-12-04T11:17:06.1576375Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1577398Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1577407Z 2025-12-04T11:17:06.1577682Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1577920Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1578037Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1578955Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1579185Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1579284Z graph_break [] 2025-12-04T11:17:06.1579583Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1580800Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1580932Z if out == self.unknown_value: 2025-12-04T11:17:06.1581662Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1581767Z warnings.warn( 2025-12-04T11:17:06.1582500Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1582638Z warnings.warn( 2025-12-04T11:17:06.1582859Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1582988Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1583224Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1584161Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1584261Z graph_break [] 2025-12-04T11:17:06.1584480Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1585225Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1585328Z warnings.warn( 2025-12-04T11:17:06.1586066Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1586167Z warnings.warn( 2025-12-04T11:17:06.1586317Z =================================== FAILURES =================================== 2025-12-04T11:17:06.1586837Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.1586962Z Traceback (most recent call last): 2025-12-04T11:17:06.1587472Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1587721Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1588183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1588363Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1588904Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1589115Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1589267Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1589275Z 2025-12-04T11:17:06.1589379Z Expected 1 but got 2. 2025-12-04T11:17:06.1589498Z Absolute difference: 1 2025-12-04T11:17:06.1589610Z Relative difference: 1.0 2025-12-04T11:17:06.1589614Z 2025-12-04T11:17:06.1589830Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1590753Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1590759Z 2025-12-04T11:17:06.1591030Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1591264Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1591379Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1592322Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1592567Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1592666Z graph_break [] 2025-12-04T11:17:06.1592882Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1594107Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1594224Z if out == self.unknown_value: 2025-12-04T11:17:06.1595573Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1595676Z warnings.warn( 2025-12-04T11:17:06.1596588Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1596791Z warnings.warn( 2025-12-04T11:17:06.1597015Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1597147Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1597378Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1598278Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1598390Z graph_break [] 2025-12-04T11:17:06.1598611Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1599358Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1599462Z warnings.warn( 2025-12-04T11:17:06.1600186Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1600301Z warnings.warn( 2025-12-04T11:17:06.1600517Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1600632Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1600874Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1602094Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1602213Z graph_break [] 2025-12-04T11:17:06.1602433Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1603167Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1603285Z warnings.warn( 2025-12-04T11:17:06.1604005Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1604117Z warnings.warn( 2025-12-04T11:17:06.1604963Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-2a4f8a9a7e80b95a.xml - 2025-12-04T11:17:06.1605137Z =========================== short test summary info ============================ 2025-12-04T11:17:06.1606200Z FAILED [0.4348s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1606208Z 2025-12-04T11:17:06.1606321Z Expected 1 but got 2. 2025-12-04T11:17:06.1606443Z Absolute difference: 1 2025-12-04T11:17:06.1606554Z Relative difference: 1.0 2025-12-04T11:17:06.1606559Z 2025-12-04T11:17:06.1606779Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1607707Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1607713Z 2025-12-04T11:17:06.1607984Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1608183Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.1608432Z ================== 1 failed, 10 deselected, 2 rerun in 19.92s ================== 2025-12-04T11:17:06.1608532Z Got exit code 1 2025-12-04T11:17:06.1608651Z Retrying single test... 2025-12-04T11:17:06.1609105Z W1204 11:10:44.949000 95164 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.1609810Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-ab0a7cd9bbef732b.xml 2025-12-04T11:17:06.1609990Z ============================= test session starts ============================== 2025-12-04T11:17:06.1610343Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.1610465Z cachedir: .pytest_cache 2025-12-04T11:17:06.1610987Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.1611115Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.1611236Z configfile: pytest.ini 2025-12-04T11:17:06.1611790Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.1612015Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.1613141Z stepcurrent: skipping 6 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1613342Z Running 1 items in this shard 2025-12-04T11:17:06.1613352Z 2025-12-04T11:17:06.1615032Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 [W1204 11:10:50.104651482 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1615042Z 2025-12-04T11:17:06.1615566Z [W1204 11:11:05.096107950 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1615576Z 2025-12-04T11:17:06.1616106Z [W1204 11:11:05.096377799 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1616113Z 2025-12-04T11:17:06.1616625Z [W1204 11:11:05.103735188 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1616630Z 2025-12-04T11:17:06.1617237Z [W1204 11:11:05.104459250 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1617243Z 2025-12-04T11:17:06.1617753Z [W1204 11:11:05.104650092 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1617761Z 2025-12-04T11:17:06.1618273Z [W1204 11:11:05.111519921 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1618293Z 2025-12-04T11:17:06.1618885Z [W1204 11:11:05.112306728 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1618893Z 2025-12-04T11:17:06.1619404Z [W1204 11:11:05.112491462 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1619409Z 2025-12-04T11:17:06.1619929Z [W1204 11:11:05.245826050 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1619933Z 2025-12-04T11:17:06.1620442Z [W1204 11:11:05.247525335 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1620477Z 2025-12-04T11:17:06.1621001Z [W1204 11:11:05.247736375 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1621006Z 2025-12-04T11:17:06.1621525Z [W1204 11:11:05.251780620 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1621559Z 2025-12-04T11:17:06.1622084Z [W1204 11:11:05.252439510 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1622089Z 2025-12-04T11:17:06.1622596Z [W1204 11:11:05.252637794 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1622601Z 2025-12-04T11:17:06.1623122Z [W1204 11:11:05.258646398 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1623127Z 2025-12-04T11:17:06.1623638Z [W1204 11:11:05.259273125 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1623643Z 2025-12-04T11:17:06.1624157Z [W1204 11:11:05.259466202 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1624176Z 2025-12-04T11:17:06.1624320Z ('RERUN', {'yellow': True}) [18.8673s] [100%] 2025-12-04T11:17:06.1625598Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 [W1204 11:11:05.656180885 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1625604Z 2025-12-04T11:17:06.1626133Z [W1204 11:11:05.656917578 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1626138Z 2025-12-04T11:17:06.1626652Z [W1204 11:11:05.657112922 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1626657Z 2025-12-04T11:17:06.1627184Z [W1204 11:11:05.661177050 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1627191Z 2025-12-04T11:17:06.1627703Z [W1204 11:11:05.661788272 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1627708Z 2025-12-04T11:17:06.1628227Z [W1204 11:11:05.661975052 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1628232Z 2025-12-04T11:17:06.1628745Z [W1204 11:11:05.668090654 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1628750Z 2025-12-04T11:17:06.1629258Z [W1204 11:11:05.668707222 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1629278Z 2025-12-04T11:17:06.1629845Z [W1204 11:11:05.668894921 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1629852Z 2025-12-04T11:17:06.1630365Z [W1204 11:11:06.756706112 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1630370Z 2025-12-04T11:17:06.1630894Z [W1204 11:11:06.757464921 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1630900Z 2025-12-04T11:17:06.1631408Z [W1204 11:11:06.757672653 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1631413Z 2025-12-04T11:17:06.1631943Z [W1204 11:11:06.761617556 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1631978Z 2025-12-04T11:17:06.1632492Z [W1204 11:11:06.762252971 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1632527Z 2025-12-04T11:17:06.1633060Z [W1204 11:11:06.762448575 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1633064Z 2025-12-04T11:17:06.1633572Z [W1204 11:11:06.768509084 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1633576Z 2025-12-04T11:17:06.1634096Z [W1204 11:11:06.769302241 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1634101Z 2025-12-04T11:17:06.1634611Z [W1204 11:11:06.769496448 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1634618Z 2025-12-04T11:17:06.1634751Z ('RERUN', {'yellow': True}) [0.4708s] [100%] 2025-12-04T11:17:06.1636044Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 [W1204 11:11:06.100146855 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1636052Z 2025-12-04T11:17:06.1636562Z [W1204 11:11:06.100870836 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1636567Z 2025-12-04T11:17:06.1637091Z [W1204 11:11:06.101069601 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1637095Z 2025-12-04T11:17:06.1637604Z [W1204 11:11:06.105062994 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1637611Z 2025-12-04T11:17:06.1638139Z [W1204 11:11:06.105676099 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1638144Z 2025-12-04T11:17:06.1638653Z [W1204 11:11:06.105863991 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1638658Z 2025-12-04T11:17:06.1639175Z [W1204 11:11:06.112093048 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1639179Z 2025-12-04T11:17:06.1639686Z [W1204 11:11:06.112711030 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1639690Z 2025-12-04T11:17:06.1640210Z [W1204 11:11:06.112898673 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1640217Z 2025-12-04T11:17:06.1640725Z [W1204 11:11:06.200654206 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1640789Z 2025-12-04T11:17:06.1641300Z [W1204 11:11:06.201398648 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1641306Z 2025-12-04T11:17:06.1641828Z [W1204 11:11:06.201606673 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1641833Z 2025-12-04T11:17:06.1642343Z [W1204 11:11:06.205523269 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1642348Z 2025-12-04T11:17:06.1642874Z [W1204 11:11:06.206151581 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1642926Z 2025-12-04T11:17:06.1643436Z [W1204 11:11:06.206347274 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1643441Z 2025-12-04T11:17:06.1643967Z [W1204 11:11:06.212433807 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1644002Z 2025-12-04T11:17:06.1644761Z [W1204 11:11:06.213227838 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1644767Z 2025-12-04T11:17:06.1645289Z [W1204 11:11:06.213420904 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1645294Z 2025-12-04T11:17:06.1645400Z FAILED [0.4415s] [100%] 2025-12-04T11:17:06.1645405Z 2025-12-04T11:17:06.1645548Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.1646078Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.1646202Z Traceback (most recent call last): 2025-12-04T11:17:06.1646735Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1646974Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1647441Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1647613Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1648168Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1648375Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1648508Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1648516Z 2025-12-04T11:17:06.1648632Z Expected 1 but got 2. 2025-12-04T11:17:06.1648737Z Absolute difference: 1 2025-12-04T11:17:06.1648845Z Relative difference: 1.0 2025-12-04T11:17:06.1648861Z 2025-12-04T11:17:06.1649081Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1649997Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1650003Z 2025-12-04T11:17:06.1650283Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1650507Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1650633Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1651529Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1651759Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1651871Z graph_break [] 2025-12-04T11:17:06.1652166Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1653383Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1653512Z if out == self.unknown_value: 2025-12-04T11:17:06.1654241Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1654356Z warnings.warn( 2025-12-04T11:17:06.1655078Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1655211Z warnings.warn( 2025-12-04T11:17:06.1655745Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.1655903Z Traceback (most recent call last): 2025-12-04T11:17:06.1656425Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1656657Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1657188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1657376Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1657909Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1658137Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1658272Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1658278Z 2025-12-04T11:17:06.1658385Z Expected 1 but got 2. 2025-12-04T11:17:06.1658505Z Absolute difference: 1 2025-12-04T11:17:06.1658618Z Relative difference: 1.0 2025-12-04T11:17:06.1658625Z 2025-12-04T11:17:06.1658840Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1659761Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1659766Z 2025-12-04T11:17:06.1660038Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1660269Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1660383Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1661277Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1661522Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1661621Z graph_break [] 2025-12-04T11:17:06.1661850Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1663068Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1663183Z if out == self.unknown_value: 2025-12-04T11:17:06.1663922Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1664028Z warnings.warn( 2025-12-04T11:17:06.1664758Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1664922Z warnings.warn( 2025-12-04T11:17:06.1665143Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1665271Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1665503Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1666401Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1666512Z graph_break [] 2025-12-04T11:17:06.1666726Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1667458Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1667594Z warnings.warn( 2025-12-04T11:17:06.1668320Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1668464Z warnings.warn( 2025-12-04T11:17:06.1668610Z =================================== FAILURES =================================== 2025-12-04T11:17:06.1669127Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.1669250Z Traceback (most recent call last): 2025-12-04T11:17:06.1669759Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1670001Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1670461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1670627Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1671183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1671391Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1671532Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1671537Z 2025-12-04T11:17:06.1671641Z Expected 1 but got 2. 2025-12-04T11:17:06.1671747Z Absolute difference: 1 2025-12-04T11:17:06.1671868Z Relative difference: 1.0 2025-12-04T11:17:06.1671873Z 2025-12-04T11:17:06.1672084Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1673009Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1673017Z 2025-12-04T11:17:06.1673289Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1673509Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1673637Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1674537Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1674777Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1674876Z graph_break [] 2025-12-04T11:17:06.1675097Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1676322Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1676440Z if out == self.unknown_value: 2025-12-04T11:17:06.1677224Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1677341Z warnings.warn( 2025-12-04T11:17:06.1678060Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1678173Z warnings.warn( 2025-12-04T11:17:06.1678394Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1678509Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1678750Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1679644Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1679787Z graph_break [] 2025-12-04T11:17:06.1680009Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1680771Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1680885Z warnings.warn( 2025-12-04T11:17:06.1681605Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1681700Z warnings.warn( 2025-12-04T11:17:06.1681928Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1682040Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1682278Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1683172Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1683271Z graph_break [] 2025-12-04T11:17:06.1683497Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1684225Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1684336Z warnings.warn( 2025-12-04T11:17:06.1685058Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1685158Z warnings.warn( 2025-12-04T11:17:06.1686014Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-ab0a7cd9bbef732b.xml - 2025-12-04T11:17:06.1686187Z =========================== short test summary info ============================ 2025-12-04T11:17:06.1687146Z FAILED [0.4415s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1687153Z 2025-12-04T11:17:06.1687261Z Expected 1 but got 2. 2025-12-04T11:17:06.1687367Z Absolute difference: 1 2025-12-04T11:17:06.1687488Z Relative difference: 1.0 2025-12-04T11:17:06.1687493Z 2025-12-04T11:17:06.1687712Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1688614Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1688634Z 2025-12-04T11:17:06.1688900Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1689082Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.1689367Z ================== 1 failed, 10 deselected, 2 rerun in 19.81s ================== 2025-12-04T11:17:06.1689468Z Got exit code 1 2025-12-04T11:17:06.1690290Z FAILED CONSISTENTLY: test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1690712Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:17:06.1691154Z W1204 11:11:17.722000 95338 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.1691823Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-81245131849bbecb.xml 2025-12-04T11:17:06.1692016Z ============================= test session starts ============================== 2025-12-04T11:17:06.1692370Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.1692490Z cachedir: .pytest_cache 2025-12-04T11:17:06.1693038Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.1693171Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.1693278Z configfile: pytest.ini 2025-12-04T11:17:06.1693819Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.1694045Z collecting ... collected 58 items / 7 deselected / 51 selected 2025-12-04T11:17:06.1694182Z stepcurrent: skipping 7 already run items. 2025-12-04T11:17:06.1694294Z Running 4 items in this shard 2025-12-04T11:17:06.1694313Z 2025-12-04T11:17:06.1695177Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 ('RERUN', {'yellow': True}) [3.8084s] [ 25%] 2025-12-04T11:17:06.1696199Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 ('RERUN', {'yellow': True}) [0.3927s] [ 25%] 2025-12-04T11:17:06.1697050Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 FAILED [0.4000s] [ 25%] 2025-12-04T11:17:06.1697057Z 2025-12-04T11:17:06.1697198Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.1697714Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.1697838Z Traceback (most recent call last): 2025-12-04T11:17:06.1698344Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1698594Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1699054Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1699230Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1699766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1699970Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1700115Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1700121Z 2025-12-04T11:17:06.1700225Z Expected 1 but got 2. 2025-12-04T11:17:06.1700329Z Absolute difference: 1 2025-12-04T11:17:06.1700448Z Relative difference: 1.0 2025-12-04T11:17:06.1700456Z 2025-12-04T11:17:06.1700667Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1701771Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1701780Z 2025-12-04T11:17:06.1702046Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1702260Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1702381Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1702913Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1703150Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1703246Z graph_break [] 2025-12-04T11:17:06.1703458Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1704252Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1704354Z warnings.warn( 2025-12-04T11:17:06.1705125Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1705223Z warnings.warn( 2025-12-04T11:17:06.1705726Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.1705860Z Traceback (most recent call last): 2025-12-04T11:17:06.1706366Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1706599Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1707067Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1707225Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1707774Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1707976Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1708103Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1708108Z 2025-12-04T11:17:06.1708222Z Expected 1 but got 2. 2025-12-04T11:17:06.1708323Z Absolute difference: 1 2025-12-04T11:17:06.1708431Z Relative difference: 1.0 2025-12-04T11:17:06.1708436Z 2025-12-04T11:17:06.1708659Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1709564Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1709571Z 2025-12-04T11:17:06.1709846Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1710065Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1710177Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1710718Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1710942Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1711046Z graph_break [] 2025-12-04T11:17:06.1711256Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1711988Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1712097Z warnings.warn( 2025-12-04T11:17:06.1712817Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1712922Z warnings.warn( 2025-12-04T11:17:06.1713192Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1713311Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1713547Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1714071Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1714167Z graph_break [] 2025-12-04T11:17:06.1714386Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1715109Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1715249Z warnings.warn( 2025-12-04T11:17:06.1715963Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1716057Z warnings.warn( 2025-12-04T11:17:06.1716238Z =================================== FAILURES =================================== 2025-12-04T11:17:06.1716742Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.1716857Z Traceback (most recent call last): 2025-12-04T11:17:06.1717368Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1717597Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1718057Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1718218Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1718749Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1718969Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1719097Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1719102Z 2025-12-04T11:17:06.1719213Z Expected 1 but got 2. 2025-12-04T11:17:06.1719315Z Absolute difference: 1 2025-12-04T11:17:06.1719420Z Relative difference: 1.0 2025-12-04T11:17:06.1719425Z 2025-12-04T11:17:06.1719653Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1720559Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1720564Z 2025-12-04T11:17:06.1720844Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1721059Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1721170Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1721713Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1721939Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1722033Z graph_break [] 2025-12-04T11:17:06.1722253Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1722982Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1723089Z warnings.warn( 2025-12-04T11:17:06.1723808Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1723905Z warnings.warn( 2025-12-04T11:17:06.1724127Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1724301Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1724525Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1725059Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1725157Z graph_break [] 2025-12-04T11:17:06.1725385Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1726108Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1726203Z warnings.warn( 2025-12-04T11:17:06.1726930Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1727058Z warnings.warn( 2025-12-04T11:17:06.1727278Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1727427Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1727650Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1728190Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1728282Z graph_break [] 2025-12-04T11:17:06.1728494Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1729222Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1729320Z warnings.warn( 2025-12-04T11:17:06.1730045Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1730144Z warnings.warn( 2025-12-04T11:17:06.1731277Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-81245131849bbecb.xml - 2025-12-04T11:17:06.1731464Z =========================== short test summary info ============================ 2025-12-04T11:17:06.1732404Z FAILED [0.4000s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1732410Z 2025-12-04T11:17:06.1732529Z Expected 1 but got 2. 2025-12-04T11:17:06.1732631Z Absolute difference: 1 2025-12-04T11:17:06.1732743Z Relative difference: 1.0 2025-12-04T11:17:06.1732750Z 2025-12-04T11:17:06.1732975Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1733882Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1733889Z 2025-12-04T11:17:06.1734175Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1734353Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.1734550Z =================== 1 failed, 7 deselected, 2 rerun in 4.63s =================== 2025-12-04T11:17:06.1734658Z Got exit code 1 2025-12-04T11:17:06.1734765Z Retrying single test... 2025-12-04T11:17:06.1735205Z W1204 11:11:37.555000 95507 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.1735880Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-4cf813d975e62c57.xml 2025-12-04T11:17:06.1736041Z ============================= test session starts ============================== 2025-12-04T11:17:06.1736495Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.1736602Z cachedir: .pytest_cache 2025-12-04T11:17:06.1737188Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.1737324Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.1737428Z configfile: pytest.ini 2025-12-04T11:17:06.1737976Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.1738201Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.1739189Z stepcurrent: skipping 7 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1739358Z Running 1 items in this shard 2025-12-04T11:17:06.1739364Z 2025-12-04T11:17:06.1740643Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 [W1204 11:11:41.792945104 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1740679Z 2025-12-04T11:17:06.1741209Z [W1204 11:11:56.315435604 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1741215Z 2025-12-04T11:17:06.1741725Z [W1204 11:11:56.315696445 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1741731Z 2025-12-04T11:17:06.1742256Z [W1204 11:11:56.323136473 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1742261Z 2025-12-04T11:17:06.1742775Z [W1204 11:11:56.323830446 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1742783Z 2025-12-04T11:17:06.1743290Z [W1204 11:11:56.324024526 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1743303Z 2025-12-04T11:17:06.1743809Z [W1204 11:11:56.330855747 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1743814Z 2025-12-04T11:17:06.1744315Z [W1204 11:11:56.331493135 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1744319Z 2025-12-04T11:17:06.1744837Z [W1204 11:11:56.331678390 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1744844Z 2025-12-04T11:17:06.1745349Z [W1204 11:11:58.334972324 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1745356Z 2025-12-04T11:17:06.1745875Z [W1204 11:11:58.336690429 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1745880Z 2025-12-04T11:17:06.1746384Z [W1204 11:11:58.336898976 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1746389Z 2025-12-04T11:17:06.1746903Z [W1204 11:11:58.340907307 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1746907Z 2025-12-04T11:17:06.1747408Z [W1204 11:11:58.341546022 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1747415Z 2025-12-04T11:17:06.1747984Z [W1204 11:11:58.341739765 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1747992Z 2025-12-04T11:17:06.1748502Z [W1204 11:11:58.347797084 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1748506Z 2025-12-04T11:17:06.1749007Z [W1204 11:11:58.348422148 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1749012Z 2025-12-04T11:17:06.1749526Z [W1204 11:11:58.348613873 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1749531Z 2025-12-04T11:17:06.1749659Z ('RERUN', {'yellow': True}) [19.3498s] [100%] 2025-12-04T11:17:06.1750981Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 [W1204 11:11:58.695458973 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1751016Z 2025-12-04T11:17:06.1751522Z [W1204 11:11:58.696180599 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1751527Z 2025-12-04T11:17:06.1752044Z [W1204 11:11:58.696384316 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1752048Z 2025-12-04T11:17:06.1752557Z [W1204 11:11:59.700314242 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1752562Z 2025-12-04T11:17:06.1753081Z [W1204 11:11:59.701081995 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1753088Z 2025-12-04T11:17:06.1753597Z [W1204 11:11:59.701271362 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1753602Z 2025-12-04T11:17:06.1754113Z [W1204 11:11:59.707222903 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1754129Z 2025-12-04T11:17:06.1754637Z [W1204 11:11:59.707827831 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1754642Z 2025-12-04T11:17:06.1755150Z [W1204 11:11:59.708015244 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1755155Z 2025-12-04T11:17:06.1755678Z [W1204 11:11:59.792533178 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1755685Z 2025-12-04T11:17:06.1756193Z [W1204 11:11:59.793209881 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1756203Z 2025-12-04T11:17:06.1756726Z [W1204 11:11:59.793410489 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1756733Z 2025-12-04T11:17:06.1757241Z [W1204 11:11:59.797272388 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1757245Z 2025-12-04T11:17:06.1757765Z [W1204 11:11:59.797882442 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1757770Z 2025-12-04T11:17:06.1758281Z [W1204 11:11:59.798075834 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1758288Z 2025-12-04T11:17:06.1758812Z [W1204 11:11:59.804083693 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1758817Z 2025-12-04T11:17:06.1759386Z [W1204 11:11:59.804862048 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1759394Z 2025-12-04T11:17:06.1759907Z [W1204 11:11:59.805055731 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1759927Z 2025-12-04T11:17:06.1760060Z ('RERUN', {'yellow': True}) [0.4162s] [100%] 2025-12-04T11:17:06.1761332Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 [W1204 11:11:59.085802296 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1761366Z 2025-12-04T11:17:06.1761893Z [W1204 11:11:59.086524521 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1761899Z 2025-12-04T11:17:06.1762416Z [W1204 11:11:59.086726898 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1762450Z 2025-12-04T11:17:06.1762980Z [W1204 11:11:59.090664309 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1762985Z 2025-12-04T11:17:06.1763496Z [W1204 11:11:59.091437905 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1763501Z 2025-12-04T11:17:06.1764025Z [W1204 11:11:59.091627399 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1764032Z 2025-12-04T11:17:06.1764545Z [W1204 11:11:59.097583280 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1764550Z 2025-12-04T11:17:06.1765066Z [W1204 11:11:59.098187410 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1765087Z 2025-12-04T11:17:06.1765597Z [W1204 11:11:59.098374648 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1765602Z 2025-12-04T11:17:06.1766114Z [W1204 11:11:59.184029564 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1766118Z 2025-12-04T11:17:06.1766644Z [W1204 11:11:59.184760122 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1766651Z 2025-12-04T11:17:06.1767161Z [W1204 11:11:59.184967415 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1767165Z 2025-12-04T11:17:06.1767695Z [W1204 11:11:59.188862209 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1767702Z 2025-12-04T11:17:06.1768217Z [W1204 11:11:59.189507666 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1768221Z 2025-12-04T11:17:06.1768749Z [W1204 11:11:59.189705005 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1768753Z 2025-12-04T11:17:06.1769264Z [W1204 11:11:59.195713978 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1769269Z 2025-12-04T11:17:06.1769794Z [W1204 11:11:59.196533202 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1769799Z 2025-12-04T11:17:06.1770367Z [W1204 11:11:59.196725107 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1770374Z 2025-12-04T11:17:06.1770480Z FAILED [0.3909s] [100%] 2025-12-04T11:17:06.1770485Z 2025-12-04T11:17:06.1770645Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.1771156Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.1771298Z Traceback (most recent call last): 2025-12-04T11:17:06.1771812Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1772044Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1772557Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1772723Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1773282Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1773533Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1773666Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1773671Z 2025-12-04T11:17:06.1773790Z Expected 1 but got 2. 2025-12-04T11:17:06.1773897Z Absolute difference: 1 2025-12-04T11:17:06.1774008Z Relative difference: 1.0 2025-12-04T11:17:06.1774013Z 2025-12-04T11:17:06.1774251Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1775155Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1775164Z 2025-12-04T11:17:06.1775443Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1775665Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1775784Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1776335Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1776561Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1776670Z graph_break [] 2025-12-04T11:17:06.1776948Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1778165Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1778298Z if out == self.unknown_value: 2025-12-04T11:17:06.1779035Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1779154Z warnings.warn( 2025-12-04T11:17:06.1779872Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1779973Z warnings.warn( 2025-12-04T11:17:06.1780496Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.1780618Z Traceback (most recent call last): 2025-12-04T11:17:06.1781130Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1781373Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1781836Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1782080Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1782618Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1782825Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1782969Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1782975Z 2025-12-04T11:17:06.1783085Z Expected 1 but got 2. 2025-12-04T11:17:06.1783192Z Absolute difference: 1 2025-12-04T11:17:06.1783313Z Relative difference: 1.0 2025-12-04T11:17:06.1783317Z 2025-12-04T11:17:06.1783536Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1784456Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1784492Z 2025-12-04T11:17:06.1784767Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1784994Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1785158Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1785694Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1785935Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1786036Z graph_break [] 2025-12-04T11:17:06.1786256Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1787485Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1787607Z if out == self.unknown_value: 2025-12-04T11:17:06.1788354Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1788456Z warnings.warn( 2025-12-04T11:17:06.1789182Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1789294Z warnings.warn( 2025-12-04T11:17:06.1789514Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1789630Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1789873Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1790408Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1790522Z graph_break [] 2025-12-04T11:17:06.1790740Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1791468Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1791581Z warnings.warn( 2025-12-04T11:17:06.1792301Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1792413Z warnings.warn( 2025-12-04T11:17:06.1792560Z =================================== FAILURES =================================== 2025-12-04T11:17:06.1793072Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.1793215Z Traceback (most recent call last): 2025-12-04T11:17:06.1793723Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1794013Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1794486Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1794654Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1795205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1795413Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1795545Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1795550Z 2025-12-04T11:17:06.1795668Z Expected 1 but got 2. 2025-12-04T11:17:06.1795776Z Absolute difference: 1 2025-12-04T11:17:06.1795884Z Relative difference: 1.0 2025-12-04T11:17:06.1795930Z 2025-12-04T11:17:06.1796345Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1797259Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1797327Z 2025-12-04T11:17:06.1797616Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1797836Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1797952Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1798506Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1798735Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1798849Z graph_break [] 2025-12-04T11:17:06.1799067Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1800293Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1800431Z if out == self.unknown_value: 2025-12-04T11:17:06.1801154Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1801266Z warnings.warn( 2025-12-04T11:17:06.1801988Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1802088Z warnings.warn( 2025-12-04T11:17:06.1802318Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1802436Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1802665Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1803219Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1803317Z graph_break [] 2025-12-04T11:17:06.1803547Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1804272Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1804373Z warnings.warn( 2025-12-04T11:17:06.1805107Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1805207Z warnings.warn( 2025-12-04T11:17:06.1805433Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1805549Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1805780Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1806409Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1806510Z graph_break [] 2025-12-04T11:17:06.1806726Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1807462Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1807563Z warnings.warn( 2025-12-04T11:17:06.1808291Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1808391Z warnings.warn( 2025-12-04T11:17:06.1809276Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-4cf813d975e62c57.xml - 2025-12-04T11:17:06.1809467Z =========================== short test summary info ============================ 2025-12-04T11:17:06.1810436Z FAILED [0.3909s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1810442Z 2025-12-04T11:17:06.1810561Z Expected 1 but got 2. 2025-12-04T11:17:06.1810668Z Absolute difference: 1 2025-12-04T11:17:06.1810778Z Relative difference: 1.0 2025-12-04T11:17:06.1810783Z 2025-12-04T11:17:06.1811013Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1811922Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1811930Z 2025-12-04T11:17:06.1812211Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1812396Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.1812598Z ================== 1 failed, 10 deselected, 2 rerun in 20.19s ================== 2025-12-04T11:17:06.1812709Z Got exit code 1 2025-12-04T11:17:06.1812815Z Retrying single test... 2025-12-04T11:17:06.1813264Z W1204 11:12:10.818000 95681 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.1813938Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-f694c40280fe3f31.xml 2025-12-04T11:17:06.1814105Z ============================= test session starts ============================== 2025-12-04T11:17:06.1814474Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.1814588Z cachedir: .pytest_cache 2025-12-04T11:17:06.1815114Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.1815255Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.1815364Z configfile: pytest.ini 2025-12-04T11:17:06.1815923Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.1816142Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.1817195Z stepcurrent: skipping 7 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1817327Z Running 1 items in this shard 2025-12-04T11:17:06.1817332Z 2025-12-04T11:17:06.1818671Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 [W1204 11:12:14.055527345 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1818679Z 2025-12-04T11:17:06.1819219Z [W1204 11:12:29.552616610 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1819225Z 2025-12-04T11:17:06.1819740Z [W1204 11:12:29.552868428 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1819745Z 2025-12-04T11:17:06.1820268Z [W1204 11:12:29.560252184 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1820273Z 2025-12-04T11:17:06.1820780Z [W1204 11:12:29.560965954 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1820835Z 2025-12-04T11:17:06.1821369Z [W1204 11:12:29.561154911 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1821374Z 2025-12-04T11:17:06.1821916Z [W1204 11:12:29.568045108 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1821921Z 2025-12-04T11:17:06.1822429Z [W1204 11:12:29.568692182 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1822434Z 2025-12-04T11:17:06.1822953Z [W1204 11:12:29.568878737 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1822958Z 2025-12-04T11:17:06.1823466Z [W1204 11:12:31.565315124 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1823473Z 2025-12-04T11:17:06.1824001Z [W1204 11:12:31.567050745 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1824005Z 2025-12-04T11:17:06.1824516Z [W1204 11:12:31.567261407 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1824521Z 2025-12-04T11:17:06.1825041Z [W1204 11:12:31.571234795 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1825046Z 2025-12-04T11:17:06.1825556Z [W1204 11:12:31.571868906 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1825561Z 2025-12-04T11:17:06.1826086Z [W1204 11:12:31.572069605 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1826093Z 2025-12-04T11:17:06.1826605Z [W1204 11:12:31.578158556 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1826615Z 2025-12-04T11:17:06.1827127Z [W1204 11:12:31.578770290 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1827145Z 2025-12-04T11:17:06.1827659Z [W1204 11:12:31.578963645 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1827665Z 2025-12-04T11:17:06.1827795Z ('RERUN', {'yellow': True}) [19.3180s] [100%] 2025-12-04T11:17:06.1829080Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 [W1204 11:12:32.934272403 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1829088Z 2025-12-04T11:17:06.1829600Z [W1204 11:12:32.935011764 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1829666Z 2025-12-04T11:17:06.1830192Z [W1204 11:12:32.935210533 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1830200Z 2025-12-04T11:17:06.1830706Z [W1204 11:12:32.939134930 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1830711Z 2025-12-04T11:17:06.1831230Z [W1204 11:12:32.939907357 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1831235Z 2025-12-04T11:17:06.1831743Z [W1204 11:12:32.940132006 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1831777Z 2025-12-04T11:17:06.1832314Z [W1204 11:12:32.946157156 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1832321Z 2025-12-04T11:17:06.1833191Z [W1204 11:12:32.946766687 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1833260Z 2025-12-04T11:17:06.1833951Z [W1204 11:12:32.946954480 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1833965Z 2025-12-04T11:17:06.1834470Z [W1204 11:12:32.033157335 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1834475Z 2025-12-04T11:17:06.1834982Z [W1204 11:12:32.033894259 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1834989Z 2025-12-04T11:17:06.1835512Z [W1204 11:12:32.034098593 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1835517Z 2025-12-04T11:17:06.1836031Z [W1204 11:12:32.038015148 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1836038Z 2025-12-04T11:17:06.1836553Z [W1204 11:12:32.038634699 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1836558Z 2025-12-04T11:17:06.1837066Z [W1204 11:12:32.038831379 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1837070Z 2025-12-04T11:17:06.1837588Z [W1204 11:12:32.044901342 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1837593Z 2025-12-04T11:17:06.1838105Z [W1204 11:12:32.045692181 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1838109Z 2025-12-04T11:17:06.1838633Z [W1204 11:12:32.045886738 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1838640Z 2025-12-04T11:17:06.1838772Z ('RERUN', {'yellow': True}) [0.4279s] [100%] 2025-12-04T11:17:06.1840039Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 [W1204 11:12:32.336875730 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1840045Z 2025-12-04T11:17:06.1840565Z [W1204 11:12:32.337617323 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1840569Z 2025-12-04T11:17:06.1841084Z [W1204 11:12:32.337816660 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1841088Z 2025-12-04T11:17:06.1841674Z [W1204 11:12:32.341751377 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1841681Z 2025-12-04T11:17:06.1842194Z [W1204 11:12:32.342523247 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1842199Z 2025-12-04T11:17:06.1842715Z [W1204 11:12:32.342713137 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1842720Z 2025-12-04T11:17:06.1843229Z [W1204 11:12:32.348728596 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1843234Z 2025-12-04T11:17:06.1843787Z [W1204 11:12:32.349341849 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1843793Z 2025-12-04T11:17:06.1844307Z [W1204 11:12:32.349531197 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1844341Z 2025-12-04T11:17:06.1844856Z [W1204 11:12:32.435377816 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1844869Z 2025-12-04T11:17:06.1845376Z [W1204 11:12:32.436117097 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1845381Z 2025-12-04T11:17:06.1845889Z [W1204 11:12:32.436319486 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1845894Z 2025-12-04T11:17:06.1846414Z [W1204 11:12:32.440238528 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1846422Z 2025-12-04T11:17:06.1846935Z [W1204 11:12:32.440860791 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1846942Z 2025-12-04T11:17:06.1847463Z [W1204 11:12:32.441055793 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1847468Z 2025-12-04T11:17:06.1847980Z [W1204 11:12:32.447030939 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1847985Z 2025-12-04T11:17:06.1848505Z [W1204 11:12:32.447805544 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1848510Z 2025-12-04T11:17:06.1849017Z [W1204 11:12:32.448000812 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1849024Z 2025-12-04T11:17:06.1849134Z FAILED [0.4002s] [100%] 2025-12-04T11:17:06.1849139Z 2025-12-04T11:17:06.1849286Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.1849803Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.1849934Z Traceback (most recent call last): 2025-12-04T11:17:06.1850444Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1850676Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1851150Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1851313Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1851854Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1852060Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1852245Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1852250Z 2025-12-04T11:17:06.1852368Z Expected 1 but got 2. 2025-12-04T11:17:06.1852468Z Absolute difference: 1 2025-12-04T11:17:06.1852575Z Relative difference: 1.0 2025-12-04T11:17:06.1852595Z 2025-12-04T11:17:06.1852809Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1853713Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1853719Z 2025-12-04T11:17:06.1854000Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1854219Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1854362Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1854910Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1855134Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1855275Z graph_break [] 2025-12-04T11:17:06.1855488Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1856703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1856833Z if out == self.unknown_value: 2025-12-04T11:17:06.1857631Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1857752Z warnings.warn( 2025-12-04T11:17:06.1858478Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1858577Z warnings.warn( 2025-12-04T11:17:06.1859097Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.1859218Z Traceback (most recent call last): 2025-12-04T11:17:06.1859734Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1859964Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1860420Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1860596Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1861132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1861335Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1861482Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1861489Z 2025-12-04T11:17:06.1861591Z Expected 1 but got 2. 2025-12-04T11:17:06.1861706Z Absolute difference: 1 2025-12-04T11:17:06.1861812Z Relative difference: 1.0 2025-12-04T11:17:06.1861817Z 2025-12-04T11:17:06.1862030Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1862947Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1862953Z 2025-12-04T11:17:06.1863219Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1863449Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1863563Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1864175Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1864416Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1864511Z graph_break [] 2025-12-04T11:17:06.1864726Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1865954Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1866068Z if out == self.unknown_value: 2025-12-04T11:17:06.1866804Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1866938Z warnings.warn( 2025-12-04T11:17:06.1867664Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1867807Z warnings.warn( 2025-12-04T11:17:06.1868024Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1868150Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1868377Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1868911Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1869021Z graph_break [] 2025-12-04T11:17:06.1869240Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1869973Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1870084Z warnings.warn( 2025-12-04T11:17:06.1870808Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1870922Z warnings.warn( 2025-12-04T11:17:06.1871062Z =================================== FAILURES =================================== 2025-12-04T11:17:06.1871568Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.1871700Z Traceback (most recent call last): 2025-12-04T11:17:06.1872206Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1872449Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1872907Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1873070Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1873619Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1873822Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1873951Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1873956Z 2025-12-04T11:17:06.1874072Z Expected 1 but got 2. 2025-12-04T11:17:06.1874176Z Absolute difference: 1 2025-12-04T11:17:06.1874295Z Relative difference: 1.0 2025-12-04T11:17:06.1874300Z 2025-12-04T11:17:06.1874511Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1875412Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1875419Z 2025-12-04T11:17:06.1875701Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1875976Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1876106Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1876639Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1876862Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1876975Z graph_break [] 2025-12-04T11:17:06.1877188Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1878400Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1878557Z if out == self.unknown_value: 2025-12-04T11:17:06.1879288Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1879432Z warnings.warn( 2025-12-04T11:17:06.1880151Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1880248Z warnings.warn( 2025-12-04T11:17:06.1880480Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1880590Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1880830Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1881360Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1881460Z graph_break [] 2025-12-04T11:17:06.1881688Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1882412Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1882514Z warnings.warn( 2025-12-04T11:17:06.1883247Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1883341Z warnings.warn( 2025-12-04T11:17:06.1883568Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1883677Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1883905Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1884441Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.1884539Z graph_break [] 2025-12-04T11:17:06.1884769Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1885488Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1885587Z warnings.warn( 2025-12-04T11:17:06.1886317Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1886410Z warnings.warn( 2025-12-04T11:17:06.1887251Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-f694c40280fe3f31.xml - 2025-12-04T11:17:06.1887434Z =========================== short test summary info ============================ 2025-12-04T11:17:06.1888498Z FAILED [0.4002s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1888506Z 2025-12-04T11:17:06.1888626Z Expected 1 but got 2. 2025-12-04T11:17:06.1888733Z Absolute difference: 1 2025-12-04T11:17:06.1888841Z Relative difference: 1.0 2025-12-04T11:17:06.1888846Z 2025-12-04T11:17:06.1889075Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1889969Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1889974Z 2025-12-04T11:17:06.1890254Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1890431Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.1890658Z ================== 1 failed, 10 deselected, 2 rerun in 20.18s ================== 2025-12-04T11:17:06.1890768Z Got exit code 1 2025-12-04T11:17:06.1891595Z FAILED CONSISTENTLY: test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.1892056Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:17:06.1892504Z W1204 11:12:43.984000 95855 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.1893164Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-a6d9b5e9da7bb417.xml 2025-12-04T11:17:06.1893340Z ============================= test session starts ============================== 2025-12-04T11:17:06.1893694Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.1893814Z cachedir: .pytest_cache 2025-12-04T11:17:06.1894340Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.1894463Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.1894581Z configfile: pytest.ini 2025-12-04T11:17:06.1895124Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.1895338Z collecting ... collected 58 items / 8 deselected / 50 selected 2025-12-04T11:17:06.1895495Z stepcurrent: skipping 8 already run items. 2025-12-04T11:17:06.1895608Z Running 3 items in this shard 2025-12-04T11:17:06.1895613Z 2025-12-04T11:17:06.1896681Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 ('RERUN', {'yellow': True}) [3.8514s] [ 33%] 2025-12-04T11:17:06.1897616Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 ('RERUN', {'yellow': True}) [0.4385s] [ 33%] 2025-12-04T11:17:06.1898410Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 FAILED [0.4409s] [ 33%] 2025-12-04T11:17:06.1898416Z 2025-12-04T11:17:06.1898560Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.1899061Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.1899197Z Traceback (most recent call last): 2025-12-04T11:17:06.1899712Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1899957Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1900421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1900689Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1901244Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1901456Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1901587Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1901592Z 2025-12-04T11:17:06.1901710Z Expected 1 but got 2. 2025-12-04T11:17:06.1901814Z Absolute difference: 1 2025-12-04T11:17:06.1901937Z Relative difference: 1.0 2025-12-04T11:17:06.1901943Z 2025-12-04T11:17:06.1902156Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1903071Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1903118Z 2025-12-04T11:17:06.1903400Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1903661Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1903787Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1904690Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1904919Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1905028Z graph_break [] 2025-12-04T11:17:06.1905243Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1905976Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1906093Z warnings.warn( 2025-12-04T11:17:06.1906825Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1906938Z warnings.warn( 2025-12-04T11:17:06.1907448Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.1907568Z Traceback (most recent call last): 2025-12-04T11:17:06.1908091Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1908325Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1908790Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1908959Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1909499Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1909722Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1909854Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1909860Z 2025-12-04T11:17:06.1909965Z Expected 1 but got 2. 2025-12-04T11:17:06.1910080Z Absolute difference: 1 2025-12-04T11:17:06.1910187Z Relative difference: 1.0 2025-12-04T11:17:06.1910192Z 2025-12-04T11:17:06.1910419Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1911328Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1911334Z 2025-12-04T11:17:06.1911607Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1911839Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1911952Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1912945Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1913175Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1913273Z graph_break [] 2025-12-04T11:17:06.1913496Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1914230Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1914339Z warnings.warn( 2025-12-04T11:17:06.1915091Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1915189Z warnings.warn( 2025-12-04T11:17:06.1915418Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1915565Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1915792Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1916707Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1916800Z graph_break [] 2025-12-04T11:17:06.1917027Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1917754Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1917856Z warnings.warn( 2025-12-04T11:17:06.1918589Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1918691Z warnings.warn( 2025-12-04T11:17:06.1918849Z =================================== FAILURES =================================== 2025-12-04T11:17:06.1919356Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.1919475Z Traceback (most recent call last): 2025-12-04T11:17:06.1919992Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1920222Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1920683Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1920856Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1921396Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1921612Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1921743Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1921748Z 2025-12-04T11:17:06.1921851Z Expected 1 but got 2. 2025-12-04T11:17:06.1921964Z Absolute difference: 1 2025-12-04T11:17:06.1922072Z Relative difference: 1.0 2025-12-04T11:17:06.1922077Z 2025-12-04T11:17:06.1922291Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1923202Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1923210Z 2025-12-04T11:17:06.1923478Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1923701Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1923877Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1924778Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1925013Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1925111Z graph_break [] 2025-12-04T11:17:06.1925333Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1926066Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1926197Z warnings.warn( 2025-12-04T11:17:06.1926926Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1927030Z warnings.warn( 2025-12-04T11:17:06.1927255Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1927399Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1927627Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1928527Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1928623Z graph_break [] 2025-12-04T11:17:06.1928838Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1929574Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1929679Z warnings.warn( 2025-12-04T11:17:06.1930413Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1930514Z warnings.warn( 2025-12-04T11:17:06.1930727Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1930851Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1931079Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1932332Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1932435Z graph_break [] 2025-12-04T11:17:06.1932653Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1933398Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1933499Z warnings.warn( 2025-12-04T11:17:06.1934221Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1934332Z warnings.warn( 2025-12-04T11:17:06.1935180Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-a6d9b5e9da7bb417.xml - 2025-12-04T11:17:06.1935368Z =========================== short test summary info ============================ 2025-12-04T11:17:06.1936308Z FAILED [0.4409s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1936316Z 2025-12-04T11:17:06.1936433Z Expected 1 but got 2. 2025-12-04T11:17:06.1936539Z Absolute difference: 1 2025-12-04T11:17:06.1936724Z Relative difference: 1.0 2025-12-04T11:17:06.1936729Z 2025-12-04T11:17:06.1937029Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1937932Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1937938Z 2025-12-04T11:17:06.1938206Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1938398Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.1938596Z =================== 1 failed, 8 deselected, 2 rerun in 4.76s =================== 2025-12-04T11:17:06.1938745Z Got exit code 1 2025-12-04T11:17:06.1938851Z Retrying single test... 2025-12-04T11:17:06.1939296Z W1204 11:13:03.708000 96024 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.1939973Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-c31955170ca46735.xml 2025-12-04T11:17:06.1940170Z ============================= test session starts ============================== 2025-12-04T11:17:06.1940537Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.1940646Z cachedir: .pytest_cache 2025-12-04T11:17:06.1941170Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.1941303Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.1941409Z configfile: pytest.ini 2025-12-04T11:17:06.1941956Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.1942185Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.1943281Z stepcurrent: skipping 8 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1943496Z Running 1 items in this shard 2025-12-04T11:17:06.1943504Z 2025-12-04T11:17:06.1945154Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 [W1204 11:13:09.862792641 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1945161Z 2025-12-04T11:17:06.1945694Z [W1204 11:13:25.936675070 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1945701Z 2025-12-04T11:17:06.1946218Z [W1204 11:13:25.936938737 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1946228Z 2025-12-04T11:17:06.1946743Z [W1204 11:13:25.944372044 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1946763Z 2025-12-04T11:17:06.1947271Z [W1204 11:13:25.945090657 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1947276Z 2025-12-04T11:17:06.1947787Z [W1204 11:13:25.945284349 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1947792Z 2025-12-04T11:17:06.1948312Z [W1204 11:13:25.952293372 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1948319Z 2025-12-04T11:17:06.1948826Z [W1204 11:13:25.953067434 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1948831Z 2025-12-04T11:17:06.1949422Z [W1204 11:13:25.953255945 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1949430Z 2025-12-04T11:17:06.1949943Z [W1204 11:13:25.087630958 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1949947Z 2025-12-04T11:17:06.1950470Z [W1204 11:13:25.089372680 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1950475Z 2025-12-04T11:17:06.1950982Z [W1204 11:13:25.089584272 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1951018Z 2025-12-04T11:17:06.1951540Z [W1204 11:13:25.093607513 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1951545Z 2025-12-04T11:17:06.1952058Z [W1204 11:13:25.094246440 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1952108Z 2025-12-04T11:17:06.1952616Z [W1204 11:13:25.094441903 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1952632Z 2025-12-04T11:17:06.1953144Z [W1204 11:13:25.100518690 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1953149Z 2025-12-04T11:17:06.1953659Z [W1204 11:13:25.101154115 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1953663Z 2025-12-04T11:17:06.1954181Z [W1204 11:13:25.101348445 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1954186Z 2025-12-04T11:17:06.1954318Z ('RERUN', {'yellow': True}) [19.9467s] [100%] 2025-12-04T11:17:06.1955605Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 [W1204 11:13:25.504737896 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1955613Z 2025-12-04T11:17:06.1956126Z [W1204 11:13:25.505499806 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1956131Z 2025-12-04T11:17:06.1956661Z [W1204 11:13:25.505701166 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1956668Z 2025-12-04T11:17:06.1957174Z [W1204 11:13:25.509755321 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1957179Z 2025-12-04T11:17:06.1957704Z [W1204 11:13:25.510439107 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1957711Z 2025-12-04T11:17:06.1958222Z [W1204 11:13:25.510635352 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1958227Z 2025-12-04T11:17:06.1958733Z [W1204 11:13:25.516764111 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1958738Z 2025-12-04T11:17:06.1959257Z [W1204 11:13:25.517370497 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1959261Z 2025-12-04T11:17:06.1959775Z [W1204 11:13:25.517559187 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1959780Z 2025-12-04T11:17:06.1960358Z [W1204 11:13:25.604381750 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1960367Z 2025-12-04T11:17:06.1960874Z [W1204 11:13:25.605133955 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1960879Z 2025-12-04T11:17:06.1961403Z [W1204 11:13:25.605341493 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1961407Z 2025-12-04T11:17:06.1961914Z [W1204 11:13:25.609290513 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1961919Z 2025-12-04T11:17:06.1962440Z [W1204 11:13:25.609915065 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1962475Z 2025-12-04T11:17:06.1962990Z [W1204 11:13:25.610145787 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1963023Z 2025-12-04T11:17:06.1963529Z [W1204 11:13:25.616222316 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1963544Z 2025-12-04T11:17:06.1964056Z [W1204 11:13:25.617005617 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1964060Z 2025-12-04T11:17:06.1964565Z [W1204 11:13:25.617200073 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1964570Z 2025-12-04T11:17:06.1964710Z ('RERUN', {'yellow': True}) [0.4761s] [100%] 2025-12-04T11:17:06.1965996Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 [W1204 11:13:26.941006601 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1966004Z 2025-12-04T11:17:06.1966518Z [W1204 11:13:26.941746938 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1966523Z 2025-12-04T11:17:06.1967032Z [W1204 11:13:26.941949018 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1967036Z 2025-12-04T11:17:06.1967554Z [W1204 11:13:26.945997796 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1967559Z 2025-12-04T11:17:06.1968068Z [W1204 11:13:26.946615602 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1968075Z 2025-12-04T11:17:06.1968598Z [W1204 11:13:26.946807997 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1968604Z 2025-12-04T11:17:06.1969114Z [W1204 11:13:26.952945603 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1969119Z 2025-12-04T11:17:06.1969626Z [W1204 11:13:26.953558593 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1969643Z 2025-12-04T11:17:06.1970152Z [W1204 11:13:26.953744508 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1970157Z 2025-12-04T11:17:06.1970668Z [W1204 11:13:26.041796213 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1970674Z 2025-12-04T11:17:06.1971196Z [W1204 11:13:26.042556500 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1971261Z 2025-12-04T11:17:06.1971781Z [W1204 11:13:26.042763056 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1971788Z 2025-12-04T11:17:06.1972311Z [W1204 11:13:26.046736256 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1972316Z 2025-12-04T11:17:06.1972823Z [W1204 11:13:26.047363533 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1972829Z 2025-12-04T11:17:06.1973353Z [W1204 11:13:26.047560042 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1973388Z 2025-12-04T11:17:06.1974138Z [W1204 11:13:26.053656242 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1974143Z 2025-12-04T11:17:06.1974661Z [W1204 11:13:26.054453851 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1974720Z 2025-12-04T11:17:06.1975235Z [W1204 11:13:26.054651652 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.1975239Z 2025-12-04T11:17:06.1975345Z FAILED [0.4360s] [100%] 2025-12-04T11:17:06.1975349Z 2025-12-04T11:17:06.1975506Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.1976020Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.1976164Z Traceback (most recent call last): 2025-12-04T11:17:06.1976679Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1976983Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1977466Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1977634Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1978179Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1978403Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1978542Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1978547Z 2025-12-04T11:17:06.1978668Z Expected 1 but got 2. 2025-12-04T11:17:06.1978777Z Absolute difference: 1 2025-12-04T11:17:06.1978890Z Relative difference: 1.0 2025-12-04T11:17:06.1978897Z 2025-12-04T11:17:06.1979131Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1980040Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1980048Z 2025-12-04T11:17:06.1980333Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1980558Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1980675Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1981595Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1981827Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1981942Z graph_break [] 2025-12-04T11:17:06.1982162Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1983437Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1983571Z if out == self.unknown_value: 2025-12-04T11:17:06.1984304Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1984418Z warnings.warn( 2025-12-04T11:17:06.1985140Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1985241Z warnings.warn( 2025-12-04T11:17:06.1985764Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.1985918Z Traceback (most recent call last): 2025-12-04T11:17:06.1986433Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.1986710Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.1987165Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.1987345Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.1987885Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.1988092Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.1988233Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.1988238Z 2025-12-04T11:17:06.1988340Z Expected 1 but got 2. 2025-12-04T11:17:06.1988447Z Absolute difference: 1 2025-12-04T11:17:06.1988567Z Relative difference: 1.0 2025-12-04T11:17:06.1988572Z 2025-12-04T11:17:06.1988786Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.1989711Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.1989719Z 2025-12-04T11:17:06.1989985Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.1990204Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1990332Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1991228Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1991472Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1991571Z graph_break [] 2025-12-04T11:17:06.1991794Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1993009Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.1993130Z if out == self.unknown_value: 2025-12-04T11:17:06.1993879Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1993981Z warnings.warn( 2025-12-04T11:17:06.1994703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1994819Z warnings.warn( 2025-12-04T11:17:06.1995033Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.1995158Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.1995443Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.1996561Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.1996676Z graph_break [] 2025-12-04T11:17:06.1996892Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.1997616Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1997728Z warnings.warn( 2025-12-04T11:17:06.1998532Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.1998645Z warnings.warn( 2025-12-04T11:17:06.1998795Z =================================== FAILURES =================================== 2025-12-04T11:17:06.1999343Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.1999480Z Traceback (most recent call last): 2025-12-04T11:17:06.1999993Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2000237Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2000692Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2000858Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2001410Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2001616Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2001752Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2001760Z 2025-12-04T11:17:06.2001880Z Expected 1 but got 2. 2025-12-04T11:17:06.2001985Z Absolute difference: 1 2025-12-04T11:17:06.2002108Z Relative difference: 1.0 2025-12-04T11:17:06.2002114Z 2025-12-04T11:17:06.2002331Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2003241Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.2003247Z 2025-12-04T11:17:06.2003529Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2003755Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2003883Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2004781Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.2005011Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2005124Z graph_break [] 2025-12-04T11:17:06.2005339Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2006562Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.2006679Z if out == self.unknown_value: 2025-12-04T11:17:06.2007408Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2007524Z warnings.warn( 2025-12-04T11:17:06.2008324Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2008428Z warnings.warn( 2025-12-04T11:17:06.2008661Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2008774Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2009013Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2009909Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.2010006Z graph_break [] 2025-12-04T11:17:06.2010269Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2011002Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2011148Z warnings.warn( 2025-12-04T11:17:06.2011871Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2011971Z warnings.warn( 2025-12-04T11:17:06.2012201Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2012315Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2012540Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2013440Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.2013541Z graph_break [] 2025-12-04T11:17:06.2013767Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2014492Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2014592Z warnings.warn( 2025-12-04T11:17:06.2015326Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2015427Z warnings.warn( 2025-12-04T11:17:06.2016278Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-c31955170ca46735.xml - 2025-12-04T11:17:06.2016449Z =========================== short test summary info ============================ 2025-12-04T11:17:06.2017457Z FAILED [0.4360s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2017465Z 2025-12-04T11:17:06.2017588Z Expected 1 but got 2. 2025-12-04T11:17:06.2017695Z Absolute difference: 1 2025-12-04T11:17:06.2017818Z Relative difference: 1.0 2025-12-04T11:17:06.2017823Z 2025-12-04T11:17:06.2018039Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2018953Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.2018958Z 2025-12-04T11:17:06.2019242Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2019421Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.2019635Z ================== 1 failed, 10 deselected, 2 rerun in 20.89s ================== 2025-12-04T11:17:06.2019734Z Got exit code 1 2025-12-04T11:17:06.2019841Z Retrying single test... 2025-12-04T11:17:06.2020375Z W1204 11:13:37.792000 96198 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.2021046Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-1b66a9993689b300.xml 2025-12-04T11:17:06.2021209Z ============================= test session starts ============================== 2025-12-04T11:17:06.2021574Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.2021684Z cachedir: .pytest_cache 2025-12-04T11:17:06.2022222Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.2022380Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.2022487Z configfile: pytest.ini 2025-12-04T11:17:06.2023048Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.2023299Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.2024307Z stepcurrent: skipping 8 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.2024421Z Running 1 items in this shard 2025-12-04T11:17:06.2024426Z 2025-12-04T11:17:06.2025707Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 [W1204 11:13:43.938036232 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2025716Z 2025-12-04T11:17:06.2026245Z [W1204 11:13:58.451513555 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2026256Z 2025-12-04T11:17:06.2026769Z [W1204 11:13:58.451775734 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2026777Z 2025-12-04T11:17:06.2027298Z [W1204 11:13:58.459146667 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2027303Z 2025-12-04T11:17:06.2027816Z [W1204 11:13:58.459876724 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2027822Z 2025-12-04T11:17:06.2028342Z [W1204 11:13:58.460106710 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2028350Z 2025-12-04T11:17:06.2028860Z [W1204 11:13:58.466960250 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2028864Z 2025-12-04T11:17:06.2029391Z [W1204 11:13:58.467712819 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2029397Z 2025-12-04T11:17:06.2029910Z [W1204 11:13:58.467899571 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2029915Z 2025-12-04T11:17:06.2030423Z [W1204 11:13:58.602633347 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2030442Z 2025-12-04T11:17:06.2030953Z [W1204 11:13:58.604352372 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2030959Z 2025-12-04T11:17:06.2031472Z [W1204 11:13:58.604561950 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2031476Z 2025-12-04T11:17:06.2032069Z [W1204 11:13:58.608460606 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2032076Z 2025-12-04T11:17:06.2032588Z [W1204 11:13:58.609106544 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2032592Z 2025-12-04T11:17:06.2033114Z [W1204 11:13:58.609303772 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2033119Z 2025-12-04T11:17:06.2033629Z [W1204 11:13:58.615340791 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2033634Z 2025-12-04T11:17:06.2034190Z [W1204 11:13:58.615986606 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2034195Z 2025-12-04T11:17:06.2034706Z [W1204 11:13:58.616193769 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2034740Z 2025-12-04T11:17:06.2034888Z ('RERUN', {'yellow': True}) [19.3773s] [100%] 2025-12-04T11:17:06.2036165Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 [W1204 11:13:59.013370776 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2036171Z 2025-12-04T11:17:06.2036681Z [W1204 11:13:59.014119404 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2036698Z 2025-12-04T11:17:06.2037212Z [W1204 11:13:59.014319838 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2037217Z 2025-12-04T11:17:06.2037733Z [W1204 11:13:59.018300818 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2037740Z 2025-12-04T11:17:06.2038256Z [W1204 11:13:59.018921486 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2038260Z 2025-12-04T11:17:06.2038768Z [W1204 11:13:59.019111370 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2038773Z 2025-12-04T11:17:06.2039291Z [W1204 11:13:59.025284814 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2039296Z 2025-12-04T11:17:06.2039805Z [W1204 11:13:59.025899285 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2039812Z 2025-12-04T11:17:06.2040334Z [W1204 11:13:59.026089095 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2040341Z 2025-12-04T11:17:06.2040849Z [W1204 11:13:59.113359665 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2040854Z 2025-12-04T11:17:06.2041364Z [W1204 11:13:59.114100213 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2041380Z 2025-12-04T11:17:06.2041891Z [W1204 11:13:59.114302436 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2041896Z 2025-12-04T11:17:06.2042405Z [W1204 11:13:59.118197793 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2042411Z 2025-12-04T11:17:06.2043003Z [W1204 11:13:59.118828021 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2043010Z 2025-12-04T11:17:06.2043522Z [W1204 11:13:59.119020128 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2043527Z 2025-12-04T11:17:06.2044045Z [W1204 11:13:59.125084123 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2044051Z 2025-12-04T11:17:06.2044562Z [W1204 11:13:59.125882819 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2044567Z 2025-12-04T11:17:06.2045091Z [W1204 11:13:59.126080538 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2045127Z 2025-12-04T11:17:06.2045257Z ('RERUN', {'yellow': True}) [0.4702s] [100%] 2025-12-04T11:17:06.2046547Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 [W1204 11:13:59.456661787 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2046587Z 2025-12-04T11:17:06.2047096Z [W1204 11:13:59.457377736 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2047101Z 2025-12-04T11:17:06.2047610Z [W1204 11:13:59.457574586 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2047632Z 2025-12-04T11:17:06.2048140Z [W1204 11:13:59.461624136 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2048147Z 2025-12-04T11:17:06.2048662Z [W1204 11:13:59.462254641 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2048666Z 2025-12-04T11:17:06.2049193Z [W1204 11:13:59.462444127 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2049198Z 2025-12-04T11:17:06.2049709Z [W1204 11:13:59.468529822 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2049714Z 2025-12-04T11:17:06.2050235Z [W1204 11:13:59.469140451 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2050240Z 2025-12-04T11:17:06.2050749Z [W1204 11:13:59.469327774 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2050756Z 2025-12-04T11:17:06.2051278Z [W1204 11:13:59.556996523 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2051288Z 2025-12-04T11:17:06.2051800Z [W1204 11:13:59.557754200 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2051808Z 2025-12-04T11:17:06.2052329Z [W1204 11:13:59.557965515 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2052334Z 2025-12-04T11:17:06.2052841Z [W1204 11:13:59.561942341 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2052846Z 2025-12-04T11:17:06.2053362Z [W1204 11:13:59.562566715 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2053382Z 2025-12-04T11:17:06.2053892Z [W1204 11:13:59.562760798 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2053897Z 2025-12-04T11:17:06.2054464Z [W1204 11:13:59.568788105 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2054471Z 2025-12-04T11:17:06.2054993Z [W1204 11:13:59.569561514 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2054998Z 2025-12-04T11:17:06.2055511Z [W1204 11:13:59.569755621 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2055515Z 2025-12-04T11:17:06.2055632Z FAILED [0.4417s] [100%] 2025-12-04T11:17:06.2055637Z 2025-12-04T11:17:06.2055782Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.2056338Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.2056474Z Traceback (most recent call last): 2025-12-04T11:17:06.2057056Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2057345Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2057809Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2057976Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2058534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2058744Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2058890Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2058897Z 2025-12-04T11:17:06.2059003Z Expected 1 but got 2. 2025-12-04T11:17:06.2059110Z Absolute difference: 1 2025-12-04T11:17:06.2059238Z Relative difference: 1.0 2025-12-04T11:17:06.2059243Z 2025-12-04T11:17:06.2059462Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2060373Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.2060393Z 2025-12-04T11:17:06.2060661Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2060883Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2061010Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2061906Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.2062135Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2062245Z graph_break [] 2025-12-04T11:17:06.2062470Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2063713Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.2063829Z if out == self.unknown_value: 2025-12-04T11:17:06.2064555Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2064667Z warnings.warn( 2025-12-04T11:17:06.2065388Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2065501Z warnings.warn( 2025-12-04T11:17:06.2066073Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.2066198Z Traceback (most recent call last): 2025-12-04T11:17:06.2066721Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2066952Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2067410Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2067588Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2068126Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2068343Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2068507Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2068512Z 2025-12-04T11:17:06.2068618Z Expected 1 but got 2. 2025-12-04T11:17:06.2068738Z Absolute difference: 1 2025-12-04T11:17:06.2068850Z Relative difference: 1.0 2025-12-04T11:17:06.2068955Z 2025-12-04T11:17:06.2069185Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2070094Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.2070100Z 2025-12-04T11:17:06.2070369Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2070602Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2070722Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2071626Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.2071861Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2071964Z graph_break [] 2025-12-04T11:17:06.2072196Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2073414Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.2073543Z if out == self.unknown_value: 2025-12-04T11:17:06.2074276Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2074382Z warnings.warn( 2025-12-04T11:17:06.2075113Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2075218Z warnings.warn( 2025-12-04T11:17:06.2075438Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2075568Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2075796Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2076707Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.2076806Z graph_break [] 2025-12-04T11:17:06.2077023Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2077763Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2077865Z warnings.warn( 2025-12-04T11:17:06.2078659Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2078763Z warnings.warn( 2025-12-04T11:17:06.2078910Z =================================== FAILURES =================================== 2025-12-04T11:17:06.2079431Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 _ 2025-12-04T11:17:06.2079553Z Traceback (most recent call last): 2025-12-04T11:17:06.2080059Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2080306Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2080762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2080968Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2081508Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2081746Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2081890Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2081896Z 2025-12-04T11:17:06.2082003Z Expected 1 but got 2. 2025-12-04T11:17:06.2082110Z Absolute difference: 1 2025-12-04T11:17:06.2082230Z Relative difference: 1.0 2025-12-04T11:17:06.2082236Z 2025-12-04T11:17:06.2082450Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2083371Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.2083379Z 2025-12-04T11:17:06.2083646Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2083868Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2083993Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2085173Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.2085419Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2085520Z graph_break [] 2025-12-04T11:17:06.2085738Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2086968Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.2087089Z if out == self.unknown_value: 2025-12-04T11:17:06.2087836Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2087939Z warnings.warn( 2025-12-04T11:17:06.2088660Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2088775Z warnings.warn( 2025-12-04T11:17:06.2088995Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2089109Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2089348Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2090249Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.2090360Z graph_break [] 2025-12-04T11:17:06.2090678Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2091405Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2091519Z warnings.warn( 2025-12-04T11:17:06.2092236Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2092345Z warnings.warn( 2025-12-04T11:17:06.2092560Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2092675Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2092915Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2093848Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2), ('pad_mm_bench', 1)] 2025-12-04T11:17:06.2093990Z graph_break [] 2025-12-04T11:17:06.2094206Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2094935Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2095048Z warnings.warn( 2025-12-04T11:17:06.2095763Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2095861Z warnings.warn( 2025-12-04T11:17:06.2096927Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-1b66a9993689b300.xml - 2025-12-04T11:17:06.2097106Z =========================== short test summary info ============================ 2025-12-04T11:17:06.2098063Z FAILED [0.4417s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2098071Z 2025-12-04T11:17:06.2098178Z Expected 1 but got 2. 2025-12-04T11:17:06.2098284Z Absolute difference: 1 2025-12-04T11:17:06.2098407Z Relative difference: 1.0 2025-12-04T11:17:06.2098412Z 2025-12-04T11:17:06.2098632Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2099555Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.2099563Z 2025-12-04T11:17:06.2099827Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2100007Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.2100224Z ================== 1 failed, 10 deselected, 2 rerun in 20.32s ================== 2025-12-04T11:17:06.2100325Z Got exit code 1 2025-12-04T11:17:06.2101159Z FAILED CONSISTENTLY: test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16 2025-12-04T11:17:06.2101575Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:17:06.2102021Z W1204 11:14:11.123000 96372 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.2102694Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-79106310b2e4ab0d.xml 2025-12-04T11:17:06.2102860Z ============================= test session starts ============================== 2025-12-04T11:17:06.2103335Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.2103447Z cachedir: .pytest_cache 2025-12-04T11:17:06.2103971Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.2104110Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.2104220Z configfile: pytest.ini 2025-12-04T11:17:06.2104768Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.2104995Z collecting ... collected 58 items / 9 deselected / 49 selected 2025-12-04T11:17:06.2105138Z stepcurrent: skipping 9 already run items. 2025-12-04T11:17:06.2105267Z Running 2 items in this shard 2025-12-04T11:17:06.2105314Z 2025-12-04T11:17:06.2106197Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 ('RERUN', {'yellow': True}) [3.8185s] [ 50%] 2025-12-04T11:17:06.2107062Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 ('RERUN', {'yellow': True}) [0.4034s] [ 50%] 2025-12-04T11:17:06.2107897Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 FAILED [0.3963s] [ 50%] 2025-12-04T11:17:06.2107902Z 2025-12-04T11:17:06.2108044Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.2108563Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.2108690Z Traceback (most recent call last): 2025-12-04T11:17:06.2109201Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2109453Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2109913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2110093Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2110631Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2110840Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2110983Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2110989Z 2025-12-04T11:17:06.2111094Z Expected 1 but got 2. 2025-12-04T11:17:06.2111215Z Absolute difference: 1 2025-12-04T11:17:06.2111324Z Relative difference: 1.0 2025-12-04T11:17:06.2111331Z 2025-12-04T11:17:06.2111543Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2112474Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.2112484Z 2025-12-04T11:17:06.2112751Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2112985Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2113098Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2113632Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2113880Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2113982Z graph_break [] 2025-12-04T11:17:06.2114200Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2114951Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2115113Z warnings.warn( 2025-12-04T11:17:06.2115856Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2115958Z warnings.warn( 2025-12-04T11:17:06.2116469Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.2116605Z Traceback (most recent call last): 2025-12-04T11:17:06.2117118Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2117364Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2117852Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2118016Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2118573Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2118813Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2118944Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2118950Z 2025-12-04T11:17:06.2119071Z Expected 1 but got 2. 2025-12-04T11:17:06.2119180Z Absolute difference: 1 2025-12-04T11:17:06.2119306Z Relative difference: 1.0 2025-12-04T11:17:06.2119311Z 2025-12-04T11:17:06.2119527Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2120435Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.2120443Z 2025-12-04T11:17:06.2120727Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2120954Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2121085Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2121627Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2121859Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2121973Z graph_break [] 2025-12-04T11:17:06.2122194Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2122937Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2123055Z warnings.warn( 2025-12-04T11:17:06.2123784Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2123900Z warnings.warn( 2025-12-04T11:17:06.2124124Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2124241Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2124480Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2125017Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2125116Z graph_break [] 2025-12-04T11:17:06.2125346Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2126075Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2126189Z warnings.warn( 2025-12-04T11:17:06.2126976Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2127077Z warnings.warn( 2025-12-04T11:17:06.2127238Z =================================== FAILURES =================================== 2025-12-04T11:17:06.2127745Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.2127886Z Traceback (most recent call last): 2025-12-04T11:17:06.2128394Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2128625Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2129095Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2129295Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2129831Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2130053Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2130233Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2130238Z 2025-12-04T11:17:06.2130355Z Expected 1 but got 2. 2025-12-04T11:17:06.2130463Z Absolute difference: 1 2025-12-04T11:17:06.2130572Z Relative difference: 1.0 2025-12-04T11:17:06.2130577Z 2025-12-04T11:17:06.2130806Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2131713Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.2131719Z 2025-12-04T11:17:06.2131998Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2132221Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2132340Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2132893Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2133126Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2133226Z graph_break [] 2025-12-04T11:17:06.2133457Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2134194Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2134310Z warnings.warn( 2025-12-04T11:17:06.2135029Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2135135Z warnings.warn( 2025-12-04T11:17:06.2135362Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2135482Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2135725Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2136262Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2136362Z graph_break [] 2025-12-04T11:17:06.2136591Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2137390Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2137492Z warnings.warn( 2025-12-04T11:17:06.2138228Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2138332Z warnings.warn( 2025-12-04T11:17:06.2138629Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2138747Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2138976Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2139520Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2139618Z graph_break [] 2025-12-04T11:17:06.2139833Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2140573Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2140704Z warnings.warn( 2025-12-04T11:17:06.2141429Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2141534Z warnings.warn( 2025-12-04T11:17:06.2142375Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-79106310b2e4ab0d.xml - 2025-12-04T11:17:06.2142592Z =========================== short test summary info ============================ 2025-12-04T11:17:06.2143529Z FAILED [0.3963s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2143535Z 2025-12-04T11:17:06.2143655Z Expected 1 but got 2. 2025-12-04T11:17:06.2143763Z Absolute difference: 1 2025-12-04T11:17:06.2143873Z Relative difference: 1.0 2025-12-04T11:17:06.2143881Z 2025-12-04T11:17:06.2144113Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2145018Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.2145026Z 2025-12-04T11:17:06.2145306Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2145487Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.2145686Z =================== 1 failed, 9 deselected, 2 rerun in 4.65s =================== 2025-12-04T11:17:06.2145798Z Got exit code 1 2025-12-04T11:17:06.2145904Z Retrying single test... 2025-12-04T11:17:06.2146370Z W1204 11:14:31.308000 96541 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.2147041Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-235ce3baacba0fb6.xml 2025-12-04T11:17:06.2147205Z ============================= test session starts ============================== 2025-12-04T11:17:06.2147573Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.2147687Z cachedir: .pytest_cache 2025-12-04T11:17:06.2148211Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.2148349Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.2148457Z configfile: pytest.ini 2025-12-04T11:17:06.2149012Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.2149232Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.2150222Z stepcurrent: skipping 9 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.2150350Z Running 1 items in this shard 2025-12-04T11:17:06.2150356Z 2025-12-04T11:17:06.2151681Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 [W1204 11:14:34.521630933 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2151690Z 2025-12-04T11:17:06.2152224Z [W1204 11:14:49.688775620 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2152229Z 2025-12-04T11:17:06.2152738Z [W1204 11:14:49.689038375 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2152743Z 2025-12-04T11:17:06.2153297Z [W1204 11:14:49.696407485 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2153302Z 2025-12-04T11:17:06.2153817Z [W1204 11:14:49.697090167 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2153852Z 2025-12-04T11:17:06.2154375Z [W1204 11:14:49.697275707 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2154379Z 2025-12-04T11:17:06.2154889Z [W1204 11:14:50.704155332 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2154895Z 2025-12-04T11:17:06.2155405Z [W1204 11:14:50.704791635 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2155421Z 2025-12-04T11:17:06.2155929Z [W1204 11:14:50.704974572 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2155936Z 2025-12-04T11:17:06.2156447Z [W1204 11:14:51.699332786 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2156453Z 2025-12-04T11:17:06.2156977Z [W1204 11:14:52.701042822 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2156982Z 2025-12-04T11:17:06.2157490Z [W1204 11:14:52.701255063 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2157495Z 2025-12-04T11:17:06.2158017Z [W1204 11:14:52.705178401 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2158021Z 2025-12-04T11:17:06.2158533Z [W1204 11:14:52.705827769 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2158540Z 2025-12-04T11:17:06.2159064Z [W1204 11:14:52.706022169 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2159071Z 2025-12-04T11:17:06.2159580Z [W1204 11:14:52.712079220 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2159585Z 2025-12-04T11:17:06.2160106Z [W1204 11:14:52.712714555 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2160111Z 2025-12-04T11:17:06.2160619Z [W1204 11:14:52.712907600 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2160624Z 2025-12-04T11:17:06.2160758Z ('RERUN', {'yellow': True}) [18.9534s] [100%] 2025-12-04T11:17:06.2162105Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 [W1204 11:14:52.061301397 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2162113Z 2025-12-04T11:17:06.2162621Z [W1204 11:14:52.062021662 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2162626Z 2025-12-04T11:17:06.2163147Z [W1204 11:14:52.062219572 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2163152Z 2025-12-04T11:17:06.2163662Z [W1204 11:14:52.066132757 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2163667Z 2025-12-04T11:17:06.2164182Z [W1204 11:14:52.066889708 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2164216Z 2025-12-04T11:17:06.2164732Z [W1204 11:14:52.067078746 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2164737Z 2025-12-04T11:17:06.2165287Z [W1204 11:14:52.073063935 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2165292Z 2025-12-04T11:17:06.2165802Z [W1204 11:14:52.073676651 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2165807Z 2025-12-04T11:17:06.2166316Z [W1204 11:14:52.073863822 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2166332Z 2025-12-04T11:17:06.2166841Z [W1204 11:14:52.159371129 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2166849Z 2025-12-04T11:17:06.2167359Z [W1204 11:14:52.160119921 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2167368Z 2025-12-04T11:17:06.2167888Z [W1204 11:14:52.160325168 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2167894Z 2025-12-04T11:17:06.2168404Z [W1204 11:14:52.164212963 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2168409Z 2025-12-04T11:17:06.2168932Z [W1204 11:14:52.164824955 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2168937Z 2025-12-04T11:17:06.2169445Z [W1204 11:14:52.165018078 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2169452Z 2025-12-04T11:17:06.2169972Z [W1204 11:14:52.170990398 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2169977Z 2025-12-04T11:17:06.2170489Z [W1204 11:14:52.171765849 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2170496Z 2025-12-04T11:17:06.2171019Z [W1204 11:14:52.171958989 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2171023Z 2025-12-04T11:17:06.2171156Z ('RERUN', {'yellow': True}) [0.4196s] [100%] 2025-12-04T11:17:06.2172429Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 [W1204 11:14:52.462746413 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2172436Z 2025-12-04T11:17:06.2172963Z [W1204 11:14:52.463496450 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2172968Z 2025-12-04T11:17:06.2173539Z [W1204 11:14:52.463698272 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2173546Z 2025-12-04T11:17:06.2174073Z [W1204 11:14:52.467621878 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2174078Z 2025-12-04T11:17:06.2174590Z [W1204 11:14:52.468410612 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2174595Z 2025-12-04T11:17:06.2175118Z [W1204 11:14:52.468604459 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2175171Z 2025-12-04T11:17:06.2175683Z [W1204 11:14:52.474616371 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2175688Z 2025-12-04T11:17:06.2176216Z [W1204 11:14:52.475234438 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2176251Z 2025-12-04T11:17:06.2176763Z [W1204 11:14:52.475421958 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2176768Z 2025-12-04T11:17:06.2177370Z [W1204 11:14:52.562098639 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2177388Z 2025-12-04T11:17:06.2177903Z [W1204 11:14:52.562844392 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2177910Z 2025-12-04T11:17:06.2178424Z [W1204 11:14:52.563048469 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2178430Z 2025-12-04T11:17:06.2178957Z [W1204 11:14:52.566940758 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2178964Z 2025-12-04T11:17:06.2179472Z [W1204 11:14:52.567561194 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2179476Z 2025-12-04T11:17:06.2180002Z [W1204 11:14:52.567756854 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2180007Z 2025-12-04T11:17:06.2180516Z [W1204 11:14:52.573775829 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2180521Z 2025-12-04T11:17:06.2181049Z [W1204 11:14:52.574557939 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2181054Z 2025-12-04T11:17:06.2181568Z [W1204 11:14:52.574750752 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2181575Z 2025-12-04T11:17:06.2181690Z FAILED [0.4015s] [100%] 2025-12-04T11:17:06.2181695Z 2025-12-04T11:17:06.2181840Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.2182357Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.2182495Z Traceback (most recent call last): 2025-12-04T11:17:06.2183008Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2191318Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2191834Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2192020Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2192727Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2192947Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2193100Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2193108Z 2025-12-04T11:17:06.2193218Z Expected 1 but got 2. 2025-12-04T11:17:06.2193327Z Absolute difference: 1 2025-12-04T11:17:06.2193457Z Relative difference: 1.0 2025-12-04T11:17:06.2193462Z 2025-12-04T11:17:06.2193681Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2194605Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.2194652Z 2025-12-04T11:17:06.2194925Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2195160Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2195336Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2195877Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2196312Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2196415Z graph_break [] 2025-12-04T11:17:06.2196641Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2197872Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.2197995Z if out == self.unknown_value: 2025-12-04T11:17:06.2198742Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2198849Z warnings.warn( 2025-12-04T11:17:06.2199574Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2199692Z warnings.warn( 2025-12-04T11:17:06.2200200Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.2200327Z Traceback (most recent call last): 2025-12-04T11:17:06.2200852Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2201089Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2201568Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2201737Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2202275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2202497Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2202632Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2202638Z 2025-12-04T11:17:06.2202761Z Expected 1 but got 2. 2025-12-04T11:17:06.2202868Z Absolute difference: 1 2025-12-04T11:17:06.2202979Z Relative difference: 1.0 2025-12-04T11:17:06.2202985Z 2025-12-04T11:17:06.2203215Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2204125Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.2204134Z 2025-12-04T11:17:06.2204420Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2204751Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2204873Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2205427Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2205657Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2205759Z graph_break [] 2025-12-04T11:17:06.2205995Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2207211Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.2207382Z if out == self.unknown_value: 2025-12-04T11:17:06.2208116Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2208284Z warnings.warn( 2025-12-04T11:17:06.2209032Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2209135Z warnings.warn( 2025-12-04T11:17:06.2209370Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2209485Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2209714Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2210262Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2210363Z graph_break [] 2025-12-04T11:17:06.2210582Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2211323Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2211428Z warnings.warn( 2025-12-04T11:17:06.2212162Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2212262Z warnings.warn( 2025-12-04T11:17:06.2212410Z =================================== FAILURES =================================== 2025-12-04T11:17:06.2212937Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.2213062Z Traceback (most recent call last): 2025-12-04T11:17:06.2213586Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2213826Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2214290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2214470Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2215008Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2215223Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2215368Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2215374Z 2025-12-04T11:17:06.2215480Z Expected 1 but got 2. 2025-12-04T11:17:06.2215601Z Absolute difference: 1 2025-12-04T11:17:06.2215711Z Relative difference: 1.0 2025-12-04T11:17:06.2215719Z 2025-12-04T11:17:06.2215932Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2217021Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.2217035Z 2025-12-04T11:17:06.2217314Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2217550Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2217665Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2218201Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2218444Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2218542Z graph_break [] 2025-12-04T11:17:06.2218760Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2220015Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.2220170Z if out == self.unknown_value: 2025-12-04T11:17:06.2220910Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2221013Z warnings.warn( 2025-12-04T11:17:06.2221738Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2221854Z warnings.warn( 2025-12-04T11:17:06.2222075Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2222205Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2222438Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2222973Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2223089Z graph_break [] 2025-12-04T11:17:06.2223305Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2224032Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2224152Z warnings.warn( 2025-12-04T11:17:06.2224871Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2224984Z warnings.warn( 2025-12-04T11:17:06.2225200Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2225319Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2225560Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2226090Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2226199Z graph_break [] 2025-12-04T11:17:06.2226412Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2227137Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2227251Z warnings.warn( 2025-12-04T11:17:06.2227966Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2228062Z warnings.warn( 2025-12-04T11:17:06.2228927Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-235ce3baacba0fb6.xml - 2025-12-04T11:17:06.2229175Z =========================== short test summary info ============================ 2025-12-04T11:17:06.2230138Z FAILED [0.4015s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2230144Z 2025-12-04T11:17:06.2230252Z Expected 1 but got 2. 2025-12-04T11:17:06.2230359Z Absolute difference: 1 2025-12-04T11:17:06.2230480Z Relative difference: 1.0 2025-12-04T11:17:06.2230486Z 2025-12-04T11:17:06.2230708Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2231630Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.2231667Z 2025-12-04T11:17:06.2231937Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2232123Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.2232371Z ================== 1 failed, 10 deselected, 2 rerun in 19.81s ================== 2025-12-04T11:17:06.2232471Z Got exit code 1 2025-12-04T11:17:06.2232591Z Retrying single test... 2025-12-04T11:17:06.2233058Z W1204 11:15:04.202000 96715 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.2234141Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-00aa73936a934045.xml 2025-12-04T11:17:06.2235744Z ============================= test session starts ============================== 2025-12-04T11:17:06.2236704Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.2237753Z cachedir: .pytest_cache 2025-12-04T11:17:06.2239063Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.2240571Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.2241227Z configfile: pytest.ini 2025-12-04T11:17:06.2242560Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.2244275Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.2246828Z stepcurrent: skipping 9 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.2248311Z Running 1 items in this shard 2025-12-04T11:17:06.2248529Z 2025-12-04T11:17:06.2249835Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 [W1204 11:15:07.426682937 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2251280Z 2025-12-04T11:17:06.2251811Z [W1204 11:15:22.602594573 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2252482Z 2025-12-04T11:17:06.2252998Z [W1204 11:15:22.602860697 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2253662Z 2025-12-04T11:17:06.2254175Z [W1204 11:15:22.610360069 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2254825Z 2025-12-04T11:17:06.2255353Z [W1204 11:15:22.611069642 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2256001Z 2025-12-04T11:17:06.2256663Z [W1204 11:15:22.611262445 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2257398Z 2025-12-04T11:17:06.2257916Z [W1204 11:15:22.618182130 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2258577Z 2025-12-04T11:17:06.2259088Z [W1204 11:15:22.618823008 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2259750Z 2025-12-04T11:17:06.2260265Z [W1204 11:15:22.619006946 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2260917Z 2025-12-04T11:17:06.2261447Z [W1204 11:15:24.613289769 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2262138Z 2025-12-04T11:17:06.2262665Z [W1204 11:15:24.614996297 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2263317Z 2025-12-04T11:17:06.2263828Z [W1204 11:15:24.615211734 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2264531Z 2025-12-04T11:17:06.2265045Z [W1204 11:15:24.619140248 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2265716Z 2025-12-04T11:17:06.2266234Z [W1204 11:15:24.619780817 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2266879Z 2025-12-04T11:17:06.2267406Z [W1204 11:15:24.619980859 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2268052Z 2025-12-04T11:17:06.2268583Z [W1204 11:15:24.626077219 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2269235Z 2025-12-04T11:17:06.2269749Z [W1204 11:15:24.626717664 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2270415Z 2025-12-04T11:17:06.2270928Z [W1204 11:15:24.626909640 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2271584Z 2025-12-04T11:17:06.2271719Z ('RERUN', {'yellow': True}) [18.9783s] [100%] 2025-12-04T11:17:06.2273252Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 [W1204 11:15:25.986265065 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2274658Z 2025-12-04T11:17:06.2275185Z [W1204 11:15:25.987025502 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2275838Z 2025-12-04T11:17:06.2276359Z [W1204 11:15:25.987226298 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2277018Z 2025-12-04T11:17:06.2277532Z [W1204 11:15:25.991224558 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2278193Z 2025-12-04T11:17:06.2278700Z [W1204 11:15:25.992010968 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2279361Z 2025-12-04T11:17:06.2280211Z [W1204 11:15:25.992212803 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2280860Z 2025-12-04T11:17:06.2281388Z [W1204 11:15:25.998241908 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2282030Z 2025-12-04T11:17:06.2282643Z [W1204 11:15:25.998861986 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2283297Z 2025-12-04T11:17:06.2283810Z [W1204 11:15:25.999050898 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2284473Z 2025-12-04T11:17:06.2284986Z [W1204 11:15:25.086316184 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2285648Z 2025-12-04T11:17:06.2286164Z [W1204 11:15:25.087089030 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2286842Z 2025-12-04T11:17:06.2287368Z [W1204 11:15:25.087295642 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2288012Z 2025-12-04T11:17:06.2288540Z [W1204 11:15:25.091296115 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2289223Z 2025-12-04T11:17:06.2289740Z [W1204 11:15:25.091936377 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2290398Z 2025-12-04T11:17:06.2290910Z [W1204 11:15:25.092143479 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2291767Z 2025-12-04T11:17:06.2292653Z [W1204 11:15:25.098186657 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2293302Z 2025-12-04T11:17:06.2293835Z [W1204 11:15:25.098974083 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2294479Z 2025-12-04T11:17:06.2295005Z [W1204 11:15:25.099168180 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2295651Z 2025-12-04T11:17:06.2295782Z ('RERUN', {'yellow': True}) [0.4334s] [100%] 2025-12-04T11:17:06.2297525Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 [W1204 11:15:25.388823233 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2298954Z 2025-12-04T11:17:06.2299471Z [W1204 11:15:25.389583651 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2300121Z 2025-12-04T11:17:06.2300659Z [W1204 11:15:25.389786977 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2301308Z 2025-12-04T11:17:06.2301838Z [W1204 11:15:25.393780515 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2302492Z 2025-12-04T11:17:06.2303004Z [W1204 11:15:25.394561726 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2303659Z 2025-12-04T11:17:06.2304172Z [W1204 11:15:25.394752878 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2304836Z 2025-12-04T11:17:06.2305347Z [W1204 11:15:25.400800973 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2306010Z 2025-12-04T11:17:06.2306520Z [W1204 11:15:25.401419661 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2307166Z 2025-12-04T11:17:06.2307810Z [W1204 11:15:25.401608160 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2308465Z 2025-12-04T11:17:06.2308995Z [W1204 11:15:25.488127672 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2309641Z 2025-12-04T11:17:06.2310153Z [W1204 11:15:25.488894719 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2310812Z 2025-12-04T11:17:06.2311326Z [W1204 11:15:25.489100226 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2311984Z 2025-12-04T11:17:06.2312495Z [W1204 11:15:25.493071468 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2313268Z 2025-12-04T11:17:06.2313799Z [W1204 11:15:25.493706564 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2314497Z 2025-12-04T11:17:06.2315025Z [W1204 11:15:25.493900936 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2315670Z 2025-12-04T11:17:06.2316185Z [W1204 11:15:25.499946117 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2316839Z 2025-12-04T11:17:06.2317349Z [W1204 11:15:25.500762103 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2318008Z 2025-12-04T11:17:06.2318520Z [W1204 11:15:25.500960989 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2319172Z 2025-12-04T11:17:06.2319289Z FAILED [0.3996s] [100%] 2025-12-04T11:17:06.2319464Z 2025-12-04T11:17:06.2319613Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.2320419Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.2321193Z Traceback (most recent call last): 2025-12-04T11:17:06.2322187Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2323064Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2323906Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2324671Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2325515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2326388Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2326871Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2327124Z 2025-12-04T11:17:06.2327247Z Expected 1 but got 2. 2025-12-04T11:17:06.2327526Z Absolute difference: 1 2025-12-04T11:17:06.2327823Z Relative difference: 1.0 2025-12-04T11:17:06.2328014Z 2025-12-04T11:17:06.2328244Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2329519Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.2330562Z 2025-12-04T11:17:06.2330833Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2331466Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2331960Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2332818Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2333715Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2334194Z graph_break [] 2025-12-04T11:17:06.2334575Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2336146Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.2337684Z if out == self.unknown_value: 2025-12-04T11:17:06.2338648Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2339666Z warnings.warn( 2025-12-04T11:17:06.2340543Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2341550Z warnings.warn( 2025-12-04T11:17:06.2342222Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.2342993Z Traceback (most recent call last): 2025-12-04T11:17:06.2343736Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2344620Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2345451Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2346222Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2347048Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2347934Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2348412Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2348664Z 2025-12-04T11:17:06.2348772Z Expected 1 but got 2. 2025-12-04T11:17:06.2349060Z Absolute difference: 1 2025-12-04T11:17:06.2349350Z Relative difference: 1.0 2025-12-04T11:17:06.2349539Z 2025-12-04T11:17:06.2349765Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2351013Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.2352060Z 2025-12-04T11:17:06.2352329Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2352964Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2353445Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2354191Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2355098Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2355568Z graph_break [] 2025-12-04T11:17:06.2355933Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2357514Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.2358969Z if out == self.unknown_value: 2025-12-04T11:17:06.2359905Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2360871Z warnings.warn( 2025-12-04T11:17:06.2361826Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2362787Z warnings.warn( 2025-12-04T11:17:06.2363169Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2363632Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2364074Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2364971Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2365723Z graph_break [] 2025-12-04T11:17:06.2366096Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2367212Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2368174Z warnings.warn( 2025-12-04T11:17:06.2369055Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2370053Z warnings.warn( 2025-12-04T11:17:06.2370362Z =================================== FAILURES =================================== 2025-12-04T11:17:06.2371149Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 _ 2025-12-04T11:17:06.2371911Z Traceback (most recent call last): 2025-12-04T11:17:06.2372654Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2373533Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2374339Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2375096Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2375926Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2376805Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2377338Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2377597Z 2025-12-04T11:17:06.2377702Z Expected 1 but got 2. 2025-12-04T11:17:06.2377987Z Absolute difference: 1 2025-12-04T11:17:06.2378270Z Relative difference: 1.0 2025-12-04T11:17:06.2378465Z 2025-12-04T11:17:06.2378682Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2379952Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.2380993Z 2025-12-04T11:17:06.2381271Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2381895Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2382374Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2383120Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2384019Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2384476Z graph_break [] 2025-12-04T11:17:06.2384847Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2386427Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.2387875Z if out == self.unknown_value: 2025-12-04T11:17:06.2388882Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2389852Z warnings.warn( 2025-12-04T11:17:06.2390737Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2391681Z warnings.warn( 2025-12-04T11:17:06.2392066Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2392543Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2392990Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2393878Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2394683Z graph_break [] 2025-12-04T11:17:06.2395049Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2396339Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2397372Z warnings.warn( 2025-12-04T11:17:06.2398263Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2399233Z warnings.warn( 2025-12-04T11:17:06.2399598Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2400072Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2400515Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2401397Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2402173Z graph_break [] 2025-12-04T11:17:06.2402545Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2403629Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2404580Z warnings.warn( 2025-12-04T11:17:06.2405458Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2406411Z warnings.warn( 2025-12-04T11:17:06.2407415Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-00aa73936a934045.xml - 2025-12-04T11:17:06.2408543Z =========================== short test summary info ============================ 2025-12-04T11:17:06.2409802Z FAILED [0.3996s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2410879Z 2025-12-04T11:17:06.2410988Z Expected 1 but got 2. 2025-12-04T11:17:06.2411274Z Absolute difference: 1 2025-12-04T11:17:06.2411553Z Relative difference: 1.0 2025-12-04T11:17:06.2411751Z 2025-12-04T11:17:06.2411963Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2413222Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.2414258Z 2025-12-04T11:17:06.2414533Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2415114Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.2415641Z ================== 1 failed, 10 deselected, 2 rerun in 19.84s ================== 2025-12-04T11:17:06.2416088Z Got exit code 1 2025-12-04T11:17:06.2417222Z FAILED CONSISTENTLY: test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16 2025-12-04T11:17:06.2418613Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:17:06.2419625Z W1204 11:15:36.946000 96889 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.2420882Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-333e44976f8ee863.xml 2025-12-04T11:17:06.2421843Z ============================= test session starts ============================== 2025-12-04T11:17:06.2422558Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.2423169Z cachedir: .pytest_cache 2025-12-04T11:17:06.2424159Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.2425009Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.2425371Z configfile: pytest.ini 2025-12-04T11:17:06.2426112Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.2427020Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.2427510Z stepcurrent: skipping 10 already run items. 2025-12-04T11:17:06.2427906Z Running 1 items in this shard 2025-12-04T11:17:06.2428118Z 2025-12-04T11:17:06.2429014Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 ('RERUN', {'yellow': True}) [3.9009s] [100%] 2025-12-04T11:17:06.2430910Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 ('RERUN', {'yellow': True}) [0.4906s] [100%] 2025-12-04T11:17:06.2432711Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 FAILED [0.4873s] [100%] 2025-12-04T11:17:06.2433646Z 2025-12-04T11:17:06.2433791Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.2434605Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.2435412Z Traceback (most recent call last): 2025-12-04T11:17:06.2436151Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2437034Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2437869Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2438628Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2439468Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2440348Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2440827Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2441076Z 2025-12-04T11:17:06.2441181Z Expected 1 but got 2. 2025-12-04T11:17:06.2441470Z Absolute difference: 1 2025-12-04T11:17:06.2441765Z Relative difference: 1.0 2025-12-04T11:17:06.2441958Z 2025-12-04T11:17:06.2442175Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2443448Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.2444516Z 2025-12-04T11:17:06.2444858Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2445503Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2445971Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2446732Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2447641Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2448117Z graph_break [] 2025-12-04T11:17:06.2448484Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2449588Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2450607Z warnings.warn( 2025-12-04T11:17:06.2451502Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2452490Z warnings.warn( 2025-12-04T11:17:06.2453170Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.2453954Z Traceback (most recent call last): 2025-12-04T11:17:06.2454697Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2455583Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2456413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2457261Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2458098Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2458990Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2459479Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2459731Z 2025-12-04T11:17:06.2459853Z Expected 1 but got 2. 2025-12-04T11:17:06.2460139Z Absolute difference: 1 2025-12-04T11:17:06.2460439Z Relative difference: 1.0 2025-12-04T11:17:06.2460630Z 2025-12-04T11:17:06.2460859Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2462113Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.2463176Z 2025-12-04T11:17:06.2463445Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2464077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2464554Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2465296Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2466205Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2466676Z graph_break [] 2025-12-04T11:17:06.2467037Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2468134Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2469104Z warnings.warn( 2025-12-04T11:17:06.2469988Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2470945Z warnings.warn( 2025-12-04T11:17:06.2471327Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2471805Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2472314Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2473223Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2473999Z graph_break [] 2025-12-04T11:17:06.2474380Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2475464Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2476432Z warnings.warn( 2025-12-04T11:17:06.2477325Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2478332Z warnings.warn( 2025-12-04T11:17:06.2478634Z =================================== FAILURES =================================== 2025-12-04T11:17:06.2479446Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.2480262Z Traceback (most recent call last): 2025-12-04T11:17:06.2481004Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2481895Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2482724Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2483491Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2484321Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2485212Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2485685Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2485940Z 2025-12-04T11:17:06.2486062Z Expected 1 but got 2. 2025-12-04T11:17:06.2486342Z Absolute difference: 1 2025-12-04T11:17:06.2486639Z Relative difference: 1.0 2025-12-04T11:17:06.2486829Z 2025-12-04T11:17:06.2487057Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2488323Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.2489382Z 2025-12-04T11:17:06.2489651Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2490281Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2490771Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2491511Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2492418Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2492895Z graph_break [] 2025-12-04T11:17:06.2493271Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2494361Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2495336Z warnings.warn( 2025-12-04T11:17:06.2496408Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2497426Z warnings.warn( 2025-12-04T11:17:06.2497816Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2498300Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2498749Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2499758Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2500539Z graph_break [] 2025-12-04T11:17:06.2500918Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2501999Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2502971Z warnings.warn( 2025-12-04T11:17:06.2503868Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2504869Z warnings.warn( 2025-12-04T11:17:06.2505242Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2505722Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2506174Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2507118Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2507874Z graph_break [] 2025-12-04T11:17:06.2508245Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2509337Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2510294Z warnings.warn( 2025-12-04T11:17:06.2511182Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2512155Z warnings.warn( 2025-12-04T11:17:06.2513172Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-333e44976f8ee863.xml - 2025-12-04T11:17:06.2514307Z =========================== short test summary info ============================ 2025-12-04T11:17:06.2515578Z FAILED [0.4873s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2516659Z 2025-12-04T11:17:06.2516769Z Expected 1 but got 2. 2025-12-04T11:17:06.2517063Z Absolute difference: 1 2025-12-04T11:17:06.2517345Z Relative difference: 1.0 2025-12-04T11:17:06.2517545Z 2025-12-04T11:17:06.2517761Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2519036Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.2520078Z 2025-12-04T11:17:06.2520362Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2520942Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.2521464Z ================== 1 failed, 10 deselected, 2 rerun in 4.91s =================== 2025-12-04T11:17:06.2521903Z Got exit code 1 2025-12-04T11:17:06.2522159Z Retrying single test... 2025-12-04T11:17:06.2522798Z W1204 11:15:56.897000 97065 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.2524044Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-acd69551823a4dec.xml 2025-12-04T11:17:06.2525006Z ============================= test session starts ============================== 2025-12-04T11:17:06.2525654Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.2526255Z cachedir: .pytest_cache 2025-12-04T11:17:06.2527033Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.2527828Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.2528168Z configfile: pytest.ini 2025-12-04T11:17:06.2528901Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.2529804Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.2531149Z stepcurrent: skipping 10 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.2532437Z Running 1 items in this shard 2025-12-04T11:17:06.2532660Z 2025-12-04T11:17:06.2533958Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 [W1204 11:16:00.207868660 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2535423Z 2025-12-04T11:17:06.2535959Z [W1204 11:16:15.393215396 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2536609Z 2025-12-04T11:17:06.2537203Z [W1204 11:16:15.393475560 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2537854Z 2025-12-04T11:17:06.2538370Z [W1204 11:16:15.400833764 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2539031Z 2025-12-04T11:17:06.2539542Z [W1204 11:16:15.401566372 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2540205Z 2025-12-04T11:17:06.2540716Z [W1204 11:16:15.401757582 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2541376Z 2025-12-04T11:17:06.2541887Z [W1204 11:16:15.408810673 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2542530Z 2025-12-04T11:17:06.2543052Z [W1204 11:16:15.409521655 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2543704Z 2025-12-04T11:17:06.2544228Z [W1204 11:16:15.409710074 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2544874Z 2025-12-04T11:17:06.2545387Z [W1204 11:16:17.411932278 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2546039Z 2025-12-04T11:17:06.2546556Z [W1204 11:16:17.413602219 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2547221Z 2025-12-04T11:17:06.2547730Z [W1204 11:16:17.413813032 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2548376Z 2025-12-04T11:17:06.2548897Z [W1204 11:16:17.417704739 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2549543Z 2025-12-04T11:17:06.2550068Z [W1204 11:16:17.418322133 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2550710Z 2025-12-04T11:17:06.2551227Z [W1204 11:16:17.418518264 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2551885Z 2025-12-04T11:17:06.2552458Z [W1204 11:16:17.424609415 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2553124Z 2025-12-04T11:17:06.2553630Z [W1204 11:16:17.425227140 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2554284Z 2025-12-04T11:17:06.2554806Z [W1204 11:16:17.425425451 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2555454Z 2025-12-04T11:17:06.2555608Z ('RERUN', {'yellow': True}) [19.0794s] [100%] 2025-12-04T11:17:06.2557143Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 [W1204 11:16:18.850491672 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2558600Z 2025-12-04T11:17:06.2559117Z [W1204 11:16:18.851192605 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2559806Z 2025-12-04T11:17:06.2560320Z [W1204 11:16:18.851394928 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2560966Z 2025-12-04T11:17:06.2561491Z [W1204 11:16:18.855277194 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2562136Z 2025-12-04T11:17:06.2562659Z [W1204 11:16:18.856021600 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2563309Z 2025-12-04T11:17:06.2563825Z [W1204 11:16:18.856225243 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2564482Z 2025-12-04T11:17:06.2564999Z [W1204 11:16:18.862204311 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2565658Z 2025-12-04T11:17:06.2566169Z [W1204 11:16:18.862808785 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2566814Z 2025-12-04T11:17:06.2567332Z [W1204 11:16:18.862996882 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2567979Z 2025-12-04T11:17:06.2568503Z [W1204 11:16:18.946472756 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2569151Z 2025-12-04T11:17:06.2569661Z [W1204 11:16:18.947149778 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2570320Z 2025-12-04T11:17:06.2570832Z [W1204 11:16:18.947350325 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2571492Z 2025-12-04T11:17:06.2572004Z [W1204 11:16:18.951242978 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2572660Z 2025-12-04T11:17:06.2573171Z [W1204 11:16:18.951864968 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2573817Z 2025-12-04T11:17:06.2574342Z [W1204 11:16:18.952057171 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2574991Z 2025-12-04T11:17:06.2575511Z [W1204 11:16:18.958011512 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2576162Z 2025-12-04T11:17:06.2576732Z [W1204 11:16:18.958776847 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2577451Z 2025-12-04T11:17:06.2577962Z [W1204 11:16:18.958970143 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2578624Z 2025-12-04T11:17:06.2578757Z ('RERUN', {'yellow': True}) [0.4941s] [100%] 2025-12-04T11:17:06.2580318Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 [W1204 11:16:18.320422120 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2581714Z 2025-12-04T11:17:06.2582244Z [W1204 11:16:18.321136169 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2582931Z 2025-12-04T11:17:06.2583451Z [W1204 11:16:18.321339731 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2584115Z 2025-12-04T11:17:06.2584656Z [W1204 11:16:18.325242825 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2585315Z 2025-12-04T11:17:06.2585827Z [W1204 11:16:18.326002466 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2586476Z 2025-12-04T11:17:06.2587001Z [W1204 11:16:18.326196372 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2587651Z 2025-12-04T11:17:06.2588176Z [W1204 11:16:18.332218689 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2588828Z 2025-12-04T11:17:06.2589341Z [W1204 11:16:18.332835682 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2590008Z 2025-12-04T11:17:06.2590525Z [W1204 11:16:18.333023779 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2591188Z 2025-12-04T11:17:06.2591700Z [W1204 11:16:18.417613452 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2592352Z 2025-12-04T11:17:06.2592882Z [W1204 11:16:18.418312541 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2593525Z 2025-12-04T11:17:06.2594052Z [W1204 11:16:18.418513106 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2594702Z 2025-12-04T11:17:06.2595217Z [W1204 11:16:18.422397980 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2595876Z 2025-12-04T11:17:06.2596542Z [W1204 11:16:18.423023517 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2597207Z 2025-12-04T11:17:06.2597719Z [W1204 11:16:18.423217799 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2598369Z 2025-12-04T11:17:06.2598895Z [W1204 11:16:18.429202289 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2599542Z 2025-12-04T11:17:06.2600067Z [W1204 11:16:18.429973813 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2600720Z 2025-12-04T11:17:06.2601234Z [W1204 11:16:18.430187832 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2601901Z 2025-12-04T11:17:06.2602096Z FAILED [0.4697s] [100%] 2025-12-04T11:17:06.2602289Z 2025-12-04T11:17:06.2602439Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.2603248Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.2604012Z Traceback (most recent call last): 2025-12-04T11:17:06.2604764Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2605646Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2606479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2607270Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2608106Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2609002Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2609504Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2609767Z 2025-12-04T11:17:06.2609871Z Expected 1 but got 2. 2025-12-04T11:17:06.2610166Z Absolute difference: 1 2025-12-04T11:17:06.2610461Z Relative difference: 1.0 2025-12-04T11:17:06.2610652Z 2025-12-04T11:17:06.2610867Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2612141Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.2613200Z 2025-12-04T11:17:06.2613468Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2614106Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2614576Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2615335Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2616243Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2616701Z graph_break [] 2025-12-04T11:17:06.2617137Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2618709Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.2620166Z if out == self.unknown_value: 2025-12-04T11:17:06.2621110Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2622087Z warnings.warn( 2025-12-04T11:17:06.2622981Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2623954Z warnings.warn( 2025-12-04T11:17:06.2624617Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.2625399Z Traceback (most recent call last): 2025-12-04T11:17:06.2626145Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2627025Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2627840Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2628602Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2629534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2630409Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2630891Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2631156Z 2025-12-04T11:17:06.2631262Z Expected 1 but got 2. 2025-12-04T11:17:06.2631553Z Absolute difference: 1 2025-12-04T11:17:06.2631835Z Relative difference: 1.0 2025-12-04T11:17:06.2632036Z 2025-12-04T11:17:06.2632248Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2633520Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.2634603Z 2025-12-04T11:17:06.2634883Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2635502Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2635984Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2636731Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2637654Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2638121Z graph_break [] 2025-12-04T11:17:06.2638494Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2640068Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.2641514Z if out == self.unknown_value: 2025-12-04T11:17:06.2642464Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2643434Z warnings.warn( 2025-12-04T11:17:06.2644326Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2645283Z warnings.warn( 2025-12-04T11:17:06.2645666Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2646143Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2646574Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2647481Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2648263Z graph_break [] 2025-12-04T11:17:06.2648971Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2650064Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2651039Z warnings.warn( 2025-12-04T11:17:06.2651924Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2652892Z warnings.warn( 2025-12-04T11:17:06.2653188Z =================================== FAILURES =================================== 2025-12-04T11:17:06.2654000Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.2654776Z Traceback (most recent call last): 2025-12-04T11:17:06.2655519Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2656400Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2657395Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2658166Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2658997Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2659991Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2660896Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2661171Z 2025-12-04T11:17:06.2661280Z Expected 1 but got 2. 2025-12-04T11:17:06.2661577Z Absolute difference: 1 2025-12-04T11:17:06.2661880Z Relative difference: 1.0 2025-12-04T11:17:06.2662072Z 2025-12-04T11:17:06.2662304Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2663613Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.2664686Z 2025-12-04T11:17:06.2664963Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2665641Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2666130Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2666880Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2667794Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2668273Z graph_break [] 2025-12-04T11:17:06.2668645Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2670237Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.2671709Z if out == self.unknown_value: 2025-12-04T11:17:06.2672664Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2673632Z warnings.warn( 2025-12-04T11:17:06.2674527Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2675494Z warnings.warn( 2025-12-04T11:17:06.2675887Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2676355Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2676808Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2677713Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2678477Z graph_break [] 2025-12-04T11:17:06.2678857Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2679951Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2680915Z warnings.warn( 2025-12-04T11:17:06.2681788Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2682758Z warnings.warn( 2025-12-04T11:17:06.2682988Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2683104Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2683337Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2683886Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2684049Z graph_break [] 2025-12-04T11:17:06.2684270Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2685013Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2685114Z warnings.warn( 2025-12-04T11:17:06.2685848Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2685948Z warnings.warn( 2025-12-04T11:17:06.2686800Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-acd69551823a4dec.xml - 2025-12-04T11:17:06.2687020Z =========================== short test summary info ============================ 2025-12-04T11:17:06.2687983Z FAILED [0.4697s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2688022Z 2025-12-04T11:17:06.2688145Z Expected 1 but got 2. 2025-12-04T11:17:06.2688253Z Absolute difference: 1 2025-12-04T11:17:06.2688363Z Relative difference: 1.0 2025-12-04T11:17:06.2688369Z 2025-12-04T11:17:06.2688600Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2689514Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.2689522Z 2025-12-04T11:17:06.2689806Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2689987Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.2690191Z ================== 1 failed, 10 deselected, 2 rerun in 20.08s ================== 2025-12-04T11:17:06.2690306Z Got exit code 1 2025-12-04T11:17:06.2690413Z Retrying single test... 2025-12-04T11:17:06.2690860Z W1204 11:16:29.831000 97246 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.2691791Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-c7f6019ad8b4bab5.xml 2025-12-04T11:17:06.2691961Z ============================= test session starts ============================== 2025-12-04T11:17:06.2692329Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.2692441Z cachedir: .pytest_cache 2025-12-04T11:17:06.2692970Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.2693110Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.2693223Z configfile: pytest.ini 2025-12-04T11:17:06.2693768Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.2694004Z collecting ... collected 58 items / 10 deselected / 48 selected 2025-12-04T11:17:06.2695006Z stepcurrent: skipping 10 already run items. Running only test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.2695132Z Running 1 items in this shard 2025-12-04T11:17:06.2695137Z 2025-12-04T11:17:06.2696593Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 [W1204 11:16:33.203111196 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2696602Z 2025-12-04T11:17:06.2697316Z [W1204 11:16:48.474515686 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2697326Z 2025-12-04T11:17:06.2697845Z [W1204 11:16:48.474777553 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2697850Z 2025-12-04T11:17:06.2698379Z [W1204 11:16:48.482167886 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2698384Z 2025-12-04T11:17:06.2698895Z [W1204 11:16:48.482863359 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2698941Z 2025-12-04T11:17:06.2699453Z [W1204 11:16:48.483057119 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2699473Z 2025-12-04T11:17:06.2699986Z [W1204 11:16:48.489965448 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2700029Z 2025-12-04T11:17:06.2700539Z [W1204 11:16:48.490643342 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2700544Z 2025-12-04T11:17:06.2701067Z [W1204 11:16:48.490836545 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2701072Z 2025-12-04T11:17:06.2701582Z [W1204 11:16:50.490161539 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2701587Z 2025-12-04T11:17:06.2702118Z [W1204 11:16:50.491867467 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2702122Z 2025-12-04T11:17:06.2702637Z [W1204 11:16:50.492076251 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2702644Z 2025-12-04T11:17:06.2703169Z [W1204 11:16:50.496005221 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2703174Z 2025-12-04T11:17:06.2703684Z [W1204 11:16:50.496662751 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2703689Z 2025-12-04T11:17:06.2704211Z [W1204 11:16:50.496861088 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2704216Z 2025-12-04T11:17:06.2704726Z [W1204 11:16:50.502918728 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2704733Z 2025-12-04T11:17:06.2705246Z [W1204 11:16:50.503539249 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2705253Z 2025-12-04T11:17:06.2705773Z [W1204 11:16:50.503733760 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2705778Z 2025-12-04T11:17:06.2705911Z ('RERUN', {'yellow': True}) [19.2069s] [100%] 2025-12-04T11:17:06.2707216Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 [W1204 11:16:51.949562801 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2707223Z 2025-12-04T11:17:06.2707729Z [W1204 11:16:51.950377149 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2707736Z 2025-12-04T11:17:06.2708395Z [W1204 11:16:51.950598760 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2708404Z 2025-12-04T11:17:06.2708916Z [W1204 11:16:51.954561371 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2708920Z 2025-12-04T11:17:06.2709442Z [W1204 11:16:51.955351367 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2709447Z 2025-12-04T11:17:06.2709957Z [W1204 11:16:51.955544495 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2709962Z 2025-12-04T11:17:06.2710468Z [W1204 11:16:51.961629231 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2710516Z 2025-12-04T11:17:06.2711029Z [W1204 11:16:51.962263483 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2711063Z 2025-12-04T11:17:06.2711572Z [W1204 11:16:51.962453140 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2711577Z 2025-12-04T11:17:06.2712100Z [W1204 11:16:51.050779212 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2712104Z 2025-12-04T11:17:06.2712615Z [W1204 11:16:51.051540767 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2712620Z 2025-12-04T11:17:06.2713144Z [W1204 11:16:51.051748961 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2713151Z 2025-12-04T11:17:06.2713670Z [W1204 11:16:51.055666629 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2713675Z 2025-12-04T11:17:06.2714198Z [W1204 11:16:51.056313127 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2714202Z 2025-12-04T11:17:06.2714713Z [W1204 11:16:51.056509324 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2714717Z 2025-12-04T11:17:06.2715233Z [W1204 11:16:51.062586959 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2715238Z 2025-12-04T11:17:06.2715744Z [W1204 11:16:51.063384465 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2715751Z 2025-12-04T11:17:06.2716258Z [W1204 11:16:51.063580068 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2716279Z 2025-12-04T11:17:06.2716409Z ('RERUN', {'yellow': True}) [0.5209s] [100%] 2025-12-04T11:17:06.2717688Z inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 [W1204 11:16:51.433952373 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2717693Z 2025-12-04T11:17:06.2718215Z [W1204 11:16:51.434656356 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2718221Z 2025-12-04T11:17:06.2718729Z [W1204 11:16:51.434859320 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2718737Z 2025-12-04T11:17:06.2719258Z [W1204 11:16:51.438742893 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2719319Z 2025-12-04T11:17:06.2719830Z [W1204 11:16:51.439503425 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2719837Z 2025-12-04T11:17:06.2720357Z [W1204 11:16:51.439694044 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2720362Z 2025-12-04T11:17:06.2720869Z [W1204 11:16:51.445712594 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2720873Z 2025-12-04T11:17:06.2721391Z [W1204 11:16:51.446319445 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2721424Z 2025-12-04T11:17:06.2721936Z [W1204 11:16:51.446506648 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2721941Z 2025-12-04T11:17:06.2722455Z [W1204 11:16:51.531644377 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2722488Z 2025-12-04T11:17:06.2723010Z [W1204 11:16:51.532358717 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2723014Z 2025-12-04T11:17:06.2723523Z [W1204 11:16:51.532562666 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2723527Z 2025-12-04T11:17:06.2724048Z [W1204 11:16:51.536439653 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2724055Z 2025-12-04T11:17:06.2724565Z [W1204 11:16:51.537048370 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2724570Z 2025-12-04T11:17:06.2725098Z [W1204 11:16:51.537244377 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2725105Z 2025-12-04T11:17:06.2725611Z [W1204 11:16:51.543244138 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2725616Z 2025-12-04T11:17:06.2726136Z [W1204 11:16:51.544009966 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2726141Z 2025-12-04T11:17:06.2726647Z [W1204 11:16:51.544215359 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T11:17:06.2726652Z 2025-12-04T11:17:06.2726757Z FAILED [0.4799s] [100%] 2025-12-04T11:17:06.2726762Z 2025-12-04T11:17:06.2726921Z ==================================== RERUNS ==================================== 2025-12-04T11:17:06.2727442Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.2727581Z Traceback (most recent call last): 2025-12-04T11:17:06.2728095Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2728327Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2728805Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2728973Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2729533Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2729744Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2729878Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2729883Z 2025-12-04T11:17:06.2730001Z Expected 1 but got 2. 2025-12-04T11:17:06.2730167Z Absolute difference: 1 2025-12-04T11:17:06.2730281Z Relative difference: 1.0 2025-12-04T11:17:06.2730286Z 2025-12-04T11:17:06.2730515Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2731436Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.2731442Z 2025-12-04T11:17:06.2731723Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2731944Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2732061Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2732643Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2732874Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2732985Z graph_break [] 2025-12-04T11:17:06.2733235Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2734452Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.2734580Z if out == self.unknown_value: 2025-12-04T11:17:06.2735310Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2735428Z warnings.warn( 2025-12-04T11:17:06.2736150Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2736252Z warnings.warn( 2025-12-04T11:17:06.2736787Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.2736985Z Traceback (most recent call last): 2025-12-04T11:17:06.2737500Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2737748Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2738213Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2738391Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2738931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2739141Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2739292Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2739297Z 2025-12-04T11:17:06.2739409Z Expected 1 but got 2. 2025-12-04T11:17:06.2739534Z Absolute difference: 1 2025-12-04T11:17:06.2739646Z Relative difference: 1.0 2025-12-04T11:17:06.2739651Z 2025-12-04T11:17:06.2739869Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2740802Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.2740808Z 2025-12-04T11:17:06.2741080Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2741314Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2741434Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2741973Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2742283Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2742393Z graph_break [] 2025-12-04T11:17:06.2742613Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2743842Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.2743960Z if out == self.unknown_value: 2025-12-04T11:17:06.2744703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2744835Z warnings.warn( 2025-12-04T11:17:06.2745560Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2745681Z warnings.warn( 2025-12-04T11:17:06.2745901Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2746083Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2746313Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2746847Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2746960Z graph_break [] 2025-12-04T11:17:06.2747179Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2747902Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2748018Z warnings.warn( 2025-12-04T11:17:06.2748746Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2748864Z warnings.warn( 2025-12-04T11:17:06.2749012Z =================================== FAILURES =================================== 2025-12-04T11:17:06.2749528Z _ TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 _ 2025-12-04T11:17:06.2749666Z Traceback (most recent call last): 2025-12-04T11:17:06.2750181Z File "/var/lib/jenkins/workspace/test/inductor/test_cuda_select_algorithm.py", line 130, in test_int8_woq_mm_cuda 2025-12-04T11:17:06.2750415Z self.assertEqual(counters["inductor"]["woq_matcher_count"], 1) 2025-12-04T11:17:06.2750887Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:17:06.2751058Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:17:06.2751615Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:17:06.2751823Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:17:06.2751961Z AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2751966Z 2025-12-04T11:17:06.2752084Z Expected 1 but got 2. 2025-12-04T11:17:06.2752193Z Absolute difference: 1 2025-12-04T11:17:06.2752303Z Relative difference: 1.0 2025-12-04T11:17:06.2752319Z 2025-12-04T11:17:06.2752532Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2753445Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.2753453Z 2025-12-04T11:17:06.2753730Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2753945Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2754127Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2754667Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2754896Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2755005Z graph_break [] 2025-12-04T11:17:06.2755221Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2756431Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/constant_folding.py:256: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T11:17:06.2756595Z if out == self.unknown_value: 2025-12-04T11:17:06.2757317Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2757440Z warnings.warn( 2025-12-04T11:17:06.2758158Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2758289Z warnings.warn( 2025-12-04T11:17:06.2758521Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2758638Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2758882Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2759415Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2759513Z graph_break [] 2025-12-04T11:17:06.2759746Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2760473Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2760576Z warnings.warn( 2025-12-04T11:17:06.2761310Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2761411Z warnings.warn( 2025-12-04T11:17:06.2761636Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:17:06.2761750Z stats [('calls_captured', 6)] 2025-12-04T11:17:06.2761977Z aot_autograd [('total', 2), ('autograd_cache_bypass', 2), ('not_ok', 2)] 2025-12-04T11:17:06.2762520Z inductor [('pattern_matcher_nodes', 16), ('woq_matcher_nodes', 12), ('pattern_matcher_count', 6), ('fxgraph_cache_miss', 2), ('woq_matcher_count', 2)] 2025-12-04T11:17:06.2762620Z graph_break [] 2025-12-04T11:17:06.2762835Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:17:06.2763582Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2763684Z warnings.warn( 2025-12-04T11:17:06.2764414Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:17:06.2764512Z warnings.warn( 2025-12-04T11:17:06.2765359Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-c7f6019ad8b4bab5.xml - 2025-12-04T11:17:06.2765543Z =========================== short test summary info ============================ 2025-12-04T11:17:06.2766492Z FAILED [0.4799s] inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 - AssertionError: Scalars are not equal! 2025-12-04T11:17:06.2766500Z 2025-12-04T11:17:06.2766674Z Expected 1 but got 2. 2025-12-04T11:17:06.2766783Z Absolute difference: 1 2025-12-04T11:17:06.2766894Z Relative difference: 1.0 2025-12-04T11:17:06.2766899Z 2025-12-04T11:17:06.2767131Z To execute this test, run the following from the base repo dir: 2025-12-04T11:17:06.2768045Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_cuda_select_algorithm.py TestSelectAlgorithmCudaCUDA.test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.2768050Z 2025-12-04T11:17:06.2768326Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:17:06.2768509Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:17:06.2768740Z ================== 1 failed, 10 deselected, 2 rerun in 20.24s ================== 2025-12-04T11:17:06.2768850Z Got exit code 1 2025-12-04T11:17:06.2769690Z FAILED CONSISTENTLY: test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16 2025-12-04T11:17:06.2770150Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:17:06.2770597Z W1204 11:17:02.998000 97428 site-packages/torch/_inductor/utils.py:1703] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:17:06.2771264Z Test results will be stored in test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-cdc355d159a50f1a.xml 2025-12-04T11:17:06.2771441Z ============================= test session starts ============================== 2025-12-04T11:17:06.2771791Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:17:06.2771918Z cachedir: .pytest_cache 2025-12-04T11:17:06.2772441Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:17:06.2772572Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:17:06.2772697Z configfile: pytest.ini 2025-12-04T11:17:06.2773243Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T11:17:06.2773465Z collecting ... collected 58 items / 11 deselected / 47 selected 2025-12-04T11:17:06.2773621Z stepcurrent: skipping 11 already run items. 2025-12-04T11:17:06.2773739Z Running 0 items in this shard 2025-12-04T11:17:06.2773744Z 2025-12-04T11:17:06.2774606Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-cdc355d159a50f1a.xml - 2025-12-04T11:17:06.2774773Z ============================ 11 deselected in 0.02s ============================ 2025-12-04T11:17:06.2784265Z The following tests failed consistently: ['test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_1_in_features_128_out_features_64_cuda_bfloat16', 'test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_1024_out_features_65_cuda_bfloat16', 'test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_1024_cuda_bfloat16', 'test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_17_mid_dim_8_in_features_128_out_features_65_cuda_bfloat16', 'test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_1_in_features_1024_out_features_64_cuda_bfloat16', 'test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_1_mid_dim_8_in_features_1024_out_features_1024_cuda_bfloat16', 'test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_128_out_features_65_cuda_bfloat16', 'test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_64_cuda_bfloat16', 'test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_1_in_features_144_out_features_65_cuda_bfloat16', 'test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_128_out_features_64_cuda_bfloat16', 'test/inductor/test_cuda_select_algorithm.py::TestSelectAlgorithmCudaCUDA::test_int8_woq_mm_cuda_batch_size_32_mid_dim_8_in_features_144_out_features_1024_cuda_bfloat16'] 2025-12-04T11:17:06.2784283Z 2025-12-04T11:17:06.2784980Z FINISHED PRINTING LOG FILE of inductor/test_cuda_select_algorithm 5/5 (test/test-reports/inductor.test_cuda_select_algorithm_5.5_387c3291169f376b_.log) 2025-12-04T11:17:06.2784985Z 2025-12-04T11:17:06.2785393Z Finished inductor/test_cuda_select_algorithm 5/5 ... [2025-12-04 11:17:05.970538][7409.660924401], took 16.08min 2025-12-04T11:17:06.2786337Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-97143c683457795e.xml 2025-12-04T11:17:06.2787273Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-c6ae8965f2e7a602.xml 2025-12-04T11:17:06.2788165Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-2f238f5f180b8489.xml 2025-12-04T11:17:06.2789078Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-df27debc7fb08425.xml 2025-12-04T11:17:06.2789970Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-39f974c40a0e5d07.xml 2025-12-04T11:17:06.2790880Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-4c5f8d97ccaac9c3.xml 2025-12-04T11:17:06.2791771Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-fda94603e2b70838.xml 2025-12-04T11:17:06.2792678Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-ace483b73b471725.xml 2025-12-04T11:17:06.3014191Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-841659f3603156bc.xml 2025-12-04T11:17:06.3323591Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-e4e48c14b07c6744.xml 2025-12-04T11:17:06.3828676Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-887ea8bf31a36206.xml 2025-12-04T11:17:06.4156141Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-d486c0fac95f9182.xml 2025-12-04T11:17:06.4554548Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-73bc989177a97c92.xml 2025-12-04T11:17:06.4872299Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-9368d7516886ec65.xml 2025-12-04T11:17:06.5276083Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-921cae3f6acd6a84.xml 2025-12-04T11:17:06.5613489Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-8d346e15c82776ef.xml 2025-12-04T11:17:06.5962540Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-1b686c821c7b9c77.xml 2025-12-04T11:17:06.6290290Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-21dcaad0846c92ef.xml 2025-12-04T11:17:06.6624261Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-633a2407c7cc9faa.xml 2025-12-04T11:17:06.7012640Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-2a4f8a9a7e80b95a.xml 2025-12-04T11:17:06.7344370Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-ab0a7cd9bbef732b.xml 2025-12-04T11:17:06.7708199Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-81245131849bbecb.xml 2025-12-04T11:17:06.7998783Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-4cf813d975e62c57.xml 2025-12-04T11:17:06.8362436Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-f694c40280fe3f31.xml 2025-12-04T11:17:06.8944686Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-a6d9b5e9da7bb417.xml 2025-12-04T11:17:06.9285845Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-c31955170ca46735.xml 2025-12-04T11:17:06.9602667Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-1b66a9993689b300.xml 2025-12-04T11:17:06.9940189Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-79106310b2e4ab0d.xml 2025-12-04T11:17:07.0429532Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-235ce3baacba0fb6.xml 2025-12-04T11:17:07.0990337Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-00aa73936a934045.xml 2025-12-04T11:17:07.1300400Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-333e44976f8ee863.xml 2025-12-04T11:17:07.1590832Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-acd69551823a4dec.xml 2025-12-04T11:17:07.2185035Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-c7f6019ad8b4bab5.xml 2025-12-04T11:17:07.2481780Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-cdc355d159a50f1a.xml 2025-12-04T11:17:07.7716053Z Uploading logs for 57119749282 to S3 2025-12-04T11:17:07.8491990Z Uploading artifacts took 0.57 seconds 2025-12-04T11:17:07.8492694Z inductor/test_cuda_select_algorithm 5/5 failed! 2025-12-04T11:17:07.8497191Z Running inductor/test_deterministic 2/8 ... [2025-12-04 11:17:07.849482][7411.539872677] 2025-12-04T11:17:07.8497952Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T11:17:07.8502774Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_deterministic.py', '--shard-id=2', '--num-shards=8', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 11:17:07.849980] 2025-12-04T11:17:26.9483459Z 2025-12-04T11:17:26.9484589Z inductor/test_deterministic 2/8 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_deterministic_2.8_649a40a09099c603_.log 2025-12-04T11:17:26.9487572Z Running 3 items in this shard: test/inductor/test_deterministic.py::DeterministicTest::test_mm_padding_deterministic_True, test/inductor/test_deterministic.py::DeterministicTest::test_reduction_coordesc_tuning_deterministic_False, test/inductor/test_deterministic.py::DeterministicTest::test_run2run_determinism_model_name_BertForMaskedLM_training_or_inference_training_precision_float32 2025-12-04T11:17:26.9489607Z 2025-12-04T11:17:26.9489982Z Finished inductor/test_deterministic 2/8 ... [2025-12-04 11:17:26.948114][7430.638507789], took 0.32min 2025-12-04T11:17:26.9563888Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_deterministic/inductor.test_deterministic-142f090bc785acb6.xml 2025-12-04T11:17:27.0352538Z Running inductor/test_deterministic 7/8 ... [2025-12-04 11:17:27.034904][7430.725295294] 2025-12-04T11:17:27.0353150Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T11:17:27.0355919Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_deterministic.py', '--shard-id=7', '--num-shards=8', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 11:17:27.035341] 2025-12-04T11:18:56.4302898Z 2025-12-04T11:18:56.4305931Z inductor/test_deterministic 7/8 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_deterministic_7.8_cf2a089c308daa96_.log 2025-12-04T11:18:56.4310248Z Running 5 items in this shard: test/inductor/test_deterministic.py::DeterministicTest::test_max_autotune_deterministic_True, test/inductor/test_deterministic.py::DeterministicTest::test_reduction_coordesc_tuning_deterministic_True, test/inductor/test_deterministic.py::DeterministicTest::test_run2run_determinism_model_name_BertForMaskedLM_training_or_inference_training_precision_amp, test/inductor/test_deterministic.py::DeterministicTest::test_run2run_determinism_model_name_DistillGPT2_training_or_inference_training_precision_amp, test/inductor/test_deterministic.py::DeterministicTest::test_run2run_determinism_model_name_GoogleFnet_training_or_inference_inference_precision_float32 2025-12-04T11:18:56.4313807Z 2025-12-04T11:18:56.4314197Z Finished inductor/test_deterministic 7/8 ... [2025-12-04 11:18:56.430069][7520.120462026], took 1.49min 2025-12-04T11:18:56.4381896Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_deterministic/inductor.test_deterministic-2940c67a6de1d3d8.xml 2025-12-04T11:18:56.5111541Z Running inductor/test_native_matmul 2/2 ... [2025-12-04 11:18:56.510839][7520.201231476] 2025-12-04T11:18:56.5112153Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T11:18:56.5115503Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_native_matmul.py', '--shard-id=2', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 11:18:56.511254] 2025-12-04T11:19:59.7231794Z 2025-12-04T11:19:59.7232927Z inductor/test_native_matmul 2/2 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_native_matmul_2.2_564ae2b75fe7723b_.log 2025-12-04T11:19:59.7234904Z Running 2 items in this shard: test/inductor/test_native_matmul.py::TestTritonDotReduction::test_batchmatmul, test/inductor/test_native_matmul.py::TestTritonDotReduction::test_reduction_mask_zeroout 2025-12-04T11:19:59.7236002Z 2025-12-04T11:19:59.7236459Z Finished inductor/test_native_matmul 2/2 ... [2025-12-04 11:19:59.722953][7583.413345986], took 1.05min 2025-12-04T11:19:59.7311889Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_native_matmul/inductor.test_native_matmul-5845c7d4a4d90a90.xml 2025-12-04T11:19:59.8080353Z Running inductor/test_decompose_mem_bound_mm 1/1 ... [2025-12-04 11:19:59.807762][7583.498155358] 2025-12-04T11:19:59.8081154Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T11:19:59.8083967Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_decompose_mem_bound_mm.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 11:19:59.808131] 2025-12-04T11:24:34.1130757Z 2025-12-04T11:24:34.1131735Z PRINTING LOG FILE of inductor/test_decompose_mem_bound_mm 1/1 (test/test-reports/inductor.test_decompose_mem_bound_mm_1.1_7b3d736361a2b2dc_.log) 2025-12-04T11:24:34.1238336Z Test results will be stored in test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-819d138cb6fe4c8f.xml 2025-12-04T11:24:34.1239630Z ============================= test session starts ============================== 2025-12-04T11:24:34.1240646Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:24:34.1241821Z cachedir: .pytest_cache 2025-12-04T11:24:34.1243161Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:24:34.1244625Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:24:34.1245382Z configfile: pytest.ini 2025-12-04T11:24:34.1246832Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:24:34.1250338Z collecting ... /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:52: PytestCollectionWarning: cannot collect test class 'TestDecomposeAddMM' because it has a __init__ constructor (from: test/inductor/test_decompose_mem_bound_mm.py) 2025-12-04T11:24:34.1253079Z class TestDecomposeAddMM(torch.nn.Module): 2025-12-04T11:24:34.1253733Z collected 37 items 2025-12-04T11:24:34.1254308Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T11:24:34.1292690Z Running 37 items in this shard: test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_check_device, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_bmm_b_10240_m_2_k_2_n_2_should_decompose_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_bmm_b_10240_m_2_k_32_n_32_should_decompose_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_bmm_b_2000_m_2_k_2_n_2_should_decompose_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_bmm_cpu_b_1_m_2_k_2_n_2_should_decompose_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_bmm_cpu_b_2_m_2_k_2_n_2_should_decompose_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_m_20480_k_32_n_2_should_decompose_False_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_m_20480_k_32_n_2_should_decompose_False_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_m_20480_k_5_n_2_should_decompose_True_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_m_20480_k_5_n_2_should_decompose_True_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_m_2048_k_2_n_2_should_decompose_False_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_m_2048_k_2_n_2_should_decompose_False_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_32_n_2_should_decompose_False_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_32_n_2_should_decompose_False_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_2048_k_2_n_2_should_decompose_False_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_2048_k_2_n_2_should_decompose_False_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_cpu_m_1_k_64_n_16_should_decompose_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_cpu_m_1_k_64_n_32_should_decompose_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_cpu_m_2_k_64_n_16_should_decompose_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_m_20480_k_32_n_2_should_decompose_False_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_m_20480_k_32_n_2_should_decompose_False_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_m_20480_k_5_n_2_should_decompose_True_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_m_20480_k_5_n_2_should_decompose_True_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_m_2048_k_2_n_2_should_decompose_False_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_m_2048_k_2_n_2_should_decompose_False_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_32_n_2_should_decompose_False_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_32_n_2_should_decompose_False_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_2048_k_2_n_2_should_decompose_False_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_2048_k_2_n_2_should_decompose_False_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_dynamic_shape_decompose_addmm, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_dynamic_shape_m_20480_k_5_n_2_should_decompose_True_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_dynamic_shape_m_20480_k_5_n_2_should_decompose_True_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_realize_input 2025-12-04T11:24:34.1329553Z 2025-12-04T11:24:34.1330381Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_check_device PASSED [1.6329s] [ 2%] 2025-12-04T11:24:34.1331794Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_bmm_b_10240_m_2_k_2_n_2_should_decompose_True PASSED [5.3531s] [ 5%] 2025-12-04T11:24:34.1333825Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_bmm_b_10240_m_2_k_32_n_32_should_decompose_False PASSED [0.4318s] [ 8%] 2025-12-04T11:24:34.1336052Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_bmm_b_2000_m_2_k_2_n_2_should_decompose_False PASSED [0.3685s] [ 10%] 2025-12-04T11:24:34.1338536Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_bmm_cpu_b_1_m_2_k_2_n_2_should_decompose_True PASSED [1.1838s] [ 13%] 2025-12-04T11:24:34.1340271Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_bmm_cpu_b_2_m_2_k_2_n_2_should_decompose_False PASSED [0.1302s] [ 16%] 2025-12-04T11:24:34.1342980Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_m_20480_k_32_n_2_should_decompose_False_has_bias_False PASSED [0.3573s] [ 18%] 2025-12-04T11:24:34.1345692Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_m_20480_k_32_n_2_should_decompose_False_has_bias_True PASSED [0.8682s] [ 21%] 2025-12-04T11:24:34.1348188Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_m_20480_k_5_n_2_should_decompose_True_has_bias_False PASSED [1.2385s] [ 24%] 2025-12-04T11:24:34.1350255Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_m_20480_k_5_n_2_should_decompose_True_has_bias_True PASSED [1.3206s] [ 27%] 2025-12-04T11:24:34.1352390Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_m_2048_k_2_n_2_should_decompose_False_has_bias_False PASSED [0.3735s] [ 29%] 2025-12-04T11:24:34.1354288Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_m_2048_k_2_n_2_should_decompose_False_has_bias_True PASSED [0.8306s] [ 32%] 2025-12-04T11:24:34.1356679Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_32_n_2_should_decompose_False_has_bias_False PASSED [0.3577s] [ 35%] 2025-12-04T11:24:34.1359185Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_32_n_2_should_decompose_False_has_bias_True PASSED [0.3607s] [ 37%] 2025-12-04T11:24:34.1361883Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False ('RERUN', {'yellow': True}) [0.3648s] [ 40%] 2025-12-04T11:24:34.1364300Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False ('RERUN', {'yellow': True}) [0.3267s] [ 40%] 2025-12-04T11:24:34.1366804Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False FAILED [0.3257s] [ 40%] 2025-12-04T11:24:34.1368171Z 2025-12-04T11:24:34.1368382Z ==================================== RERUNS ==================================== 2025-12-04T11:24:34.1369460Z _ TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False _ 2025-12-04T11:24:34.1370737Z Traceback (most recent call last): 2025-12-04T11:24:34.1372211Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 275, in test_decompose_linear_mixed_precision 2025-12-04T11:24:34.1373713Z self.assertEqual( 2025-12-04T11:24:34.1374839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:34.1376529Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:34.1378167Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:34.1379827Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:34.1380426Z AssertionError: Scalars are not equal! 2025-12-04T11:24:34.1380679Z 2025-12-04T11:24:34.1380803Z Expected 1 but got 0. 2025-12-04T11:24:34.1381230Z Absolute difference: 1 2025-12-04T11:24:34.1381702Z Relative difference: 1.0 2025-12-04T11:24:34.1382068Z 2025-12-04T11:24:34.1382405Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:34.1384018Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:34.1385322Z 2025-12-04T11:24:34.1385607Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:34.1386501Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:34.1387224Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:34.1387724Z stats [('calls_captured', 1)] 2025-12-04T11:24:34.1388301Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:34.1390044Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_nodes', 3), ('pad_mm_bench', 3), ('pattern_matcher_count', 2), ('fxgraph_cache_miss', 1), ('decompose_mm', 1)] 2025-12-04T11:24:34.1391535Z graph_break [] 2025-12-04T11:24:34.1392113Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:34.1393318Z V1204 11:20:23.820000 99036 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpg6zqt4nj 2025-12-04T11:24:34.1394866Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.1396841Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:34.1399341Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.1402628Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.1405290Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:34.1407534Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:34.1409189Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.1411471Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.1414714Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, None); l_x_ = l_self_modules_linear_parameters_weight_ = None 2025-12-04T11:24:34.1416692Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:34.1418354Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.1419749Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.1421769Z V1204 11:20:24.088000 99036 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:34.1424405Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v: 2025-12-04T11:24:34.1426625Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [trz4xcrmguerwj2nwxhu4t6zwtig5vb4aexoqw3sqeb3z2aw5kv] gm: GraphModule() 2025-12-04T11:24:34.1428420Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.1429490Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.1430690Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.1432181Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:34.1434222Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.1436868Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.1439464Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.1442009Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, permute) 2025-12-04T11:24:34.1443981Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_3 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.1445993Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, convert_element_type_1, permute_3) 2025-12-04T11:24:34.1447406Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.1448920Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:34.1453355Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.1460657Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.1465759Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:34.1468366Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:34.1470961Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:34.1473465Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:34.1476079Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:34.1478595Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:34.1480726Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:34.1483015Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:34.1485252Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:34.1487471Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [reylvtt3rqfvzq5b3fjekfvick4dxsjoog5svxckgogwqosgp6z] fx_kwargs[static_input_idxs]: [0] 2025-12-04T11:24:34.1489950Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[0]: 1 2025-12-04T11:24:34.1492879Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:34.1495800Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:34.1499961Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:34.1503000Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:34.1547928Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:34.1597462Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:34.1600931Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:34.1603895Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:34.1606370Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:34.1608781Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:34.1611955Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:34.1614951Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:34.1618136Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:34.1620822Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:34.1623435Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:34.1626211Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:34.1628985Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:34.1631674Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:34.1634022Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:34.1635919Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:34.1639008Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:34.1641558Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:34.1644280Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:34.1646883Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:34.1649864Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:34.1652960Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:34.1656215Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:34.1659424Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:34.1662093Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:34.1664454Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:34.1667099Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:34.1669725Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:34.1672358Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:34.1674982Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:34.1677421Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:34.1679440Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:34.1682179Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:34.1684928Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:34.1687346Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:34.1689587Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:34.1692408Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:34.1695643Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:34.1698918Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:34.1702009Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:34.1704908Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:34.1707530Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:34.1710297Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:34.1713092Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:34.1715510Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:34.1718278Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:34.1720711Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:34.1723980Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:34.1727278Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:34.1730411Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:34.1733344Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:34.1736412Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:34.1739911Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:34.1743281Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:34.1746411Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:34.1751119Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:34.1754416Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:34.1766304Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:34.1768791Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:34.1771530Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:34.1774347Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:34.1776999Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:34.1779205Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:34.1782033Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:34.1784729Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:34.1923419Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:34.1926770Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.1929680Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:34.1931662Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.1958042Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:34.1961238Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.1964382Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:34.1967295Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:34.1970004Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:34.1973072Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:34.1997006Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:34.2000157Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:34.2002867Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:34.2005421Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:34.2008295Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:34.2011135Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:34.2013949Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:34.2017214Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:34.2020397Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:34.2022945Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:34.2025451Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:34.2028473Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:34.2031499Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:34.2034451Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:34.2037437Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:34.2040591Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:34.2043668Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:34.2046846Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:34.2049994Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:34.2052970Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:34.2055672Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:34.2058304Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:34.2061156Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:34.2063455Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:34.2065644Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:34.2068359Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:34.2071102Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:34.2073708Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:34.2076104Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:34.2078476Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:34.2081426Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:34.2084521Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:34.2087607Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:34.2090406Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:34.2092180Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:34.2093962Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:34.2096166Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:34.2098274Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:34.2100180Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:34.2101944Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:34.2103655Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:34.2105495Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:34.2107297Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:34.2109328Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:34.2111025Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:34.2112772Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:34.2114533Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:34.2116223Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:34.2117944Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:34.2119699Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:34.2121563Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:34.2123393Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:34.2125165Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:34.2126947Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:34.2128740Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:34.2130510Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:34.2132400Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:34.2134129Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:34.2135814Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:34.2137543Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:34.2139272Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:34.2141012Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:34.2142850Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:34.2144655Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:34.2146439Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:34.2148229Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:34.2149975Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:34.2151755Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:34.2153627Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:34.2155415Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:34.2157143Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:34.2158900Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:34.2160636Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:34.2162300Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:34.2163970Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:34.2165708Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:34.2167353Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:34.2169016Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:34.2170773Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:34.2172598Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:34.2174412Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:34.2176185Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:34.2177952Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:34.2179669Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:34.2181405Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:34.2183195Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:34.2185037Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:34.2186862Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:34.2188644Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:34.2190541Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:34.2192493Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:34.2194714Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:34.2196695Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:34.2198494Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:34.2200482Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:34.2202525Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:34.2204416Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:34.2206306Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:34.2208224Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:34.2210134Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:34.2211992Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:34.2213900Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:34.2215714Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:34.2217669Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:34.2219699Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:34.2221985Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:34.2224209Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:34.2226377Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:34.2228749Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:34.2230996Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:34.2233233Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:34.2235547Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:34.2237754Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:34.2239972Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:34.2241865Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:34.2243764Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:34.2245760Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:34.2247811Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:34.2249692Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:34.2251524Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:34.2253343Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:34.2255152Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:34.2257114Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:34.2258951Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:34.2260721Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:34.2262713Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:34.2264685Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:34.2266539Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:34.2268454Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:34.2270223Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:34.2272042Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:34.2273912Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:34.2275845Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:34.2277819Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:34.2279660Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:34.2281523Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:34.2283372Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:34.2285204Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:34.2287089Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:34.2289042Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:34.2291015Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:34.2293033Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:34.2294866Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:34.2296950Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:34.2298853Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:34.2300804Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:34.2302738Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:34.2304708Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:34.2306801Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:34.2308879Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:34.2310780Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:34.2312921Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:34.2314877Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:34.2316874Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:34.2318729Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:34.2320531Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:34.2322525Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:34.2324540Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:34.2326595Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:34.2328612Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:34.2330350Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:34.2332267Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:34.2334136Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:34.2336019Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:34.2338023Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:34.2339885Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:34.2341762Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:34.2343753Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.2345699Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.2347627Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:34.2349588Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:34.2351627Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:34.2353568Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:34.2355552Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:34.2357560Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:34.2359566Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:34.2361638Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:34.2363562Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:34.2365354Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:34.2367324Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:34.2369227Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:34.2371191Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:34.2373116Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:34.2374975Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:34.2377009Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:34.2378983Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:34.2380880Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:34.2382795Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:34.2384748Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:34.2386699Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:34.2388839Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:34.2390880Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:34.2392945Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:34.2395129Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:34.2397416Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:34.2399484Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:34.2401523Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:34.2403457Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:34.2405489Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:34.2407560Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:34.2409487Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:34.2411433Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:34.2413339Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:34.2415228Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:34.2417309Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:34.2419164Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:34.2421016Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:34.2422976Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:34.2425011Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:34.2426928Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:34.2428947Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:34.2430894Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:34.2432738Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:34.2434865Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:34.2436941Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:34.2438839Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:34.2440877Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:34.2442853Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:34.2444854Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:34.2446804Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:34.2448776Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:34.2450674Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:34.2452537Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:34.2454428Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:34.2456377Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:34.2458285Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:34.2460085Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:34.2462141Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:34.2464186Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:34.2466238Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.2468274Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:34.2470272Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:34.2472235Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:34.2474342Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:34.2476228Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:34.2478274Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:34.2480540Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:34.2482403Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:34.2484326Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:34.2486211Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:34.2488185Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:34.2490311Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:34.2492339Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:34.2494309Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:34.2496437Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:34.2498446Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:34.2500537Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:34.2502619Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:34.2504643Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:34.2506581Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:34.2508633Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:34.2510679Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:34.2512908Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:34.2514837Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:34.2516762Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:34.2518701Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:34.2520672Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:34.2522543Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:34.2524521Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:34.2526584Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:34.2528498Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:34.2530387Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:34.2532306Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:34.2534299Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:34.2536504Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:34.2538625Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:34.2540653Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:34.2542775Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:34.2544775Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:34.2546692Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:34.2548847Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:34.2550869Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:34.2552872Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:34.2554796Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:34.2556634Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:34.2558616Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:34.2560605Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:34.2562582Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:34.2564636Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:34.2566503Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:34.2568213Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:34.2570090Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:34.2571949Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:34.2573851Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:34.2575797Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:34.2577664Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:34.2579551Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:34.2581751Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:34.2583840Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:34.2585841Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:34.2587625Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:34.2589498Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:34.2591483Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:34.2593445Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:34.2595400Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:34.2597551Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:34.2599465Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:34.2601412Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:34.2603387Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:34.2605290Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:34.2607224Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:34.2609243Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:34.2611312Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:34.2613102Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:34.2615075Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:34.2617138Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:34.2618957Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:34.2620841Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:34.2622691Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:34.2624546Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:34.2626478Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:34.2628318Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:34.2630120Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:34.2631963Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:34.2633870Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:34.2635846Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:34.2637767Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:34.2639734Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:34.2641681Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:34.2643503Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:34.2645326Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:34.2647263Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:34.2648973Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:34.2650706Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:34.2652587Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:34.2654425Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:34.2656409Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:34.2658420Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:34.2660265Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:34.2662077Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:34.2663896Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:34.2666922Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:34.2671610Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:34.2675034Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:34.2676909Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:34.2678907Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:34.2680784Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:34.2682853Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:34.2684995Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:34.2686987Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:34.2688960Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:34.2691070Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:34.2693130Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:34.2695265Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:34.2697626Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:34.2699688Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:34.2701710Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:34.2703725Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:34.2705801Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:34.2707913Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:34.2709900Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:34.2711931Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:34.2713941Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:34.2715619Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:34.2717283Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:34.2719066Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:34.2720796Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:34.2722465Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:34.2724391Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:34.2726362Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:34.2728322Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:34.2730304Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:34.2731986Z V1204 11:20:24.098000 99036 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.2733605Z I1204 11:20:24.098000 99036 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.2735037Z V1204 11:20:24.098000 99036 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:34.2736240Z V1204 11:20:24.098000 99036 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:34.2737555Z I1204 11:20:24.099000 99036 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 13 2025-12-04T11:24:34.2739109Z V1204 11:20:24.118000 99036 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:34.2740572Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.2741839Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:34.2743477Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.2745411Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.2747422Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.2749606Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.2751767Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.2754061Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.2756027Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:34.2757879Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:34.2759997Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:34.2762145Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:34.2764065Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_3: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.2765742Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.2767057Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.2768106Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.2769333Z V1204 11:20:24.125000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:34.2770722Z V1204 11:20:24.125000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:34.2772609Z V1204 11:20:24.126000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.2774323Z V1204 11:20:24.126000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.2776170Z V1204 11:20:24.127000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.2778093Z V1204 11:20:24.127000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.2779751Z V1204 11:20:24.129000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.2781312Z V1204 11:20:24.129000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.2782963Z V1204 11:20:24.132000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:34.2784613Z V1204 11:20:24.132000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.2786251Z V1204 11:20:24.133000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:34.2787945Z V1204 11:20:24.134000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.2789575Z V1204 11:20:24.134000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:34.2791190Z V1204 11:20:24.135000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.2792697Z V1204 11:20:24.137000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:34.2794337Z V1204 11:20:24.137000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.2795905Z V1204 11:20:24.139000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_3 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.2797926Z V1204 11:20:24.139000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.2799155Z V1204 11:20:24.140000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.2800618Z V1204 11:20:24.140000 99036 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 13 2025-12-04T11:24:34.2802207Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:34.2803304Z warnings.warn( 2025-12-04T11:24:34.2804049Z V1204 11:20:24.141000 99036 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:34.2805269Z _ TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False _ 2025-12-04T11:24:34.2806130Z Traceback (most recent call last): 2025-12-04T11:24:34.2807119Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 275, in test_decompose_linear_mixed_precision 2025-12-04T11:24:34.2808082Z self.assertEqual( 2025-12-04T11:24:34.2808853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:34.2809783Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:34.2810735Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:34.2811695Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:34.2812339Z AssertionError: Scalars are not equal! 2025-12-04T11:24:34.2812631Z 2025-12-04T11:24:34.2812839Z Expected 1 but got 0. 2025-12-04T11:24:34.2813217Z Absolute difference: 1 2025-12-04T11:24:34.2813650Z Relative difference: 1.0 2025-12-04T11:24:34.2813926Z 2025-12-04T11:24:34.2814192Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:34.2815510Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:34.2816604Z 2025-12-04T11:24:34.2817123Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:34.2817846Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:34.2818421Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:34.2818956Z stats [('calls_captured', 1)] 2025-12-04T11:24:34.2819470Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:34.2820838Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_nodes', 3), ('pad_mm_bench', 3), ('pattern_matcher_count', 2), ('fxgraph_cache_miss', 1), ('decompose_mm', 1)] 2025-12-04T11:24:34.2822069Z graph_break [] 2025-12-04T11:24:34.2822563Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:34.2823406Z V1204 11:20:23.820000 99036 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpg6zqt4nj 2025-12-04T11:24:34.2824623Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.2825797Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:34.2827670Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.2829774Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.2831582Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:34.2833204Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:34.2843414Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.2845176Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.2847543Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, None); l_x_ = l_self_modules_linear_parameters_weight_ = None 2025-12-04T11:24:34.2849240Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:34.2850390Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.2851490Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.2852852Z V1204 11:20:24.088000 99036 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:34.2854506Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v: 2025-12-04T11:24:34.2856160Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [trz4xcrmguerwj2nwxhu4t6zwtig5vb4aexoqw3sqeb3z2aw5kv] gm: GraphModule() 2025-12-04T11:24:34.2857479Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.2858262Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.2859163Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.2860113Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:34.2861838Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.2863810Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.2865614Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.2867197Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, permute) 2025-12-04T11:24:34.2868624Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_3 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.2870044Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, convert_element_type_1, permute_3) 2025-12-04T11:24:34.2871164Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.2872298Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:34.2875404Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.2880341Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.2883621Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:34.2885370Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:34.2887330Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:34.2889105Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:34.2890947Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:34.2892764Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:34.2894388Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:34.2896202Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:34.2898173Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:34.2899899Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [reylvtt3rqfvzq5b3fjekfvick4dxsjoog5svxckgogwqosgp6z] fx_kwargs[static_input_idxs]: [0] 2025-12-04T11:24:34.2901610Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[0]: 1 2025-12-04T11:24:34.2903439Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:34.2905405Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:34.2907689Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:34.2909624Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:34.2939670Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:34.2970217Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:34.2972184Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:34.2974128Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:34.2975993Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:34.2977962Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:34.2979939Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:34.2981787Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:34.2983633Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:34.2985562Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:34.2987432Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:34.2989409Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:34.2991330Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:34.2993288Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:34.2995278Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:34.2997422Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:34.2999501Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:34.3001482Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:34.3007771Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:34.3010041Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:34.3011927Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:34.3013762Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:34.3015682Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:34.3017741Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:34.3019436Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:34.3021314Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:34.3023179Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:34.3024921Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:34.3026924Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:34.3028775Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:34.3030595Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:34.3032411Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:34.3034153Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:34.3035980Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:34.3192483Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:34.3194401Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:34.3196819Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:34.3198569Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:34.3200513Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:34.3202176Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:34.3203891Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:34.3205605Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:34.3207341Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:34.3209062Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:34.3210809Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:34.3212578Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:34.3214335Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:34.3216188Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:34.3218108Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:34.3219754Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:34.3221396Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:34.3223082Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:34.3224945Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:34.3226806Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:34.3228512Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:34.3230275Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:34.3232283Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:34.3234364Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:34.3236582Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:34.3238367Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:34.3240238Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:34.3242100Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:34.3243863Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:34.3245639Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:34.3247479Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:34.3249324Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:34.3251191Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.3253212Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:34.3255119Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.3257066Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:34.3258936Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.3260830Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:34.3262650Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:34.3264409Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:34.3266036Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:34.3267760Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:34.3269644Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:34.3271385Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:34.3273095Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:34.3274811Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:34.3276601Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:34.3278443Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:34.3280366Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:34.3282279Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:34.3284062Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:34.3285786Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:34.3287662Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:34.3289471Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:34.3291175Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:34.3292960Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:34.3294850Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:34.3296948Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:34.3298818Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:34.3300725Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:34.3302444Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:34.3304205Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:34.3305844Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:34.3307558Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:34.3309286Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:34.3311197Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:34.3313000Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:34.3314787Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:34.3316543Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:34.3318221Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:34.3319926Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:34.3321813Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:34.3323786Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:34.3325772Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:34.3327645Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:34.3329395Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:34.3331188Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:34.3333093Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:34.3334963Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:34.3336938Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:34.3338758Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:34.3340487Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:34.3342253Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:34.3344043Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:34.3345811Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:34.3347496Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:34.3349248Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:34.3351007Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:34.3352691Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:34.3354423Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:34.3356223Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:34.3358003Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:34.3359820Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:34.3361610Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:34.3363388Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:34.3365179Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:34.3366962Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:34.3368769Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:34.3370508Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:34.3372176Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:34.3373988Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:34.3375682Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:34.3377502Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:34.3379363Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:34.3381162Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:34.3382951Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:34.3384768Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:34.3386515Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:34.3388337Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:34.3390263Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:34.3392053Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:34.3393799Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:34.3395562Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:34.3397447Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:34.3399103Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:34.3400766Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:34.3402439Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:34.3404080Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:34.3405736Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:34.3407556Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:34.3409387Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:34.3411169Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:34.3412911Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:34.3414642Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:34.3416343Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:34.3418157Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:34.3419935Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:34.3421839Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:34.3423652Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:34.3425470Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:34.3427239Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:34.3428984Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:34.3430765Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:34.3432523Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:34.3434265Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:34.3436051Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:34.3437839Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:34.3439584Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:34.3441371Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:34.3443231Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:34.3444928Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:34.3446637Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:34.3448418Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:34.3450122Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:34.3451865Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:34.3453776Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:34.3455861Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:34.3458079Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:34.3460190Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:34.3462243Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:34.3464338Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:34.3466466Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:34.3468580Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:34.3470686Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:34.3472635Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:34.3474432Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:34.3476207Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:34.3478116Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:34.3480000Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:34.3481734Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:34.3483600Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:34.3485269Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:34.3486955Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:34.3488670Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:34.3490364Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:34.3492106Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:34.3493946Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:34.3495785Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:34.3497746Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:34.3499438Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:34.3501136Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:34.3502860Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:34.3504529Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:34.3506259Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:34.3508050Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:34.3509816Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:34.3511648Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:34.3513335Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:34.3515009Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:34.3516710Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:34.3518511Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:34.3520419Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:34.3522300Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:34.3523971Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:34.3525729Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:34.3527527Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:34.3529376Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:34.3531174Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:34.3532922Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:34.3534789Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:34.3536754Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:34.3538681Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:34.3540449Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:34.3542276Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:34.3544106Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:34.3545848Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:34.3547574Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:34.3549341Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:34.3551156Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:34.3552996Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:34.3554699Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:34.3556378Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:34.3558112Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:34.3559823Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:34.3561504Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:34.3563234Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:34.3564979Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:34.3566726Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:34.3568492Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.3570353Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.3572120Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:34.3573866Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:34.3575685Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:34.3577554Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:34.3579436Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:34.3581330Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:34.3583197Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:34.3585037Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:34.3586868Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:34.3588618Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:34.3590388Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:34.3592151Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:34.3593950Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:34.3595702Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:34.3597663Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:34.3599456Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:34.3601253Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:34.3603030Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:34.3604742Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:34.3606580Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:34.3608455Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:34.3610311Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:34.3612232Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:34.3614229Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:34.3616183Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:34.3618203Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:34.3620088Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:34.3621964Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:34.3623772Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:34.3625661Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:34.3627592Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:34.3629497Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:34.3631354Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:34.3633117Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:34.3635046Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:34.3636850Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:34.3638627Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:34.3640366Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:34.3642135Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:34.3643921Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:34.3645973Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:34.3647852Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:34.3649746Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:34.3651483Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:34.3653389Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:34.3655308Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:34.3657283Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:34.3659105Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:34.3660952Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:34.3662799Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:34.3664676Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:34.3666604Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:34.3668402Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:34.3670107Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:34.3671883Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:34.3673650Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:34.3675367Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:34.3677090Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:34.3678859Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:34.3680816Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:34.3682774Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.3684672Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:34.3686522Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:34.3688385Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:34.3690278Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:34.3692084Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:34.3693998Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:34.3695946Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:34.3698652Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:34.3700535Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:34.3702348Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:34.3704213Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:34.3706143Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:34.3708110Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:34.3709973Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:34.3711746Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:34.3713549Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:34.3715398Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:34.3717329Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:34.3719316Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:34.3721155Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:34.3722975Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:34.3724870Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:34.3726841Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:34.3728694Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:34.3730503Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:34.3732292Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:34.3734267Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:34.3736036Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:34.3737974Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:34.3739909Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:34.3741714Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:34.3743444Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:34.3745243Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:34.3747153Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:34.3749117Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:34.3751036Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:34.3752979Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:34.3754973Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:34.3756820Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:34.3758698Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:34.3760663Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:34.3762580Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:34.3764411Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:34.3766198Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:34.3767970Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:34.3769852Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:34.3771752Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:34.3773609Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:34.3775489Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:34.3777302Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:34.3778945Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:34.3780652Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:34.3782404Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:34.3784132Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:34.3785892Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:34.3787626Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:34.3789471Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:34.3791438Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:34.3793386Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:34.3795250Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:34.3797170Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:34.3798927Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:34.3800746Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:34.3802553Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:34.3804494Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:34.3806405Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:34.3808256Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:34.3810088Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:34.3811846Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:34.3813639Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:34.3815491Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:34.3817464Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:34.3819309Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:34.3820987Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:34.3822766Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:34.3824711Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:34.3826426Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:34.3828092Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:34.3829831Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:34.3831545Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:34.3833343Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:34.3835112Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:34.3836765Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:34.3838497Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:34.3840349Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:34.3842171Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:34.3844032Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:34.3845892Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:34.3847656Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:34.3849373Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:34.3851104Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:34.3852801Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:34.3854450Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:34.3856091Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:34.3857898Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:34.3859636Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:34.3861438Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:34.3863258Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:34.3865004Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:34.3866678Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:34.3868370Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:34.3871277Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:34.3875856Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:34.3879186Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:34.3880998Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:34.3882820Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:34.3884580Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:34.3886501Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:34.3888424Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:34.3890309Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:34.3892277Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:34.3894201Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:34.3896334Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:34.3898415Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:34.3900438Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:34.3902434Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:34.3904285Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:34.3906170Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:34.3908208Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:34.3910189Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:34.3912056Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:34.3914027Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:34.3915875Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:34.3917450Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:34.3919020Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:34.3920591Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:34.3922151Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:34.3923725Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:34.3925455Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:34.3927349Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:34.3929154Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:34.3930846Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:34.3932355Z V1204 11:20:24.098000 99036 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.3933851Z I1204 11:20:24.098000 99036 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.3935142Z V1204 11:20:24.098000 99036 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:34.3936224Z V1204 11:20:24.098000 99036 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:34.3937404Z I1204 11:20:24.099000 99036 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 13 2025-12-04T11:24:34.3938804Z V1204 11:20:24.118000 99036 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:34.3940219Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.3941280Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:34.3942903Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.3944719Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.3946555Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.3948631Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.3950688Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.3952685Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.3954544Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:34.3956276Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:34.3958232Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:34.3960226Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:34.3961935Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_3: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.3963513Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.3964718Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.3965667Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.3966761Z V1204 11:20:24.125000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:34.3968013Z V1204 11:20:24.125000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:34.3969754Z V1204 11:20:24.126000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.3971428Z V1204 11:20:24.126000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.3973071Z V1204 11:20:24.127000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.3974712Z V1204 11:20:24.127000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.3976228Z V1204 11:20:24.129000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.3977778Z V1204 11:20:24.129000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.3979302Z V1204 11:20:24.132000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:34.3980827Z V1204 11:20:24.132000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.3982286Z V1204 11:20:24.133000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:34.3983742Z V1204 11:20:24.134000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.3985247Z V1204 11:20:24.134000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:34.3986720Z V1204 11:20:24.135000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.3988178Z V1204 11:20:24.137000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:34.3989579Z V1204 11:20:24.137000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.3990985Z V1204 11:20:24.139000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_3 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.3992402Z V1204 11:20:24.139000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.3993604Z V1204 11:20:24.140000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.3994903Z V1204 11:20:24.140000 99036 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 13 2025-12-04T11:24:34.3996778Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:34.3997882Z warnings.warn( 2025-12-04T11:24:34.3998531Z V1204 11:20:24.141000 99036 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:34.3999474Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:34.3999944Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:34.4000287Z stats [('calls_captured', 1)] 2025-12-04T11:24:34.4000728Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:34.4001896Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_nodes', 3), ('pad_mm_bench', 3), ('pattern_matcher_count', 2), ('fxgraph_cache_miss', 1), ('decompose_mm', 1)] 2025-12-04T11:24:34.4002907Z graph_break [] 2025-12-04T11:24:34.4003283Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:34.4004086Z V1204 11:20:24.187000 99036 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpmayp0j_i 2025-12-04T11:24:34.4005082Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.4006152Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:34.4007730Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.4009655Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.4011373Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:34.4012704Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:34.4013719Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.4015256Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.4017532Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, None); l_x_ = l_self_modules_linear_parameters_weight_ = None 2025-12-04T11:24:34.4019148Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:34.4020181Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.4021123Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.4022407Z V1204 11:20:24.450000 99036 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:34.4023978Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v: 2025-12-04T11:24:34.4025476Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [trz4xcrmguerwj2nwxhu4t6zwtig5vb4aexoqw3sqeb3z2aw5kv] gm: GraphModule() 2025-12-04T11:24:34.4026542Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.4027238Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.4027913Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.4028824Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:34.4030242Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.4032045Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.4033731Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.4035160Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, permute) 2025-12-04T11:24:34.4036497Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_3 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.4037751Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, convert_element_type_1, permute_3) 2025-12-04T11:24:34.4038675Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.4039667Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:34.4042659Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.4047429Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.4050522Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:34.4052205Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:34.4054032Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:34.4055706Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:34.4057484Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:34.4059116Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:34.4060675Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:34.4062294Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:34.4063898Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:34.4065477Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [reylvtt3rqfvzq5b3fjekfvick4dxsjoog5svxckgogwqosgp6z] fx_kwargs[static_input_idxs]: [0] 2025-12-04T11:24:34.4067038Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[0]: 1 2025-12-04T11:24:34.4068730Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:34.4070522Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:34.4072671Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:34.4074554Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:34.4104799Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:34.4135061Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:34.4137070Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:34.4138749Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:34.4140541Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:34.4142341Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:34.4144035Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:34.4145667Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:34.4147352Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:34.4149047Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:34.4150801Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:34.4152621Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:34.4154452Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:34.4156287Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:34.4158152Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:34.4160011Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:34.4161895Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:34.4163735Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:34.4165685Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:34.4167579Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:34.4169294Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:34.4171043Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:34.4172756Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:34.4174522Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:34.4176120Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:34.4177844Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:34.4179516Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:34.4181151Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:34.4182873Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:34.4184588Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:34.4186252Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:34.4187934Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:34.4189609Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:34.4191337Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:34.4193008Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:34.4194673Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:34.4196507Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:34.4198268Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:34.4199957Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:34.4201606Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:34.4203313Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:34.4205065Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:34.4206731Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:34.4208479Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:34.4210211Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:34.4211967Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:34.4213732Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:34.4215518Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:34.4217351Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:34.4219003Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:34.4220654Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:34.4222351Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:34.4224216Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:34.4226137Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:34.4227839Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:34.4229608Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:34.4231376Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:34.4233448Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:34.4235591Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:34.4237375Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:34.4239300Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:34.4241139Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:34.4242935Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:34.4244721Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:34.4246573Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:34.4248393Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:34.4250270Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.4252167Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:34.4254076Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.4255948Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:34.4257886Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.4259804Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:34.4261619Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:34.4263391Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:34.4265022Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:34.4266775Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:34.4268593Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:34.4270331Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:34.4272032Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:34.4273784Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:34.4275539Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:34.4277412Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:34.4279330Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:34.4281242Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:34.4283000Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:34.4284724Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:34.4286569Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:34.4288386Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:34.4290086Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:34.4291873Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:34.4293786Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:34.4295633Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:34.4297731Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:34.4299538Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:34.4301314Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:34.4302976Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:34.4304612Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:34.4306334Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:34.4308104Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:34.4309847Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:34.4311689Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:34.4313468Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:34.4315220Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:34.4316901Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:34.4318558Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:34.4320392Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:34.4322367Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:34.4324354Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:34.4326234Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:34.4328037Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:34.4329832Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:34.4331638Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:34.4333498Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:34.4335424Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:34.4337242Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:34.4338974Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:34.4340759Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:34.4342579Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:34.4344323Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:34.4346052Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:34.4347806Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:34.4349550Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:34.4351250Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:34.4352953Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:34.4354712Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:34.4356495Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:34.4358375Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:34.4360260Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:34.4362369Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:34.4364354Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:34.4366203Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:34.4368147Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:34.4370057Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:34.4371852Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:34.4373693Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:34.4375493Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:34.4377370Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:34.4379403Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:34.4381371Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:34.4383255Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:34.4385229Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:34.4387027Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:34.4388912Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:34.4390940Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:34.4392858Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:34.4394802Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:34.4396799Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:34.4398666Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:34.4400545Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:34.4402366Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:34.4404148Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:34.4405861Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:34.4407716Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:34.4409593Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:34.4411566Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:34.4413474Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:34.4415392Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:34.4417294Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:34.4419176Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:34.4421011Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:34.4422978Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:34.4424899Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:34.4426812Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:34.4428766Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:34.4430634Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:34.4432515Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:34.4434421Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:34.4436347Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:34.4438189Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:34.4440105Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:34.4442032Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:34.4443963Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:34.4445840Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:34.4447774Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:34.4448628Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:34.4449497Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:34.4450416Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:34.4451272Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:34.4452126Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:34.4453144Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:34.4454219Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:34.4455356Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:34.4456355Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:34.4457489Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:34.4458521Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:34.4459623Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:34.4460703Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:34.4461728Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:34.4462652Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:34.4463540Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:34.4464420Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:34.4465420Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:34.4466360Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:34.4467208Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:34.4468022Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:34.4468898Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:34.4469708Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:34.4470638Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:34.4471469Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:34.4472346Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:34.4473267Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:34.4474157Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:34.4475117Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:34.4475927Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:34.4476820Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:34.4477665Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:34.4478528Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:34.4479372Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:34.4480362Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:34.4481212Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:34.4482072Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:34.4482891Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:34.4483706Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:34.4484628Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:34.4485509Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:34.4486578Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:34.4487411Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:34.4488229Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:34.4489085Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:34.4489966Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:34.4490939Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:34.4491777Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:34.4492735Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:34.4493709Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:34.4494713Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:34.4495624Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:34.4496609Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:34.4497666Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:34.4498575Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:34.4499425Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:34.4500337Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:34.4501263Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:34.4502206Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:34.4503109Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:34.4503936Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:34.4504755Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:34.4505686Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:34.4506483Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:34.4507357Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:34.4508178Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:34.4509077Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:34.4509933Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:34.4510851Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.4511929Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.4512738Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:34.4513684Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:34.4514526Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:34.4515562Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:34.4516465Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:34.4517444Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:34.4518317Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:34.4519243Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:34.4520140Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:34.4521037Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:34.4521948Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:34.4522797Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:34.4523713Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:34.4524537Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:34.4525459Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:34.4526379Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:34.4527257Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:34.4528130Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:34.4529003Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:34.4530079Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:34.4530999Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:34.4531947Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:34.4532932Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:34.4533863Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:34.4534842Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:34.4535826Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:34.4536805Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:34.4537764Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:34.4538720Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:34.4539689Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:34.4540673Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:34.4541578Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:34.4542437Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:34.4543340Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:34.4544183Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:34.4545089Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:34.4546053Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:34.4547013Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:34.4547876Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:34.4548720Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:34.4549746Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:34.4550663Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:34.4551629Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:34.4552462Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:34.4553454Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:34.4554425Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:34.4555347Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:34.4556286Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:34.4557211Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:34.4558141Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:34.4559029Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:34.4560000Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:34.4560796Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:34.4561745Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:34.4562627Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:34.4563502Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:34.4564372Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:34.4565208Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:34.4566162Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:34.4567202Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:34.4568240Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.4569116Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:34.4570077Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:34.4570955Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:34.4571935Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:34.4572874Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:34.4573880Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:34.4574798Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:34.4575676Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:34.4576597Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:34.4577559Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:34.4578542Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:34.4579471Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:34.4580444Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:34.4581358Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:34.4582279Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:34.4583210Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:34.4584124Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:34.4585162Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:34.4586080Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:34.4587096Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:34.4587985Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:34.4588961Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:34.4589923Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:34.4590849Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:34.4591716Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:34.4592641Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:34.4593587Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:34.4594419Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:34.4595386Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:34.4596526Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:34.4597346Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:34.4598303Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:34.4599295Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:34.4600304Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:34.4601249Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:34.4602192Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:34.4603253Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:34.4604178Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:34.4605127Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:34.4606036Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:34.4607159Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:34.4608083Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:34.4609075Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:34.4609927Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:34.4610835Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:34.4611744Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:34.4612646Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:34.4613617Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:34.4614551Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:34.4615378Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:34.4616170Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:34.4617177Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:34.4617993Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:34.4618891Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:34.4619840Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:34.4620660Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:34.4621639Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:34.4622629Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:34.4623620Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:34.4624519Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:34.4625488Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:34.4626379Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:34.4627236Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:34.4628157Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:34.4629094Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:34.4630051Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:34.4630954Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:34.4631864Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:34.4632718Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:34.4633651Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:34.4634624Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:34.4635540Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:34.4636456Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:34.4637245Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:34.4638239Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:34.4639155Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:34.4640030Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:34.4640840Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:34.4641693Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:34.4642609Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:34.4643563Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:34.4644474Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:34.4645258Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:34.4646171Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:34.4647052Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:34.4647940Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:34.4648913Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:34.4649820Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:34.4650699Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:34.4651563Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:34.4652469Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:34.4653226Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:34.4654121Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:34.4654937Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:34.4655771Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:34.4656678Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:34.4657644Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:34.4658612Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:34.4659435Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:34.4660319Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:34.4661207Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:34.4663246Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:34.4665680Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:34.4666584Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:34.4667542Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:34.4668458Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:34.4669329Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:34.4670313Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:34.4671248Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:34.4672246Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:34.4673237Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:34.4674155Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:34.4675138Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:34.4676085Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:34.4677257Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:34.4678171Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:34.4679034Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:34.4680034Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:34.4680986Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:34.4681978Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:34.4682897Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:34.4684011Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:34.4684786Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:34.4685519Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:34.4686364Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:34.4687157Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:34.4687975Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:34.4688760Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:34.4689721Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:34.4690596Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:34.4691544Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:34.4692312Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:34.4693049Z V1204 11:20:24.460000 99036 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.4693812Z I1204 11:20:24.460000 99036 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.4694363Z V1204 11:20:24.460000 99036 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:34.4694903Z V1204 11:20:24.460000 99036 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:34.4695488Z I1204 11:20:24.461000 99036 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 14 2025-12-04T11:24:34.4696483Z V1204 11:20:24.473000 99036 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:34.4697103Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.4697641Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:34.4698681Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.4713235Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.4714417Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.4715416Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.4716559Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.4717467Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.4718367Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:34.4719128Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:34.4720182Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:34.4721002Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:34.4721847Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_3: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.4722482Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.4722972Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.4723357Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.4723927Z V1204 11:20:24.480000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:34.4724505Z V1204 11:20:24.481000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:34.4725516Z V1204 11:20:24.481000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.4726034Z V1204 11:20:24.482000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.4727042Z V1204 11:20:24.483000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.4727556Z V1204 11:20:24.483000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.4728439Z V1204 11:20:24.484000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.4728891Z V1204 11:20:24.485000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.4729874Z V1204 11:20:24.486000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:34.4730326Z V1204 11:20:24.486000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.4731211Z V1204 11:20:24.488000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:34.4731662Z V1204 11:20:24.488000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.4732606Z V1204 11:20:24.489000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:34.4733038Z V1204 11:20:24.489000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.4733882Z V1204 11:20:24.491000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:34.4734327Z V1204 11:20:24.491000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.4735157Z V1204 11:20:24.493000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_3 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.4735650Z V1204 11:20:24.493000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.4736259Z V1204 11:20:24.494000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.4736924Z V1204 11:20:24.494000 99036 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 14 2025-12-04T11:24:34.4737671Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:34.4737774Z warnings.warn( 2025-12-04T11:24:34.4738264Z V1204 11:20:24.495000 99036 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:34.4738416Z =================================== FAILURES =================================== 2025-12-04T11:24:34.4738901Z _ TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False _ 2025-12-04T11:24:34.4739024Z Traceback (most recent call last): 2025-12-04T11:24:34.4739614Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 275, in test_decompose_linear_mixed_precision 2025-12-04T11:24:34.4739734Z self.assertEqual( 2025-12-04T11:24:34.4740190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:34.4740353Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:34.4740903Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:34.4741108Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:34.4741252Z AssertionError: Scalars are not equal! 2025-12-04T11:24:34.4741261Z 2025-12-04T11:24:34.4741367Z Expected 1 but got 0. 2025-12-04T11:24:34.4741475Z Absolute difference: 1 2025-12-04T11:24:34.4741598Z Relative difference: 1.0 2025-12-04T11:24:34.4741604Z 2025-12-04T11:24:34.4741823Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:34.4742735Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:34.4742754Z 2025-12-04T11:24:34.4743023Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:34.4743248Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:34.4743373Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:34.4743486Z stats [('calls_captured', 1)] 2025-12-04T11:24:34.4743710Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:34.4744525Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_nodes', 3), ('pad_mm_bench', 3), ('pattern_matcher_count', 2), ('fxgraph_cache_miss', 1), ('decompose_mm', 1)] 2025-12-04T11:24:34.4744628Z graph_break [] 2025-12-04T11:24:34.4744860Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:34.4745298Z V1204 11:20:23.820000 99036 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpg6zqt4nj 2025-12-04T11:24:34.4745731Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.4746235Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:34.4747173Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.4748066Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.4748843Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:34.4749292Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:34.4749714Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.4750676Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.4751717Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, None); l_x_ = l_self_modules_linear_parameters_weight_ = None 2025-12-04T11:24:34.4752188Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:34.4752614Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.4752988Z V1204 11:20:23.848000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.4753697Z V1204 11:20:24.088000 99036 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:34.4754431Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v: 2025-12-04T11:24:34.4755086Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [trz4xcrmguerwj2nwxhu4t6zwtig5vb4aexoqw3sqeb3z2aw5kv] gm: GraphModule() 2025-12-04T11:24:34.4755399Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.4755670Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.4755936Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.4756414Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:34.4757226Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.4758082Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.4758826Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.4759400Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, permute) 2025-12-04T11:24:34.4760026Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_3 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.4760566Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, convert_element_type_1, permute_3) 2025-12-04T11:24:34.4760868Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.4761437Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:34.4763785Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.4766129Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.4766784Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:34.4767688Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:34.4768415Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:34.4769228Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:34.4770021Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:34.4770761Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:34.4771474Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:34.4772197Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:34.4772937Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:34.4773683Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [reylvtt3rqfvzq5b3fjekfvick4dxsjoog5svxckgogwqosgp6z] fx_kwargs[static_input_idxs]: [0] 2025-12-04T11:24:34.4774365Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[0]: 1 2025-12-04T11:24:34.4775243Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:34.4776035Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:34.4777365Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:34.4778161Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:34.4807373Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:34.4808416Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:34.4809180Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:34.4809949Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:34.4810854Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:34.4811695Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:34.4812416Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:34.4813188Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:34.4814019Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:34.4814786Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:34.4815682Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:34.4816522Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:34.4817454Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:34.4818312Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:34.4819153Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:34.4820063Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:34.4820913Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:34.4821720Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:34.4822735Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:34.4823527Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:34.4824375Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:34.4825114Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:34.4825957Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:34.4826747Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:34.4827452Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:34.4828251Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:34.4828993Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:34.4829752Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:34.4830619Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:34.4831417Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:34.4832185Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:34.4832953Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:34.4833718Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:34.4834493Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:34.4835280Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:34.4836036Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:34.4836833Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:34.4837614Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:34.4838378Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:34.4839178Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:34.4839976Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:34.4840748Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:34.4841501Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:34.4842371Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:34.4843169Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:34.4844009Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:34.4844804Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:34.4845699Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:34.4846475Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:34.4847261Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:34.4848012Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:34.4848811Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:34.4849730Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:34.4850530Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:34.4851317Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:34.4852159Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:34.4852897Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:34.4854105Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:34.4854929Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:34.4855803Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:34.4856685Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:34.4857553Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:34.4858414Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:34.4859258Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:34.4860144Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:34.4860940Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:34.4861923Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.4862746Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:34.4863736Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.4864540Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:34.4865468Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.4866296Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:34.4867165Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:34.4867933Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:34.4868674Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:34.4869503Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:34.4870365Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:34.4871149Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:34.4871967Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:34.4872746Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:34.4873598Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:34.4874482Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:34.4875431Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:34.4876280Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:34.4877041Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:34.4877913Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:34.4878791Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:34.4879636Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:34.4880408Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:34.4881311Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:34.4882167Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:34.4883041Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:34.4883921Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:34.4884711Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:34.4885513Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:34.4886232Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:34.4887061Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:34.4887865Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:34.4888650Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:34.4889491Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:34.4890360Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:34.4891189Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:34.4891996Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:34.4892745Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:34.4893567Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:34.4894473Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:34.4895440Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:34.4896559Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:34.4897447Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:34.4898271Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:34.4899120Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:34.4899948Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:34.4900864Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:34.4901711Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:34.4902493Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:34.4903390Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:34.4904215Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:34.4905052Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:34.4905821Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:34.4906663Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:34.4907499Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:34.4908275Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:34.4909056Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:34.4909905Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:34.4910737Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:34.4911609Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:34.4912485Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:34.4913258Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:34.4914127Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:34.4914919Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:34.4915781Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:34.4916599Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:34.4917378Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:34.4918145Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:34.4918921Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:34.4919732Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:34.4920562Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:34.4921400Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:34.4922266Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:34.4923090Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:34.4923948Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:34.4924695Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:34.4925580Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:34.4926465Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:34.4927312Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:34.4928120Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:34.4928946Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:34.4929725Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:34.4930488Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:34.4931260Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:34.4932031Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:34.4932768Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:34.4933552Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:34.4934404Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:34.4935294Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:34.4936107Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:34.4936945Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:34.4937721Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:34.4938568Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:34.4939385Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:34.4940219Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:34.4941103Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:34.4941941Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:34.4942768Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:34.4943609Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:34.4944421Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:34.4945248Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:34.4946038Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:34.4946834Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:34.4947700Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:34.4948487Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:34.4949305Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:34.4950161Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:34.4951046Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:34.4951795Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:34.4952623Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:34.4953418Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:34.4954232Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:34.4955061Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:34.4956003Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:34.4957011Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:34.4958017Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:34.4958991Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:34.4960014Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:34.4961021Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:34.4962014Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:34.4962983Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:34.4963980Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:34.4964814Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:34.4965647Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:34.4966460Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:34.4967423Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:34.4968255Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:34.4969047Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:34.4969816Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:34.4970600Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:34.4971404Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:34.4972185Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:34.4972979Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:34.4973752Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:34.4974757Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:34.4975603Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:34.4976406Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:34.4977231Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:34.4978043Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:34.4978826Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:34.4979595Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:34.4980440Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:34.4981272Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:34.4982059Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:34.4982893Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:34.4983654Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:34.4984448Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:34.4985233Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:34.4986072Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:34.4987041Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:34.4987837Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:34.4988578Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:34.4989408Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:34.4990263Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:34.4991144Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:34.4991961Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:34.4992755Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:34.4993688Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:34.4994564Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:34.4995402Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:34.4996352Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:34.4997253Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:34.4998054Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:34.4998830Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:34.4999702Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:34.5000530Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:34.5001385Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:34.5002209Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:34.5002993Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:34.5003815Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:34.5004603Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:34.5005350Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:34.5006188Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:34.5006956Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:34.5007838Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:34.5008620Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:34.5009487Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.5010354Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.5011132Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:34.5011981Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:34.5012807Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:34.5013663Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:34.5014517Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:34.5015460Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:34.5016285Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:34.5017221Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:34.5018022Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:34.5018876Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:34.5019714Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:34.5020525Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:34.5021363Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:34.5022220Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:34.5023079Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:34.5023926Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:34.5024774Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:34.5025576Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:34.5026385Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:34.5027300Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:34.5028150Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:34.5029027Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:34.5029932Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:34.5030847Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:34.5031799Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:34.5032701Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:34.5033547Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:34.5034427Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:34.5035266Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:34.5036193Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:34.5037078Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:34.5037917Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:34.5038764Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:34.5039632Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:34.5040439Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:34.5041302Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:34.5042085Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:34.5042908Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:34.5043736Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:34.5044549Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:34.5045421Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:34.5046291Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:34.5047121Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:34.5047954Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:34.5048862Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:34.5049748Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:34.5050618Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:34.5051479Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:34.5052341Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:34.5053183Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:34.5054028Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:34.5054953Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:34.5055774Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:34.5056588Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:34.5057477Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:34.5058270Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:34.5059067Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:34.5059865Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:34.5060718Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:34.5061659Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:34.5062552Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.5063388Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:34.5064300Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:34.5065169Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:34.5066025Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:34.5066871Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:34.5067849Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:34.5068693Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:34.5069513Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:34.5070337Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:34.5071160Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:34.5072091Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:34.5072988Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:34.5073919Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:34.5074730Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:34.5075558Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:34.5076398Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:34.5077276Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:34.5078183Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:34.5079067Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:34.5079936Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:34.5080786Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:34.5081686Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:34.5082553Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:34.5083443Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:34.5084251Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:34.5085108Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:34.5085938Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:34.5086771Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:34.5087702Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:34.5088605Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:34.5089402Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:34.5090198Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:34.5091081Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:34.5092002Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:34.5092922Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:34.5093794Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:34.5094737Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:34.5095587Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:34.5096673Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:34.5097599Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:34.5098519Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:34.5099450Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:34.5100294Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:34.5101128Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:34.5101953Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:34.5102885Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:34.5103737Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:34.5104651Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:34.5105542Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:34.5106278Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:34.5107051Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:34.5107863Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:34.5108673Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:34.5109484Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:34.5110301Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:34.5111091Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:34.5111966Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:34.5112954Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:34.5113829Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:34.5114645Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:34.5115439Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:34.5116320Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:34.5117142Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:34.5117997Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:34.5118891Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:34.5119733Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:34.5120649Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:34.5121470Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:34.5122302Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:34.5123138Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:34.5124007Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:34.5124903Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:34.5125729Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:34.5126456Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:34.5127391Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:34.5128246Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:34.5129006Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:34.5129775Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:34.5130561Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:34.5131380Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:34.5132260Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:34.5133024Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:34.5133777Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:34.5134593Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:34.5135468Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:34.5136355Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:34.5137280Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:34.5138141Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:34.5138929Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:34.5139728Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:34.5140535Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:34.5141289Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:34.5142043Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:34.5142796Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:34.5143602Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:34.5144466Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:34.5145329Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:34.5146148Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:34.5146928Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:34.5147709Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:34.5148509Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:34.5150482Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:34.5152931Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:34.5153737Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:34.5154646Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:34.5155452Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:34.5156287Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:34.5157263Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:34.5158096Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:34.5159008Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:34.5159916Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:34.5160833Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:34.5161742Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:34.5162636Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:34.5163667Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:34.5164549Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:34.5165384Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:34.5166299Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:34.5167258Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:34.5168133Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:34.5169619Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:34.5170595Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:34.5171329Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:34.5172044Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:34.5172785Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:34.5173512Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:34.5174222Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:34.5174939Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:34.5175822Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:34.5176663Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:34.5177640Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:34.5178357Z V1204 11:20:24.097000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:34.5179028Z V1204 11:20:24.098000 99036 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.5179700Z I1204 11:20:24.098000 99036 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.5180220Z V1204 11:20:24.098000 99036 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:34.5180721Z V1204 11:20:24.098000 99036 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:34.5181226Z I1204 11:20:24.099000 99036 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 13 2025-12-04T11:24:34.5182006Z V1204 11:20:24.118000 99036 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:34.5182446Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.5182978Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:34.5183940Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.5184702Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.5185699Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.5186655Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.5187656Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.5188545Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.5189397Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:34.5190153Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:34.5191227Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:34.5192048Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:34.5192822Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_3: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.5193475Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.5193924Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.5194330Z V1204 11:20:24.122000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.5194897Z V1204 11:20:24.125000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:34.5195456Z V1204 11:20:24.125000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:34.5196655Z V1204 11:20:24.126000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.5197220Z V1204 11:20:24.126000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.5198300Z V1204 11:20:24.127000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.5198799Z V1204 11:20:24.127000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.5199699Z V1204 11:20:24.129000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.5200154Z V1204 11:20:24.129000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.5201107Z V1204 11:20:24.132000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:34.5201562Z V1204 11:20:24.132000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.5202438Z V1204 11:20:24.133000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:34.5202894Z V1204 11:20:24.134000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.5203804Z V1204 11:20:24.134000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:34.5204243Z V1204 11:20:24.135000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.5205132Z V1204 11:20:24.137000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:34.5205557Z V1204 11:20:24.137000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.5206397Z V1204 11:20:24.139000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_3 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.5206836Z V1204 11:20:24.139000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.5207469Z V1204 11:20:24.140000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.5208044Z V1204 11:20:24.140000 99036 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 13 2025-12-04T11:24:34.5208800Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:34.5208903Z warnings.warn( 2025-12-04T11:24:34.5209380Z V1204 11:20:24.141000 99036 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:34.5209617Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:34.5209730Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:34.5209875Z stats [('calls_captured', 1)] 2025-12-04T11:24:34.5210111Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:34.5210878Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_nodes', 3), ('pad_mm_bench', 3), ('pattern_matcher_count', 2), ('fxgraph_cache_miss', 1), ('decompose_mm', 1)] 2025-12-04T11:24:34.5211021Z graph_break [] 2025-12-04T11:24:34.5211243Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:34.5211677Z V1204 11:20:24.187000 99036 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpmayp0j_i 2025-12-04T11:24:34.5212124Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.5212610Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:34.5213557Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.5214406Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.5215161Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:34.5215615Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:34.5216027Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.5217078Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.5218149Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, None); l_x_ = l_self_modules_linear_parameters_weight_ = None 2025-12-04T11:24:34.5218635Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:34.5219052Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.5219429Z V1204 11:20:24.212000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.5220154Z V1204 11:20:24.450000 99036 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:34.5220971Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v: 2025-12-04T11:24:34.5221645Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [trz4xcrmguerwj2nwxhu4t6zwtig5vb4aexoqw3sqeb3z2aw5kv] gm: GraphModule() 2025-12-04T11:24:34.5221918Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.5222200Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.5222468Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.5222930Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:34.5223788Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.5224602Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.5225385Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.5225948Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, permute) 2025-12-04T11:24:34.5226574Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_3 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.5227080Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, convert_element_type_1, permute_3) 2025-12-04T11:24:34.5227371Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.5227950Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:34.5230244Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.5232643Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.5233287Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:34.5234206Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:34.5234925Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:34.5235772Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:34.5236561Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:34.5237270Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:34.5237996Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:34.5238708Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:34.5239469Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:34.5240240Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [reylvtt3rqfvzq5b3fjekfvick4dxsjoog5svxckgogwqosgp6z] fx_kwargs[static_input_idxs]: [0] 2025-12-04T11:24:34.5240919Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[0]: 1 2025-12-04T11:24:34.5241801Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:34.5242584Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:34.5243769Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:34.5244563Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:34.5273579Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:34.5274611Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:34.5275370Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:34.5276198Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:34.5277041Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:34.5277927Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:34.5278636Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:34.5279414Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:34.5280189Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:34.5280956Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:34.5281819Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:34.5282644Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:34.5283508Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:34.5284353Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:34.5285239Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:34.5286130Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:34.5286979Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:34.5287800Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:34.5288832Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:34.5289602Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:34.5290433Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:34.5291181Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:34.5292051Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:34.5292822Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:34.5293558Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:34.5294344Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:34.5295101Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:34.5295851Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:34.5297049Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:34.5297814Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:34.5298595Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:34.5299356Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:34.5300117Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:34.5300905Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:34.5301744Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:34.5302511Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:34.5303298Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:34.5304086Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:34.5304896Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:34.5305657Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:34.5306472Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:34.5307231Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:34.5308041Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:34.5308847Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:34.5309696Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:34.5310512Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:34.5311294Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:34.5312173Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:34.5312939Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:34.5313713Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:34.5314454Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:34.5315264Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:34.5316182Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:34.5316972Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:34.5317805Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:34.5318631Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:34.5319384Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:34.5320597Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:34.5321401Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:34.5322260Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:34.5323135Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:34.5323989Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:34.5324798Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:34.5325686Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:34.5326558Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:34.5327364Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:34.5328297Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.5329121Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:34.5330084Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.5330877Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:34.5331830Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.5332644Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:34.5333559Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:34.5334319Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:34.5335070Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:34.5335903Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:34.5336794Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:34.5337632Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:34.5338445Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:34.5339243Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:34.5340110Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:34.5340984Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:34.5341937Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:34.5342790Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:34.5343566Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:34.5344389Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:34.5345275Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:34.5346074Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:34.5346857Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:34.5347740Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:34.5348594Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:34.5349509Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:34.5350373Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:34.5351176Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:34.5351969Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:34.5352730Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:34.5353521Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:34.5354319Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:34.5355121Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:34.5355975Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:34.5356815Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:34.5357654Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:34.5358466Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:34.5359198Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:34.5359990Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:34.5360909Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:34.5361824Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:34.5362765Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:34.5363564Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:34.5364403Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:34.5365268Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:34.5366105Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:34.5367002Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:34.5367849Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:34.5368666Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:34.5369471Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:34.5370304Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:34.5371128Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:34.5371945Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:34.5372730Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:34.5373590Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:34.5374378Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:34.5375148Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:34.5375967Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:34.5376779Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:34.5377676Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:34.5378542Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:34.5379319Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:34.5380210Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:34.5381031Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:34.5381901Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:34.5382706Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:34.5383497Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:34.5384290Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:34.5385064Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:34.5385847Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:34.5386668Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:34.5387521Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:34.5388381Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:34.5389242Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:34.5390092Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:34.5390840Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:34.5391743Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:34.5392581Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:34.5393409Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:34.5394199Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:34.5395034Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:34.5395802Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:34.5396820Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:34.5397697Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:34.5398457Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:34.5399212Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:34.5399987Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:34.5400887Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:34.5401731Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:34.5402542Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:34.5403336Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:34.5404140Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:34.5404965Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:34.5405803Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:34.5406655Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:34.5407526Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:34.5408339Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:34.5409186Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:34.5409974Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:34.5410803Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:34.5411628Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:34.5412431Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:34.5413250Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:34.5414127Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:34.5414922Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:34.5415728Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:34.5416625Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:34.5417519Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:34.5418286Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:34.5419103Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:34.5419940Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:34.5420727Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:34.5421571Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:34.5422532Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:34.5423522Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:34.5424506Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:34.5425472Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:34.5426455Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:34.5427445Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:34.5428440Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:34.5429454Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:34.5430438Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:34.5431286Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:34.5432103Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:34.5432961Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:34.5433871Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:34.5434711Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:34.5435485Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:34.5436293Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:34.5437061Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:34.5437895Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:34.5438686Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:34.5439466Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:34.5440261Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:34.5441145Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:34.5441965Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:34.5442789Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:34.5443541Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:34.5444365Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:34.5445132Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:34.5445945Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:34.5446770Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:34.5447601Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:34.5448427Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:34.5449253Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:34.5449989Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:34.5450769Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:34.5451562Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:34.5452418Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:34.5453398Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:34.5454201Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:34.5454932Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:34.5455784Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:34.5456590Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:34.5457520Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:34.5458321Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:34.5459116Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:34.5460063Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:34.5460930Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:34.5461825Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:34.5462614Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:34.5463528Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:34.5464347Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:34.5465131Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:34.5465948Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:34.5466762Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:34.5467627Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:34.5468478Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:34.5469260Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:34.5470063Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:34.5470869Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:34.5471614Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:34.5472403Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:34.5473185Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:34.5474011Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:34.5474801Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:34.5475665Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.5476547Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.5477359Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:34.5478201Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:34.5479041Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:34.5479883Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:34.5480786Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:34.5481674Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:34.5482517Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:34.5483379Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:34.5484206Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:34.5485032Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:34.5485882Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:34.5486709Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:34.5487529Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:34.5488355Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:34.5489199Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:34.5490010Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:34.5490874Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:34.5491660Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:34.5492485Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:34.5493478Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:34.5494345Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:34.5495205Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:34.5496277Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:34.5497316Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:34.5498235Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:34.5499151Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:34.5499979Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:34.5500888Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:34.5501768Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:34.5502695Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:34.5503570Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:34.5504410Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:34.5505241Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:34.5506059Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:34.5506879Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:34.5507726Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:34.5508524Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:34.5509336Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:34.5510201Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:34.5511026Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:34.5511885Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:34.5512797Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:34.5513610Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:34.5514421Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:34.5515313Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:34.5516204Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:34.5517082Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:34.5517967Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:34.5518841Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:34.5519673Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:34.5520535Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:34.5521410Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:34.5522202Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:34.5523001Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:34.5523830Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:34.5524625Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:34.5525418Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:34.5526262Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:34.5527101Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:34.5528060Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:34.5528974Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.5529811Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:34.5530701Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:34.5531562Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:34.5532454Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:34.5533252Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:34.5534270Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:34.5535102Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:34.5535926Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:34.5536729Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:34.5537604Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:34.5538521Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:34.5539406Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:34.5540348Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:34.5541146Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:34.5542025Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:34.5542858Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:34.5543730Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:34.5544648Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:34.5545548Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:34.5546400Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:34.5547241Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:34.5548157Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:34.5549045Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:34.5549929Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:34.5550742Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:34.5551588Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:34.5552434Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:34.5553231Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:34.5554176Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:34.5555031Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:34.5555830Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:34.5556625Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:34.5557493Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:34.5558455Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:34.5559361Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:34.5560242Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:34.5561201Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:34.5562069Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:34.5562929Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:34.5563801Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:34.5564752Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:34.5565608Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:34.5566484Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:34.5567301Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:34.5568134Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:34.5569006Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:34.5569847Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:34.5570731Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:34.5571611Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:34.5572360Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:34.5573120Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:34.5573970Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:34.5574764Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:34.5575566Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:34.5576386Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:34.5577282Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:34.5578177Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:34.5579122Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:34.5580021Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:34.5580858Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:34.5581636Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:34.5582532Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:34.5583358Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:34.5584220Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:34.5585072Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:34.5585926Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:34.5586797Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:34.5587623Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:34.5588437Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:34.5589268Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:34.5590188Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:34.5591067Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:34.5591902Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:34.5592621Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:34.5593590Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:34.5594401Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:34.5595156Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:34.5595941Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:34.5596899Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:34.5597696Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:34.5598613Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:34.5599390Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:34.5600132Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:34.5600951Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:34.5601803Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:34.5602654Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:34.5603527Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:34.5604372Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:34.5605179Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:34.5605971Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:34.5606825Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:34.5607582Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:34.5608339Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:34.5609149Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:34.5609936Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:34.5610755Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:34.5611600Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:34.5612416Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:34.5613243Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:34.5614032Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:34.5614832Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:34.5616796Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:34.5619278Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:34.5620037Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:34.5620952Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:34.5621745Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:34.5622627Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:34.5623576Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:34.5624423Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:34.5625358Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:34.5626257Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:34.5627152Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:34.5628042Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:34.5628947Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:34.5629955Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:34.5630871Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:34.5631698Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:34.5632613Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:34.5633543Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:34.5634416Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:34.5635304Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:34.5636280Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:34.5637022Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:34.5637721Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:34.5638517Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:34.5639231Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:34.5639944Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:34.5640675Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:34.5641575Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:34.5642434Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:34.5643274Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:34.5644000Z V1204 11:20:24.459000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:34.5644653Z V1204 11:20:24.460000 99036 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.5645358Z I1204 11:20:24.460000 99036 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.5645877Z V1204 11:20:24.460000 99036 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:34.5646357Z V1204 11:20:24.460000 99036 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:34.5646876Z I1204 11:20:24.461000 99036 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 14 2025-12-04T11:24:34.5647636Z V1204 11:20:24.473000 99036 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:34.5648079Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.5648587Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:34.5649539Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.5650278Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.5651264Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.5652240Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.5653249Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.5654144Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.5654981Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:34.5655773Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:34.5656829Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:34.5657709Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:34.5658497Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_3: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.5659168Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.5659630Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.5660014Z V1204 11:20:24.478000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.5660580Z V1204 11:20:24.480000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:34.5661157Z V1204 11:20:24.481000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:34.5662156Z V1204 11:20:24.481000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.5662675Z V1204 11:20:24.482000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.5663685Z V1204 11:20:24.483000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.5664200Z V1204 11:20:24.483000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.5665087Z V1204 11:20:24.484000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.5665535Z V1204 11:20:24.485000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.5666485Z V1204 11:20:24.486000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:34.5666968Z V1204 11:20:24.486000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.5667852Z V1204 11:20:24.488000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:34.5668301Z V1204 11:20:24.488000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.5669219Z V1204 11:20:24.489000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:34.5669672Z V1204 11:20:24.489000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.5670517Z V1204 11:20:24.491000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:34.5670959Z V1204 11:20:24.491000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.5671784Z V1204 11:20:24.493000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_3 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.5672267Z V1204 11:20:24.493000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.5672840Z V1204 11:20:24.494000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.5673456Z V1204 11:20:24.494000 99036 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 14 2025-12-04T11:24:34.5674198Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:34.5674299Z warnings.warn( 2025-12-04T11:24:34.5674787Z V1204 11:20:24.495000 99036 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:34.5675012Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:34.5675138Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:34.5675256Z stats [('calls_captured', 1)] 2025-12-04T11:24:34.5675479Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:34.5676262Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_nodes', 3), ('pad_mm_bench', 3), ('pattern_matcher_count', 2), ('fxgraph_cache_miss', 1), ('decompose_mm', 1)] 2025-12-04T11:24:34.5676366Z graph_break [] 2025-12-04T11:24:34.5676589Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:34.5677036Z V1204 11:20:24.515000 99036 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpdpddhhqo 2025-12-04T11:24:34.5677473Z V1204 11:20:24.539000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.5677977Z V1204 11:20:24.539000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:34.5678914Z V1204 11:20:24.539000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.5679788Z V1204 11:20:24.539000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.5680544Z V1204 11:20:24.539000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:34.5680993Z V1204 11:20:24.539000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:34.5681418Z V1204 11:20:24.539000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.5682414Z V1204 11:20:24.539000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.5683452Z V1204 11:20:24.539000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, None); l_x_ = l_self_modules_linear_parameters_weight_ = None 2025-12-04T11:24:34.5683918Z V1204 11:20:24.539000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:34.5684331Z V1204 11:20:24.539000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.5684718Z V1204 11:20:24.539000 99036 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.5685462Z V1204 11:20:24.777000 99036 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:34.5686185Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v: 2025-12-04T11:24:34.5686886Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [trz4xcrmguerwj2nwxhu4t6zwtig5vb4aexoqw3sqeb3z2aw5kv] gm: GraphModule() 2025-12-04T11:24:34.5687171Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.5687441Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.5687706Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.5688176Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:34.5688991Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.5689813Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.5690549Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.5691112Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, permute) 2025-12-04T11:24:34.5691755Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_3 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.5692248Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, convert_element_type_1, permute_3) 2025-12-04T11:24:34.5692550Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.5693144Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:34.5695463Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.5698231Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.5698900Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:34.5699815Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:34.5700597Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:34.5701410Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:34.5702235Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:34.5702959Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:34.5703665Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:34.5704394Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:34.5705101Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:34.5705842Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [reylvtt3rqfvzq5b3fjekfvick4dxsjoog5svxckgogwqosgp6z] fx_kwargs[static_input_idxs]: [0] 2025-12-04T11:24:34.5706535Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[0]: 1 2025-12-04T11:24:34.5707405Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:34.5708197Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:34.5709356Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:34.5710201Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:34.5739109Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:34.5740188Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:34.5740962Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:34.5741733Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:34.5742585Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:34.5743423Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:34.5744154Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:34.5744917Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:34.5745693Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:34.5746515Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:34.5747366Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:34.5748210Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:34.5749063Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:34.5749944Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:34.5750788Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:34.5751678Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:34.5752538Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:34.5753370Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:34.5754382Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:34.5755241Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:34.5756083Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:34.5756818Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:34.5757653Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:34.5758425Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:34.5759131Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:34.5759935Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:34.5760678Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:34.5761436Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:34.5762261Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:34.5763048Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:34.5763827Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:34.5764596Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:34.5765399Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:34.5766173Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:34.5766958Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:34.5767717Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:34.5768510Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:34.5769364Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:34.5770125Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:34.5770933Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:34.5771729Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:34.5772503Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:34.5773259Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:34.5774067Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:34.5774875Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:34.5775696Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:34.5776489Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:34.5777427Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:34.5778244Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:34.5779002Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:34.5779749Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:34.5780567Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:34.5781508Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:34.5782314Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:34.5783100Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:34.5783934Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:34.5784672Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:34.5785897Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:34.5786723Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:34.5787586Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:34.5788472Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:34.5789286Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:34.5790103Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:34.5790941Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:34.5791828Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:34.5792619Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:34.5793557Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.5794421Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:34.5795380Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.5796372Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:34.5797310Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.5798199Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:34.5799069Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:34.5799827Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:34.5800582Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:34.5801458Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:34.5802323Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:34.5803124Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:34.5803950Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:34.5804733Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:34.5805571Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:34.5806438Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:34.5807359Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:34.5808227Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:34.5808996Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:34.5809834Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:34.5810750Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:34.5811562Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:34.5812332Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:34.5813214Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:34.5814113Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:34.5814983Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:34.5815861Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:34.5816653Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:34.5817568Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:34.5818287Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:34.5819103Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:34.5819916Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:34.5820702Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:34.5821547Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:34.5822379Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:34.5823205Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:34.5824002Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:34.5824737Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:34.5825540Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:34.5826481Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:34.5827412Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:34.5828331Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:34.5829152Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:34.5830011Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:34.5830847Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:34.5831689Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:34.5832597Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:34.5833488Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:34.5834268Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:34.5835112Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:34.5835930Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:34.5836764Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:34.5837536Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:34.5838321Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:34.5839160Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:34.5839932Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:34.5840712Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:34.5841516Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:34.5842373Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:34.5843197Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:34.5844061Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:34.5844841Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:34.5845738Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:34.5846532Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:34.5847383Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:34.5848201Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:34.5849011Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:34.5849762Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:34.5850594Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:34.5851364Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:34.5852195Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:34.5853033Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:34.5853871Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:34.5854690Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:34.5855539Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:34.5856298Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:34.5857248Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:34.5858105Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:34.5858950Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:34.5859752Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:34.5860577Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:34.5861372Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:34.5862144Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:34.5862910Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:34.5863682Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:34.5864418Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:34.5865232Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:34.5866105Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:34.5866937Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:34.5867756Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:34.5868535Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:34.5869304Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:34.5870111Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:34.5870916Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:34.5871751Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:34.5872612Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:34.5873437Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:34.5874299Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:34.5875103Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:34.5875917Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:34.5876747Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:34.5877561Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:34.5878342Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:34.5879218Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:34.5880003Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:34.5880853Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:34.5881714Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:34.5882562Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:34.5883310Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:34.5884121Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:34.5884919Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:34.5885703Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:34.5886537Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:34.5887483Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:34.5888490Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:34.5889463Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:34.5890476Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:34.5891445Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:34.5892435Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:34.5893467Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:34.5894441Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:34.5895435Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:34.5896410Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:34.5897374Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:34.5898190Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:34.5899165Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:34.5899995Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:34.5900773Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:34.5901564Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:34.5902315Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:34.5903130Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:34.5903909Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:34.5904708Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:34.5905488Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:34.5906421Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:34.5907253Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:34.5908062Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:34.5908824Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:34.5909671Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:34.5910460Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:34.5911223Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:34.5912053Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:34.5912902Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:34.5913719Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:34.5914582Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:34.5915307Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:34.5916103Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:34.5916887Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:34.5917706Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:34.5918666Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:34.5919459Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:34.5920200Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:34.5921031Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:34.5921848Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:34.5922730Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:34.5923531Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:34.5924337Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:34.5925309Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:34.5926189Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:34.5927029Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:34.5927825Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:34.5928728Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:34.5929547Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:34.5930367Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:34.5931181Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:34.5932017Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:34.5932871Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:34.5933708Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:34.5934450Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:34.5935249Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:34.5936054Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:34.5936803Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:34.5937662Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:34.5938477Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:34.5939322Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:34.5940103Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:34.5940967Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.5941901Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.5942669Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:34.5943526Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:34.5944353Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:34.5945242Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:34.5946101Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:34.5947049Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:34.5947876Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:34.5948738Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:34.5949550Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:34.5950371Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:34.5951198Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:34.5952012Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:34.5952843Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:34.5953651Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:34.5954525Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:34.5955350Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:34.5956192Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:34.5956993Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:34.5957827Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:34.5958728Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:34.5959579Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:34.5960444Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:34.5961394Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:34.5962333Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:34.5963263Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:34.5964159Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:34.5965008Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:34.5965862Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:34.5966711Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:34.5967632Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:34.5968497Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:34.5969346Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:34.5970163Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:34.5971026Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:34.5971831Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:34.5972688Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:34.5973492Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:34.5974306Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:34.5975140Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:34.5975956Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:34.5976830Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:34.5977792Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:34.5978644Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:34.5979439Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:34.5980327Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:34.5981236Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:34.5982059Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:34.5982932Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:34.5983799Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:34.5984641Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:34.5985486Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:34.5986361Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:34.5987178Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:34.5987975Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:34.5988814Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:34.5989624Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:34.5990428Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:34.5991222Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:34.5992069Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:34.5993017Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:34.5993921Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.5994795Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:34.5995672Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:34.5996692Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:34.5997543Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:34.5998353Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:34.5999325Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:34.6000149Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:34.6000988Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:34.6001775Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:34.6002680Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:34.6003574Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:34.6004471Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:34.6005405Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:34.6006245Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:34.6007096Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:34.6007923Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:34.6008816Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:34.6009854Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:34.6010738Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:34.6011615Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:34.6012463Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:34.6013371Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:34.6014226Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:34.6015100Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:34.6015900Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:34.6016759Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:34.6017651Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:34.6018459Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:34.6019441Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:34.6020299Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:34.6021113Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:34.6021929Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:34.6022822Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:34.6023740Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:34.6024658Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:34.6025531Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:34.6026492Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:34.6027393Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:34.6028254Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:34.6029145Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:34.6030066Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:34.6030932Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:34.6031776Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:34.6032602Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:34.6033423Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:34.6034298Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:34.6035194Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:34.6036069Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:34.6036967Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:34.6037701Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:34.6038505Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:34.6039317Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:34.6040108Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:34.6040928Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:34.6041757Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:34.6042552Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:34.6043454Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:34.6044416Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:34.6045293Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:34.6046100Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:34.6046874Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:34.6047737Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:34.6048573Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:34.6049429Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:34.6050291Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:34.6051164Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:34.6052052Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:34.6052870Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:34.6053679Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:34.6054558Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:34.6055429Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:34.6056319Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:34.6057228Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:34.6058004Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:34.6058934Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:34.6059777Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:34.6060544Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:34.6061320Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:34.6062112Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:34.6062901Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:34.6063775Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:34.6064540Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:34.6065283Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:34.6066121Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:34.6066959Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:34.6067854Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:34.6068725Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:34.6069583Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:34.6070428Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:34.6071221Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:34.6072040Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:34.6072775Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:34.6073544Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:34.6074339Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:34.6075174Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:34.6075983Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:34.6076824Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:34.6077658Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:34.6078427Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:34.6079197Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:34.6079991Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:34.6081965Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:34.6084443Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:34.6085216Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:34.6086125Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:34.6086962Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:34.6087790Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:34.6088745Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:34.6089594Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:34.6090521Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:34.6091725Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:34.6092718Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:34.6093663Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:34.6106362Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:34.6107534Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:34.6108461Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:34.6109301Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:34.6110267Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:34.6111194Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:34.6112228Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:34.6113104Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:34.6114072Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:34.6114821Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:34.6115577Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:34.6116334Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:34.6117045Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:34.6117769Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:34.6118485Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:34.6119417Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:34.6120320Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:34.6121164Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:34.6121898Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:34.6122559Z V1204 11:20:24.786000 99036 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.6123248Z I1204 11:20:24.787000 99036 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.6123735Z V1204 11:20:24.787000 99036 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:34.6124217Z V1204 11:20:24.787000 99036 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:34.6124734Z I1204 11:20:24.788000 99036 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 15 2025-12-04T11:24:34.6125498Z V1204 11:20:24.800000 99036 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:34.6125962Z V1204 11:20:24.805000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.6126459Z V1204 11:20:24.805000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:34.6127442Z V1204 11:20:24.805000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.6128182Z V1204 11:20:24.805000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.6129161Z V1204 11:20:24.805000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.6130170Z V1204 11:20:24.805000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.6131156Z V1204 11:20:24.805000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.6132059Z V1204 11:20:24.805000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.6132894Z V1204 11:20:24.805000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:34.6133731Z V1204 11:20:24.805000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:34.6138777Z V1204 11:20:24.805000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:34.6139623Z V1204 11:20:24.805000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:34.6140421Z V1204 11:20:24.805000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_3: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.6141063Z V1204 11:20:24.805000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.6141503Z V1204 11:20:24.805000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.6141890Z V1204 11:20:24.805000 99036 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.6142456Z V1204 11:20:24.808000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:34.6143030Z V1204 11:20:24.808000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:34.6144037Z V1204 11:20:24.809000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.6144548Z V1204 11:20:24.809000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.6145599Z V1204 11:20:24.810000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.6146096Z V1204 11:20:24.810000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.6146994Z V1204 11:20:24.811000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.6147443Z V1204 11:20:24.812000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.6148425Z V1204 11:20:24.815000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:34.6148879Z V1204 11:20:24.815000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.6149772Z V1204 11:20:24.816000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:34.6150220Z V1204 11:20:24.816000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.6151161Z V1204 11:20:24.817000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:34.6151628Z V1204 11:20:24.817000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.6152467Z V1204 11:20:24.819000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:34.6152898Z V1204 11:20:24.820000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.6153721Z V1204 11:20:24.822000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_3 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.6154166Z V1204 11:20:24.822000 99036 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.6154742Z V1204 11:20:24.823000 99036 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.6155327Z V1204 11:20:24.823000 99036 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 15 2025-12-04T11:24:34.6156065Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:34.6156162Z warnings.warn( 2025-12-04T11:24:34.6156639Z V1204 11:20:24.824000 99036 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:34.6157474Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-819d138cb6fe4c8f.xml - 2025-12-04T11:24:34.6157649Z =========================== short test summary info ============================ 2025-12-04T11:24:34.6158576Z FAILED [0.3257s] inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False - AssertionError: Scalars are not equal! 2025-12-04T11:24:34.6158617Z 2025-12-04T11:24:34.6158724Z Expected 1 but got 0. 2025-12-04T11:24:34.6158834Z Absolute difference: 1 2025-12-04T11:24:34.6158939Z Relative difference: 1.0 2025-12-04T11:24:34.6158944Z 2025-12-04T11:24:34.6159160Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:34.6160030Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:34.6160039Z 2025-12-04T11:24:34.6160302Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:34.6160488Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:24:34.6160699Z ==================== 1 failed, 14 passed, 2 rerun in 15.93s ==================== 2025-12-04T11:24:34.6160803Z --- Logging error --- 2025-12-04T11:24:34.6160927Z Traceback (most recent call last): 2025-12-04T11:24:34.6161233Z File "/opt/conda/envs/py_3.10/lib/python3.10/logging/__init__.py", line 1103, in emit 2025-12-04T11:24:34.6161355Z stream.write(msg + self.terminator) 2025-12-04T11:24:34.6161495Z ValueError: I/O operation on closed file. 2025-12-04T11:24:34.6161583Z Call stack: 2025-12-04T11:24:34.6162104Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/remote_cache.py", line 432, in dump_cache_stats 2025-12-04T11:24:34.6162247Z log.info("Cache Metrics:%s", out.getvalue()) 2025-12-04T11:24:34.6162408Z Message: 'Cache Metrics:%s' 2025-12-04T11:24:34.6163031Z Arguments: ('\n LocalAutotuneCache: {hit: 0, miss: 5, put: 9, exception: 0}\n backend:_LocalAutotuneCacheBackend: {hit: 0, miss: 5, put: 9, exception: 0}\n',) 2025-12-04T11:24:34.6163127Z Got exit code 1 2025-12-04T11:24:34.6163229Z Retrying single test... 2025-12-04T11:24:34.6163932Z Test results will be stored in test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-3d7372eabf0c20ef.xml 2025-12-04T11:24:34.6164091Z ============================= test session starts ============================== 2025-12-04T11:24:34.6164438Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:24:34.6164569Z cachedir: .pytest_cache 2025-12-04T11:24:34.6165094Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:24:34.6165219Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:24:34.6165345Z configfile: pytest.ini 2025-12-04T11:24:34.6165939Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:24:34.6187778Z collecting ... /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:52: PytestCollectionWarning: cannot collect test class 'TestDecomposeAddMM' because it has a __init__ constructor (from: test/inductor/test_decompose_mem_bound_mm.py) 2025-12-04T11:24:34.6187991Z class TestDecomposeAddMM(torch.nn.Module): 2025-12-04T11:24:34.6188148Z collected 37 items / 36 deselected / 1 selected 2025-12-04T11:24:34.6189125Z stepcurrent: skipping 14 already run items. Running only test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:34.6189241Z Running 1 items in this shard 2025-12-04T11:24:34.6189247Z 2025-12-04T11:24:34.6190092Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False ('RERUN', {'yellow': True}) [5.9306s] [100%] 2025-12-04T11:24:34.6190915Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False ('RERUN', {'yellow': True}) [0.3399s] [100%] 2025-12-04T11:24:34.6191725Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False FAILED [0.3368s] [100%] 2025-12-04T11:24:34.6191747Z 2025-12-04T11:24:34.6191891Z ==================================== RERUNS ==================================== 2025-12-04T11:24:34.6192359Z _ TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False _ 2025-12-04T11:24:34.6192495Z Traceback (most recent call last): 2025-12-04T11:24:34.6193087Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 275, in test_decompose_linear_mixed_precision 2025-12-04T11:24:34.6193193Z self.assertEqual( 2025-12-04T11:24:34.6193712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:34.6193882Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:34.6194434Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:34.6194639Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:34.6194773Z AssertionError: Scalars are not equal! 2025-12-04T11:24:34.6194779Z 2025-12-04T11:24:34.6194896Z Expected 1 but got 0. 2025-12-04T11:24:34.6195003Z Absolute difference: 1 2025-12-04T11:24:34.6195112Z Relative difference: 1.0 2025-12-04T11:24:34.6195118Z 2025-12-04T11:24:34.6195345Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:34.6196442Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:34.6196451Z 2025-12-04T11:24:34.6196735Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:34.6197025Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:34.6197139Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:34.6197266Z stats [('calls_captured', 1)] 2025-12-04T11:24:34.6198039Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_nodes', 3), ('pad_mm_bench', 3), ('pattern_matcher_count', 2), ('fxgraph_cache_miss', 1), ('decompose_mm', 1)] 2025-12-04T11:24:34.6198278Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:34.6198376Z graph_break [] 2025-12-04T11:24:34.6198595Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:34.6199105Z I1204 11:20:38.756000 99616 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:34.6199697Z I1204 11:20:39.001000 99616 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:34.6200149Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.6200640Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:34.6201576Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.6202439Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.6203179Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:34.6203688Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:34.6204103Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.6205083Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.6206111Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, None); l_x_ = l_self_modules_linear_parameters_weight_ = None 2025-12-04T11:24:34.6206623Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:34.6207055Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.6207430Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.6207995Z I1204 11:20:41.212000 99616 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:34.6208699Z V1204 11:20:41.699000 99616 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:34.6209457Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v: 2025-12-04T11:24:34.6210129Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [trz4xcrmguerwj2nwxhu4t6zwtig5vb4aexoqw3sqeb3z2aw5kv] gm: GraphModule() 2025-12-04T11:24:34.6210440Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.6210724Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.6210990Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.6211455Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:34.6212282Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.6213099Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.6213857Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.6214422Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, permute) 2025-12-04T11:24:34.6215063Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_3 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.6215554Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, convert_element_type_1, permute_3) 2025-12-04T11:24:34.6215850Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.6216428Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:34.6218813Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.6221207Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.6221853Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:34.6222769Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:34.6223483Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:34.6224340Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:34.6225119Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:34.6225869Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:34.6226575Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:34.6227291Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:34.6228011Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:34.6228752Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [reylvtt3rqfvzq5b3fjekfvick4dxsjoog5svxckgogwqosgp6z] fx_kwargs[static_input_idxs]: [0] 2025-12-04T11:24:34.6229449Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[0]: 1 2025-12-04T11:24:34.6230313Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:34.6231115Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:34.6232295Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:34.6233082Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:34.6262147Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:34.6263310Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:34.6264089Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:34.6264850Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:34.6265698Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:34.6266562Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:34.6267278Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:34.6268068Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:34.6268846Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:34.6269637Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:34.6270518Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:34.6271346Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:34.6272211Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:34.6273055Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:34.6273936Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:34.6274828Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:34.6275692Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:34.6276499Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:34.6277543Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:34.6278329Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:34.6279168Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:34.6279920Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:34.6280751Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:34.6281525Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:34.6282231Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:34.6283034Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:34.6283780Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:34.6284526Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:34.6285374Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:34.6286159Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:34.6286946Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:34.6287709Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:34.6288482Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:34.6289292Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:34.6290071Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:34.6290834Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:34.6291621Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:34.6292402Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:34.6293201Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:34.6294004Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:34.6294809Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:34.6295570Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:34.6296485Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:34.6297344Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:34.6298165Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:34.6298982Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:34.6299780Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:34.6300640Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:34.6301413Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:34.6302237Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:34.6302987Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:34.6303803Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:34.6304711Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:34.6305557Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:34.6306345Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:34.6307165Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:34.6307919Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:34.6309143Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:34.6309950Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:34.6310845Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:34.6311736Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:34.6312547Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:34.6313352Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:34.6314212Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:34.6315083Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:34.6315893Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:34.6316822Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.6317666Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:34.6318677Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.6319496Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:34.6320428Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.6321279Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:34.6322167Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:34.6322922Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:34.6323679Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:34.6324511Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:34.6325401Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:34.6326191Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:34.6326997Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:34.6327789Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:34.6328622Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:34.6329499Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:34.6330422Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:34.6331286Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:34.6332049Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:34.6332875Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:34.6333762Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:34.6334600Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:34.6335389Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:34.6336277Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:34.6337244Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:34.6338116Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:34.6338994Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:34.6339804Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:34.6340593Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:34.6341364Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:34.6342184Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:34.6342994Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:34.6343774Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:34.6344597Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:34.6345434Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:34.6346253Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:34.6347072Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:34.6347806Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:34.6348612Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:34.6349520Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:34.6350492Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:34.6351418Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:34.6352220Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:34.6353048Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:34.6353911Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:34.6354758Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:34.6355657Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:34.6356510Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:34.6357318Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:34.6358154Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:34.6358986Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:34.6359808Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:34.6360583Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:34.6361369Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:34.6362213Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:34.6362986Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:34.6363756Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:34.6364562Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:34.6365380Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:34.6366268Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:34.6367130Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:34.6367910Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:34.6368788Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:34.6369597Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:34.6370480Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:34.6371291Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:34.6372086Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:34.6372870Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:34.6373655Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:34.6374479Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:34.6375304Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:34.6376152Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:34.6377034Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:34.6377863Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:34.6378719Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:34.6379481Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:34.6380364Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:34.6381199Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:34.6382058Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:34.6382848Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:34.6383689Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:34.6384451Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:34.6385261Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:34.6386038Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:34.6386794Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:34.6387552Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:34.6388327Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:34.6389217Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:34.6390079Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:34.6390905Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:34.6391694Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:34.6392465Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:34.6393278Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:34.6394075Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:34.6394926Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:34.6395789Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:34.6396845Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:34.6397672Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:34.6398553Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:34.6399370Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:34.6400188Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:34.6400989Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:34.6401808Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:34.6402688Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:34.6403479Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:34.6404300Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:34.6405204Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:34.6406049Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:34.6406814Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:34.6407626Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:34.6408425Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:34.6409216Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:34.6410053Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:34.6411011Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:34.6412011Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:34.6412993Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:34.6413954Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:34.6414969Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:34.6415966Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:34.6417029Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:34.6418045Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:34.6419057Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:34.6419894Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:34.6420715Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:34.6421577Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:34.6422485Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:34.6423360Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:34.6424137Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:34.6424922Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:34.6425677Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:34.6426473Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:34.6427269Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:34.6428052Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:34.6428843Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:34.6429729Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:34.6430587Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:34.6431394Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:34.6432148Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:34.6432963Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:34.6433770Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:34.6434550Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:34.6435377Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:34.6436224Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:34.6437049Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:34.6437863Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:34.6438639Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:34.6439417Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:34.6440228Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:34.6441050Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:34.6442007Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:34.6442807Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:34.6443552Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:34.6444386Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:34.6445207Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:34.6446062Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:34.6446906Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:34.6447723Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:34.6448655Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:34.6449573Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:34.6450426Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:34.6451217Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:34.6452118Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:34.6452940Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:34.6453732Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:34.6454580Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:34.6455411Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:34.6456266Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:34.6457167Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:34.6457906Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:34.6458713Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:34.6459521Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:34.6460270Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:34.6461067Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:34.6461840Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:34.6462732Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:34.6463516Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:34.6464372Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.6465289Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.6466057Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:34.6466924Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:34.6467758Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:34.6468620Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:34.6469513Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:34.6470440Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:34.6471265Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:34.6472130Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:34.6472938Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:34.6473766Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:34.6474605Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:34.6475414Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:34.6476244Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:34.6477057Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:34.6477892Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:34.6478752Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:34.6479595Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:34.6480387Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:34.6481184Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:34.6482709Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:34.6483571Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:34.6484432Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:34.6485351Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:34.6486279Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:34.6487241Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:34.6488138Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:34.6488985Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:34.6489834Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:34.6490657Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:34.6491589Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:34.6492454Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:34.6493303Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:34.6494118Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:34.6494985Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:34.6495792Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:34.6496817Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:34.6497663Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:34.6498570Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:34.6499409Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:34.6500219Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:34.6501085Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:34.6502005Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:34.6502830Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:34.6503669Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:34.6504560Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:34.6505457Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:34.6506281Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:34.6507144Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:34.6508010Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:34.6508859Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:34.6509710Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:34.6510596Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:34.6511436Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:34.6512234Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:34.6513083Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:34.6513865Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:34.6514713Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:34.6515521Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:34.6516363Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:34.6517327Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:34.6518310Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.6519168Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:34.6520079Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:34.6520959Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:34.6521815Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:34.6522634Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:34.6523614Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:34.6524445Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:34.6525289Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:34.6526071Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:34.6526909Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:34.6527845Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:34.6528746Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:34.6529670Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:34.6530467Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:34.6531344Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:34.6532179Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:34.6533067Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:34.6533972Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:34.6534886Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:34.6535754Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:34.6536599Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:34.6537574Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:34.6538434Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:34.6539299Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:34.6540106Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:34.6540972Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:34.6541800Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:34.6542607Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:34.6543547Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:34.6544442Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:34.6545255Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:34.6546042Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:34.6546956Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:34.6547873Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:34.6548788Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:34.6549656Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:34.6550615Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:34.6551474Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:34.6552367Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:34.6553254Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:34.6554166Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:34.6555043Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:34.6555881Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:34.6556711Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:34.6557533Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:34.6558408Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:34.6559268Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:34.6560196Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:34.6561089Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:34.6561826Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:34.6562590Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:34.6563426Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:34.6564226Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:34.6565044Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:34.6565849Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:34.6566676Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:34.6567549Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:34.6568533Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:34.6569412Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:34.6570210Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:34.6570985Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:34.6571845Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:34.6572673Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:34.6573532Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:34.6574390Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:34.6575230Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:34.6576141Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:34.6577038Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:34.6577853Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:34.6578698Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:34.6579601Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:34.6580490Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:34.6581313Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:34.6582044Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:34.6583004Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:34.6583844Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:34.6584614Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:34.6585390Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:34.6586174Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:34.6586963Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:34.6587837Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:34.6588596Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:34.6589339Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:34.6590174Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:34.6591009Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:34.6591906Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:34.6592771Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:34.6593629Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:34.6594414Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:34.6595227Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:34.6596197Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:34.6596940Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:34.6597717Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:34.6598531Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:34.6599331Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:34.6600189Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:34.6601037Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:34.6601869Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:34.6602631Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:34.6603400Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:34.6604195Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:34.6606169Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:34.6608620Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:34.6609394Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:34.6610296Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:34.6611141Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:34.6611972Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:34.6612929Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:34.6613783Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:34.6614688Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:34.6615648Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:34.6616584Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:34.6617553Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:34.6618450Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:34.6619449Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:34.6620335Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:34.6621155Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:34.6622087Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:34.6623001Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:34.6623896Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:34.6624827Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:34.6625808Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:34.6626543Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:34.6627247Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:34.6628030Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:34.6628744Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:34.6629470Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:34.6630184Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:34.6631092Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:34.6631925Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:34.6632800Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:34.6633522Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:34.6634173Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.6634854Z I1204 11:20:41.710000 99616 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.6635331Z V1204 11:20:41.710000 99616 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:34.6635827Z V1204 11:20:41.710000 99616 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:34.6636326Z I1204 11:20:41.711000 99616 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:34.6637082Z V1204 11:20:41.750000 99616 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:34.6637533Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.6638025Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:34.6638985Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.6639745Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.6640735Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.6641690Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.6642707Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.6643611Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.6644436Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:34.6645228Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:34.6646270Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:34.6647137Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:34.6647912Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_3: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.6648558Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.6648978Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.6649365Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.6649948Z V1204 11:20:41.758000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:34.6650508Z V1204 11:20:41.759000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:34.6651529Z V1204 11:20:41.760000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.6652034Z V1204 11:20:41.760000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.6653076Z V1204 11:20:41.791000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.6653576Z V1204 11:20:41.792000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.6654457Z V1204 11:20:41.794000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.6654916Z V1204 11:20:41.794000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.6655880Z V1204 11:20:41.800000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:34.6656343Z V1204 11:20:41.800000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.6657308Z V1204 11:20:41.802000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:34.6657774Z V1204 11:20:41.802000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.6658683Z V1204 11:20:41.803000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:34.6659145Z V1204 11:20:41.803000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.6659995Z V1204 11:20:41.806000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:34.6660455Z V1204 11:20:41.806000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.6661302Z V1204 11:20:41.809000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_3 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.6661744Z V1204 11:20:41.809000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.6662336Z V1204 11:20:41.810000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.6662910Z V1204 11:20:41.810000 99616 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:34.6663651Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:34.6663767Z warnings.warn( 2025-12-04T11:24:34.6664247Z V1204 11:20:41.811000 99616 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:34.6664734Z _ TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False _ 2025-12-04T11:24:34.6664858Z Traceback (most recent call last): 2025-12-04T11:24:34.6665445Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 275, in test_decompose_linear_mixed_precision 2025-12-04T11:24:34.6665566Z self.assertEqual( 2025-12-04T11:24:34.6666032Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:34.6666200Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:34.6666780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:34.6666990Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:34.6667137Z AssertionError: Scalars are not equal! 2025-12-04T11:24:34.6667145Z 2025-12-04T11:24:34.6667251Z Expected 1 but got 0. 2025-12-04T11:24:34.6667360Z Absolute difference: 1 2025-12-04T11:24:34.6667484Z Relative difference: 1.0 2025-12-04T11:24:34.6667489Z 2025-12-04T11:24:34.6667707Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:34.6668580Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:34.6668628Z 2025-12-04T11:24:34.6668900Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:34.6669124Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:34.6669254Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:34.6669369Z stats [('calls_captured', 1)] 2025-12-04T11:24:34.6670138Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_nodes', 3), ('pad_mm_bench', 3), ('pattern_matcher_count', 2), ('fxgraph_cache_miss', 1), ('decompose_mm', 1)] 2025-12-04T11:24:34.6670371Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:34.6670471Z graph_break [] 2025-12-04T11:24:34.6670749Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:34.6671246Z I1204 11:20:38.756000 99616 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:34.6671838Z I1204 11:20:39.001000 99616 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:34.6672315Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.6672802Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:34.6673750Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.6674598Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.6675344Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:34.6675810Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:34.6676226Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.6677206Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.6678233Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, None); l_x_ = l_self_modules_linear_parameters_weight_ = None 2025-12-04T11:24:34.6678715Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:34.6679159Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.6679537Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.6680103Z I1204 11:20:41.212000 99616 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:34.6680813Z V1204 11:20:41.699000 99616 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:34.6681573Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v: 2025-12-04T11:24:34.6682229Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [trz4xcrmguerwj2nwxhu4t6zwtig5vb4aexoqw3sqeb3z2aw5kv] gm: GraphModule() 2025-12-04T11:24:34.6682525Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.6682792Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.6683058Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.6683535Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:34.6684347Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.6685201Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.6685974Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.6686537Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, permute) 2025-12-04T11:24:34.6687172Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_3 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.6687662Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, convert_element_type_1, permute_3) 2025-12-04T11:24:34.6687965Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.6688528Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:34.6690838Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.6693157Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.6693842Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:34.6694746Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:34.6695459Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:34.6696495Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:34.6697357Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:34.6698081Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:34.6698793Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:34.6699523Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:34.6700281Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:34.6701024Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [reylvtt3rqfvzq5b3fjekfvick4dxsjoog5svxckgogwqosgp6z] fx_kwargs[static_input_idxs]: [0] 2025-12-04T11:24:34.6701768Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[0]: 1 2025-12-04T11:24:34.6702639Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:34.6703436Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:34.6704594Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:34.6705394Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:34.6734333Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:34.6735342Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:34.6736116Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:34.6736951Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:34.6737847Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:34.6738714Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:34.6739440Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:34.6740210Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:34.6740987Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:34.6741776Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:34.6742636Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:34.6743475Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:34.6744329Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:34.6745188Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:34.6746028Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:34.6746943Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:34.6747802Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:34.6748612Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:34.6749681Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:34.6750443Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:34.6751285Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:34.6752025Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:34.6752858Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:34.6753659Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:34.6754398Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:34.6755198Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:34.6755939Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:34.6756702Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:34.6757529Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:34.6758292Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:34.6759078Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:34.6759843Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:34.6760626Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:34.6761403Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:34.6762221Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:34.6762982Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:34.6763773Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:34.6764561Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:34.6765356Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:34.6766137Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:34.6766943Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:34.6767724Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:34.6768481Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:34.6769320Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:34.6770165Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:34.6770985Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:34.6771781Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:34.6772644Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:34.6773428Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:34.6774186Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:34.6774925Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:34.6775737Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:34.6776659Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:34.6777536Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:34.6778353Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:34.6779197Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:34.6779938Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:34.6781241Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:34.6782034Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:34.6782897Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:34.6783792Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:34.6784600Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:34.6785448Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:34.6786320Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:34.6787205Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:34.6787993Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:34.6788929Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.6789767Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:34.6790716Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.6791522Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:34.6792449Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.6793271Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:34.6794165Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:34.6794922Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:34.6795679Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:34.6796647Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:34.6797568Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:34.6798330Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:34.6799148Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:34.6799920Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:34.6800753Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:34.6801662Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:34.6802624Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:34.6803489Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:34.6804253Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:34.6805086Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:34.6805955Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:34.6806769Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:34.6807535Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:34.6808421Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:34.6809292Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:34.6810202Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:34.6811074Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:34.6811861Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:34.6812662Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:34.6813407Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:34.6814188Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:34.6814999Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:34.6815782Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:34.6816613Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:34.6817524Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:34.6818387Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:34.6819183Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:34.6819916Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:34.6820715Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:34.6821624Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:34.6822554Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:34.6823484Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:34.6824298Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:34.6825120Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:34.6825951Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:34.6826828Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:34.6827724Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:34.6828582Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:34.6829389Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:34.6830212Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:34.6831036Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:34.6831874Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:34.6832644Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:34.6833467Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:34.6834339Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:34.6835116Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:34.6835895Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:34.6836691Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:34.6837516Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:34.6838345Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:34.6839201Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:34.6839992Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:34.6840867Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:34.6841659Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:34.6842547Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:34.6843363Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:34.6844143Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:34.6844899Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:34.6845715Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:34.6846494Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:34.6847328Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:34.6848167Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:34.6849058Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:34.6849879Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:34.6850763Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:34.6851526Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:34.6852410Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:34.6853263Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:34.6854073Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:34.6854875Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:34.6855702Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:34.6856463Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:34.6857295Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:34.6858101Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:34.6858874Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:34.6859612Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:34.6860395Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:34.6861268Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:34.6862114Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:34.6862937Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:34.6863722Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:34.6864522Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:34.6865328Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:34.6866178Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:34.6867008Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:34.6867874Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:34.6868696Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:34.6869522Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:34.6870327Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:34.6871138Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:34.6871967Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:34.6872750Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:34.6873530Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:34.6874433Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:34.6875223Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:34.6876041Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:34.6876931Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:34.6877752Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:34.6878502Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:34.6879311Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:34.6880107Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:34.6880923Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:34.6881781Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:34.6882730Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:34.6883739Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:34.6884706Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:34.6885679Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:34.6886642Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:34.6887632Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:34.6888636Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:34.6889607Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:34.6890643Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:34.6891475Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:34.6892303Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:34.6893138Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:34.6894058Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:34.6894888Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:34.6895661Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:34.6896615Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:34.6897495Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:34.6898362Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:34.6899140Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:34.6899934Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:34.6900712Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:34.6901603Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:34.6902440Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:34.6903254Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:34.6904019Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:34.6904824Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:34.6905614Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:34.6906425Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:34.6907260Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:34.6908106Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:34.6908892Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:34.6909761Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:34.6910494Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:34.6911286Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:34.6912077Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:34.6912929Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:34.6913888Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:34.6914712Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:34.6915460Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:34.6916293Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:34.6917118Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:34.6917970Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:34.6918773Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:34.6919586Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:34.6920519Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:34.6921400Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:34.6922280Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:34.6923077Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:34.6923980Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:34.6924768Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:34.6925584Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:34.6926399Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:34.6927229Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:34.6928081Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:34.6928954Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:34.6929690Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:34.6930523Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:34.6931322Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:34.6932065Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:34.6932862Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:34.6933633Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:34.6934477Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:34.6935258Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:34.6936120Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.6937054Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.6937819Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:34.6938719Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:34.6939550Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:34.6940407Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:34.6941290Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:34.6942194Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:34.6943023Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:34.6943879Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:34.6944721Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:34.6945531Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:34.6946391Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:34.6947201Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:34.6948025Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:34.6948833Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:34.6949674Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:34.6950506Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:34.6951344Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:34.6952136Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:34.6952944Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:34.6953840Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:34.6954731Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:34.6955598Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:34.6956515Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:34.6957438Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:34.6958371Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:34.6959272Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:34.6960121Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:34.6960996Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:34.6961837Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:34.6962782Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:34.6963654Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:34.6964504Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:34.6965322Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:34.6966155Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:34.6966953Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:34.6967813Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:34.6968591Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:34.6969413Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:34.6970280Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:34.6971090Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:34.6971957Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:34.6972824Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:34.6973679Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:34.6974481Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:34.6975378Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:34.6976283Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:34.6977194Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:34.6978059Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:34.6978960Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:34.6979807Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:34.6980655Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:34.6981534Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:34.6982334Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:34.6983137Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:34.6983984Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:34.6984765Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:34.6985568Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:34.6986406Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:34.6987261Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:34.6988200Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:34.6989075Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.6989969Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:34.6990849Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:34.6991722Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:34.6992570Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:34.6993408Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:34.6994413Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:34.6995237Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:34.6996224Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:34.6997012Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:34.6997846Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:34.6998746Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:34.6999653Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:34.7000580Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:34.7001378Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:34.7002224Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:34.7003118Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:34.7004007Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:34.7004915Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:34.7005846Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:34.7006679Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:34.7007524Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:34.7008424Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:34.7009322Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:34.7010188Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:34.7011027Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:34.7011884Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:34.7012716Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:34.7013519Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:34.7014451Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:34.7015308Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:34.7016121Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:34.7016962Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:34.7017853Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:34.7018810Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:34.7019725Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:34.7020593Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:34.7021527Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:34.7022437Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:34.7023301Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:34.7024190Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:34.7025100Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:34.7025996Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:34.7026862Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:34.7027691Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:34.7028508Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:34.7029380Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:34.7030247Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:34.7031118Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:34.7032006Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:34.7032744Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:34.7033516Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:34.7034319Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:34.7035226Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:34.7036044Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:34.7036849Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:34.7037645Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:34.7038545Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:34.7039504Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:34.7040385Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:34.7041186Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:34.7041996Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:34.7042891Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:34.7043731Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:34.7044577Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:34.7045438Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:34.7046276Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:34.7047165Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:34.7047978Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:34.7048789Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:34.7049656Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:34.7050526Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:34.7051461Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:34.7052290Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:34.7053022Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:34.7053983Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:34.7054800Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:34.7055573Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:34.7056345Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:34.7057193Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:34.7058024Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:34.7058926Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:34.7059688Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:34.7060430Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:34.7061267Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:34.7062108Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:34.7062972Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:34.7063840Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:34.7064702Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:34.7065486Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:34.7066276Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:34.7067138Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:34.7067880Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:34.7068649Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:34.7069405Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:34.7070232Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:34.7071049Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:34.7071890Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:34.7072721Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:34.7073523Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:34.7074287Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:34.7075115Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:34.7077092Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:34.7079510Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:34.7080287Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:34.7081195Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:34.7082003Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:34.7082869Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:34.7083819Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:34.7084669Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:34.7085567Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:34.7086508Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:34.7087399Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:34.7088308Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:34.7089197Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:34.7090223Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:34.7091126Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:34.7091948Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:34.7092879Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:34.7093790Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:34.7094688Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:34.7095568Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:34.7096676Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:34.7097465Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:34.7098181Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:34.7098919Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:34.7099684Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:34.7100411Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:34.7101125Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:34.7102001Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:34.7102872Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:34.7103722Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:34.7104449Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:34.7105103Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.7105827Z I1204 11:20:41.710000 99616 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.7106309Z V1204 11:20:41.710000 99616 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:34.7106858Z V1204 11:20:41.710000 99616 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:34.7107358Z I1204 11:20:41.711000 99616 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:34.7108118Z V1204 11:20:41.750000 99616 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:34.7108569Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.7109060Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:34.7110021Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.7110756Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.7111742Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.7112695Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.7113683Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.7114615Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.7115447Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:34.7116215Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:34.7117284Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:34.7118118Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:34.7118888Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_3: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.7119566Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.7119986Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.7120401Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.7120981Z V1204 11:20:41.758000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:34.7121537Z V1204 11:20:41.759000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:34.7122550Z V1204 11:20:41.760000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.7123056Z V1204 11:20:41.760000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.7124081Z V1204 11:20:41.791000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.7124580Z V1204 11:20:41.792000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.7125462Z V1204 11:20:41.794000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.7125921Z V1204 11:20:41.794000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.7126860Z V1204 11:20:41.800000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:34.7127366Z V1204 11:20:41.800000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.7128241Z V1204 11:20:41.802000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:34.7128703Z V1204 11:20:41.802000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.7129607Z V1204 11:20:41.803000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:34.7130066Z V1204 11:20:41.803000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.7130923Z V1204 11:20:41.806000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:34.7131353Z V1204 11:20:41.806000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.7132191Z V1204 11:20:41.809000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_3 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.7132633Z V1204 11:20:41.809000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.7133254Z V1204 11:20:41.810000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.7133828Z V1204 11:20:41.810000 99616 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:34.7134595Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:34.7134710Z warnings.warn( 2025-12-04T11:24:34.7135185Z V1204 11:20:41.811000 99616 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:34.7135420Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:34.7135535Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:34.7135650Z stats [('calls_captured', 1)] 2025-12-04T11:24:34.7135887Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:34.7136658Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_nodes', 3), ('pad_mm_bench', 3), ('pattern_matcher_count', 2), ('fxgraph_cache_miss', 1), ('decompose_mm', 1)] 2025-12-04T11:24:34.7136757Z graph_break [] 2025-12-04T11:24:34.7137049Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:34.7137490Z V1204 11:20:42.045000 99616 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpvknbptgu 2025-12-04T11:24:34.7137940Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.7138427Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:34.7139360Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.7140225Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.7141019Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:34.7141485Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:34.7141898Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.7142881Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.7143935Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, None); l_x_ = l_self_modules_linear_parameters_weight_ = None 2025-12-04T11:24:34.7144411Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:34.7144836Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.7145213Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.7145934Z V1204 11:20:42.319000 99616 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:34.7146683Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v: 2025-12-04T11:24:34.7147385Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [trz4xcrmguerwj2nwxhu4t6zwtig5vb4aexoqw3sqeb3z2aw5kv] gm: GraphModule() 2025-12-04T11:24:34.7147658Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.7147925Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.7148205Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.7148667Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:34.7149482Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.7150303Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.7151050Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.7151626Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, permute) 2025-12-04T11:24:34.7152249Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_3 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.7152750Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, convert_element_type_1, permute_3) 2025-12-04T11:24:34.7153043Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.7153601Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:34.7155927Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.7158297Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.7158938Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:34.7159842Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:34.7160571Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:34.7161421Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:34.7162241Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:34.7162950Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:34.7163679Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:34.7164398Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:34.7165105Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:34.7165859Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [reylvtt3rqfvzq5b3fjekfvick4dxsjoog5svxckgogwqosgp6z] fx_kwargs[static_input_idxs]: [0] 2025-12-04T11:24:34.7166541Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[0]: 1 2025-12-04T11:24:34.7167424Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:34.7168207Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:34.7169350Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:34.7170163Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:34.7199308Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:34.7200414Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:34.7201176Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:34.7201955Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:34.7202793Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:34.7203641Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:34.7204352Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:34.7205117Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:34.7205904Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:34.7206676Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:34.7207585Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:34.7208411Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:34.7209275Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:34.7210149Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:34.7210991Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:34.7211902Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:34.7212746Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:34.7213565Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:34.7214596Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:34.7215395Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:34.7216223Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:34.7217016Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:34.7217863Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:34.7218622Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:34.7219339Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:34.7220132Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:34.7220889Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:34.7221637Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:34.7222462Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:34.7223271Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:34.7224045Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:34.7224819Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:34.7225578Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:34.7226404Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:34.7227184Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:34.7227936Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:34.7228743Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:34.7229543Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:34.7230326Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:34.7231124Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:34.7231946Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:34.7232705Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:34.7233458Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:34.7234280Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:34.7235080Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:34.7235917Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:34.7236696Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:34.7237570Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:34.7238337Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:34.7239125Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:34.7239888Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:34.7240688Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:34.7241617Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:34.7242443Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:34.7243248Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:34.7244073Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:34.7244817Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:34.7246039Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:34.7246861Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:34.7247735Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:34.7248606Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:34.7249428Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:34.7250227Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:34.7251084Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:34.7251956Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:34.7252752Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:34.7253700Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.7254512Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:34.7255509Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.7256307Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:34.7257340Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.7258187Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:34.7259058Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:34.7259827Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:34.7260576Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:34.7261450Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:34.7262287Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:34.7263083Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:34.7263885Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:34.7264665Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:34.7265511Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:34.7266368Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:34.7267302Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:34.7268152Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:34.7268927Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:34.7269761Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:34.7270666Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:34.7271478Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:34.7272249Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:34.7273145Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:34.7274031Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:34.7274915Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:34.7275775Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:34.7276578Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:34.7277402Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:34.7278115Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:34.7278935Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:34.7279732Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:34.7280523Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:34.7281346Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:34.7282181Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:34.7282998Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:34.7283796Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:34.7284540Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:34.7285326Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:34.7286244Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:34.7287189Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:34.7288121Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:34.7288928Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:34.7289776Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:34.7290632Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:34.7291461Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:34.7292374Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:34.7293250Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:34.7294043Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:34.7294956Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:34.7295774Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:34.7296766Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:34.7297599Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:34.7298407Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:34.7299240Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:34.7300032Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:34.7300802Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:34.7301606Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:34.7302432Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:34.7303332Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:34.7304208Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:34.7304977Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:34.7305901Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:34.7306678Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:34.7307549Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:34.7308357Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:34.7309143Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:34.7309957Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:34.7310774Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:34.7311556Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:34.7312375Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:34.7313230Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:34.7314057Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:34.7314878Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:34.7315738Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:34.7316486Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:34.7317384Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:34.7318227Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:34.7319074Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:34.7319864Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:34.7320691Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:34.7321465Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:34.7322249Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:34.7323033Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:34.7323790Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:34.7324538Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:34.7325336Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:34.7326183Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:34.7327060Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:34.7327869Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:34.7328660Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:34.7329420Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:34.7330241Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:34.7331041Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:34.7331875Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:34.7332746Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:34.7333555Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:34.7334399Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:34.7335213Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:34.7336043Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:34.7336914Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:34.7337738Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:34.7338531Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:34.7339400Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:34.7340208Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:34.7342004Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:34.7343840Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:34.7345685Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:34.7347385Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:34.7349094Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:34.7350848Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:34.7352555Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:34.7354335Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:34.7356241Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:34.7358321Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:34.7360427Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:34.7362526Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:34.7364596Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:34.7366698Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:34.7368843Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:34.7370984Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:34.7373090Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:34.7375081Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:34.7376934Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:34.7378767Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:34.7380705Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:34.7382607Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:34.7384371Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:34.7386078Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:34.7387754Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:34.7389457Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:34.7391183Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:34.7392913Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:34.7394640Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:34.7396619Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:34.7398581Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:34.7400354Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:34.7402082Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:34.7403806Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:34.7405602Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:34.7407302Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:34.7409040Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:34.7410861Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:34.7412687Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:34.7414428Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:34.7416179Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:34.7417877Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:34.7419587Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:34.7421348Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:34.7423260Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:34.7425153Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:34.7426801Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:34.7428513Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:34.7430303Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:34.7432167Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:34.7433966Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:34.7435704Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:34.7437590Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:34.7439580Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:34.7492133Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:34.7494032Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:34.7495914Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:34.7498177Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:34.7499909Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:34.7501817Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:34.7503600Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:34.7505412Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:34.7507239Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:34.7508954Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:34.7510638Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:34.7512361Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:34.7514031Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:34.7515715Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:34.7517409Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:34.7519205Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:34.7520960Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:34.7522729Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.7524660Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.7526468Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:34.7528230Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:34.7530056Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:34.7531877Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:34.7533772Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:34.7535712Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:34.7537639Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:34.7539474Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:34.7541265Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:34.7543014Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:34.7544810Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:34.7546581Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:34.7548350Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:34.7550114Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:34.7551900Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:34.7553726Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:34.7555536Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:34.7557307Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:34.7559055Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:34.7560892Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:34.7562778Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:34.7564640Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:34.7566594Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:34.7568537Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:34.7570521Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:34.7572465Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:34.7574345Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:34.7576188Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:34.7578068Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:34.7579973Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:34.7581909Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:34.7583794Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:34.7585598Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:34.7587410Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:34.7589186Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:34.7590972Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:34.7592743Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:34.7594501Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:34.7596460Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:34.7598240Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:34.7600054Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:34.7601994Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:34.7603817Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:34.7605601Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:34.7607432Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:34.7609362Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:34.7611226Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:34.7613053Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:34.7614899Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:34.7616747Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:34.7618633Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:34.7620541Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:34.7622396Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:34.7624127Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:34.7625886Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:34.7627651Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:34.7629415Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:34.7631158Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:34.7632926Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:34.7634861Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:34.7636856Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.7638744Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:34.7640597Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:34.7642481Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:34.7644322Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:34.7646117Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:34.7648046Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:34.7649991Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:34.7651797Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:34.7653539Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:34.7655289Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:34.7657258Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:34.7659189Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:34.7661155Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:34.7663044Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:34.7664818Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:34.7666620Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:34.7668475Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:34.7670402Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:34.7672370Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:34.7674247Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:34.7676069Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:34.7677964Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:34.7679871Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:34.7681701Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:34.7683517Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:34.7685307Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:34.7687138Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:34.7688909Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:34.7690795Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:34.7692721Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:34.7694523Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:34.7696423Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:34.7698358Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:34.7700313Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:34.7702268Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:34.7704192Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:34.7706200Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:34.7708184Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:34.7710048Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:34.7711911Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:34.7713838Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:34.7715767Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:34.7717614Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:34.7719423Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:34.7721194Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:34.7723041Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:34.7724917Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:34.7726833Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:34.7728733Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:34.7730477Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:34.7732151Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:34.7733864Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:34.7735617Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:34.7737419Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:34.7739162Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:34.7740930Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:34.7742763Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:34.7744722Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:34.7746687Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:34.7748507Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:34.7750209Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:34.7751968Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:34.7753798Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:34.7755613Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:34.7757450Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:34.7759275Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:34.7761161Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:34.7762981Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:34.7764768Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:34.7766584Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:34.7768460Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:34.7770341Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:34.7772191Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:34.7773889Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:34.7775724Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:34.7777695Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:34.7779400Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:34.7781077Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:34.7782776Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:34.7784489Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:34.7786285Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:34.7788047Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:34.7789701Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:34.7791424Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:34.7793243Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:34.7795124Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:34.7797187Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:34.7799059Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:34.7800832Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:34.7802646Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:34.7804400Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:34.7806076Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:34.7807721Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:34.7809435Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:34.7811113Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:34.7812895Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:34.7814675Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:34.7816489Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:34.7818286Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:34.7819952Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:34.7821645Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:34.7824530Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:34.7829176Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:34.7832469Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:34.7834278Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:34.7836162Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:34.7837923Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:34.7839837Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:34.7841770Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:34.7843674Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:34.7845614Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:34.7847567Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:34.7849482Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:34.7851416Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:34.7853421Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:34.7855429Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:34.7857350Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:34.7859234Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:34.7861208Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:34.7863135Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:34.7865069Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:34.7867060Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:34.7868895Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:34.7870498Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:34.7872057Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:34.7873655Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:34.7875218Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:34.7876788Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:34.7878554Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:34.7880424Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:34.7882244Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:34.7883936Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:34.7885444Z V1204 11:20:42.329000 99616 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.7886909Z I1204 11:20:42.329000 99616 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.7888191Z V1204 11:20:42.329000 99616 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:34.7889296Z V1204 11:20:42.330000 99616 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:34.7890414Z I1204 11:20:42.330000 99616 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:34.7891828Z V1204 11:20:42.343000 99616 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:34.7893159Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.7894246Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:34.7895882Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.7897924Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.7899769Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.7901913Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.7903981Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.7905994Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.7907854Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:34.7909638Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:34.7911637Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:34.7913614Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:34.7915352Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_3: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.7916911Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.7918107Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.7919067Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.7920148Z V1204 11:20:42.350000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:34.7921412Z V1204 11:20:42.350000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:34.7923121Z V1204 11:20:42.351000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.7924768Z V1204 11:20:42.351000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.7926485Z V1204 11:20:42.352000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.7928117Z V1204 11:20:42.352000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.7929638Z V1204 11:20:42.354000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.7931135Z V1204 11:20:42.354000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.7932693Z V1204 11:20:42.356000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:34.7934221Z V1204 11:20:42.357000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.7935686Z V1204 11:20:42.358000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:34.7937218Z V1204 11:20:42.358000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.7938716Z V1204 11:20:42.359000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:34.7940228Z V1204 11:20:42.359000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.7941665Z V1204 11:20:42.361000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:34.7943084Z V1204 11:20:42.361000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.7944481Z V1204 11:20:42.363000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_3 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.7945899Z V1204 11:20:42.364000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.7947067Z V1204 11:20:42.364000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.7948349Z V1204 11:20:42.365000 99616 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:34.7949803Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:34.7950784Z warnings.warn( 2025-12-04T11:24:34.7951432Z V1204 11:20:42.365000 99616 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:34.7952184Z =================================== FAILURES =================================== 2025-12-04T11:24:34.7952952Z _ TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False _ 2025-12-04T11:24:34.7953692Z Traceback (most recent call last): 2025-12-04T11:24:34.7954507Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 275, in test_decompose_linear_mixed_precision 2025-12-04T11:24:34.7955339Z self.assertEqual( 2025-12-04T11:24:34.7956023Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:34.7956790Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:34.7957620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:34.7958507Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:34.7958988Z AssertionError: Scalars are not equal! 2025-12-04T11:24:34.7959239Z 2025-12-04T11:24:34.7959360Z Expected 1 but got 0. 2025-12-04T11:24:34.7959642Z Absolute difference: 1 2025-12-04T11:24:34.7959940Z Relative difference: 1.0 2025-12-04T11:24:34.7960130Z 2025-12-04T11:24:34.7960355Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:34.7961597Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:34.7962619Z 2025-12-04T11:24:34.7962889Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:34.7963524Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:34.7964004Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:34.7964333Z stats [('calls_captured', 1)] 2025-12-04T11:24:34.7965327Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_nodes', 3), ('pad_mm_bench', 3), ('pattern_matcher_count', 2), ('fxgraph_cache_miss', 1), ('decompose_mm', 1)] 2025-12-04T11:24:34.7966486Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:34.7966939Z graph_break [] 2025-12-04T11:24:34.7967321Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:34.7968177Z I1204 11:20:38.756000 99616 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:34.7969437Z I1204 11:20:39.001000 99616 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:34.7970604Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.7971675Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:34.7973241Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.7975180Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.7976982Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:34.7978307Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:34.7979316Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.7980848Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.7982991Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, None); l_x_ = l_self_modules_linear_parameters_weight_ = None 2025-12-04T11:24:34.7984706Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:34.7985772Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.7986706Z V1204 11:20:39.003000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.7987782Z I1204 11:20:41.212000 99616 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:34.7989198Z V1204 11:20:41.699000 99616 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:34.7990829Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v: 2025-12-04T11:24:34.7992339Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [trz4xcrmguerwj2nwxhu4t6zwtig5vb4aexoqw3sqeb3z2aw5kv] gm: GraphModule() 2025-12-04T11:24:34.7993416Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.7994111Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.7994798Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.7995662Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:34.7997430Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.7999267Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.8000972Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.8002428Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, permute) 2025-12-04T11:24:34.8003761Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_3 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.8005036Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, convert_element_type_1, permute_3) 2025-12-04T11:24:34.8005970Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.8006997Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:34.8009991Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.8014798Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.8017999Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:34.8019674Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:34.8021505Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:34.8023189Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:34.8024931Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:34.8026554Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:34.8028127Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:34.8029824Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:34.8031405Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:34.8033055Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [reylvtt3rqfvzq5b3fjekfvick4dxsjoog5svxckgogwqosgp6z] fx_kwargs[static_input_idxs]: [0] 2025-12-04T11:24:34.8034604Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[0]: 1 2025-12-04T11:24:34.8036306Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:34.8038113Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:34.8040196Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:34.8041939Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:34.8071946Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:34.8102516Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:34.8104419Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:34.8106170Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:34.8107957Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:34.8109779Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:34.8111456Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:34.8113081Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:34.8114777Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:34.8116466Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:34.8118231Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:34.8120043Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:34.8121869Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:34.8123707Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:34.8125571Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:34.8127450Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:34.8129322Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:34.8131200Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:34.8133155Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:34.8135054Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:34.8136778Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:34.8138533Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:34.8140284Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:34.8142044Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:34.8143644Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:34.8145281Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:34.8146947Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:34.8148586Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:34.8150316Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:34.8152041Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:34.8153694Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:34.8155381Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:34.8157058Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:34.8158769Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:34.8160462Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:34.8162118Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:34.8163804Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:34.8165546Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:34.8167231Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:34.8168898Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:34.8170596Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:34.8172332Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:34.8174033Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:34.8175768Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:34.8177583Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:34.8179334Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:34.8181089Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:34.8182876Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:34.8184649Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:34.8186310Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:34.8187937Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:34.8189621Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:34.8191484Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:34.8193381Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:34.8195124Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:34.8197042Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:34.8198766Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:34.8200907Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:34.8203032Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:34.8204826Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:34.8206704Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:34.8208600Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:34.8210396Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:34.8212189Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:34.8214047Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:34.8215848Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:34.8217770Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.8219670Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:34.8221583Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.8223469Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:34.8225349Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.8227240Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:34.8229105Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:34.8230864Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:34.8232503Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:34.8234246Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:34.8236079Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:34.8237827Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:34.8239540Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:34.8241276Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:34.8243058Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:34.8244933Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:34.8246846Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:34.8248761Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:34.8250524Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:34.8252246Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:34.8254084Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:34.8255890Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:34.8257660Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:34.8259473Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:34.8261344Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:34.8263236Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:34.8265107Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:34.8266902Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:34.8268661Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:34.8270297Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:34.8271930Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:34.8273654Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:34.8275377Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:34.8277188Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:34.8279575Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:34.8281360Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:34.8283118Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:34.8284790Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:34.8286464Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:34.8288296Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:34.8290256Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:34.8292235Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:34.8294103Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:34.8295865Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:34.8297936Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:34.8299741Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:34.8301611Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:34.8303552Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:34.8305324Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:34.8307036Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:34.8308799Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:34.8310590Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:34.8312365Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:34.8314102Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:34.8314929Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:34.8315711Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:34.8316493Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:34.8317299Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:34.8318128Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:34.8318954Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:34.8319824Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:34.8320600Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:34.8321472Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:34.8322296Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:34.8323158Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:34.8323982Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:34.8324767Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:34.8325582Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:34.8326366Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:34.8327136Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:34.8327967Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:34.8328843Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:34.8329704Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:34.8330559Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:34.8331422Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:34.8332172Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:34.8333064Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:34.8333920Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:34.8334728Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:34.8335525Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:34.8336353Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:34.8337192Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:34.8337994Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:34.8338764Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:34.8339538Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:34.8340277Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:34.8341095Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:34.8341944Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:34.8342796Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:34.8343607Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:34.8344391Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:34.8345196Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:34.8346038Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:34.8346846Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:34.8347680Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:34.8348561Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:34.8349375Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:34.8350214Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:34.8351016Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:34.8351831Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:34.8352667Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:34.8353459Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:34.8354290Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:34.8355153Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:34.8355943Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:34.8356792Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:34.8357650Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:34.8358467Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:34.8359216Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:34.8360046Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:34.8360870Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:34.8361687Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:34.8362516Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:34.8363467Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:34.8364479Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:34.8365453Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:34.8366429Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:34.8367401Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:34.8368407Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:34.8369399Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:34.8370400Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:34.8371402Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:34.8372237Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:34.8373103Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:34.8373908Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:34.8374832Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:34.8375670Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:34.8376460Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:34.8377322Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:34.8378111Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:34.8378916Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:34.8379696Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:34.8380495Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:34.8381283Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:34.8382182Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:34.8382999Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:34.8383823Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:34.8384590Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:34.8385397Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:34.8386233Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:34.8387001Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:34.8387845Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:34.8388677Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:34.8389493Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:34.8390334Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:34.8391060Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:34.8391857Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:34.8392679Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:34.8393522Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:34.8394495Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:34.8395288Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:34.8396178Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:34.8397018Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:34.8397850Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:34.8398701Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:34.8399531Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:34.8400330Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:34.8401265Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:34.8402209Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:34.8403041Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:34.8403837Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:34.8404742Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:34.8405593Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:34.8406376Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:34.8407190Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:34.8408023Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:34.8408921Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:34.8409757Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:34.8410536Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:34.8411353Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:34.8412147Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:34.8412901Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:34.8413702Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:34.8414480Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:34.8415324Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:34.8416106Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:34.8417034Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.8417901Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.8418734Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:34.8419584Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:34.8420416Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:34.8421301Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:34.8422158Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:34.8423056Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:34.8423887Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:34.8424754Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:34.8425578Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:34.8426421Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:34.8427249Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:34.8428055Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:34.8428882Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:34.8429694Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:34.8430550Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:34.8431367Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:34.8432209Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:34.8433003Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:34.8433804Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:34.8434732Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:34.8435587Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:34.8436459Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:34.8437402Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:34.8438303Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:34.8439228Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:34.8440124Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:34.8441017Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:34.8441871Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:34.8442756Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:34.8443673Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:34.8444560Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:34.8445405Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:34.8446220Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:34.8447054Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:34.8447859Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:34.8448718Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:34.8449502Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:34.8450365Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:34.8451189Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:34.8451997Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:34.8452867Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:34.8453845Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:34.8454675Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:34.8455471Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:34.8456379Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:34.8457360Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:34.8458190Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:34.8459103Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:34.8459963Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:34.8460816Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:34.8461671Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:34.8462576Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:34.8463361Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:34.8464171Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:34.8465004Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:34.8465794Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:34.8466651Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:34.8467455Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:34.8468393Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:34.8469418Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:34.8470354Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.8471193Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:34.8472071Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:34.8472940Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:34.8473826Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:34.8474670Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:34.8475645Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:34.8476486Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:34.8477311Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:34.8478101Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:34.8478944Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:34.8479846Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:34.8480751Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:34.8481685Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:34.8482490Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:34.8483360Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:34.8484197Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:34.8485075Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:34.8486011Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:34.8486902Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:34.8487736Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:34.8488586Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:34.8489488Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:34.8490389Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:34.8491274Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:34.8492074Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:34.8492934Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:34.8493765Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:34.8494568Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:34.8495502Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:34.8496545Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:34.8497440Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:34.8498234Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:34.8499195Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:34.8500114Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:34.8501029Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:34.8501911Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:34.8502900Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:34.8503757Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:34.8504632Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:34.8505509Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:34.8506494Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:34.8507412Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:34.8508248Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:34.8509080Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:34.8509905Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:34.8510798Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:34.8511650Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:34.8512525Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:34.8513417Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:34.8514158Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:34.8514926Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:34.8515775Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:34.8516587Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:34.8517394Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:34.8518209Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:34.8519038Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:34.8519920Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:34.8520876Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:34.8521756Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:34.8522601Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:34.8523394Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:34.8524272Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:34.8525103Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:34.8525959Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:34.8526827Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:34.8527677Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:34.8528564Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:34.8529385Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:34.8530213Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:34.8531043Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:34.8531950Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:34.8532867Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:34.8533692Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:34.8534428Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:34.8535398Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:34.8536219Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:34.8537040Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:34.8537816Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:34.8538646Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:34.8539469Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:34.8540353Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:34.8541116Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:34.8541879Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:34.8542710Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:34.8543559Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:34.8544416Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:34.8545283Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:34.8546148Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:34.8546937Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:34.8547773Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:34.8548582Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:34.8549326Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:34.8550104Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:34.8550884Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:34.8551687Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:34.8552494Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:34.8553356Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:34.8554204Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:34.8554969Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:34.8555766Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:34.8556552Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:34.8558521Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:34.8560937Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:34.8561737Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:34.8562647Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:34.8563448Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:34.8564308Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:34.8565270Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:34.8566102Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:34.8567056Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:34.8567953Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:34.8568825Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:34.8569727Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:34.8570647Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:34.8571648Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:34.8572553Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:34.8573387Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:34.8574302Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:34.8575220Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:34.8576119Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:34.8577049Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:34.8578023Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:34.8578768Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:34.8579485Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:34.8580259Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:34.8580988Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:34.8581703Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:34.8582420Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:34.8583329Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:34.8584171Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:34.8585027Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:34.8585743Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:34.8586440Z V1204 11:20:41.709000 99616 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.8587113Z I1204 11:20:41.710000 99616 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.8587622Z V1204 11:20:41.710000 99616 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:34.8588114Z V1204 11:20:41.710000 99616 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:34.8588615Z I1204 11:20:41.711000 99616 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:34.8589389Z V1204 11:20:41.750000 99616 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:34.8589837Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.8590455Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:34.8591417Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.8592145Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.8593132Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.8594089Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.8595129Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.8596182Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.8597036Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:34.8597860Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:34.8598903Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:34.8599735Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:34.8600558Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_3: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.8601211Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.8601677Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.8602075Z V1204 11:20:41.755000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.8602640Z V1204 11:20:41.758000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:34.8603198Z V1204 11:20:41.759000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:34.8604222Z V1204 11:20:41.760000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.8604728Z V1204 11:20:41.760000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.8605749Z V1204 11:20:41.791000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.8606249Z V1204 11:20:41.792000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.8607149Z V1204 11:20:41.794000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.8607601Z V1204 11:20:41.794000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.8608576Z V1204 11:20:41.800000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:34.8609042Z V1204 11:20:41.800000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.8609917Z V1204 11:20:41.802000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:34.8610377Z V1204 11:20:41.802000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.8611327Z V1204 11:20:41.803000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:34.8611773Z V1204 11:20:41.803000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.8612614Z V1204 11:20:41.806000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:34.8613044Z V1204 11:20:41.806000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.8613888Z V1204 11:20:41.809000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_3 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.8614360Z V1204 11:20:41.809000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.8614947Z V1204 11:20:41.810000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.8615549Z V1204 11:20:41.810000 99616 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:34.8616307Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:34.8616414Z warnings.warn( 2025-12-04T11:24:34.8616950Z V1204 11:20:41.811000 99616 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:34.8617199Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:34.8617315Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:34.8617432Z stats [('calls_captured', 1)] 2025-12-04T11:24:34.8617675Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:34.8618451Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_nodes', 3), ('pad_mm_bench', 3), ('pattern_matcher_count', 2), ('fxgraph_cache_miss', 1), ('decompose_mm', 1)] 2025-12-04T11:24:34.8618565Z graph_break [] 2025-12-04T11:24:34.8618785Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:34.8619221Z V1204 11:20:42.045000 99616 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpvknbptgu 2025-12-04T11:24:34.8619668Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.8620159Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:34.8621114Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.8622016Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.8622757Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:34.8623218Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:34.8623633Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.8624643Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.8625667Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, None); l_x_ = l_self_modules_linear_parameters_weight_ = None 2025-12-04T11:24:34.8626149Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:34.8626562Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.8626973Z V1204 11:20:42.072000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.8627697Z V1204 11:20:42.319000 99616 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:34.8628475Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v: 2025-12-04T11:24:34.8629144Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [trz4xcrmguerwj2nwxhu4t6zwtig5vb4aexoqw3sqeb3z2aw5kv] gm: GraphModule() 2025-12-04T11:24:34.8629416Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.8629686Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.8629965Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.8630430Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:34.8631260Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.8632076Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.8632827Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.8633392Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, permute) 2025-12-04T11:24:34.8634018Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_3 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.8634526Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, convert_element_type_1, permute_3) 2025-12-04T11:24:34.8634819Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.8635421Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:34.8637725Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.8640086Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.8640730Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:34.8641651Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:34.8642396Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:34.8643248Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:34.8644027Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:34.8644737Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:34.8645472Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:34.8646194Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:34.8646920Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:34.8647662Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [reylvtt3rqfvzq5b3fjekfvick4dxsjoog5svxckgogwqosgp6z] fx_kwargs[static_input_idxs]: [0] 2025-12-04T11:24:34.8648341Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[0]: 1 2025-12-04T11:24:34.8649224Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:34.8650013Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:34.8651185Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:34.8652004Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:34.8681009Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:34.8682111Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:34.8685382Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:34.8686181Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:34.8687018Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:34.8687859Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:34.8688588Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:34.8689395Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:34.8690189Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:34.8691012Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:34.8691863Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:34.8692705Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:34.8693565Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:34.8694459Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:34.8695309Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:34.8696384Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:34.8697300Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:34.8698190Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:34.8699205Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:34.8699956Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:34.8700803Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:34.8701545Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:34.8702486Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:34.8703245Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:34.8703958Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:34.8704745Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:34.8705492Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:34.8706262Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:34.8707134Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:34.8707904Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:34.8708676Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:34.8709462Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:34.8710264Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:34.8711042Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:34.8711819Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:34.8712578Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:34.8713414Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:34.8714193Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:34.8714953Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:34.8715733Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:34.8716535Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:34.8717313Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:34.8718180Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:34.8719006Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:34.8719802Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:34.8720621Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:34.8721421Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:34.8722276Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:34.8723082Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:34.8723833Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:34.8724583Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:34.8725389Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:34.8726347Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:34.8727140Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:34.8727924Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:34.8728759Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:34.8729534Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:34.8730737Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:34.8731527Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:34.8732397Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:34.8733276Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:34.8734123Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:34.8734942Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:34.8735777Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:34.8736663Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:34.8737539Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:34.8738483Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.8739331Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:34.8740281Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.8741091Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:34.8742052Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.8742877Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:34.8743742Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:34.8744512Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:34.8745255Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:34.8746131Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:34.8746976Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:34.8747729Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:34.8748546Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:34.8749323Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:34.8750201Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:34.8751056Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:34.8751987Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:34.8752837Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:34.8753601Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:34.8754432Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:34.8755333Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:34.8756146Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:34.8756913Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:34.8757856Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:34.8758715Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:34.8759575Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:34.8760447Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:34.8761268Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:34.8762074Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:34.8762793Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:34.8763594Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:34.8764389Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:34.8765174Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:34.8766049Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:34.8766871Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:34.8767697Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:34.8768493Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:34.8769244Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:34.8770033Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:34.8770970Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:34.8771905Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:34.8772826Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:34.8773675Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:34.8774493Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:34.8775340Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:34.8776159Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:34.8777128Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:34.8778024Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:34.8778801Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:34.8779620Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:34.8780448Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:34.8781289Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:34.8782089Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:34.8782886Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:34.8783716Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:34.8784489Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:34.8785273Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:34.8786072Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:34.8786927Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:34.8787750Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:34.8788620Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:34.8789393Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:34.8790291Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:34.8791086Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:34.8791945Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:34.8792764Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:34.8793580Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:34.8794340Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:34.8795117Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:34.8795886Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:34.8796907Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:34.8797808Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:34.8798650Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:34.8799462Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:34.8800321Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:34.8801070Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:34.8801966Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:34.8802858Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:34.8803666Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:34.8804473Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:34.8805306Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:34.8806127Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:34.8806888Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:34.8807664Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:34.8808439Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:34.8809223Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:34.8810014Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:34.8810859Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:34.8811706Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:34.8812515Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:34.8813336Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:34.8814107Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:34.8814919Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:34.8815728Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:34.8816568Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:34.8817527Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:34.8818337Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:34.8819221Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:34.8830112Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:34.8831098Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:34.8832057Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:34.8832858Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:34.8833648Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:34.8834516Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:34.8835302Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:34.8836179Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:34.8837041Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:34.8837865Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:34.8838614Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:34.8839439Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:34.8840283Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:34.8841066Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:34.8841893Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:34.8842839Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:34.8843849Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:34.8844820Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:34.8845838Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:34.8846815Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:34.8847816Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:34.8848843Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:34.8849814Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:34.8850813Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:34.8851644Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:34.8852514Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:34.8853326Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:34.8854245Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:34.8855083Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:34.8855866Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:34.8856696Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:34.8857525Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:34.8858336Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:34.8859114Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:34.8859916Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:34.8860701Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:34.8861614Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:34.8862440Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:34.8863251Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:34.8864019Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:34.8864857Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:34.8865641Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:34.8866403Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:34.8867229Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:34.8868107Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:34.8868890Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:34.8869715Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:34.8870433Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:34.8871224Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:34.8872051Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:34.8872873Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:34.8873832Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:34.8874624Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:34.8875365Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:34.8876207Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:34.8877055Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:34.8877902Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:34.8878706Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:34.8879513Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:34.8880479Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:34.8881357Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:34.8882198Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:34.8882987Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:34.8883922Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:34.8884723Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:34.8885498Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:34.8886311Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:34.8887139Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:34.8888053Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:34.8888880Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:34.8889615Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:34.8890427Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:34.8891216Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:34.8891968Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:34.8892762Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:34.8893559Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:34.8894400Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:34.8895181Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:34.8896299Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.8897245Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.8898014Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:34.8898868Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:34.8899701Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:34.8900615Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:34.8901472Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:34.8902370Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:34.8903194Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:34.8904103Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:34.8904913Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:34.8905726Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:34.8906558Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:34.8907373Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:34.8908207Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:34.8909016Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:34.8909890Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:34.8910714Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:34.8911556Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:34.8912385Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:34.8913185Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:34.8914074Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:34.8914922Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:34.8915796Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:34.8916734Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:34.8917627Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:34.8918550Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:34.8919444Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:34.8920326Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:34.8921168Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:34.8921995Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:34.8922913Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:34.8923779Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:34.8924634Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:34.8925482Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:34.8926311Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:34.8927114Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:34.8927975Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:34.8928784Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:34.8929602Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:34.8930434Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:34.8931241Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:34.8932144Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:34.8933016Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:34.8933842Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:34.8934632Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:34.8935537Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:34.8936466Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:34.8937357Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:34.8938223Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:34.8939077Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:34.8939925Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:34.8940776Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:34.8941700Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:34.8942477Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:34.8943272Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:34.8944118Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:34.8944932Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:34.8945738Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:34.8946532Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:34.8947378Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:34.8948372Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:34.8949250Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.8950098Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:34.8950966Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:34.8951840Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:34.8952724Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:34.8953531Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:34.8954503Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:34.8955334Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:34.8956175Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:34.8956958Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:34.8957833Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:34.8958725Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:34.8959623Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:34.8960589Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:34.8961397Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:34.8962227Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:34.8963056Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:34.8963937Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:34.8964877Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:34.8965769Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:34.8966599Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:34.8967443Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:34.8968403Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:34.8969253Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:34.8970117Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:34.8970914Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:34.8971770Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:34.8972610Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:34.8973406Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:34.8974371Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:34.8975238Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:34.8976038Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:34.8976932Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:34.8977830Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:34.8978746Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:34.8979662Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:34.8980646Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:34.8981598Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:34.8982448Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:34.8983310Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:34.8984199Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:34.8985151Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:34.8986022Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:34.8986866Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:34.8987700Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:34.8988521Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:34.8989403Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:34.8990297Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:34.8991176Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:34.8992067Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:34.8992841Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:34.8993611Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:34.8994423Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:34.8995210Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:34.8996184Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:34.8997061Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:34.8997862Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:34.8998733Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:34.8999691Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:34.9000565Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:34.9001424Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:34.9002190Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:34.9003046Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:34.9003882Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:34.9004743Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:34.9005615Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:34.9006499Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:34.9007392Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:34.9008216Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:34.9009087Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:34.9009927Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:34.9010800Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:34.9011693Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:34.9012512Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:34.9013287Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:34.9014213Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:34.9015027Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:34.9015802Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:34.9016571Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:34.9017462Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:34.9018250Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:34.9019124Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:34.9019894Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:34.9020638Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:34.9021479Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:34.9022355Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:34.9023204Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:34.9024075Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:34.9024941Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:34.9025752Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:34.9026557Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:34.9027368Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:34.9028109Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:34.9028910Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:34.9029661Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:34.9030454Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:34.9031256Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:34.9032100Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:34.9032961Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:34.9033724Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:34.9034477Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:34.9035260Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:34.9037233Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:34.9039678Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:34.9040450Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:34.9041387Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:34.9042192Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:34.9043016Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:34.9043979Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:34.9044846Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:34.9045746Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:34.9046658Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:34.9047531Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:34.9048436Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:34.9049370Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:34.9050363Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:34.9051242Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:34.9052073Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:34.9052988Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:34.9053908Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:34.9054835Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:34.9055703Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:34.9056679Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:34.9057505Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:34.9058216Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:34.9058954Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:34.9059665Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:34.9060394Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:34.9061145Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:34.9062027Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:34.9062861Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:34.9063717Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:34.9064431Z V1204 11:20:42.328000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:34.9065126Z V1204 11:20:42.329000 99616 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.9065799Z I1204 11:20:42.329000 99616 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.9066284Z V1204 11:20:42.329000 99616 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:34.9066771Z V1204 11:20:42.330000 99616 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:34.9067269Z I1204 11:20:42.330000 99616 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:34.9068030Z V1204 11:20:42.343000 99616 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:34.9068486Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.9068976Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:34.9069982Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.9070707Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.9071692Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.9072687Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.9073679Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.9074560Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.9075424Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:34.9076185Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:34.9077224Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:34.9078050Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:34.9078862Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_3: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.9079510Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.9079930Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.9080325Z V1204 11:20:42.347000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.9080888Z V1204 11:20:42.350000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:34.9081450Z V1204 11:20:42.350000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:34.9082473Z V1204 11:20:42.351000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.9083004Z V1204 11:20:42.351000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.9084024Z V1204 11:20:42.352000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.9084519Z V1204 11:20:42.352000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.9085413Z V1204 11:20:42.354000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.9085896Z V1204 11:20:42.354000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.9086831Z V1204 11:20:42.356000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:34.9087295Z V1204 11:20:42.357000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.9088168Z V1204 11:20:42.358000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:34.9088660Z V1204 11:20:42.358000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.9089570Z V1204 11:20:42.359000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:34.9090005Z V1204 11:20:42.359000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.9090848Z V1204 11:20:42.361000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:34.9091272Z V1204 11:20:42.361000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.9092115Z V1204 11:20:42.363000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_3 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.9092592Z V1204 11:20:42.364000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.9093180Z V1204 11:20:42.364000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.9093754Z V1204 11:20:42.365000 99616 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:34.9094493Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:34.9094609Z warnings.warn( 2025-12-04T11:24:34.9095086Z V1204 11:20:42.365000 99616 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:34.9095327Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:34.9095441Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:34.9095560Z stats [('calls_captured', 1)] 2025-12-04T11:24:34.9095795Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:34.9096781Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_nodes', 3), ('pad_mm_bench', 3), ('pattern_matcher_count', 2), ('fxgraph_cache_miss', 1), ('decompose_mm', 1)] 2025-12-04T11:24:34.9096965Z graph_break [] 2025-12-04T11:24:34.9097205Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:34.9097641Z V1204 11:20:42.387000 99616 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpev40aa41 2025-12-04T11:24:34.9098089Z V1204 11:20:42.413000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.9098581Z V1204 11:20:42.413000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:34.9099573Z V1204 11:20:42.413000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.9100439Z V1204 11:20:42.413000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.9101176Z V1204 11:20:42.413000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:34.9101641Z V1204 11:20:42.413000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:34.9102102Z V1204 11:20:42.413000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.9103087Z V1204 11:20:42.413000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.9104116Z V1204 11:20:42.413000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, None); l_x_ = l_self_modules_linear_parameters_weight_ = None 2025-12-04T11:24:34.9104596Z V1204 11:20:42.413000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:34.9105010Z V1204 11:20:42.413000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.9105432Z V1204 11:20:42.413000 99616 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.9106158Z V1204 11:20:42.659000 99616 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:34.9106870Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v: 2025-12-04T11:24:34.9107535Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [trz4xcrmguerwj2nwxhu4t6zwtig5vb4aexoqw3sqeb3z2aw5kv] gm: GraphModule() 2025-12-04T11:24:34.9107812Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.9108081Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.9108364Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.9108827Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:34.9109653Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.9110491Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.9111232Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.9111804Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, permute) 2025-12-04T11:24:34.9112436Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_3 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.9112970Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, convert_element_type_1, permute_3) 2025-12-04T11:24:34.9113264Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.9113831Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:34.9116133Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.9118504Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.9119139Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:34.9120069Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:34.9120836Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:34.9121650Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:34.9122439Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:34.9123148Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:34.9123876Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:34.9124597Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:34.9125353Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:34.9126091Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [reylvtt3rqfvzq5b3fjekfvick4dxsjoog5svxckgogwqosgp6z] fx_kwargs[static_input_idxs]: [0] 2025-12-04T11:24:34.9126772Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[0]: 1 2025-12-04T11:24:34.9127660Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:34.9128483Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:34.9129671Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:34.9130446Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:34.9159427Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:34.9160527Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:34.9161290Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:34.9162070Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:34.9162939Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:34.9163792Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:34.9164507Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:34.9165310Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:34.9166101Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:34.9166872Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:34.9167738Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:34.9168567Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:34.9169471Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:34.9170315Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:34.9171175Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:34.9172067Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:34.9172915Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:34.9173775Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:34.9174773Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:34.9175549Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:34.9176384Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:34.9177209Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:34.9178047Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:34.9178844Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:34.9179565Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:34.9180354Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:34.9181118Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:34.9181890Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:34.9182738Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:34.9183492Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:34.9184262Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:34.9185075Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:34.9185840Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:34.9186625Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:34.9187397Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:34.9188168Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:34.9188989Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:34.9189763Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:34.9190537Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:34.9191295Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:34.9192107Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:34.9192875Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:34.9193644Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:34.9194474Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:34.9195275Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:34.9196275Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:34.9197057Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:34.9197982Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:34.9198748Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:34.9199511Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:34.9200256Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:34.9201115Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:34.9202039Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:34.9202828Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:34.9203624Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:34.9204445Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:34.9205240Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:34.9206422Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:34.9207214Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:34.9208085Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:34.9208971Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:34.9209794Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:34.9210635Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:34.9211489Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:34.9212356Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:34.9213185Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:34.9214129Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.9214952Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:34.9215913Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.9216745Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:34.9217756Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.9218568Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:34.9219452Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:34.9220205Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:34.9221001Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:34.9221850Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:34.9222697Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:34.9223468Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:34.9224277Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:34.9225066Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:34.9225901Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:34.9226784Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:34.9227717Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:34.9228571Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:34.9229384Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:34.9230205Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:34.9231093Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:34.9231888Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:34.9232661Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:34.9233589Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:34.9234447Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:34.9235321Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:34.9236179Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:34.9237034Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:34.9237826Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:34.9238540Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:34.9239331Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:34.9240127Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:34.9240928Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:34.9241752Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:34.9242611Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:34.9243427Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:34.9244225Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:34.9245070Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:34.9245858Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:34.9246780Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:34.9247696Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:34.9248630Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:34.9249468Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:34.9250295Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:34.9251126Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:34.9251953Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:34.9252900Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:34.9253758Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:34.9254548Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:34.9255349Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:34.9256182Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:34.9257071Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:34.9257841Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:34.9258670Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:34.9259501Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:34.9260291Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:34.9261105Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:34.9261913Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:34.9262727Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:34.9263549Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:34.9264418Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:34.9265225Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:34.9266106Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:34.9266885Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:34.9267759Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:34.9268565Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:34.9269382Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:34.9270148Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:34.9270922Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:34.9271705Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:34.9272525Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:34.9273380Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:34.9274236Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:34.9275059Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:34.9275922Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:34.9276676Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:34.9277617Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:34.9278457Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:34.9279278Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:34.9280070Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:34.9280928Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:34.9281712Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:34.9282471Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:34.9283253Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:34.9284009Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:34.9284792Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:34.9285575Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:34.9286421Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:34.9287271Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:34.9288084Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:34.9288884Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:34.9289636Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:34.9290484Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:34.9291281Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:34.9292134Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:34.9293033Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:34.9293843Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:34.9294688Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:34.9295473Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:34.9296432Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:34.9297372Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:34.9298176Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:34.9298956Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:34.9299821Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:34.9300625Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:34.9301484Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:34.9302354Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:34.9303160Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:34.9303918Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:34.9304737Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:34.9305531Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:34.9306374Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:34.9307187Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:34.9308159Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:34.9309200Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:34.9310190Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:34.9311150Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:34.9312117Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:34.9313158Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:34.9314148Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:34.9315136Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:34.9316121Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:34.9316972Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:34.9317823Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:34.9318642Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:34.9319544Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:34.9320378Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:34.9321170Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:34.9321946Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:34.9322759Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:34.9323556Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:34.9324349Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:34.9325141Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:34.9325954Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:34.9326859Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:34.9327668Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:34.9328490Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:34.9329274Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:34.9330092Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:34.9330873Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:34.9331636Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:34.9332486Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:34.9333359Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:34.9334153Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:34.9334971Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:34.9335703Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:34.9336480Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:34.9337330Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:34.9338208Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:34.9339147Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:34.9339957Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:34.9340685Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:34.9341566Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:34.9342374Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:34.9343218Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:34.9344031Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:34.9344863Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:34.9345807Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:34.9346672Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:34.9347516Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:34.9348296Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:34.9349246Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:34.9350034Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:34.9350811Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:34.9351640Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:34.9352461Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:34.9353331Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:34.9354181Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:34.9354933Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:34.9355731Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:34.9356518Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:34.9357316Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:34.9358098Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:34.9358877Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:34.9359706Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:34.9360497Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:34.9361387Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.9362252Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.9363033Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:34.9363876Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:34.9364753Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:34.9365597Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:34.9366463Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:34.9367350Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:34.9368182Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:34.9369058Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:34.9369848Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:34.9370704Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:34.9371516Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:34.9372337Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:34.9373186Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:34.9373997Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:34.9374845Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:34.9375655Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:34.9376509Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:34.9377397Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:34.9378219Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:34.9379098Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:34.9379960Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:34.9380878Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:34.9381789Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:34.9382696Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:34.9383608Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:34.9384517Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:34.9385360Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:34.9386250Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:34.9387076Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:34.9387999Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:34.9388894Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:34.9389766Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:34.9390594Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:34.9391413Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:34.9392226Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:34.9393109Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:34.9393882Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:34.9394711Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:34.9395533Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:34.9396533Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:34.9397454Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:34.9398347Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:34.9399156Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:34.9399953Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:34.9400864Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:34.9401756Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:34.9402631Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:34.9403486Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:34.9404359Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:34.9405194Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:34.9406106Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:34.9406986Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:34.9407760Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:34.9408578Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:34.9409450Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:34.9410252Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:34.9411043Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:34.9411850Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:34.9412688Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:34.9413661Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:34.9414556Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.9415399Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:34.9416314Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:34.9417233Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:34.9418105Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:34.9418948Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:34.9419970Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:34.9420816Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:34.9421680Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:34.9422475Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:34.9423297Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:34.9424216Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:34.9425094Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:34.9426077Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:34.9426874Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:34.9427705Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:34.9428548Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:34.9429429Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:34.9430380Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:34.9431250Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:34.9432091Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:34.9432922Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:34.9433831Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:34.9434723Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:34.9435574Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:34.9436391Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:34.9437239Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:34.9438119Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:34.9438907Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:34.9439843Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:34.9440714Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:34.9441556Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:34.9442357Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:34.9443228Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:34.9444152Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:34.9445054Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:34.9445986Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:34.9446925Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:34.9447780Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:34.9448652Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:34.9449533Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:34.9450463Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:34.9451349Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:34.9452200Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:34.9453014Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:34.9453868Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:34.9454753Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:34.9455604Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:34.9456491Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:34.9457433Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:34.9458230Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:34.9458990Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:34.9459812Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:34.9460614Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:34.9461420Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:34.9462277Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:34.9463060Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:34.9463947Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:34.9464887Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:34.9465791Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:34.9466597Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:34.9467388Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:34.9468265Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:34.9469086Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:34.9469981Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:34.9470830Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:34.9471679Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:34.9472548Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:34.9473370Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:34.9474229Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:34.9475057Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:34.9475942Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:34.9476817Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:34.9477655Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:34.9478409Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:34.9479336Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:34.9480155Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:34.9480909Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:34.9481695Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:34.9482472Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:34.9483301Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:34.9484160Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:34.9484917Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:34.9485673Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:34.9486523Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:34.9487373Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:34.9488218Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:34.9489099Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:34.9489984Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:34.9490780Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:34.9491571Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:34.9492388Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:34.9493143Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:34.9493941Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:34.9494709Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:34.9495493Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:34.9496462Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:34.9497365Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:34.9498191Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:34.9499098Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:34.9499852Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:34.9500655Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:34.9502671Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:34.9505111Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:34.9505926Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:34.9506847Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:34.9507642Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:34.9508484Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:34.9509445Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:34.9510320Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:34.9511236Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:34.9512140Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:34.9513030Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:34.9513926Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:34.9514836Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:34.9515845Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:34.9516731Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:34.9517549Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:34.9518501Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:34.9519434Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:34.9520310Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:34.9521192Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:34.9522153Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:34.9522941Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:34.9523640Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:34.9524372Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:34.9525092Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:34.9525803Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:34.9526568Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:34.9527438Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:34.9528288Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:34.9529129Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:34.9529845Z V1204 11:20:42.668000 99616 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:34.9530514Z V1204 11:20:42.669000 99616 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.9531212Z I1204 11:20:42.669000 99616 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:34.9531708Z V1204 11:20:42.669000 99616 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:34.9532184Z V1204 11:20:42.669000 99616 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:34.9532680Z I1204 11:20:42.670000 99616 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 2 2025-12-04T11:24:34.9533455Z V1204 11:20:42.682000 99616 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:34.9533923Z V1204 11:20:42.687000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.9534431Z V1204 11:20:42.687000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:34.9535376Z V1204 11:20:42.687000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.9536114Z V1204 11:20:42.687000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.9537178Z V1204 11:20:42.687000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.9538150Z V1204 11:20:42.687000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.9539131Z V1204 11:20:42.687000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.9540009Z V1204 11:20:42.687000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.9540888Z V1204 11:20:42.687000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:34.9541640Z V1204 11:20:42.687000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:34.9542690Z V1204 11:20:42.687000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:34.9543512Z V1204 11:20:42.687000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:34.9544307Z V1204 11:20:42.687000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_3: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.9544978Z V1204 11:20:42.687000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.9545414Z V1204 11:20:42.687000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.9545797Z V1204 11:20:42.687000 99616 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:34.9546366Z V1204 11:20:42.689000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:34.9546937Z V1204 11:20:42.690000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:34.9547969Z V1204 11:20:42.690000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.9548490Z V1204 11:20:42.691000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.9549499Z V1204 11:20:42.692000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:34.9550010Z V1204 11:20:42.692000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.9550923Z V1204 11:20:42.693000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.9551371Z V1204 11:20:42.694000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.9552318Z V1204 11:20:42.696000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:34.9552766Z V1204 11:20:42.697000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.9553656Z V1204 11:20:42.698000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:34.9554138Z V1204 11:20:42.698000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.9555059Z V1204 11:20:42.699000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:34.9555485Z V1204 11:20:42.699000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.9556326Z V1204 11:20:42.701000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:34.9556763Z V1204 11:20:42.701000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.9557593Z V1204 11:20:42.703000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_3 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:34.9558055Z V1204 11:20:42.704000 99616 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:34.9558672Z V1204 11:20:42.704000 99616 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:34.9559245Z V1204 11:20:42.705000 99616 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 2 2025-12-04T11:24:34.9559997Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:34.9560101Z warnings.warn( 2025-12-04T11:24:34.9560585Z V1204 11:20:42.705000 99616 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:34.9561462Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-3d7372eabf0c20ef.xml - 2025-12-04T11:24:34.9561639Z =========================== short test summary info ============================ 2025-12-04T11:24:34.9562582Z FAILED [0.3368s] inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False - AssertionError: Scalars are not equal! 2025-12-04T11:24:34.9562591Z 2025-12-04T11:24:34.9562701Z Expected 1 but got 0. 2025-12-04T11:24:34.9562821Z Absolute difference: 1 2025-12-04T11:24:34.9562930Z Relative difference: 1.0 2025-12-04T11:24:34.9562936Z 2025-12-04T11:24:34.9563156Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:34.9564037Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:34.9564075Z 2025-12-04T11:24:34.9564350Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:34.9564542Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:24:34.9564744Z ================== 1 failed, 36 deselected, 2 rerun in 6.65s =================== 2025-12-04T11:24:34.9564853Z --- Logging error --- 2025-12-04T11:24:34.9564992Z Traceback (most recent call last): 2025-12-04T11:24:34.9565301Z File "/opt/conda/envs/py_3.10/lib/python3.10/logging/__init__.py", line 1103, in emit 2025-12-04T11:24:34.9565444Z stream.write(msg + self.terminator) 2025-12-04T11:24:34.9565581Z ValueError: I/O operation on closed file. 2025-12-04T11:24:34.9565676Z Call stack: 2025-12-04T11:24:34.9566209Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/remote_cache.py", line 432, in dump_cache_stats 2025-12-04T11:24:34.9566393Z log.info("Cache Metrics:%s", out.getvalue()) 2025-12-04T11:24:34.9566507Z Message: 'Cache Metrics:%s' 2025-12-04T11:24:34.9566626Z Arguments: (' None\n',) 2025-12-04T11:24:34.9566727Z Got exit code 1 2025-12-04T11:24:34.9566833Z Retrying single test... 2025-12-04T11:24:34.9567517Z Test results will be stored in test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-e6eefab2abf634a5.xml 2025-12-04T11:24:34.9567680Z ============================= test session starts ============================== 2025-12-04T11:24:34.9568048Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:24:34.9568156Z cachedir: .pytest_cache 2025-12-04T11:24:34.9568678Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:24:34.9568817Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:24:34.9568926Z configfile: pytest.ini 2025-12-04T11:24:34.9569522Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:24:34.9571299Z collecting ... /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:52: PytestCollectionWarning: cannot collect test class 'TestDecomposeAddMM' because it has a __init__ constructor (from: test/inductor/test_decompose_mem_bound_mm.py) 2025-12-04T11:24:34.9571458Z class TestDecomposeAddMM(torch.nn.Module): 2025-12-04T11:24:34.9571621Z collected 37 items / 36 deselected / 1 selected 2025-12-04T11:24:34.9572574Z stepcurrent: skipping 14 already run items. Running only test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:34.9572688Z Running 1 items in this shard 2025-12-04T11:24:34.9572711Z 2025-12-04T11:24:34.9573537Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False ('RERUN', {'yellow': True}) [6.0127s] [100%] 2025-12-04T11:24:34.9574399Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False ('RERUN', {'yellow': True}) [0.3448s] [100%] 2025-12-04T11:24:34.9575154Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False FAILED [0.3445s] [100%] 2025-12-04T11:24:34.9575160Z 2025-12-04T11:24:34.9575302Z ==================================== RERUNS ==================================== 2025-12-04T11:24:34.9575788Z _ TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False _ 2025-12-04T11:24:34.9575911Z Traceback (most recent call last): 2025-12-04T11:24:34.9576522Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 275, in test_decompose_linear_mixed_precision 2025-12-04T11:24:34.9576648Z self.assertEqual( 2025-12-04T11:24:34.9577190Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:34.9577376Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:34.9577918Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:34.9578129Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:34.9578280Z AssertionError: Scalars are not equal! 2025-12-04T11:24:34.9578286Z 2025-12-04T11:24:34.9578393Z Expected 1 but got 0. 2025-12-04T11:24:34.9578500Z Absolute difference: 1 2025-12-04T11:24:34.9578626Z Relative difference: 1.0 2025-12-04T11:24:34.9578631Z 2025-12-04T11:24:34.9578847Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:34.9579774Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:34.9579779Z 2025-12-04T11:24:34.9580050Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:34.9580274Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:34.9580403Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:34.9580522Z stats [('calls_captured', 1)] 2025-12-04T11:24:34.9581312Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_nodes', 3), ('pad_mm_bench', 3), ('pattern_matcher_count', 2), ('fxgraph_cache_miss', 1), ('decompose_mm', 1)] 2025-12-04T11:24:34.9581532Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:34.9581631Z graph_break [] 2025-12-04T11:24:34.9581869Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:34.9582365Z I1204 11:20:59.024000 99813 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:34.9582962Z I1204 11:20:59.270000 99813 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:34.9583442Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:34.9583935Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:34.9584885Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:34.9585734Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:34.9586523Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:34.9586977Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:34.9587391Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.9588377Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:34.9589441Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, None); l_x_ = l_self_modules_linear_parameters_weight_ = None 2025-12-04T11:24:34.9589926Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:34.9590341Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.9590732Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:34.9591288Z I1204 11:21:01.477000 99813 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:34.9592000Z V1204 11:21:01.965000 99813 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:34.9592762Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v: 2025-12-04T11:24:34.9593421Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [trz4xcrmguerwj2nwxhu4t6zwtig5vb4aexoqw3sqeb3z2aw5kv] gm: GraphModule() 2025-12-04T11:24:34.9593713Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.9593988Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.9594259Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.9594740Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:34.9595560Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:34.9596567Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:34.9597371Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:34.9597935Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, permute) 2025-12-04T11:24:34.9598576Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_3 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:34.9599067Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, convert_element_type_1, permute_3) 2025-12-04T11:24:34.9599374Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:34.9599979Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:34.9602281Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.9604608Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:34.9605327Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:34.9606240Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:34.9606966Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:34.9607836Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:34.9608612Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:34.9609340Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:34.9610057Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:34.9610787Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:34.9611500Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:34.9612252Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [reylvtt3rqfvzq5b3fjekfvick4dxsjoog5svxckgogwqosgp6z] fx_kwargs[static_input_idxs]: [0] 2025-12-04T11:24:34.9612965Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[0]: 1 2025-12-04T11:24:34.9613837Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:34.9614632Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:34.9615777Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:34.9616607Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:34.9645618Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:34.9646731Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:34.9647500Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:34.9648266Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:34.9649125Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:34.9649995Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:34.9650726Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:34.9651492Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:34.9652284Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:34.9653092Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:34.9653950Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:34.9654792Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:34.9655648Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:34.9656536Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:34.9657438Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:34.9658337Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:34.9659182Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:34.9659992Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:34.9661061Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:34.9661819Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:34.9662669Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:34.9663410Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:34.9664259Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:34.9665018Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:34.9665754Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:34.9666556Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:34.9667302Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:34.9668064Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:34.9668928Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:34.9669697Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:34.9670471Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:34.9671239Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:34.9672014Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:34.9672833Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:34.9673620Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:34.9674380Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:34.9675180Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:34.9675956Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:34.9676755Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:34.9677526Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:34.9678329Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:34.9679109Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:34.9679862Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:34.9680694Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:34.9681531Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:34.9682351Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:34.9683141Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:34.9684084Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:34.9684920Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:34.9685670Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:34.9686426Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:34.9687234Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:34.9688198Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:34.9689006Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:34.9689794Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:34.9690640Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:34.9691383Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:34.9692628Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:34.9693423Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:34.9694282Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:34.9695176Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:34.9696146Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:34.9697039Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:34.9697944Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:34.9698844Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:34.9699644Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:34.9700599Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.9701468Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:34.9702436Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.9703249Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:34.9704180Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:34.9705057Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:34.9705925Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:34.9706696Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:34.9707443Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:34.9708278Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:34.9709187Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:34.9709938Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:34.9710763Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:34.9711546Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:34.9712395Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:34.9713262Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:34.9714217Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:34.9715089Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:34.9715852Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:34.9716699Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:34.9717620Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:34.9718445Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:34.9719213Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:34.9720102Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:34.9721015Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:34.9721887Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:34.9722765Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:34.9723560Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:34.9724364Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:34.9725118Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:34.9725913Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:34.9726715Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:34.9727508Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:34.9728343Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:34.9729168Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:34.9730033Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:34.9730837Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:34.9731582Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:34.9732375Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:34.9733319Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:34.9734264Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:34.9735180Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:34.9736001Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:34.9736937Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:34.9737789Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:34.9738617Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:34.9739516Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:34.9740378Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:34.9741210Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:34.9742033Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:34.9742856Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:34.9743695Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:34.9744463Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:34.9745257Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:34.9746138Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:34.9746916Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:34.9747700Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:34.9748499Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:34.9749377Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:34.9750207Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:34.9751066Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:34.9751861Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:34.9752765Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:34.9753561Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:34.9754424Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:34.9755247Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:34.9756028Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:34.9756919Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:34.9757712Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:34.9758491Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:34.9759325Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:34.9760170Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:34.9761020Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:34.9761843Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:34.9762748Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:34.9763502Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:34.9764394Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:34.9765294Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:34.9766102Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:34.9766905Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:34.9767734Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:34.9768522Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:34.9769325Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:34.9770093Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:34.9770867Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:34.9771606Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:34.9772391Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:34.9773274Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:34.9774126Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:34.9774937Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:34.9775720Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:34.9776494Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:34.9777362Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:34.9778201Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:34.9779036Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:34.9779914Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:34.9780725Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:34.9781591Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:34.9782396Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:34.9783207Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:34.9784037Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:34.9784867Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:34.9785659Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:34.9786518Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:34.9787309Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:34.9788131Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:34.9789021Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:34.9789838Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:34.9790586Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:34.9791414Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:34.9792204Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:34.9792994Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:34.9793855Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:34.9794805Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:34.9795817Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:34.9796952Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:34.9798001Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:34.9798971Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:34.9799979Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:34.9800967Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:34.9801998Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:34.9802999Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:34.9803831Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:34.9804665Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:34.9805525Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:34.9806445Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:34.9807274Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:34.9808047Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:34.9808841Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:34.9809597Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:34.9810464Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:34.9811247Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:34.9812051Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:34.9812829Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:34.9813745Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:34.9814574Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:34.9815382Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:34.9816147Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:34.9817051Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:34.9817846Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:34.9818612Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:34.9819441Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:34.9820288Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:34.9821113Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:34.9821951Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:34.9822681Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:34.9823475Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:34.9824260Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:34.9825090Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:34.9826042Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:34.9826865Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:34.9827608Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:34.9828445Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:34.9829293Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:34.9830141Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:34.9830956Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:34.9831753Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:34.9832685Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:34.9833593Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:34.9834437Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:34.9835229Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:34.9836125Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:34.9836927Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:34.9837734Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:34.9838544Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:34.9839379Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:34.9840237Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:34.9841074Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:34.9841819Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:34.9842670Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:34.9843459Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:34.9844209Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:34.9845007Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:34.9845812Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:34.9846653Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:34.9847439Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:34.9848312Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.9849217Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.9849987Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:34.9850845Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:34.9851670Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:34.9852528Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:34.9853423Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:34.9854323Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:34.9855151Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:34.9856013Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:34.9856825Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:34.9857700Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:34.9858576Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:34.9859392Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:34.9860231Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:34.9861040Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:34.9861927Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:34.9862740Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:34.9863586Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:34.9864377Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:34.9865217Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:34.9866117Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:34.9866971Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:34.9867854Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:34.9868760Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:34.9869712Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:34.9870639Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:34.9871536Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:34.9872380Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:34.9873232Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:34.9874070Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:34.9875026Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:34.9875898Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:34.9876745Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:34.9877594Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:34.9878428Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:34.9879228Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:34.9880089Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:34.9880865Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:34.9881729Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:34.9882554Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:34.9883365Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:34.9884245Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:34.9885117Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:34.9885974Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:34.9886779Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:34.9887688Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:34.9888575Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:34.9889402Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:34.9890267Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:34.9891164Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:34.9892018Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:34.9892866Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:34.9893789Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:34.9894575Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:34.9895374Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:34.9896377Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:34.9897298Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:34.9898112Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:34.9898914Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:34.9899773Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:34.9900722Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:34.9901653Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:34.9902506Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:34.9903382Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:34.9904259Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:34.9905113Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:34.9905935Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:34.9906953Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:34.9907797Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:34.9908617Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:34.9909407Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:34.9910289Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:34.9911193Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:34.9912097Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:34.9913027Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:34.9913876Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:34.9914704Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:34.9915529Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:34.9916423Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:34.9917335Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:34.9918265Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:34.9919101Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:34.9919951Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:34.9920854Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:34.9921708Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:34.9922578Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:34.9923409Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:34.9924265Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:34.9925095Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:34.9925946Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:34.9926880Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:34.9927752Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:34.9928551Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:34.9929335Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:34.9930281Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:34.9931199Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:34.9932110Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:34.9932984Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:34.9933962Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:34.9934815Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:34.9935678Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:34.9936563Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:34.9937537Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:34.9938413Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:34.9939284Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:34.9940114Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:34.9940929Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:34.9941811Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:34.9942701Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:34.9943579Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:34.9944465Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:34.9945200Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:34.9946003Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:34.9946809Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:34.9947625Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:34.9948430Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:34.9949235Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:34.9950059Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:34.9950933Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:34.9951884Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:34.9952766Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:34.9953579Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:34.9954343Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:34.9955230Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:34.9956069Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:34.9956916Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:34.9957774Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:34.9958644Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:34.9959525Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:34.9960342Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:34.9961154Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:34.9962029Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:34.9962900Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:34.9963794Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:34.9964626Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:34.9965356Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:34.9966326Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:34.9967130Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:34.9967900Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:34.9968674Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:34.9969461Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:34.9970254Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:34.9971167Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:34.9971932Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:34.9972687Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:34.9973504Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:34.9974372Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:34.9975231Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:34.9976100Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:34.9977024Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:34.9977844Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:34.9978648Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:34.9979455Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:34.9980198Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:34.9980976Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:34.9981767Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:34.9982573Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:34.9983385Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:34.9984241Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:34.9985059Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:34.9985829Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:34.9986599Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:34.9987421Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:34.9989399Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:34.9991853Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:34.9992623Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:34.9993529Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:34.9994364Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:34.9995190Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:34.9996345Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:34.9997184Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:34.9998086Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:34.9999066Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:34.9999944Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:35.0000852Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:35.0001749Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:35.0002751Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:35.0003676Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:35.0004514Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:35.0005432Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:35.0006346Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:35.0007279Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:35.0008145Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:35.0009117Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:35.0009849Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:35.0010678Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:35.0011412Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:35.0012121Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:35.0012846Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:35.0013566Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:35.0014499Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:35.0015339Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:35.0016193Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:35.0016961Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:35.0017624Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:35.0018311Z I1204 11:21:01.976000 99813 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:35.0018792Z V1204 11:21:01.976000 99813 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:35.0019320Z V1204 11:21:01.976000 99813 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:35.0019821Z I1204 11:21:01.977000 99813 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:35.0020579Z V1204 11:21:02.018000 99813 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:35.0021032Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.0021526Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:35.0022519Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.0023249Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.0024242Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.0025195Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.0026220Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.0027099Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:35.0027935Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:35.0028733Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:35.0029765Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:35.0030591Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:35.0031366Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_3: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.0032019Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:35.0032438Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.0032867Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.0033437Z V1204 11:21:02.026000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:35.0033995Z V1204 11:21:02.027000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:35.0035016Z V1204 11:21:02.028000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.0035567Z V1204 11:21:02.028000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0036588Z V1204 11:21:02.060000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.0037085Z V1204 11:21:02.060000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0037985Z V1204 11:21:02.062000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.0038432Z V1204 11:21:02.062000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0039397Z V1204 11:21:02.068000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:35.0039862Z V1204 11:21:02.068000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0040736Z V1204 11:21:02.070000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:35.0041198Z V1204 11:21:02.070000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0042102Z V1204 11:21:02.071000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:35.0042587Z V1204 11:21:02.072000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0043432Z V1204 11:21:02.074000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:35.0043858Z V1204 11:21:02.074000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0044703Z V1204 11:21:02.077000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_3 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.0045146Z V1204 11:21:02.077000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0045746Z V1204 11:21:02.078000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:35.0046323Z V1204 11:21:02.078000 99813 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:35.0047112Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:35.0047229Z warnings.warn( 2025-12-04T11:24:35.0047709Z V1204 11:21:02.079000 99813 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:35.0048196Z _ TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False _ 2025-12-04T11:24:35.0048326Z Traceback (most recent call last): 2025-12-04T11:24:35.0048916Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 275, in test_decompose_linear_mixed_precision 2025-12-04T11:24:35.0049041Z self.assertEqual( 2025-12-04T11:24:35.0049529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:35.0049712Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:35.0050251Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:35.0050461Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:35.0050611Z AssertionError: Scalars are not equal! 2025-12-04T11:24:35.0050618Z 2025-12-04T11:24:35.0050730Z Expected 1 but got 0. 2025-12-04T11:24:35.0050836Z Absolute difference: 1 2025-12-04T11:24:35.0050964Z Relative difference: 1.0 2025-12-04T11:24:35.0050970Z 2025-12-04T11:24:35.0051186Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:35.0052109Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:35.0052116Z 2025-12-04T11:24:35.0052391Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:35.0052619Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:35.0052747Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:35.0052863Z stats [('calls_captured', 1)] 2025-12-04T11:24:35.0053649Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_nodes', 3), ('pad_mm_bench', 3), ('pattern_matcher_count', 2), ('fxgraph_cache_miss', 1), ('decompose_mm', 1)] 2025-12-04T11:24:35.0053873Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:35.0053973Z graph_break [] 2025-12-04T11:24:35.0054211Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:35.0054732Z I1204 11:20:59.024000 99813 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:35.0055325Z I1204 11:20:59.270000 99813 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:35.0055779Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.0056267Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:35.0057290Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.0058146Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.0058901Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:35.0059386Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:35.0059799Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.0060783Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.0061847Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, None); l_x_ = l_self_modules_linear_parameters_weight_ = None 2025-12-04T11:24:35.0062330Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:35.0062743Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.0063133Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.0063686Z I1204 11:21:01.477000 99813 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:35.0064391Z V1204 11:21:01.965000 99813 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:35.0065149Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v: 2025-12-04T11:24:35.0065802Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [trz4xcrmguerwj2nwxhu4t6zwtig5vb4aexoqw3sqeb3z2aw5kv] gm: GraphModule() 2025-12-04T11:24:35.0066089Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.0066361Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.0066627Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.0067102Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:35.0067912Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.0068770Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.0069510Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:35.0070086Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, permute) 2025-12-04T11:24:35.0070710Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_3 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.0071202Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, convert_element_type_1, permute_3) 2025-12-04T11:24:35.0071507Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.0072069Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:35.0074402Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.0076746Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.0077408Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:35.0078311Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:35.0079041Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:35.0079884Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:35.0080663Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:35.0081386Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:35.0082097Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:35.0082820Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:35.0083565Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:35.0084312Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [reylvtt3rqfvzq5b3fjekfvick4dxsjoog5svxckgogwqosgp6z] fx_kwargs[static_input_idxs]: [0] 2025-12-04T11:24:35.0084990Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[0]: 1 2025-12-04T11:24:35.0085855Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:35.0086654Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:35.0087834Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:35.0088619Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:35.0117804Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:35.0118877Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:35.0119642Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:35.0120411Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:35.0121320Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:35.0122161Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:35.0122882Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:35.0123651Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:35.0124437Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:35.0125211Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:35.0126092Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:35.0126936Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:35.0127788Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:35.0128647Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:35.0129518Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:35.0130423Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:35.0131273Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:35.0132088Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:35.0133140Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:35.0133893Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:35.0134739Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:35.0135478Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:35.0136319Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:35.0137169Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:35.0137874Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:35.0138676Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:35.0139420Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:35.0140182Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:35.0141023Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:35.0141826Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:35.0142597Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:35.0143361Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:35.0144140Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:35.0144953Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:35.0145740Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:35.0146496Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:35.0147295Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:35.0148063Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:35.0148858Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:35.0149635Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:35.0150441Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:35.0151217Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:35.0151969Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:35.0152826Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:35.0153623Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:35.0154447Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:35.0155241Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:35.0156097Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:35.0156878Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:35.0157657Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:35.0158413Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:35.0159211Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:35.0160129Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:35.0160959Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:35.0161748Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:35.0162585Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:35.0163332Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:35.0164559Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:35.0165349Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:35.0166216Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:35.0167095Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:35.0167911Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:35.0168771Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:35.0169611Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:35.0170493Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:35.0171287Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:35.0172238Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.0173058Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:35.0174036Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.0174848Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:35.0175781Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.0176635Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:35.0177570Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:35.0178339Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:35.0179077Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:35.0179911Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:35.0180818Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:35.0181570Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:35.0188774Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:35.0189635Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:35.0190508Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:35.0191466Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:35.0192393Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:35.0193261Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:35.0194030Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:35.0194870Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:35.0195748Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:35.0196808Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:35.0197589Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:35.0198476Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:35.0199404Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:35.0200271Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:35.0201143Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:35.0201940Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:35.0202751Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:35.0203530Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:35.0204313Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:35.0205125Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:35.0205915Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:35.0206757Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:35.0207633Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:35.0208463Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:35.0209269Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:35.0210007Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:35.0210810Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:35.0211723Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:35.0212686Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:35.0213611Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:35.0214430Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:35.0215291Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:35.0216122Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:35.0217034Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:35.0217935Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:35.0218799Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:35.0219620Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:35.0220440Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:35.0221260Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:35.0222100Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:35.0222866Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:35.0223681Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:35.0224521Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:35.0225301Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:35.0226081Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:35.0226879Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:35.0227704Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:35.0228558Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:35.0229412Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:35.0230192Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:35.0231061Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:35.0231876Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:35.0232736Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:35.0233552Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:35.0234333Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:35.0235122Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:35.0235906Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:35.0236679Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:35.0237504Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:35.0238340Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:35.0239208Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:35.0240031Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:35.0240877Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:35.0241633Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:35.0242519Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:35.0243374Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:35.0244209Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:35.0245012Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:35.0245838Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:35.0246603Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:35.0247425Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:35.0248198Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:35.0248974Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:35.0249713Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:35.0250491Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:35.0251371Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:35.0252209Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:35.0253032Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:35.0253816Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:35.0254611Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:35.0255421Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:35.0256227Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:35.0257132Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:35.0257999Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:35.0258816Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:35.0259645Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:35.0260483Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:35.0261299Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:35.0262133Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:35.0262961Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:35.0263743Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:35.0264620Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:35.0265412Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:35.0266234Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:35.0267122Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:35.0267950Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:35.0268700Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:35.0269512Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:35.0270313Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:35.0271132Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:35.0271963Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:35.0272915Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:35.0273923Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:35.0274897Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:35.0275869Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:35.0276863Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:35.0277860Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:35.0278868Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:35.0279946Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:35.0280961Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:35.0281787Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:35.0282621Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:35.0283460Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:35.0284374Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:35.0285210Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:35.0285986Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:35.0286767Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:35.0287569Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:35.0288379Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:35.0289158Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:35.0289955Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:35.0290737Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:35.0291618Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:35.0292468Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:35.0293276Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:35.0294038Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:35.0294842Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:35.0295653Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:35.0296575Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:35.0297462Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:35.0298309Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:35.0299152Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:35.0299979Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:35.0300705Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:35.0301496Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:35.0302283Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:35.0303146Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:35.0304101Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:35.0304894Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:35.0305632Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:35.0306466Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:35.0307291Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:35.0308194Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:35.0308998Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:35.0309814Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:35.0310747Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:35.0311671Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:35.0312506Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:35.0313301Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:35.0314206Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:35.0315024Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:35.0315817Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:35.0316635Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:35.0317468Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:35.0318322Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:35.0319182Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:35.0319924Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:35.0320729Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:35.0321531Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:35.0322278Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:35.0323076Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:35.0323853Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:35.0324718Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:35.0325504Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:35.0326376Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.0327273Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.0328040Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:35.0328904Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:35.0329727Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:35.0330584Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:35.0331471Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:35.0332373Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:35.0333201Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:35.0334053Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:35.0334886Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:35.0335705Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:35.0336533Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:35.0337439Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:35.0338275Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:35.0339086Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:35.0339927Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:35.0340797Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:35.0341643Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:35.0342440Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:35.0343275Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:35.0344177Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:35.0345027Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:35.0345889Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:35.0346806Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:35.0347741Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:35.0348665Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:35.0349562Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:35.0350413Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:35.0351297Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:35.0352132Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:35.0353057Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:35.0353927Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:35.0354779Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:35.0355600Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:35.0356459Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:35.0357259Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:35.0358117Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:35.0358898Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:35.0359749Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:35.0360587Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:35.0361399Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:35.0362268Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:35.0363172Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:35.0363997Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:35.0364796Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:35.0365687Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:35.0366589Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:35.0367462Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:35.0368330Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:35.0369187Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:35.0370036Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:35.0370885Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:35.0371776Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:35.0372583Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:35.0373387Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:35.0374231Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:35.0375020Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:35.0375849Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:35.0376650Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:35.0377554Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:35.0378495Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:35.0379414Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.0380255Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:35.0381136Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:35.0382001Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:35.0382849Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:35.0383690Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:35.0384660Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:35.0385486Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:35.0386318Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:35.0387101Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:35.0387941Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:35.0388866Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:35.0389758Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:35.0390686Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:35.0391513Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:35.0392351Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:35.0393179Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:35.0394065Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:35.0394969Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:35.0395887Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:35.0397015Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:35.0397869Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:35.0398770Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:35.0399698Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:35.0400567Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:35.0401368Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:35.0402221Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:35.0403056Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:35.0403863Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:35.0404836Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:35.0405694Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:35.0406495Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:35.0407276Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:35.0408202Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:35.0409118Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:35.0410031Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:35.0410902Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:35.0411893Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:35.0412746Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:35.0413610Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:35.0414496Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:35.0415412Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:35.0416316Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:35.0417221Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:35.0418054Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:35.0418876Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:35.0419749Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:35.0420615Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:35.0421529Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:35.0422424Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:35.0423160Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:35.0423928Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:35.0424764Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:35.0425558Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:35.0426371Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:35.0427179Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:35.0428025Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:35.0428901Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:35.0429859Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:35.0430732Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:35.0431528Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:35.0432330Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:35.0433188Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:35.0434016Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:35.0434871Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:35.0435730Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:35.0436568Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:35.0437475Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:35.0438290Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:35.0439100Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:35.0439937Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:35.0440829Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:35.0441719Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:35.0442541Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:35.0443271Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:35.0444234Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:35.0445039Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:35.0445806Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:35.0446570Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:35.0447359Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:35.0448175Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:35.0449050Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:35.0449810Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:35.0450546Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:35.0451379Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:35.0452220Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:35.0453102Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:35.0453976Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:35.0454834Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:35.0455626Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:35.0456445Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:35.0457333Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:35.0458071Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:35.0458839Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:35.0459621Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:35.0460412Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:35.0461218Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:35.0462066Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:35.0462887Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:35.0463684Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:35.0464448Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:35.0465239Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:35.0467225Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:35.0469676Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:35.0470445Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:35.0471351Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:35.0472184Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:35.0473007Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:35.0473966Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:35.0474802Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:35.0475736Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:35.0476644Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:35.0477518Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:35.0478420Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:35.0479319Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:35.0480355Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:35.0481229Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:35.0482051Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:35.0482979Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:35.0483898Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:35.0484779Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:35.0485675Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:35.0486651Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:35.0487384Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:35.0488139Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:35.0488875Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:35.0489577Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:35.0490295Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:35.0491015Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:35.0491925Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:35.0492776Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:35.0493631Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:35.0494340Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:35.0494997Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:35.0495719Z I1204 11:21:01.976000 99813 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:35.0496350Z V1204 11:21:01.976000 99813 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:35.0496918Z V1204 11:21:01.976000 99813 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:35.0497421Z I1204 11:21:01.977000 99813 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:35.0498189Z V1204 11:21:02.018000 99813 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:35.0498640Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.0499140Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:35.0500103Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.0500891Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.0501891Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.0502851Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.0503885Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.0504786Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:35.0505622Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:35.0506435Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:35.0507476Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:35.0508310Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:35.0509086Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_3: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.0509788Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:35.0510214Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.0510603Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.0511191Z V1204 11:21:02.026000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:35.0511751Z V1204 11:21:02.027000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:35.0512777Z V1204 11:21:02.028000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.0513283Z V1204 11:21:02.028000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0514331Z V1204 11:21:02.060000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.0514827Z V1204 11:21:02.060000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0515712Z V1204 11:21:02.062000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.0516170Z V1204 11:21:02.062000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0517131Z V1204 11:21:02.068000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:35.0517594Z V1204 11:21:02.068000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0518465Z V1204 11:21:02.070000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:35.0518925Z V1204 11:21:02.070000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0519823Z V1204 11:21:02.071000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:35.0520286Z V1204 11:21:02.072000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0521141Z V1204 11:21:02.074000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:35.0521568Z V1204 11:21:02.074000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0522414Z V1204 11:21:02.077000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_3 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.0522960Z V1204 11:21:02.077000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0523609Z V1204 11:21:02.078000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:35.0524182Z V1204 11:21:02.078000 99813 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:35.0524924Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:35.0525046Z warnings.warn( 2025-12-04T11:24:35.0525521Z V1204 11:21:02.079000 99813 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:35.0525761Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:35.0525875Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:35.0525992Z stats [('calls_captured', 1)] 2025-12-04T11:24:35.0526230Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:35.0526998Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_nodes', 3), ('pad_mm_bench', 3), ('pattern_matcher_count', 2), ('fxgraph_cache_miss', 1), ('decompose_mm', 1)] 2025-12-04T11:24:35.0527096Z graph_break [] 2025-12-04T11:24:35.0527447Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:35.0527879Z V1204 11:21:02.321000 99813 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmp5tt8qcoa 2025-12-04T11:24:35.0528327Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.0528812Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:35.0529741Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.0530637Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.0531390Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:35.0531851Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:35.0532265Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.0533242Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.0534306Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, None); l_x_ = l_self_modules_linear_parameters_weight_ = None 2025-12-04T11:24:35.0534775Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:35.0535195Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.0535573Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.0536297Z V1204 11:21:02.599000 99813 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:35.0537127Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v: 2025-12-04T11:24:35.0537805Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [trz4xcrmguerwj2nwxhu4t6zwtig5vb4aexoqw3sqeb3z2aw5kv] gm: GraphModule() 2025-12-04T11:24:35.0538080Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.0538351Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.0538629Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.0539090Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:35.0539921Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.0540740Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.0541528Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:35.0542106Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, permute) 2025-12-04T11:24:35.0542733Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_3 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.0543238Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, convert_element_type_1, permute_3) 2025-12-04T11:24:35.0543559Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.0544129Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:35.0546428Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.0548831Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.0549468Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:35.0550369Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:35.0551100Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:35.0551939Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:35.0552734Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:35.0553441Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:35.0554165Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:35.0554879Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:35.0555593Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:35.0556382Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [reylvtt3rqfvzq5b3fjekfvick4dxsjoog5svxckgogwqosgp6z] fx_kwargs[static_input_idxs]: [0] 2025-12-04T11:24:35.0557069Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[0]: 1 2025-12-04T11:24:35.0557946Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:35.0558733Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:35.0559967Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:35.0560756Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:35.0589769Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:35.0590866Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:35.0591625Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:35.0592413Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:35.0593247Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:35.0594119Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:35.0594830Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:35.0595599Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:35.0596575Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:35.0597404Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:35.0598267Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:35.0599093Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:35.0599959Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:35.0600848Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:35.0601690Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:35.0602594Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:35.0603442Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:35.0604257Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:35.0605298Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:35.0606066Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:35.0606902Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:35.0607641Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:35.0608486Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:35.0609246Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:35.0609997Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:35.0610790Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:35.0611543Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:35.0612295Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:35.0613146Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:35.0613918Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:35.0614684Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:35.0615468Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:35.0616262Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:35.0617111Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:35.0617888Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:35.0618643Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:35.0619443Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:35.0620264Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:35.0621038Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:35.0621799Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:35.0622617Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:35.0623373Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:35.0624134Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:35.0624948Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:35.0625775Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:35.0626610Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:35.0627387Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:35.0628264Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:35.0629054Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:35.0629809Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:35.0630566Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:35.0631370Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:35.0632331Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:35.0633118Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:35.0633924Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:35.0634749Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:35.0635495Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:35.0636725Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:35.0637521Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:35.0638397Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:35.0639274Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:35.0640104Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:35.0640902Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:35.0641784Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:35.0642661Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:35.0643455Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:35.0644428Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.0645248Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:35.0646211Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.0647006Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:35.0647947Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.0648789Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:35.0649649Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:35.0650412Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:35.0651157Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:35.0652034Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:35.0652882Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:35.0653649Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:35.0654456Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:35.0655233Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:35.0656076Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:35.0656997Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:35.0657967Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:35.0658823Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:35.0659606Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:35.0660464Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:35.0661356Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:35.0662156Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:35.0662930Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:35.0663833Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:35.0664729Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:35.0665609Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:35.0666471Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:35.0667272Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:35.0668099Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:35.0668814Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:35.0669612Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:35.0670408Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:35.0671207Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:35.0672035Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:35.0672875Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:35.0673722Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:35.0674526Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:35.0675277Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:35.0676072Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:35.0677020Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:35.0677947Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:35.0678883Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:35.0679684Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:35.0680549Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:35.0681399Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:35.0682219Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:35.0683131Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:35.0684020Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:35.0684816Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:35.0685625Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:35.0686452Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:35.0687290Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:35.0688064Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:35.0688866Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:35.0689722Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:35.0690512Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:35.0691279Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:35.0692083Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:35.0692946Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:35.0693771Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:35.0694640Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:35.0695414Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:35.0696489Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:35.0697318Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:35.0698192Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:35.0698998Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:35.0699779Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:35.0700611Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:35.0701391Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:35.0702177Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:35.0702998Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:35.0703852Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:35.0704686Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:35.0705552Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:35.0706418Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:35.0707168Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:35.0708071Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:35.0708955Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:35.0709783Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:35.0710572Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:35.0711399Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:35.0712218Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:35.0712980Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:35.0713760Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:35.0714520Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:35.0715272Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:35.0716078Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:35.0716920Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:35.0717776Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:35.0718590Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:35.0719385Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:35.0720148Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:35.0720971Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:35.0721788Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:35.0722625Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:35.0723502Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:35.0724344Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:35.0725184Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:35.0725968Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:35.0726795Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:35.0727611Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:35.0728437Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:35.0729228Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:35.0730087Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:35.0730892Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:35.0731697Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:35.0732597Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:35.0733407Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:35.0734157Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:35.0734983Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:35.0735775Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:35.0736570Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:35.0737498Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:35.0738468Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:35.0739473Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:35.0740535Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:35.0741500Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:35.0742472Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:35.0743476Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:35.0744505Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:35.0745490Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:35.0746473Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:35.0747323Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:35.0748139Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:35.0748986Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:35.0749909Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:35.0750739Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:35.0751527Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:35.0752306Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:35.0753069Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:35.0753900Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:35.0754685Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:35.0755481Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:35.0756263Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:35.0757191Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:35.0758009Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:35.0758837Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:35.0759588Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:35.0760443Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:35.0761209Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:35.0761975Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:35.0762813Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:35.0763641Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:35.0764478Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:35.0765295Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:35.0766036Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:35.0766816Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:35.0767606Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:35.0768448Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:35.0769413Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:35.0770216Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:35.0770950Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:35.0771794Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:35.0772636Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:35.0773493Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:35.0774310Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:35.0775110Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:35.0776093Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:35.0777032Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:35.0777880Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:35.0778659Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:35.0779559Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:35.0780403Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:35.0781177Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:35.0782003Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:35.0782820Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:35.0783689Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:35.0784515Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:35.0785282Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:35.0786094Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:35.0786881Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:35.0787642Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:35.0788456Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:35.0789247Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:35.0790076Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:35.0790858Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:35.0791729Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.0792628Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.0793403Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:35.0794252Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:35.0795096Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:35.0795941Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:35.0797027Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:35.0797929Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:35.0798755Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:35.0799630Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:35.0800428Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:35.0801255Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:35.0802158Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:35.0802985Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:35.0803825Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:35.0804679Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:35.0805539Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:35.0806363Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:35.0807219Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:35.0808006Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:35.0808884Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:35.0809774Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:35.0810627Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:35.0811508Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:35.0812471Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:35.0813382Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:35.0814298Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:35.0815212Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:35.0816051Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:35.0816958Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:35.0817844Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:35.0818762Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:35.0819640Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:35.0820484Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:35.0821349Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:35.0822174Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:35.0822992Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:35.0823841Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:35.0824658Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:35.0825492Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:35.0826317Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:35.0827147Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:35.0827997Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:35.0828918Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:35.0829724Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:35.0830527Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:35.0831428Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:35.0832323Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:35.0833169Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:35.0834066Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:35.0834945Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:35.0835777Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:35.0836626Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:35.0837558Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:35.0838342Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:35.0839154Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:35.0839989Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:35.0840821Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:35.0841615Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:35.0842410Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:35.0843254Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:35.0844202Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:35.0845126Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.0845963Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:35.0846846Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:35.0847704Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:35.0848568Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:35.0849368Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:35.0850373Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:35.0851207Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:35.0852026Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:35.0852819Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:35.0853670Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:35.0854581Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:35.0855469Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:35.0856400Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:35.0857370Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:35.0858205Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:35.0859041Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:35.0859917Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:35.0860839Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:35.0861766Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:35.0862606Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:35.0863441Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:35.0864345Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:35.0865216Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:35.0866092Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:35.0866904Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:35.0867751Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:35.0868592Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:35.0869418Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:35.0870353Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:35.0871218Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:35.0872017Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:35.0872853Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:35.0873725Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:35.0874657Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:35.0875564Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:35.0876432Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:35.0877411Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:35.0878261Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:35.0879129Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:35.0879999Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:35.0880928Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:35.0881785Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:35.0882670Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:35.0883486Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:35.0884308Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:35.0885195Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:35.0886074Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:35.0886962Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:35.0887841Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:35.0888591Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:35.0889378Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:35.0890180Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:35.0890987Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:35.0891796Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:35.0892612Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:35.0893425Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:35.0894309Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:35.0895246Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:35.0896268Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:35.0897143Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:35.0897915Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:35.0898852Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:35.0899677Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:35.0900545Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:35.0901404Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:35.0902287Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:35.0903167Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:35.0903987Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:35.0904818Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:35.0905692Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:35.0906574Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:35.0907451Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:35.0908288Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:35.0909006Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:35.0909988Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:35.0910817Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:35.0911575Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:35.0912358Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:35.0913141Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:35.0913947Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:35.0914851Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:35.0915610Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:35.0916368Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:35.0917193Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:35.0918077Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:35.0918926Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:35.0919806Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:35.0920653Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:35.0921489Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:35.0922288Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:35.0923093Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:35.0923846Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:35.0924602Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:35.0925406Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:35.0926192Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:35.0926999Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:35.0927862Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:35.0928679Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:35.0929467Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:35.0930219Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:35.0931058Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:35.0933029Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:35.0935490Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:35.0936250Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:35.0937221Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:35.0938054Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:35.0938886Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:35.0939850Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:35.0940679Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:35.0941625Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:35.0942519Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:35.0943409Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:35.0944303Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:35.0945210Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:35.0946197Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:35.0947108Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:35.0947948Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:35.0948861Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:35.0949830Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:35.0950705Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:35.0951587Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:35.0952548Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:35.0953279Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:35.0954024Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:35.0954758Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:35.0955483Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:35.0956195Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:35.0956925Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:35.0957829Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:35.0958666Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:35.0959517Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:35.0960230Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:35.0960899Z V1204 11:21:02.609000 99813 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:35.0961578Z I1204 11:21:02.609000 99813 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:35.0962070Z V1204 11:21:02.609000 99813 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:35.0962581Z V1204 11:21:02.610000 99813 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:35.0963082Z I1204 11:21:02.610000 99813 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:35.0963862Z V1204 11:21:02.623000 99813 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:35.0964303Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.0964845Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:35.0965798Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.0966541Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.0967519Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.0968511Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.0969501Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.0970386Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:35.0971230Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:35.0972023Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:35.0973082Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:35.0973899Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:35.0974686Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_3: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.0975326Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:35.0975750Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.0976180Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.0976749Z V1204 11:21:02.630000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:35.0977414Z V1204 11:21:02.630000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:35.0978423Z V1204 11:21:02.631000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.0978991Z V1204 11:21:02.631000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0980003Z V1204 11:21:02.632000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.0980503Z V1204 11:21:02.632000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0981397Z V1204 11:21:02.634000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.0981878Z V1204 11:21:02.634000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0982832Z V1204 11:21:02.637000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:35.0983288Z V1204 11:21:02.637000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0984180Z V1204 11:21:02.638000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:35.0984631Z V1204 11:21:02.638000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0985536Z V1204 11:21:02.639000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:35.0986010Z V1204 11:21:02.639000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0986849Z V1204 11:21:02.641000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:35.0987293Z V1204 11:21:02.642000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0988122Z V1204 11:21:02.644000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_3 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.0988584Z V1204 11:21:02.644000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.0989160Z V1204 11:21:02.645000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:35.0989732Z V1204 11:21:02.645000 99813 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:35.0990515Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:35.0990619Z warnings.warn( 2025-12-04T11:24:35.0991107Z V1204 11:21:02.646000 99813 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:35.0991259Z =================================== FAILURES =================================== 2025-12-04T11:24:35.0991729Z _ TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False _ 2025-12-04T11:24:35.0991865Z Traceback (most recent call last): 2025-12-04T11:24:35.0992477Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 275, in test_decompose_linear_mixed_precision 2025-12-04T11:24:35.0992587Z self.assertEqual( 2025-12-04T11:24:35.0993061Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:35.0993225Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:35.0993772Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:35.0993977Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:35.0994108Z AssertionError: Scalars are not equal! 2025-12-04T11:24:35.0994118Z 2025-12-04T11:24:35.0994236Z Expected 1 but got 0. 2025-12-04T11:24:35.0994376Z Absolute difference: 1 2025-12-04T11:24:35.0994487Z Relative difference: 1.0 2025-12-04T11:24:35.0994504Z 2025-12-04T11:24:35.0994721Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:35.0995594Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:35.0995601Z 2025-12-04T11:24:35.0995888Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:35.0996245Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:35.0996356Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:35.0996485Z stats [('calls_captured', 1)] 2025-12-04T11:24:35.0997254Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_nodes', 3), ('pad_mm_bench', 3), ('pattern_matcher_count', 2), ('fxgraph_cache_miss', 1), ('decompose_mm', 1)] 2025-12-04T11:24:35.0997554Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:35.0997661Z graph_break [] 2025-12-04T11:24:35.0997883Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:35.0998392Z I1204 11:20:59.024000 99813 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:35.0998987Z I1204 11:20:59.270000 99813 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:35.0999433Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.0999920Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:35.1000853Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.1001721Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.1002509Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:35.1002971Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:35.1003384Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.1004371Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.1005438Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, None); l_x_ = l_self_modules_linear_parameters_weight_ = None 2025-12-04T11:24:35.1005911Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:35.1006339Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.1006715Z V1204 11:20:59.273000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.1007273Z I1204 11:21:01.477000 99813 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:35.1008029Z V1204 11:21:01.965000 99813 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:35.1008754Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v: 2025-12-04T11:24:35.1009410Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [trz4xcrmguerwj2nwxhu4t6zwtig5vb4aexoqw3sqeb3z2aw5kv] gm: GraphModule() 2025-12-04T11:24:35.1009683Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.1009966Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.1010234Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.1010697Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:35.1011555Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.1012368Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.1013122Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:35.1013688Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, permute) 2025-12-04T11:24:35.1014325Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_3 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.1014822Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, convert_element_type_1, permute_3) 2025-12-04T11:24:35.1015112Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.1015719Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:35.1018084Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.1020469Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.1021109Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:35.1022030Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:35.1022780Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:35.1023610Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:35.1024389Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:35.1025107Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:35.1025817Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:35.1026572Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:35.1027288Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:35.1028028Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [reylvtt3rqfvzq5b3fjekfvick4dxsjoog5svxckgogwqosgp6z] fx_kwargs[static_input_idxs]: [0] 2025-12-04T11:24:35.1028719Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[0]: 1 2025-12-04T11:24:35.1029585Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:35.1030381Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:35.1031533Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:35.1032366Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:35.1061383Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:35.1062507Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:35.1063289Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:35.1064093Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:35.1064946Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:35.1065786Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:35.1066499Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:35.1067277Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:35.1068057Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:35.1068869Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:35.1069723Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:35.1070565Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:35.1071418Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:35.1072297Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:35.1073157Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:35.1074045Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:35.1074907Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:35.1075743Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:35.1076750Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:35.1077501Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:35.1078328Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:35.1079081Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:35.1079944Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:35.1080709Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:35.1081409Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:35.1082206Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:35.1082947Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:35.1083701Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:35.1084568Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:35.1085322Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:35.1086104Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:35.1086871Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:35.1087680Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:35.1088460Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:35.1089232Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:35.1089998Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:35.1090796Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:35.1091620Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:35.1092385Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:35.1093155Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:35.1093957Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:35.1094720Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:35.1095526Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:35.1096480Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:35.1097340Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:35.1098162Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:35.1098955Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:35.1099817Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:35.1100639Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:35.1101406Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:35.1102144Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:35.1102954Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:35.1103914Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:35.1104720Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:35.1105508Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:35.1106331Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:35.1107133Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:35.1108321Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:35.1109124Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:35.1109985Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:35.1110887Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:35.1111757Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:35.1112577Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:35.1113414Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:35.1114289Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:35.1115106Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:35.1116033Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.1116906Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:35.1117862Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.1118678Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:35.1119640Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.1120459Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:35.1121341Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:35.1122097Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:35.1122853Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:35.1123722Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:35.1124579Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:35.1125339Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:35.1126147Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:35.1126939Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:35.1127807Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:35.1128676Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:35.1129595Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:35.1130462Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:35.1131229Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:35.1132056Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:35.1132975Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:35.1133780Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:35.1134566Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:35.1135483Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:35.1136352Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:35.1137274Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:35.1138156Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:35.1138945Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:35.1139777Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:35.1140505Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:35.1141284Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:35.1142098Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:35.1142888Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:35.1143766Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:35.1144590Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:35.1145406Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:35.1146223Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:35.1146959Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:35.1147768Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:35.1148709Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:35.1149644Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:35.1150569Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:35.1151403Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:35.1152234Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:35.1153063Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:35.1153895Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:35.1154787Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:35.1155681Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:35.1156458Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:35.1157259Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:35.1158099Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:35.1158923Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:35.1159745Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:35.1160537Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:35.1161371Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:35.1162145Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:35.1162912Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:35.1163727Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:35.1164577Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:35.1165415Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:35.1166277Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:35.1167066Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:35.1167967Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:35.1168749Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:35.1169619Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:35.1170427Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:35.1171270Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:35.1172028Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:35.1172818Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:35.1173592Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:35.1174414Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:35.1175299Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:35.1176123Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:35.1177014Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:35.1177866Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:35.1178631Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:35.1179522Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:35.1180406Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:35.1181215Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:35.1182004Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:35.1182846Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:35.1183651Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:35.1184428Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:35.1185194Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:35.1185970Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:35.1186711Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:35.1187520Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:35.1188382Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:35.1189219Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:35.1190041Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:35.1190827Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:35.1191628Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:35.1192437Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:35.1193237Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:35.1194087Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:35.1194956Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:35.1195781Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:35.1196846Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:35.1197655Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:35.1198466Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:35.1199334Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:35.1200136Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:35.1200914Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:35.1201785Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:35.1202574Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:35.1203444Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:35.1204299Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:35.1205105Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:35.1205866Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:35.1206677Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:35.1207531Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:35.1208325Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:35.1209155Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:35.1210102Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:35.1211108Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:35.1212089Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:35.1213087Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:35.1214074Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:35.1215068Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:35.1216107Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:35.1217146Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:35.1218147Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:35.1218981Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:35.1219840Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:35.1220665Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:35.1221569Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:35.1222415Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:35.1223193Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:35.1224024Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:35.1224782Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:35.1225583Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:35.1226379Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:35.1227165Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:35.1227964Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:35.1228893Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:35.1229715Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:35.1230524Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:35.1231279Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:35.1232155Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:35.1232932Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:35.1233716Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:35.1234548Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:35.1235431Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:35.1236218Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:35.1237054Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:35.1237777Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:35.1238557Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:35.1239399Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:35.1240221Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:35.1241172Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:35.1241966Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:35.1242707Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:35.1243546Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:35.1244359Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:35.1245255Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:35.1246057Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:35.1246867Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:35.1247839Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:35.1248727Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:35.1249564Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:35.1250346Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:35.1251292Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:35.1252089Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:35.1252877Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:35.1253692Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:35.1254527Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:35.1255425Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:35.1256244Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:35.1257043Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:35.1257847Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:35.1258651Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:35.1259405Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:35.1260205Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:35.1261024Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:35.1261856Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:35.1262656Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:35.1263551Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.1264434Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.1265201Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:35.1266063Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:35.1266901Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:35.1267791Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:35.1268653Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:35.1269539Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:35.1270391Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:35.1271249Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:35.1272093Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:35.1272915Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:35.1273743Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:35.1274556Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:35.1275383Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:35.1276209Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:35.1277081Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:35.1277912Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:35.1278756Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:35.1279593Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:35.1280399Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:35.1281290Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:35.1282154Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:35.1283024Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:35.1283979Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:35.1284876Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:35.1285810Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:35.1286706Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:35.1287583Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:35.1288442Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:35.1289272Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:35.1290215Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:35.1291087Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:35.1291949Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:35.1292819Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:35.1293639Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:35.1294459Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:35.1295308Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:35.1296332Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:35.1297211Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:35.1298040Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:35.1298855Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:35.1299786Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:35.1300660Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:35.1301466Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:35.1302274Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:35.1303163Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:35.1304108Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:35.1304928Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:35.1305794Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:35.1306653Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:35.1307485Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:35.1308346Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:35.1309260Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:35.1310052Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:35.1310851Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:35.1311698Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:35.1312582Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:35.1313375Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:35.1314185Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:35.1315023Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:35.1316006Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:35.1316884Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.1317738Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:35.1318608Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:35.1319469Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:35.1320369Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:35.1321164Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:35.1322150Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:35.1322979Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:35.1323822Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:35.1324609Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:35.1325475Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:35.1326374Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:35.1327265Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:35.1328239Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:35.1329033Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:35.1329873Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:35.1330702Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:35.1331585Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:35.1332528Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:35.1333402Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:35.1334241Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:35.1335071Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:35.1335993Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:35.1336936Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:35.1337809Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:35.1338609Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:35.1339467Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:35.1340360Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:35.1341154Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:35.1342151Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:35.1343010Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:35.1343830Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:35.1344652Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:35.1345545Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:35.1346459Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:35.1347361Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:35.1348275Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:35.1349204Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:35.1350070Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:35.1350925Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:35.1351805Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:35.1352768Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:35.1353629Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:35.1354478Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:35.1355287Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:35.1356118Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:35.1356999Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:35.1357891Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:35.1358768Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:35.1359656Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:35.1360393Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:35.1361185Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:35.1362009Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:35.1362806Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:35.1363630Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:35.1364469Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:35.1365268Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:35.1366139Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:35.1367082Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:35.1367973Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:35.1368815Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:35.1369591Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:35.1370449Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:35.1371283Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:35.1372136Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:35.1372994Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:35.1373870Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:35.1374736Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:35.1375570Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:35.1376386Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:35.1377310Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:35.1378181Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:35.1379055Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:35.1379893Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:35.1380651Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:35.1381598Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:35.1382402Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:35.1383172Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:35.1383942Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:35.1384762Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:35.1385557Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:35.1386420Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:35.1387198Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:35.1387950Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:35.1388788Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:35.1389656Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:35.1390511Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:35.1391388Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:35.1392236Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:35.1393061Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:35.1393855Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:35.1394674Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:35.1395410Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:35.1396340Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:35.1397096Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:35.1397883Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:35.1398705Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:35.1399556Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:35.1400446Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:35.1401209Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:35.1401976Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:35.1402763Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:35.1404721Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:35.1407184Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:35.1407955Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:35.1408912Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:35.1409704Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:35.1410552Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:35.1411500Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:35.1412351Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:35.1413312Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:35.1414241Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:35.1415119Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:35.1416013Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:35.1417020Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:35.1418005Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:35.1418897Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:35.1419727Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:35.1420666Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:35.1421589Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:35.1422518Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:35.1423392Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:35.1424355Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:35.1425138Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:35.1425839Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:35.1426592Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:35.1427301Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:35.1428028Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:35.1428784Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:35.1429652Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:35.1430510Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:35.1431351Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:35.1432084Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:35.1432776Z V1204 11:21:01.975000 99813 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:35.1433465Z I1204 11:21:01.976000 99813 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:35.1433948Z V1204 11:21:01.976000 99813 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:35.1434427Z V1204 11:21:01.976000 99813 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:35.1434941Z I1204 11:21:01.977000 99813 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:35.1435700Z V1204 11:21:02.018000 99813 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:35.1436162Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.1436655Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:35.1437634Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.1438372Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.1439352Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.1440352Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.1441334Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.1442224Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:35.1443060Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:35.1443864Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:35.1444905Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:35.1445736Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:35.1446553Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_3: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.1447189Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:35.1447620Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.1448004Z V1204 11:21:02.023000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.1448587Z V1204 11:21:02.026000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:35.1449144Z V1204 11:21:02.027000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:35.1450164Z V1204 11:21:02.028000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.1450665Z V1204 11:21:02.028000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.1451698Z V1204 11:21:02.060000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.1452204Z V1204 11:21:02.060000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.1453085Z V1204 11:21:02.062000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.1453582Z V1204 11:21:02.062000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.1454517Z V1204 11:21:02.068000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:35.1454977Z V1204 11:21:02.068000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.1455851Z V1204 11:21:02.070000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:35.1456297Z V1204 11:21:02.070000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.1457309Z V1204 11:21:02.071000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:35.1457737Z V1204 11:21:02.072000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.1458587Z V1204 11:21:02.074000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:35.1459014Z V1204 11:21:02.074000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.1459854Z V1204 11:21:02.077000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_3 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.1460341Z V1204 11:21:02.077000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.1460914Z V1204 11:21:02.078000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:35.1461504Z V1204 11:21:02.078000 99813 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:35.1462245Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:35.1462362Z warnings.warn( 2025-12-04T11:24:35.1462838Z V1204 11:21:02.079000 99813 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:35.1463069Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:35.1463196Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:35.1463315Z stats [('calls_captured', 1)] 2025-12-04T11:24:35.1463539Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:35.1464353Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_nodes', 3), ('pad_mm_bench', 3), ('pattern_matcher_count', 2), ('fxgraph_cache_miss', 1), ('decompose_mm', 1)] 2025-12-04T11:24:35.1464454Z graph_break [] 2025-12-04T11:24:35.1464688Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:35.1465120Z V1204 11:21:02.321000 99813 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmp5tt8qcoa 2025-12-04T11:24:35.1465553Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.1466053Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:35.1467034Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.1467897Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.1468639Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:35.1469101Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:35.1469553Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.1470524Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.1471560Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, None); l_x_ = l_self_modules_linear_parameters_weight_ = None 2025-12-04T11:24:35.1472028Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:35.1472452Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.1472867Z V1204 11:21:02.347000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.1473575Z V1204 11:21:02.599000 99813 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:35.1474305Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v: 2025-12-04T11:24:35.1474962Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [trz4xcrmguerwj2nwxhu4t6zwtig5vb4aexoqw3sqeb3z2aw5kv] gm: GraphModule() 2025-12-04T11:24:35.1475249Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.1475517Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.1475786Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.1476266Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:35.1477081Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.1477945Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.1478685Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:35.1479261Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, permute) 2025-12-04T11:24:35.1479891Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_3 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.1480411Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, convert_element_type_1, permute_3) 2025-12-04T11:24:35.1480714Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.1481283Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:35.1483583Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.1485937Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.1486588Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:35.1487495Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:35.1488259Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:35.1489066Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:35.1489854Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:35.1490557Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:35.1491267Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:35.1492003Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:35.1492748Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:35.1493499Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [reylvtt3rqfvzq5b3fjekfvick4dxsjoog5svxckgogwqosgp6z] fx_kwargs[static_input_idxs]: [0] 2025-12-04T11:24:35.1494178Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[0]: 1 2025-12-04T11:24:35.1495058Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:35.1495876Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:35.1497277Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:35.1498072Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:35.1526976Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:35.1528118Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:35.1528890Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:35.1529651Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:35.1530539Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:35.1531381Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:35.1532104Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:35.1532876Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:35.1533704Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:35.1534477Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:35.1535326Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:35.1536158Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:35.1537117Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:35.1537975Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:35.1538818Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:35.1539721Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:35.1540566Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:35.1541421Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:35.1542418Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:35.1543170Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:35.1544011Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:35.1544749Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:35.1545594Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:35.1546398Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:35.1547112Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:35.1547897Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:35.1548642Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:35.1549432Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:35.1550264Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:35.1551031Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:35.1551801Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:35.1552617Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:35.1553386Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:35.1554159Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:35.1554942Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:35.1555698Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:35.1556502Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:35.1557307Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:35.1558083Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:35.1558846Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:35.1559651Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:35.1560425Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:35.1561180Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:35.1562030Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:35.1562827Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:35.1563657Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:35.1564443Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:35.1565326Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:35.1574330Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:35.1575223Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:35.1575993Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:35.1577041Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:35.1577972Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:35.1578783Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:35.1579573Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:35.1580416Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:35.1581221Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:35.1582415Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:35.1583206Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:35.1584086Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:35.1584966Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:35.1585780Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:35.1586713Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:35.1587553Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:35.1588443Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:35.1589276Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:35.1590224Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.1591045Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:35.1591999Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.1592807Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:35.1593774Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.1594605Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:35.1595475Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:35.1596416Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:35.1597167Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:35.1598065Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:35.1598924Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:35.1599680Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:35.1600506Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:35.1601285Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:35.1602137Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:35.1603025Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:35.1603946Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:35.1604815Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:35.1605623Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:35.1606461Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:35.1607331Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:35.1608146Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:35.1608915Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:35.1609852Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:35.1610710Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:35.1611573Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:35.1612448Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:35.1613237Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:35.1614096Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:35.1614809Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:35.1615607Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:35.1616403Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:35.1617255Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:35.1618090Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:35.1618941Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:35.1619768Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:35.1620561Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:35.1621315Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:35.1622132Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:35.1623034Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:35.1623969Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:35.1624886Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:35.1625748Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:35.1626567Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:35.1627408Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:35.1628231Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:35.1629124Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:35.1630021Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:35.1630802Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:35.1631624Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:35.1632449Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:35.1633291Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:35.1634059Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:35.1634875Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:35.1635714Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:35.1636489Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:35.1637277Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:35.1638102Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:35.1638931Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:35.1639748Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:35.1640601Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:35.1641420Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:35.1642295Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:35.1643084Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:35.1643942Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:35.1644758Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:35.1645571Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:35.1646335Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:35.1647108Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:35.1647882Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:35.1648717Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:35.1649566Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:35.1650438Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:35.1651256Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:35.1652118Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:35.1652865Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:35.1653786Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:35.1654640Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:35.1655448Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:35.1656255Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:35.1657176Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:35.1657961Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:35.1658727Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:35.1659494Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:35.1660266Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:35.1661046Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:35.1661835Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:35.1662678Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:35.1663528Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:35.1664334Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:35.1665118Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:35.1665890Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:35.1666718Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:35.1667527Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:35.1668366Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:35.1669274Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:35.1670075Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:35.1670908Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:35.1671709Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:35.1672523Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:35.1673404Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:35.1674190Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:35.1674982Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:35.1675839Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:35.1676623Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:35.1677477Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:35.1678337Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:35.1679152Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:35.1679900Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:35.1680733Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:35.1681521Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:35.1682340Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:35.1683163Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:35.1684111Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:35.1685147Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:35.1686111Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:35.1687080Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:35.1688044Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:35.1689084Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:35.1690082Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:35.1691052Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:35.1692050Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:35.1692877Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:35.1693743Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:35.1694556Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:35.1695478Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:35.1696430Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:35.1697276Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:35.1698064Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:35.1698895Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:35.1699705Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:35.1700498Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:35.1701302Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:35.1702127Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:35.1703008Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:35.1703835Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:35.1704643Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:35.1705452Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:35.1706258Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:35.1707039Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:35.1707803Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:35.1708644Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:35.1709518Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:35.1710307Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:35.1711134Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:35.1711856Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:35.1712654Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:35.1713446Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:35.1714280Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:35.1715249Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:35.1716041Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:35.1716781Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:35.1717646Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:35.1718470Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:35.1719322Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:35.1720140Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:35.1720932Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:35.1721904Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:35.1722784Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:35.1723617Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:35.1724407Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:35.1725332Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:35.1726136Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:35.1726906Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:35.1727715Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:35.1728542Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:35.1729400Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:35.1730234Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:35.1730996Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:35.1731816Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:35.1732606Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:35.1733404Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:35.1734196Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:35.1734965Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:35.1735812Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:35.1736590Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:35.1737553Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.1738417Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.1739181Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:35.1740043Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:35.1740865Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:35.1741763Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:35.1742616Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:35.1743514Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:35.1744346Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:35.1745231Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:35.1746021Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:35.1746864Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:35.1747697Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:35.1748506Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:35.1749371Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:35.1750175Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:35.1751029Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:35.1751840Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:35.1752679Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:35.1753522Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:35.1754322Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:35.1755218Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:35.1756066Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:35.1756942Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:35.1757887Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:35.1758778Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:35.1759695Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:35.1760596Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:35.1761453Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:35.1762296Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:35.1763156Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:35.1764068Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:35.1764934Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:35.1765818Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:35.1766628Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:35.1767453Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:35.1768246Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:35.1769141Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:35.1769916Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:35.1770743Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:35.1771550Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:35.1772354Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:35.1773250Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:35.1774114Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:35.1774929Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:35.1775719Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:35.1776622Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:35.1777568Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:35.1778426Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:35.1779291Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:35.1780148Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:35.1780991Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:35.1781867Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:35.1782755Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:35.1783532Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:35.1784327Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:35.1785202Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:35.1785987Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:35.1786792Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:35.1787583Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:35.1788428Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:35.1789407Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:35.1790286Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.1791123Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:35.1791994Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:35.1792860Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:35.1793715Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:35.1794562Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:35.1795530Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:35.1796540Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:35.1797374Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:35.1798228Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:35.1799066Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:35.1799963Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:35.1800858Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:35.1801840Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:35.1802652Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:35.1803472Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:35.1804295Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:35.1805179Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:35.1806131Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:35.1807020Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:35.1807846Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:35.1808685Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:35.1809590Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:35.1810452Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:35.1811348Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:35.1812147Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:35.1813006Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:35.1813870Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:35.1814670Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:35.1815599Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:35.1816467Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:35.1817352Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:35.1818143Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:35.1819026Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:35.1819946Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:35.1820864Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:35.1821775Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:35.1822717Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:35.1823564Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:35.1824427Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:35.1825309Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:35.1826225Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:35.1827131Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:35.1827959Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:35.1828783Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:35.1829604Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:35.1830513Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:35.1831366Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:35.1832238Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:35.1833133Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:35.1833898Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:35.1834669Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:35.1835476Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:35.1836285Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:35.1837086Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:35.1837996Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:35.1838797Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:35.1839664Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:35.1840619Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:35.1841501Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:35.1842318Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:35.1843111Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:35.1843967Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:35.1844799Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:35.1845650Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:35.1846537Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:35.1847372Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:35.1848250Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:35.1849070Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:35.1849920Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:35.1850755Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:35.1851621Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:35.1852511Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:35.1853334Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:35.1854112Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:35.1855037Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:35.1855859Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:35.1856618Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:35.1857453Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:35.1858244Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:35.1859087Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:35.1859963Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:35.1860728Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:35.1861487Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:35.1862341Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:35.1863177Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:35.1864038Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:35.1864906Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:35.1865804Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:35.1866585Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:35.1867388Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:35.1868195Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:35.1868938Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:35.1869747Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:35.1870495Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:35.1871292Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:35.1872102Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:35.1872963Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:35.1873784Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:35.1874543Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:35.1875356Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:35.1876147Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:35.1878160Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:35.1880582Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:35.1881359Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:35.1882301Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:35.1883101Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:35.1883928Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:35.1884889Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:35.1885758Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:35.1886661Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:35.1887574Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:35.1888453Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:35.1889364Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:35.1890264Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:35.1891296Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:35.1892177Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:35.1893016Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:35.1893935Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:35.1894890Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:35.1895782Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:35.1896796Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:35.1897843Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:35.1898650Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:35.1899365Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:35.1900103Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:35.1900807Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:35.1901534Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:35.1902321Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:35.1903244Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:35.1904084Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:35.1904936Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:35.1905652Z V1204 11:21:02.608000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:35.1906317Z V1204 11:21:02.609000 99813 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:35.1907000Z I1204 11:21:02.609000 99813 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:35.1907536Z V1204 11:21:02.609000 99813 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:35.1908033Z V1204 11:21:02.610000 99813 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:35.1908534Z I1204 11:21:02.610000 99813 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:35.1909292Z V1204 11:21:02.623000 99813 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:35.1909797Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.1910293Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:35.1911257Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.1911984Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.1912973Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.1913986Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.1914977Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.1915859Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:35.1916727Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:35.1917498Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:35.1918530Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:35.1919354Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:35.1920132Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_3: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.1920776Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:35.1921238Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.1921635Z V1204 11:21:02.627000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.1922204Z V1204 11:21:02.630000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:35.1922768Z V1204 11:21:02.630000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:35.1923824Z V1204 11:21:02.631000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.1924328Z V1204 11:21:02.631000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.1925347Z V1204 11:21:02.632000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.1925845Z V1204 11:21:02.632000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.1926774Z V1204 11:21:02.634000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.1927224Z V1204 11:21:02.634000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.1928160Z V1204 11:21:02.637000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:35.1928621Z V1204 11:21:02.637000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.1929497Z V1204 11:21:02.638000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:35.1929992Z V1204 11:21:02.638000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.1930896Z V1204 11:21:02.639000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:35.1931339Z V1204 11:21:02.639000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.1932182Z V1204 11:21:02.641000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:35.1932606Z V1204 11:21:02.642000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.1933453Z V1204 11:21:02.644000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_3 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.1933902Z V1204 11:21:02.644000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.1934490Z V1204 11:21:02.645000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:35.1935097Z V1204 11:21:02.645000 99813 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:35.1935845Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:35.1935963Z warnings.warn( 2025-12-04T11:24:35.1936437Z V1204 11:21:02.646000 99813 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:35.1936677Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:35.1936789Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:35.1937013Z stats [('calls_captured', 1)] 2025-12-04T11:24:35.1937254Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:35.1938026Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_nodes', 3), ('pad_mm_bench', 3), ('pattern_matcher_count', 2), ('fxgraph_cache_miss', 1), ('decompose_mm', 1)] 2025-12-04T11:24:35.1938141Z graph_break [] 2025-12-04T11:24:35.1938365Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:35.1938800Z V1204 11:21:02.667000 99813 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmp1dymz4ks 2025-12-04T11:24:35.1939254Z V1204 11:21:02.694000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.1939777Z V1204 11:21:02.694000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:35.1940729Z V1204 11:21:02.694000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.1941574Z V1204 11:21:02.694000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.1942313Z V1204 11:21:02.694000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:35.1942774Z V1204 11:21:02.694000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:35.1943222Z V1204 11:21:02.694000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.1944201Z V1204 11:21:02.694000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.1945232Z V1204 11:21:02.694000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, None); l_x_ = l_self_modules_linear_parameters_weight_ = None 2025-12-04T11:24:35.1945717Z V1204 11:21:02.694000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:35.1946129Z V1204 11:21:02.694000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.1946509Z V1204 11:21:02.694000 99813 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.1947235Z V1204 11:21:02.947000 99813 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:35.1947985Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v: 2025-12-04T11:24:35.1948654Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [trz4xcrmguerwj2nwxhu4t6zwtig5vb4aexoqw3sqeb3z2aw5kv] gm: GraphModule() 2025-12-04T11:24:35.1948927Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.1949197Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.1949481Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.1949949Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:35.1950808Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.1951618Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.1952356Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:35.1952929Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, permute) 2025-12-04T11:24:35.1953601Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_3 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.1954103Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, convert_element_type_1, permute_3) 2025-12-04T11:24:35.1954393Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.1954970Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:35.1957266Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.1959650Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.1960285Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:35.1961200Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:35.1961920Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:35.1962759Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:35.1963551Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:35.1964262Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:35.1964991Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:35.1965756Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:35.1966476Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:35.1967215Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [reylvtt3rqfvzq5b3fjekfvick4dxsjoog5svxckgogwqosgp6z] fx_kwargs[static_input_idxs]: [0] 2025-12-04T11:24:35.1967892Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[0]: 1 2025-12-04T11:24:35.1968775Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:35.1969591Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:35.1970775Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:35.1971555Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:35.2000743Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:35.2001816Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:35.2002637Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:35.2003416Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:35.2004255Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:35.2005101Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:35.2005816Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:35.2006636Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:35.2007422Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:35.2008190Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:35.2009054Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:35.2009883Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:35.2010793Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:35.2011635Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:35.2012490Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:35.2013382Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:35.2014239Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:35.2015067Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:35.2016108Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:35.2016925Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:35.2017756Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:35.2018512Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:35.2019382Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:35.2020140Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:35.2020853Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:35.2021638Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:35.2022427Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:35.2023185Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:35.2024035Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:35.2024788Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:35.2025552Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:35.2026367Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:35.2027128Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:35.2027916Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:35.2028688Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:35.2029452Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:35.2030247Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:35.2031020Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:35.2031820Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:35.2032583Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:35.2033397Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:35.2034191Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:35.2034963Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:35.2035771Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:35.2036565Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:35.2037399Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:35.2038215Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:35.2039084Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:35.2039853Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:35.2040614Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:35.2041353Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:35.2042194Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:35.2043126Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:35.2043919Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:35.2044718Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:35.2045541Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:35.2046308Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:35.2047551Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:35.2048348Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:35.2049225Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:35.2050139Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:35.2050963Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:35.2051769Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:35.2052622Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:35.2053498Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:35.2054332Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:35.2055278Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.2056093Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:35.2057112Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.2057917Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:35.2058901Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.2059718Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:35.2060598Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:35.2061351Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:35.2062097Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:35.2062947Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:35.2063827Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:35.2064593Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:35.2065400Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:35.2066226Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:35.2067061Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:35.2067924Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:35.2068854Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:35.2069713Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:35.2070530Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:35.2071350Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:35.2072237Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:35.2073036Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:35.2073809Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:35.2074747Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:35.2075601Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:35.2076472Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:35.2077330Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:35.2078141Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:35.2078928Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:35.2079674Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:35.2080467Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:35.2081267Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:35.2082100Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:35.2082925Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:35.2083759Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:35.2084577Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:35.2085392Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:35.2086165Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:35.2086956Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:35.2087874Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:35.2088786Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:35.2089717Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:35.2090561Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:35.2091385Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:35.2092224Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:35.2093041Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:35.2093958Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:35.2094804Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:35.2095622Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:35.2096586Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:35.2097470Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:35.2098374Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:35.2099143Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:35.2099935Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:35.2100755Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:35.2101540Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:35.2102422Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:35.2103235Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:35.2104048Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:35.2104874Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:35.2105745Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:35.2106571Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:35.2107455Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:35.2108235Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:35.2109102Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:35.2109911Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:35.2110701Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:35.2111504Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:35.2112278Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:35.2113066Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:35.2113892Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:35.2114775Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:35.2115609Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:35.2116424Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:35.2117282Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:35.2118061Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:35.2118966Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:35.2119806Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:35.2120626Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:35.2121414Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:35.2122268Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:35.2123044Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:35.2123804Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:35.2124586Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:35.2125348Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:35.2126102Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:35.2126874Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:35.2127739Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:35.2128586Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:35.2129395Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:35.2130222Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:35.2130975Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:35.2131795Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:35.2132588Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:35.2133442Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:35.2134344Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:35.2135153Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:35.2135999Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:35.2136788Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:35.2137670Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:35.2138528Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:35.2139330Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:35.2140106Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:35.2140962Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:35.2141772Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:35.2142584Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:35.2143875Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:35.2144685Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:35.2145447Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:35.2146264Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:35.2147079Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:35.2147879Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:35.2148699Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:35.2149660Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:35.2150679Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:35.2151664Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:35.2152620Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:35.2153587Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:35.2154629Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:35.2155621Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:35.2156606Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:35.2157596Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:35.2158444Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:35.2159265Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:35.2160118Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:35.2161018Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:35.2161846Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:35.2162668Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:35.2163444Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:35.2164214Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:35.2165006Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:35.2165799Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:35.2166639Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:35.2167417Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:35.2168310Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:35.2169125Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:35.2169951Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:35.2170733Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:35.2171551Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:35.2172320Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:35.2173080Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:35.2173922Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:35.2174764Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:35.2175586Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:35.2176401Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:35.2177251Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:35.2178029Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:35.2178855Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:35.2179696Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:35.2180636Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:35.2181444Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:35.2182203Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:35.2183050Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:35.2183857Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:35.2184707Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:35.2185520Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:35.2186352Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:35.2187306Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:35.2188174Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:35.2189038Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:35.2189820Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:35.2190747Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:35.2191565Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:35.2192334Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:35.2193157Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:35.2193974Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:35.2194883Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:35.2195704Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:35.2196612Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:35.2197419Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:35.2198212Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:35.2199040Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:35.2199828Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:35.2200612Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:35.2201443Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:35.2202241Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:35.2203144Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.2204006Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.2204788Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:35.2205639Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:35.2206488Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:35.2207330Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:35.2208228Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:35.2209118Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:35.2209950Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:35.2210874Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:35.2211666Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:35.2212494Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:35.2213311Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:35.2214135Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:35.2214989Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:35.2215798Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:35.2216647Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:35.2217523Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:35.2218383Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:35.2219204Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:35.2220019Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:35.2220902Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:35.2221772Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:35.2222637Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:35.2223550Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:35.2224483Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:35.2225402Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:35.2226311Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:35.2227193Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:35.2228055Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:35.2228878Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:35.2229794Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:35.2230705Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:35.2231550Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:35.2232374Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:35.2233190Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:35.2234003Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:35.2234887Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:35.2235664Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:35.2236496Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:35.2237311Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:35.2238136Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:35.2238996Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:35.2239905Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:35.2240715Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:35.2241506Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:35.2242403Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:35.2243322Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:35.2244160Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:35.2245012Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:35.2245883Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:35.2246746Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:35.2247605Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:35.2248486Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:35.2249261Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:35.2250072Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:35.2250936Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:35.2251731Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:35.2252520Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:35.2253322Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:35.2254161Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:35.2255103Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:35.2256014Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.2256843Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:35.2257789Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:35.2258655Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:35.2259559Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:35.2260360Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:35.2261326Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:35.2262166Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:35.2263027Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:35.2263828Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:35.2264650Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:35.2265559Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:35.2266446Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:35.2267425Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:35.2268219Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:35.2269053Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:35.2269890Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:35.2270767Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:35.2271683Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:35.2272585Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:35.2273427Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:35.2274262Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:35.2275194Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:35.2276061Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:35.2276915Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:35.2277726Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:35.2278571Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:35.2279445Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:35.2280235Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:35.2281162Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:35.2282022Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:35.2282850Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:35.2283647Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:35.2284523Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:35.2285449Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:35.2286363Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:35.2287250Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:35.2288220Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:35.2289072Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:35.2289944Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:35.2290816Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:35.2291771Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:35.2292629Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:35.2293480Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:35.2294300Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:35.2295159Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:35.2296173Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:35.2297067Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:35.2297959Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:35.2298836Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:35.2299652Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:35.2300408Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:35.2301213Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:35.2302021Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:35.2302823Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:35.2303646Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:35.2304471Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:35.2305352Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:35.2306291Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:35.2307190Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:35.2308031Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:35.2308796Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:35.2309662Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:35.2310485Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:35.2311399Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:35.2312245Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:35.2313097Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:35.2313961Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:35.2314779Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:35.2315644Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:35.2316479Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:35.2317361Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:35.2318238Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:35.2319078Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:35.2319801Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:35.2320758Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:35.2321580Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:35.2322341Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:35.2323133Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:35.2324906Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:35.2326610Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:35.2328403Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:35.2330184Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:35.2331882Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:35.2333607Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:35.2335411Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:35.2337360Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:35.2339231Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:35.2341138Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:35.2342928Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:35.2344647Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:35.2346375Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:35.2348059Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:35.2349710Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:35.2351393Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:35.2353081Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:35.2354807Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:35.2356611Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:35.2358452Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:35.2360197Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:35.2361869Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:35.2363547Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:35.2366450Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:35.2371020Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:35.2374384Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:35.2376201Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:35.2378115Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:35.2379872Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:35.2381793Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:35.2383738Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:35.2385614Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:35.2387596Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:35.2389529Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:35.2391436Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:35.2393478Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:35.2395507Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:35.2397694Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:35.2399554Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:35.2401529Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:35.2403499Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:35.2405444Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:35.2407361Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:35.2409336Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:35.2411235Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:35.2412791Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:35.2414370Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:35.2415956Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:35.2417583Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:35.2419153Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:35.2420929Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:35.2422772Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:35.2424596Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:35.2426298Z V1204 11:21:02.956000 99813 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:35.2427844Z V1204 11:21:02.957000 99813 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:35.2429316Z I1204 11:21:02.957000 99813 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fm2p4qdzbntqthrm4xfjjjpant2c2uamhwpdjxgfbrysicabcc6v 2025-12-04T11:24:35.2430605Z V1204 11:21:02.957000 99813 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:35.2431703Z V1204 11:21:02.957000 99813 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:35.2432818Z I1204 11:21:02.958000 99813 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 2 2025-12-04T11:24:35.2434245Z V1204 11:21:02.970000 99813 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:35.2435587Z V1204 11:21:02.975000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.2436667Z V1204 11:21:02.975000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:35.2438245Z V1204 11:21:02.975000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.2440048Z V1204 11:21:02.975000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.2441932Z V1204 11:21:02.975000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.2444020Z V1204 11:21:02.975000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.2446106Z V1204 11:21:02.975000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.2448118Z V1204 11:21:02.975000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:35.2449990Z V1204 11:21:02.975000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:35.2451739Z V1204 11:21:02.975000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:35.2453670Z V1204 11:21:02.975000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:35.2455668Z V1204 11:21:02.975000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:35.2457521Z V1204 11:21:02.975000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_3: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.2459100Z V1204 11:21:02.975000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:35.2460287Z V1204 11:21:02.975000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.2461245Z V1204 11:21:02.975000 99813 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.2462343Z V1204 11:21:02.977000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:35.2463651Z V1204 11:21:02.978000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:35.2465371Z V1204 11:21:02.978000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.2467021Z V1204 11:21:02.979000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.2468676Z V1204 11:21:02.980000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.2470325Z V1204 11:21:02.980000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.2471923Z V1204 11:21:02.981000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.2473412Z V1204 11:21:02.982000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.2474936Z V1204 11:21:02.984000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:35.2476470Z V1204 11:21:02.985000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.2477952Z V1204 11:21:02.986000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:35.2479427Z V1204 11:21:02.986000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.2480921Z V1204 11:21:02.987000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:35.2482443Z V1204 11:21:02.987000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.2483867Z V1204 11:21:02.989000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:35.2485285Z V1204 11:21:02.989000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.2486707Z V1204 11:21:02.991000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_3 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.2488137Z V1204 11:21:02.992000 99813 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.2489311Z V1204 11:21:02.992000 99813 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, convert_element_type_1, permute_3) 2025-12-04T11:24:35.2490601Z V1204 11:21:02.993000 99813 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 2 2025-12-04T11:24:35.2492060Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:35.2493019Z warnings.warn( 2025-12-04T11:24:35.2493716Z V1204 11:21:02.993000 99813 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:35.2495182Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-e6eefab2abf634a5.xml - 2025-12-04T11:24:35.2496514Z =========================== short test summary info ============================ 2025-12-04T11:24:35.2497811Z FAILED [0.3445s] inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False - AssertionError: Scalars are not equal! 2025-12-04T11:24:35.2498866Z 2025-12-04T11:24:35.2498974Z Expected 1 but got 0. 2025-12-04T11:24:35.2499269Z Absolute difference: 1 2025-12-04T11:24:35.2499573Z Relative difference: 1.0 2025-12-04T11:24:35.2499765Z 2025-12-04T11:24:35.2499981Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:35.2501209Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:35.2502305Z 2025-12-04T11:24:35.2502579Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:35.2503175Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:24:35.2503691Z ================== 1 failed, 36 deselected, 2 rerun in 6.74s =================== 2025-12-04T11:24:35.2504143Z --- Logging error --- 2025-12-04T11:24:35.2504449Z Traceback (most recent call last): 2025-12-04T11:24:35.2504992Z File "/opt/conda/envs/py_3.10/lib/python3.10/logging/__init__.py", line 1103, in emit 2025-12-04T11:24:35.2505577Z stream.write(msg + self.terminator) 2025-12-04T11:24:35.2505970Z ValueError: I/O operation on closed file. 2025-12-04T11:24:35.2506330Z Call stack: 2025-12-04T11:24:35.2506983Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/remote_cache.py", line 432, in dump_cache_stats 2025-12-04T11:24:35.2507794Z log.info("Cache Metrics:%s", out.getvalue()) 2025-12-04T11:24:35.2508191Z Message: 'Cache Metrics:%s' 2025-12-04T11:24:35.2508497Z Arguments: (' None\n',) 2025-12-04T11:24:35.2508781Z Got exit code 1 2025-12-04T11:24:35.2509788Z FAILED CONSISTENTLY: test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:35.2511129Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:24:35.2512341Z Test results will be stored in test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-fd020a6d441a5738.xml 2025-12-04T11:24:35.2513325Z ============================= test session starts ============================== 2025-12-04T11:24:35.2513992Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:24:35.2514612Z cachedir: .pytest_cache 2025-12-04T11:24:35.2515374Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:24:35.2516171Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:24:35.2516527Z configfile: pytest.ini 2025-12-04T11:24:35.2517295Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:24:35.2519224Z collecting ... /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:52: PytestCollectionWarning: cannot collect test class 'TestDecomposeAddMM' because it has a __init__ constructor (from: test/inductor/test_decompose_mem_bound_mm.py) 2025-12-04T11:24:35.2520698Z class TestDecomposeAddMM(torch.nn.Module): 2025-12-04T11:24:35.2521135Z collected 37 items / 15 deselected / 22 selected 2025-12-04T11:24:35.2521603Z stepcurrent: skipping 15 already run items. 2025-12-04T11:24:35.2521996Z Running 22 items in this shard 2025-12-04T11:24:35.2522207Z 2025-12-04T11:24:35.2523051Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True ('RERUN', {'yellow': True}) [5.9579s] [ 4%] 2025-12-04T11:24:35.2524840Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True ('RERUN', {'yellow': True}) [0.3702s] [ 4%] 2025-12-04T11:24:35.2526525Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True FAILED [0.3597s] [ 4%] 2025-12-04T11:24:35.2527406Z 2025-12-04T11:24:35.2527549Z ==================================== RERUNS ==================================== 2025-12-04T11:24:35.2528300Z _ TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True _ 2025-12-04T11:24:35.2529069Z Traceback (most recent call last): 2025-12-04T11:24:35.2529881Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 275, in test_decompose_linear_mixed_precision 2025-12-04T11:24:35.2530701Z self.assertEqual( 2025-12-04T11:24:35.2531353Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:35.2532123Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:35.2532950Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:35.2533837Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:35.2534312Z AssertionError: Scalars are not equal! 2025-12-04T11:24:35.2534561Z 2025-12-04T11:24:35.2534667Z Expected 1 but got 0. 2025-12-04T11:24:35.2534955Z Absolute difference: 1 2025-12-04T11:24:35.2535260Z Relative difference: 1.0 2025-12-04T11:24:35.2535452Z 2025-12-04T11:24:35.2535683Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:35.2536965Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:35.2537989Z 2025-12-04T11:24:35.2538300Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:35.2538942Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:35.2539421Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:35.2539745Z stats [('calls_captured', 1)] 2025-12-04T11:24:35.2540744Z inductor [('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_count', 4), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_addmm', 1)] 2025-12-04T11:24:35.2541886Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:35.2542337Z graph_break [] 2025-12-04T11:24:35.2542714Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:35.2543608Z I1204 11:21:18.753000 100010 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:35.2544851Z I1204 11:21:19.000000 100010 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:35.2546021Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.2547104Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:35.2548693Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.2550906Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_self_modules_linear_parameters_bias_: "f32[2][1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.2552870Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:35.2554478Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_bias_ = L_self_modules_linear_parameters_bias_ 2025-12-04T11:24:35.2555810Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:35.2556874Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.2558425Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.2560925Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, l_self_modules_linear_parameters_bias_); l_x_ = l_self_modules_linear_parameters_weight_ = l_self_modules_linear_parameters_bias_ = None 2025-12-04T11:24:35.2562894Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:35.2563939Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.2564899Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.2565995Z I1204 11:21:21.335000 100010 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:35.2567427Z V1204 11:21:21.714000 100010 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:35.2568999Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw: 2025-12-04T11:24:35.2570508Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4ral453goioq2ytnrbxj5hhqjyfnxjkats7ca6dgxms6usqpldw] gm: GraphModule() 2025-12-04T11:24:35.2571592Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.2572308Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.2573026Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.2573965Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2, primals_3): 2025-12-04T11:24:35.2575446Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.2577302Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.2579061Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_2 = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.2580823Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.2582567Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] addmm = torch.ops.aten.addmm.default(convert_element_type, convert_element_type_2, permute); convert_element_type = None 2025-12-04T11:24:35.2584183Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.2585456Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] return (addmm, convert_element_type_2, permute_1) 2025-12-04T11:24:35.2586390Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.2587442Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:35.2590451Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.2595151Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [di4vpbifhxsheleamv47gds75iwa6xnkz5malaf2qsqkrhojljw] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2]), stride=(1,), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=8, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.2600105Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[2]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.2603204Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:35.2604922Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:35.2606759Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:35.2608436Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:35.2610174Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:35.2611811Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:35.2613437Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:35.2615022Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:35.2616589Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:35.2618252Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [nnqi2dye4qdwjdgeujlxh7d67y2n67kat5ehf3472pn5avlyqbk] fx_kwargs[static_input_idxs]: [0, 1] 2025-12-04T11:24:35.2619817Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inputs_to_check[0]: 2 2025-12-04T11:24:35.2621507Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:35.2623354Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:35.2625467Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:35.2627209Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:35.2657121Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:35.2687362Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:35.2689330Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:35.2691000Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:35.2692755Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:35.2694590Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:35.2696435Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:35.2698196Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:35.2699887Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:35.2701595Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:35.2703381Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:35.2705227Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:35.2707082Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:35.2708997Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:35.2710843Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:35.2712732Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:35.2714678Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:35.2716491Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:35.2718451Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:35.2720351Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:35.2722094Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:35.2723869Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:35.2725600Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:35.2727326Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:35.2728945Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:35.2730584Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:35.2732318Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:35.2733966Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:35.2735689Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:35.2737486Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:35.2739166Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:35.2740871Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:35.2742589Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:35.2744269Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:35.2745969Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:35.2747649Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:35.2749385Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:35.2751107Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:35.2752787Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:35.2754460Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:35.2756187Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:35.2757946Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:35.2759618Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:35.2761332Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:35.2763102Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:35.2764875Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:35.2766673Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:35.2768464Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:35.2770253Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:35.2771936Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:35.2773587Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:35.2775299Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:35.2777320Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:35.2779189Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:35.2780926Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:35.2782700Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:35.2784469Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:35.2786545Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:35.2788688Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:35.2790494Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:35.2792435Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:35.2794295Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:35.2796215Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:35.2798017Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:35.2799950Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:35.2801784Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:35.2803680Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.2805574Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:35.2807500Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.2809409Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:35.2811331Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.2813223Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:35.2815052Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:35.2816819Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:35.2818578Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:35.2820315Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:35.2822149Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:35.2823899Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:35.2825646Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:35.2827394Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:35.2829162Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:35.2831001Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:35.2832914Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:35.2834878Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:35.2836650Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:35.2838402Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:35.2840255Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:35.2842076Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:35.2843795Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:35.2845651Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:35.2847543Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:35.2849414Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:35.2851289Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:35.2853116Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:35.2854865Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:35.2856530Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:35.2858241Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:35.2859994Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:35.2861738Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:35.2863506Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:35.2865316Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:35.2867106Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:35.2868896Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:35.2870587Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:35.2872262Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:35.2874111Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:35.2876096Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:35.2878086Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:35.2879985Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:35.2881762Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:35.2883569Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:35.2885377Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:35.2887270Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:35.2889175Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:35.2890958Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:35.2892689Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:35.2894512Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:35.2896454Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:35.2898256Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:35.2899962Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:35.2901730Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:35.2903584Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:35.2905281Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:35.2907015Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:35.2908782Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:35.2910587Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:35.2912439Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:35.2914214Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:35.2916052Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:35.2925206Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:35.2927015Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:35.2928985Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:35.2930744Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:35.2932438Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:35.2934126Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:35.2935832Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:35.2937730Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:35.2939546Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:35.2941372Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:35.2943183Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:35.2945065Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:35.2946810Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:35.2948610Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:35.2950495Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:35.2952303Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:35.2954080Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:35.2955848Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:35.2957704Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:35.2959391Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:35.2961082Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:35.2962799Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:35.2964455Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:35.2966127Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:35.2967899Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:35.2969734Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:35.2971564Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:35.2973301Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:35.2975004Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:35.2976940Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:35.2978706Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:35.2980558Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:35.2982417Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:35.2984258Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:35.2986049Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:35.2987828Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:35.2989588Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:35.2991399Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:35.2993162Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:35.2994884Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:35.2996860Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:35.2998735Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:35.3000483Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:35.3002296Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:35.3004119Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:35.3005872Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:35.3007588Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:35.3009332Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:35.3011056Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:35.3012813Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:35.3014793Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:35.3016951Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:35.3019050Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:35.3021133Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:35.3023232Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:35.3025390Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:35.3027521Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:35.3029621Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:35.3031740Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:35.3033749Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:35.3035559Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:35.3037343Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:35.3039214Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:35.3041153Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:35.3042917Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:35.3044621Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:35.3046301Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:35.3047990Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:35.3049770Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:35.3051495Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:35.3053219Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:35.3055034Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:35.3056932Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:35.3058719Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:35.3060465Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:35.3062173Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:35.3063904Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:35.3065587Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:35.3067385Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:35.3069218Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:35.3070980Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:35.3072737Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:35.3074473Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:35.3076134Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:35.3077855Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:35.3079622Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:35.3081531Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:35.3083430Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:35.3085112Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:35.3086832Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:35.3088627Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:35.3090439Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:35.3092240Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:35.3093990Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:35.3095924Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:35.3098095Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:35.3099968Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:35.3101803Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:35.3103654Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:35.3105511Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:35.3107232Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:35.3109029Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:35.3110822Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:35.3112660Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:35.3114491Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:35.3116194Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:35.3117936Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:35.3119680Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:35.3121380Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:35.3123063Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:35.3124772Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:35.3126527Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:35.3128281Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:35.3130111Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.3131990Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.3133778Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:35.3135577Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:35.3137460Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:35.3139296Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:35.3141145Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:35.3143035Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:35.3144950Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:35.3146788Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:35.3148591Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:35.3150356Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:35.3152176Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:35.3153957Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:35.3155737Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:35.3157521Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:35.3159320Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:35.3161131Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:35.3162935Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:35.3164736Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:35.3166473Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:35.3168311Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:35.3170260Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:35.3172131Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:35.3174053Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:35.3176010Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:35.3178067Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:35.3180043Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:35.3181933Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:35.3183761Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:35.3185596Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:35.3187538Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:35.3189486Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:35.3191367Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:35.3193165Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:35.3194957Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:35.3196904Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:35.3198772Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:35.3200555Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:35.3202281Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:35.3204080Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:35.3205907Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:35.3207737Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:35.3209615Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:35.3211436Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:35.3213260Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:35.3215104Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:35.3216005Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:35.3216842Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:35.3217751Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:35.3218687Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:35.3219533Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:35.3220395Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:35.3221278Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:35.3222060Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:35.3222872Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:35.3223741Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:35.3224544Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:35.3225340Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:35.3226158Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:35.3227032Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:35.3228000Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:35.3228878Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.3229717Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:35.3230654Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:35.3231523Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:35.3232389Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:35.3233191Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:35.3234185Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:35.3235052Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:35.3235880Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:35.3236675Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:35.3237504Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:35.3238428Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:35.3239346Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:35.3240296Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:35.3241100Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:35.3241933Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:35.3242812Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:35.3243695Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:35.3244622Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:35.3245498Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:35.3246377Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:35.3247220Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:35.3248142Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:35.3248996Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:35.3249859Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:35.3250717Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:35.3251566Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:35.3252405Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:35.3253199Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:35.3254165Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:35.3255030Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:35.3255874Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:35.3256678Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:35.3257609Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:35.3258589Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:35.3259507Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:35.3260406Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:35.3261339Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:35.3262212Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:35.3263112Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:35.3263988Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:35.3264922Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:35.3265787Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:35.3266680Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:35.3267499Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:35.3268346Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:35.3269224Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:35.3270081Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:35.3270986Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:35.3271904Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:35.3272661Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:35.3273425Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:35.3274253Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:35.3275093Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:35.3275914Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:35.3276738Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:35.3277532Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:35.3278454Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:35.3279410Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:35.3280312Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:35.3281120Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:35.3281903Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:35.3282804Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:35.3283635Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:35.3284510Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:35.3285355Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:35.3286211Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:35.3287094Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:35.3287957Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:35.3288786Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:35.3289618Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:35.3290507Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:35.3291419Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:35.3292268Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:35.3292992Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:35.3293930Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:35.3294776Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:35.3295539Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:35.3296480Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:35.3297314Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:35.3298125Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:35.3299054Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:35.3299834Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:35.3300586Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:35.3301413Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:35.3302273Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:35.3303127Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:35.3304060Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:35.3304909Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:35.3305708Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:35.3306501Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:35.3307359Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:35.3308117Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:35.3308879Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:35.3309649Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:35.3310433Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:35.3311310Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:35.3312168Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:35.3312993Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:35.3313779Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:35.3314543Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:35.3315400Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:35.3317355Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:35.3319792Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:35.3320589Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:35.3321513Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:35.3322307Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:35.3323205Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:35.3324169Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:35.3325020Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:35.3325929Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:35.3326828Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:35.3327765Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:35.3328671Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:35.3329594Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:35.3330582Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:35.3331515Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:35.3332348Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:35.3333280Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:35.3334197Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:35.3335083Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:35.3335974Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:35.3337028Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:35.3337788Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:35.3338493Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:35.3339248Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:35.3339990Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:35.3340711Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:35.3341446Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:35.3342317Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:35.3343201Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:35.3344053Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:35.3344784Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:35.3345435Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.3346101Z I1204 11:21:21.725000 100010 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.3346695Z V1204 11:21:21.725000 100010 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:35.3347183Z V1204 11:21:21.725000 100010 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:35.3347709Z I1204 11:21:21.726000 100010 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:35.3348574Z V1204 11:21:21.768000 100010 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.addmm.default with input shape: torch.Size([2]), torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:35.3349018Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.3349529Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:35.3350483Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.3351351Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[2][1]cuda:0", primals_3: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.3352332Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.3353302Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2][1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.3354300Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.3355294Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_2: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.3356197Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.3357049Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_2, 2) 2025-12-04T11:24:35.3357843Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:35.3358887Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:35.3359724Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:35.3360728Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] add_tensor: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.add.Tensor(sum_dim_int_list, convert_element_type); sum_dim_int_list = convert_element_type = None 2025-12-04T11:24:35.3361555Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.3362169Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.3362605Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.3362994Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.3363564Z V1204 11:21:21.778000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:35.3364147Z V1204 11:21:21.778000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:35.3364706Z V1204 11:21:21.779000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_3 : [num_users=1] = placeholder[target=primals_3] 2025-12-04T11:24:35.3365756Z V1204 11:21:21.779000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.3366263Z V1204 11:21:21.780000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.3367283Z V1204 11:21:21.781000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.3367816Z V1204 11:21:21.782000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.3368829Z V1204 11:21:21.813000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_2 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_3, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.3369339Z V1204 11:21:21.814000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.3370226Z V1204 11:21:21.815000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.3370737Z V1204 11:21:21.816000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.3371675Z V1204 11:21:21.821000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_2, 2), kwargs = {}) 2025-12-04T11:24:35.3372141Z V1204 11:21:21.822000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.3373019Z V1204 11:21:21.823000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:35.3373472Z V1204 11:21:21.823000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.3374396Z V1204 11:21:21.824000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:35.3374861Z V1204 11:21:21.825000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.3375723Z V1204 11:21:21.827000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:35.3376156Z V1204 11:21:21.827000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.3377152Z V1204 11:21:21.830000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %add_tensor : [num_users=1] = call_function[target=torch.ops.aten.add.Tensor](args = (%sum_dim_int_list, %convert_element_type), kwargs = {}) 2025-12-04T11:24:35.3377694Z V1204 11:21:21.830000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via .inner at 0x7f11500feb00> 2025-12-04T11:24:35.3378533Z V1204 11:21:21.832000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.3379040Z V1204 11:21:21.833000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.3379592Z V1204 11:21:21.833000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.3380182Z V1204 11:21:21.838000 100010 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:35.3380918Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:35.3381043Z warnings.warn( 2025-12-04T11:24:35.3381563Z V1204 11:21:21.839000 100010 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:35.3382031Z _ TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True _ 2025-12-04T11:24:35.3382167Z Traceback (most recent call last): 2025-12-04T11:24:35.3382752Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 275, in test_decompose_linear_mixed_precision 2025-12-04T11:24:35.3382858Z self.assertEqual( 2025-12-04T11:24:35.3383329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:35.3383493Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:35.3384038Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:35.3384282Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:35.3384415Z AssertionError: Scalars are not equal! 2025-12-04T11:24:35.3384426Z 2025-12-04T11:24:35.3384547Z Expected 1 but got 0. 2025-12-04T11:24:35.3384654Z Absolute difference: 1 2025-12-04T11:24:35.3384764Z Relative difference: 1.0 2025-12-04T11:24:35.3384770Z 2025-12-04T11:24:35.3385000Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:35.3385868Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:35.3385874Z 2025-12-04T11:24:35.3386159Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:35.3386383Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:35.3386496Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:35.3386659Z stats [('calls_captured', 1)] 2025-12-04T11:24:35.3387445Z inductor [('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_count', 4), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_addmm', 1)] 2025-12-04T11:24:35.3387679Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:35.3387777Z graph_break [] 2025-12-04T11:24:35.3387999Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:35.3388513Z I1204 11:21:18.753000 100010 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:35.3389109Z I1204 11:21:19.000000 100010 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:35.3389561Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.3390056Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:35.3390999Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.3392133Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_self_modules_linear_parameters_bias_: "f32[2][1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.3392883Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:35.3393630Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_bias_ = L_self_modules_linear_parameters_bias_ 2025-12-04T11:24:35.3394122Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:35.3394557Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.3398449Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.3399854Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, l_self_modules_linear_parameters_bias_); l_x_ = l_self_modules_linear_parameters_weight_ = l_self_modules_linear_parameters_bias_ = None 2025-12-04T11:24:35.3400432Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:35.3400854Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.3401254Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.3401811Z I1204 11:21:21.335000 100010 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:35.3402567Z V1204 11:21:21.714000 100010 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:35.3403275Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw: 2025-12-04T11:24:35.3404001Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4ral453goioq2ytnrbxj5hhqjyfnxjkats7ca6dgxms6usqpldw] gm: GraphModule() 2025-12-04T11:24:35.3404283Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.3404574Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.3404852Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.3405370Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2, primals_3): 2025-12-04T11:24:35.3406214Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.3407032Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.3407904Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_2 = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.3408672Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.3409516Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] addmm = torch.ops.aten.addmm.default(convert_element_type, convert_element_type_2, permute); convert_element_type = None 2025-12-04T11:24:35.3410159Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.3410714Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] return (addmm, convert_element_type_2, permute_1) 2025-12-04T11:24:35.3411023Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.3411602Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:35.3413985Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.3416317Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [di4vpbifhxsheleamv47gds75iwa6xnkz5malaf2qsqkrhojljw] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2]), stride=(1,), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=8, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.3418734Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[2]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.3419430Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:35.3420359Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:35.3421083Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:35.3421912Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:35.3422697Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:35.3423411Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:35.3424136Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:35.3424857Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:35.3425581Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:35.3426337Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [nnqi2dye4qdwjdgeujlxh7d67y2n67kat5ehf3472pn5avlyqbk] fx_kwargs[static_input_idxs]: [0, 1] 2025-12-04T11:24:35.3427080Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inputs_to_check[0]: 2 2025-12-04T11:24:35.3427955Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:35.3428786Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:35.3429997Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:35.3430822Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:35.3459799Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:35.3460867Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:35.3461650Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:35.3462419Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:35.3463264Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:35.3464153Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:35.3464874Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:35.3465659Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:35.3466475Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:35.3467270Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:35.3468159Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:35.3468989Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:35.3469869Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:35.3470721Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:35.3471579Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:35.3472511Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:35.3473379Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:35.3474196Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:35.3475205Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:35.3475977Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:35.3476810Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:35.3477572Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:35.3478411Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:35.3479185Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:35.3479927Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:35.3480719Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:35.3481482Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:35.3482274Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:35.3483128Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:35.3483930Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:35.3484717Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:35.3485488Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:35.3486259Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:35.3487058Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:35.3487871Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:35.3488653Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:35.3489453Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:35.3490250Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:35.3491028Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:35.3491804Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:35.3492622Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:35.3493387Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:35.3494165Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:35.3494973Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:35.3495826Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:35.3496821Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:35.3497659Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:35.3498605Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:35.3499384Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:35.3500203Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:35.3500947Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:35.3501765Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:35.3502685Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:35.3503479Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:35.3504349Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:35.3505178Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:35.3505936Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:35.3507127Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:35.3507941Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:35.3508804Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:35.3509699Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:35.3510513Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:35.3511329Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:35.3512235Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:35.3513116Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:35.3513964Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:35.3514904Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.3515780Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:35.3516738Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.3517537Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:35.3518492Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.3519305Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:35.3520221Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:35.3520982Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:35.3521742Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:35.3522579Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:35.3523424Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:35.3524204Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:35.3525020Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:35.3525812Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:35.3526646Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:35.3527516Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:35.3528472Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:35.3529339Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:35.3530141Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:35.3530965Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:35.3531888Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:35.3532685Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:35.3533470Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:35.3534357Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:35.3535233Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:35.3536131Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:35.3537050Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:35.3537861Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:35.3538659Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:35.3539388Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:35.3540180Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:35.3540996Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:35.3541791Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:35.3542624Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:35.3543469Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:35.3544330Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:35.3545151Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:35.3545920Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:35.3546733Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:35.3547673Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:35.3548599Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:35.3549540Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:35.3550346Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:35.3551180Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:35.3552053Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:35.3552892Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:35.3553798Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:35.3554658Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:35.3555444Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:35.3556255Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:35.3557093Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:35.3557927Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:35.3558711Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:35.3559502Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:35.3560372Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:35.3561156Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:35.3561960Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:35.3562782Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:35.3563659Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:35.3564498Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:35.3565378Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:35.3566172Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:35.3567046Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:35.3567878Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:35.3568747Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:35.3569576Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:35.3570368Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:35.3571137Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:35.3571922Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:35.3572702Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:35.3573544Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:35.3574389Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:35.3575238Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:35.3576099Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:35.3577024Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:35.3577826Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:35.3578727Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:35.3579621Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:35.3580435Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:35.3581243Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:35.3582073Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:35.3582856Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:35.3583663Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:35.3584434Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:35.3585217Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:35.3585965Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:35.3586760Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:35.3587612Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:35.3588472Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:35.3589292Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:35.3590079Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:35.3590855Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:35.3591695Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:35.3592516Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:35.3593359Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:35.3594280Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:35.3595096Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:35.3595966Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:35.3596943Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:35.3597761Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:35.3598602Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:35.3599392Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:35.3600254Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:35.3601123Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:35.3601922Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:35.3602749Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:35.3603609Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:35.3604434Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:35.3605182Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:35.3606014Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:35.3606806Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:35.3607597Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:35.3608468Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:35.3609424Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:35.3610557Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:35.3611533Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:35.3612549Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:35.3613513Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:35.3614518Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:35.3615514Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:35.3616518Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:35.3617584Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:35.3618421Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:35.3619257Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:35.3620077Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:35.3621004Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:35.3621834Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:35.3622632Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:35.3623407Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:35.3624205Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:35.3625020Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:35.3625801Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:35.3626653Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:35.3627441Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:35.3628369Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:35.3629188Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:35.3630003Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:35.3630777Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:35.3631581Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:35.3632402Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:35.3633173Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:35.3634017Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:35.3634856Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:35.3635644Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:35.3636485Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:35.3637215Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:35.3638013Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:35.3638804Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:35.3639644Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:35.3640618Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:35.3641414Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:35.3642192Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:35.3643035Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:35.3643893Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:35.3644752Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:35.3645576Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:35.3646376Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:35.3647314Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:35.3648236Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:35.3649077Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:35.3649874Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:35.3650786Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:35.3651588Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:35.3652369Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:35.3653184Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:35.3654023Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:35.3654880Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:35.3655719Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:35.3656491Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:35.3657377Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:35.3658210Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:35.3658973Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:35.3659793Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:35.3660567Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:35.3661410Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:35.3662191Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:35.3663069Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.3663967Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.3664755Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:35.3665604Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:35.3666442Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:35.3667313Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:35.3668172Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:35.3669078Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:35.3669913Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:35.3670797Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:35.3671588Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:35.3672438Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:35.3673274Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:35.3674120Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:35.3674959Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:35.3675803Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:35.3676657Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:35.3677477Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:35.3678327Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:35.3679126Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:35.3679963Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:35.3680866Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:35.3681720Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:35.3682600Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:35.3683505Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:35.3684421Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:35.3685332Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:35.3686231Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:35.3687079Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:35.3687933Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:35.3688815Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:35.3689733Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:35.3690670Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:35.3691518Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:35.3692370Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:35.3693209Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:35.3694019Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:35.3694884Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:35.3695664Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:35.3696710Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:35.3697601Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:35.3698420Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:35.3699297Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:35.3700177Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:35.3701005Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:35.3701806Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:35.3702720Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:35.3703621Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:35.3704518Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:35.3705390Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:35.3706259Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:35.3707151Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:35.3708006Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:35.3708948Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:35.3709732Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:35.3710550Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:35.3711386Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:35.3712220Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:35.3713037Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:35.3713837Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:35.3714697Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:35.3715638Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:35.3716539Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.3717378Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:35.3718259Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:35.3719135Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:35.3719992Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:35.3720852Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:35.3721831Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:35.3722707Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:35.3723535Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:35.3724353Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:35.3725197Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:35.3726101Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:35.3727008Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:35.3727944Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:35.3728797Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:35.3729632Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:35.3730478Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:35.3731361Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:35.3732275Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:35.3733178Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:35.3734020Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:35.3734875Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:35.3735785Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:35.3736694Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:35.3737624Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:35.3738431Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:35.3739330Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:35.3740164Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:35.3741018Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:35.3741956Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:35.3742827Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:35.3743633Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:35.3744439Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:35.3745371Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:35.3746290Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:35.3747214Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:35.3748093Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:35.3749046Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:35.3749906Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:35.3750778Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:35.3751665Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:35.3752586Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:35.3753489Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:35.3754332Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:35.3755194Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:35.3756021Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:35.3756939Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:35.3757799Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:35.3758677Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:35.3759574Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:35.3760315Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:35.3761124Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:35.3761930Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:35.3762740Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:35.3763554Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:35.3764375Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:35.3765164Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:35.3766044Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:35.3767005Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:35.3767888Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:35.3768713Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:35.3769508Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:35.3770388Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:35.3771249Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:35.3772112Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:35.3773011Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:35.3773856Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:35.3774743Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:35.3775570Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:35.3776400Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:35.3777321Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:35.3778196Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:35.3779088Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:35.3779923Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:35.3780659Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:35.3781591Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:35.3782419Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:35.3783177Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:35.3783955Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:35.3784751Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:35.3785584Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:35.3786463Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:35.3787260Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:35.3788021Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:35.3788882Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:35.3789728Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:35.3790590Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:35.3791468Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:35.3792334Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:35.3793161Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:35.3793965Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:35.3794777Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:35.3795535Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:35.3796442Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:35.3797207Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:35.3798017Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:35.3798832Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:35.3799713Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:35.3800539Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:35.3801413Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:35.3802174Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:35.3803008Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:35.3804987Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:35.3807470Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:35.3808253Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:35.3809209Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:35.3810029Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:35.3810862Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:35.3811844Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:35.3812693Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:35.3813613Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:35.3814515Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:35.3815404Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:35.3816314Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:35.3817277Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:35.3818328Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:35.3819215Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:35.3820090Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:35.3821015Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:35.3821987Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:35.3822872Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:35.3823750Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:35.3824729Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:35.3825501Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:35.3826215Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:35.3826956Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:35.3827680Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:35.3828395Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:35.3829119Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:35.3830003Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:35.3830843Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:35.3831699Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:35.3832417Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:35.3833082Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.3833777Z I1204 11:21:21.725000 100010 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.3834258Z V1204 11:21:21.725000 100010 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:35.3834753Z V1204 11:21:21.725000 100010 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:35.3835288Z I1204 11:21:21.726000 100010 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:35.3836158Z V1204 11:21:21.768000 100010 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.addmm.default with input shape: torch.Size([2]), torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:35.3836635Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.3837127Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:35.3838089Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.3838914Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[2][1]cuda:0", primals_3: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.3839908Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.3840903Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2][1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.3841875Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.3842857Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_2: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.3843771Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.3844611Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_2, 2) 2025-12-04T11:24:35.3845384Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:35.3846434Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:35.3847306Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:35.3848313Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] add_tensor: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.add.Tensor(sum_dim_int_list, convert_element_type); sum_dim_int_list = convert_element_type = None 2025-12-04T11:24:35.3849124Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.3849760Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.3850218Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.3850620Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.3851194Z V1204 11:21:21.778000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:35.3851772Z V1204 11:21:21.778000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:35.3852341Z V1204 11:21:21.779000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_3 : [num_users=1] = placeholder[target=primals_3] 2025-12-04T11:24:35.3853353Z V1204 11:21:21.779000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.3853904Z V1204 11:21:21.780000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.3854907Z V1204 11:21:21.781000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.3855419Z V1204 11:21:21.782000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.3856427Z V1204 11:21:21.813000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_2 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_3, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.3856992Z V1204 11:21:21.814000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.3857894Z V1204 11:21:21.815000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.3858343Z V1204 11:21:21.816000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.3859299Z V1204 11:21:21.821000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_2, 2), kwargs = {}) 2025-12-04T11:24:35.3859753Z V1204 11:21:21.822000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.3860640Z V1204 11:21:21.823000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:35.3861209Z V1204 11:21:21.823000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.3862133Z V1204 11:21:21.824000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:35.3862564Z V1204 11:21:21.825000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.3863453Z V1204 11:21:21.827000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:35.3863903Z V1204 11:21:21.827000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.3864832Z V1204 11:21:21.830000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %add_tensor : [num_users=1] = call_function[target=torch.ops.aten.add.Tensor](args = (%sum_dim_int_list, %convert_element_type), kwargs = {}) 2025-12-04T11:24:35.3865399Z V1204 11:21:21.830000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via .inner at 0x7f11500feb00> 2025-12-04T11:24:35.3866237Z V1204 11:21:21.832000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.3866696Z V1204 11:21:21.833000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.3867249Z V1204 11:21:21.833000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.3867856Z V1204 11:21:21.838000 100010 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:35.3868608Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:35.3868713Z warnings.warn( 2025-12-04T11:24:35.3869211Z V1204 11:21:21.839000 100010 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:35.3869439Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:35.3869557Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:35.3869688Z stats [('calls_captured', 1)] 2025-12-04T11:24:35.3869913Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:35.3870702Z inductor [('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_count', 4), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_addmm', 1)] 2025-12-04T11:24:35.3870816Z graph_break [] 2025-12-04T11:24:35.3871042Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:35.3871500Z V1204 11:21:22.073000 100010 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpq92dyh2h 2025-12-04T11:24:35.3871941Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.3872436Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:35.3873394Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.3874512Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_self_modules_linear_parameters_bias_: "f32[2][1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.3875275Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:35.3876594Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_bias_ = L_self_modules_linear_parameters_bias_ 2025-12-04T11:24:35.3877069Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:35.3877491Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.3878508Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.3879898Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, l_self_modules_linear_parameters_bias_); l_x_ = l_self_modules_linear_parameters_weight_ = l_self_modules_linear_parameters_bias_ = None 2025-12-04T11:24:35.3880380Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:35.3880817Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.3881235Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.3881966Z V1204 11:21:22.368000 100010 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:35.3882680Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw: 2025-12-04T11:24:35.3883339Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4ral453goioq2ytnrbxj5hhqjyfnxjkats7ca6dgxms6usqpldw] gm: GraphModule() 2025-12-04T11:24:35.3883639Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.3883917Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.3884208Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.3884722Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2, primals_3): 2025-12-04T11:24:35.3885536Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.3886368Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.3887181Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_2 = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.3887958Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.3888831Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] addmm = torch.ops.aten.addmm.default(convert_element_type, convert_element_type_2, permute); convert_element_type = None 2025-12-04T11:24:35.3889479Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.3889986Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] return (addmm, convert_element_type_2, permute_1) 2025-12-04T11:24:35.3890313Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.3890909Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:35.3893239Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.3895530Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [di4vpbifhxsheleamv47gds75iwa6xnkz5malaf2qsqkrhojljw] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2]), stride=(1,), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=8, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.3898126Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[2]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.3898778Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:35.3899687Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:35.3900428Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:35.3901236Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:35.3902031Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:35.3902742Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:35.3903468Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:35.3904193Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:35.3904982Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:35.3905750Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [nnqi2dye4qdwjdgeujlxh7d67y2n67kat5ehf3472pn5avlyqbk] fx_kwargs[static_input_idxs]: [0, 1] 2025-12-04T11:24:35.3906467Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inputs_to_check[0]: 2 2025-12-04T11:24:35.3907354Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:35.3908184Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:35.3909385Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:35.3910164Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:35.3939113Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:35.3940216Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:35.3940985Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:35.3941819Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:35.3942657Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:35.3943512Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:35.3944258Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:35.3945034Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:35.3945862Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:35.3946644Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:35.3947509Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:35.3948343Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:35.3949214Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:35.3950098Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:35.3950945Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:35.3951855Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:35.3952710Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:35.3953534Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:35.3954542Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:35.3955313Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:35.3956154Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:35.3956903Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:35.3957785Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:35.3958547Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:35.3959269Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:35.3960116Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:35.3960884Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:35.3961678Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:35.3962514Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:35.3963287Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:35.3964067Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:35.3964853Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:35.3965656Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:35.3966442Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:35.3967217Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:35.3967984Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:35.3968794Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:35.3969582Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:35.3970362Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:35.3971128Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:35.3971951Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:35.3972720Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:35.3973512Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:35.3974333Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:35.3975138Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:35.3976006Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:35.3976794Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:35.3977802Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:35.3978572Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:35.3979330Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:35.3980095Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:35.3980905Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:35.3981879Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:35.3982673Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:35.3983485Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:35.3984315Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:35.3985070Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:35.3986266Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:35.3987064Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:35.3987938Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:35.3988822Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:35.3989686Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:35.3990498Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:35.3991391Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:35.3992265Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:35.3993102Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:35.3994052Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.3994872Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:35.3995841Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.3996799Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:35.3997826Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.3998646Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:35.3999519Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:35.4000299Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:35.4001044Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:35.4001895Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:35.4002744Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:35.4003513Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:35.4004333Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:35.4005119Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:35.4006016Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:35.4006881Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:35.4007861Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:35.4008723Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:35.4009549Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:35.4010373Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:35.4011264Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:35.4012066Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:35.4012836Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:35.4013773Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:35.4014632Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:35.4015508Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:35.4016375Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:35.4017243Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:35.4018048Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:35.4018764Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:35.4019570Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:35.4020374Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:35.4021183Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:35.4022045Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:35.4022889Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:35.4023744Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:35.4024544Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:35.4025324Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:35.4026121Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:35.4027048Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:35.4027974Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:35.4028915Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:35.4029762Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:35.4030577Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:35.4031419Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:35.4032248Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:35.4033164Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:35.4034022Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:35.4034817Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:35.4035633Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:35.4036470Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:35.4037302Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:35.4038114Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:35.4038924Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:35.4039785Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:35.4040591Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:35.4041398Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:35.4042219Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:35.4043033Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:35.4043858Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:35.4044736Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:35.4045548Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:35.4046432Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:35.4047213Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:35.4048094Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:35.4048911Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:35.4049698Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:35.4050469Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:35.4051250Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:35.4052051Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:35.4052879Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:35.4053774Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:35.4054610Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:35.4055461Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:35.4056335Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:35.4057199Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:35.4058120Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:35.4058960Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:35.4059788Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:35.4060589Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:35.4061464Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:35.4062255Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:35.4063021Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:35.4063810Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:35.4064576Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:35.4065337Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:35.4066122Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:35.4066975Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:35.4067834Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:35.4068653Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:35.4069499Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:35.4070266Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:35.4071096Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:35.4071930Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:35.4072774Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:35.4073690Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:35.4074501Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:35.4075345Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:35.4076138Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:35.4076974Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:35.4077830Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:35.4078616Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:35.4079411Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:35.4080280Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:35.4081089Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:35.4081902Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:35.4082773Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:35.4083586Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:35.4084347Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:35.4085166Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:35.4085989Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:35.4086790Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:35.4087640Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:35.4088614Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:35.4089648Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:35.4090628Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:35.4091597Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:35.4092570Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:35.4093621Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:35.4094623Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:35.4095614Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:35.4096749Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:35.4097653Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:35.4098484Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:35.4099310Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:35.4100223Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:35.4101059Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:35.4101928Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:35.4102708Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:35.4103476Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:35.4104321Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:35.4105120Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:35.4105962Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:35.4106752Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:35.4107652Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:35.4108471Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:35.4109301Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:35.4110098Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:35.4110922Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:35.4111700Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:35.4112468Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:35.4113314Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:35.4114157Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:35.4114961Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:35.4115783Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:35.4116529Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:35.4117313Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:35.4118137Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:35.4118981Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:35.4119969Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:35.4120776Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:35.4121544Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:35.4122400Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:35.4123213Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:35.4124064Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:35.4124880Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:35.4125783Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:35.4126742Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:35.4127615Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:35.4128478Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:35.4129266Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:35.4130191Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:35.4130984Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:35.4131761Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:35.4132589Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:35.4133407Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:35.4134312Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:35.4135135Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:35.4135916Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:35.4136722Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:35.4137602Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:35.4138374Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:35.4139162Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:35.4139950Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:35.4140781Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:35.4141613Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:35.4142476Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.4143350Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.4144133Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:35.4144980Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:35.4145831Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:35.4146677Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:35.4147539Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:35.4148436Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:35.4149267Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:35.4150180Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:35.4150976Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:35.4151835Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:35.4152656Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:35.4153511Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:35.4154339Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:35.4155152Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:35.4156000Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:35.4156818Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:35.4157714Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:35.4158503Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:35.4159319Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:35.4160207Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:35.4161075Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:35.4161947Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:35.4162855Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:35.4163772Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:35.4164687Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:35.4165609Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:35.4166472Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:35.4167346Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:35.4168204Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:35.4169123Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:35.4170041Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:35.4170883Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:35.4171713Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:35.4172536Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:35.4173359Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:35.4174247Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:35.4175030Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:35.4175860Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:35.4176681Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:35.4177573Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:35.4178438Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:35.4179326Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:35.4180142Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:35.4180957Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:35.4181910Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:35.4182811Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:35.4183655Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:35.4184540Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:35.4185421Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:35.4186308Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:35.4187174Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:35.4188054Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:35.4188837Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:35.4189654Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:35.4190520Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:35.4191326Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:35.4192125Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:35.4192946Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:35.4193792Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:35.4194741Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:35.4195636Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.4196637Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:35.4197539Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:35.4198475Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:35.4199340Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:35.4200184Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:35.4201176Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:35.4202077Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:35.4202907Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:35.4203709Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:35.4204543Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:35.4205465Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:35.4206405Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:35.4207359Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:35.4208161Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:35.4209002Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:35.4209855Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:35.4210741Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:35.4211663Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:35.4212542Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:35.4213391Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:35.4214233Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:35.4215171Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:35.4216037Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:35.4216995Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:35.4217822Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:35.4218709Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:35.4219562Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:35.4220360Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:35.4221318Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:35.4222180Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:35.4223017Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:35.4223819Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:35.4224702Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:35.4225630Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:35.4226543Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:35.4227434Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:35.4228367Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:35.4229221Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:35.4230101Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:35.4231015Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:35.4231949Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:35.4232833Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:35.4233692Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:35.4234544Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:35.4235375Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:35.4236253Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:35.4237112Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:35.4238003Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:35.4238932Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:35.4239683Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:35.4240445Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:35.4241271Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:35.4242071Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:35.4242887Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:35.4243715Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:35.4244500Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:35.4245391Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:35.4246338Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:35.4247270Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:35.4248076Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:35.4248874Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:35.4249756Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:35.4250620Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:35.4251496Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:35.4252348Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:35.4253205Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:35.4254082Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:35.4254945Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:35.4255773Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:35.4256609Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:35.4257573Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:35.4258453Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:35.4259302Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:35.4260026Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:35.4260970Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:35.4261785Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:35.4262557Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:35.4263386Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:35.4264166Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:35.4264993Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:35.4265859Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:35.4266675Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:35.4267430Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:35.4268264Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:35.4269115Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:35.4269962Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:35.4270892Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:35.4271745Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:35.4272548Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:35.4273341Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:35.4274156Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:35.4274915Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:35.4275676Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:35.4276445Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:35.4277236Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:35.4278066Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:35.4278955Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:35.4279780Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:35.4280593Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:35.4281354Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:35.4282163Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:35.4284150Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:35.4286582Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:35.4287372Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:35.4288297Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:35.4289095Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:35.4289942Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:35.4290905Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:35.4291760Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:35.4292658Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:35.4293567Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:35.4294463Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:35.4295392Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:35.4296426Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:35.4297538Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:35.4298441Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:35.4299335Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:35.4300256Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:35.4301188Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:35.4302076Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:35.4302963Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:35.4303973Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:35.4304723Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:35.4305420Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:35.4306161Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:35.4306884Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:35.4307599Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:35.4308334Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:35.4309209Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:35.4310061Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:35.4310908Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:35.4311661Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:35.4312324Z V1204 11:21:22.379000 100010 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.4313019Z I1204 11:21:22.379000 100010 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.4313516Z V1204 11:21:22.379000 100010 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:35.4314000Z V1204 11:21:22.379000 100010 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:35.4314552Z I1204 11:21:22.380000 100010 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:35.4315417Z V1204 11:21:22.394000 100010 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.addmm.default with input shape: torch.Size([2]), torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:35.4315868Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.4316385Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:35.4317333Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.4318198Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[2][1]cuda:0", primals_3: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.4319176Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.4320144Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2][1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.4321108Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.4322102Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_2: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.4323007Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.4323846Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_2, 2) 2025-12-04T11:24:35.4324614Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:35.4325691Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:35.4326530Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:35.4327567Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] add_tensor: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.add.Tensor(sum_dim_int_list, convert_element_type); sum_dim_int_list = convert_element_type = None 2025-12-04T11:24:35.4328390Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.4329005Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.4329431Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.4336391Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.4337113Z V1204 11:21:22.404000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:35.4337701Z V1204 11:21:22.405000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:35.4338375Z V1204 11:21:22.405000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_3 : [num_users=1] = placeholder[target=primals_3] 2025-12-04T11:24:35.4339404Z V1204 11:21:22.406000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.4339907Z V1204 11:21:22.406000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.4340918Z V1204 11:21:22.407000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.4341435Z V1204 11:21:22.407000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.4342453Z V1204 11:21:22.408000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_2 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_3, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.4342966Z V1204 11:21:22.408000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.4343867Z V1204 11:21:22.410000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.4344332Z V1204 11:21:22.410000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.4345270Z V1204 11:21:22.411000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_2, 2), kwargs = {}) 2025-12-04T11:24:35.4345769Z V1204 11:21:22.412000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.4346665Z V1204 11:21:22.413000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:35.4347115Z V1204 11:21:22.413000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.4348074Z V1204 11:21:22.414000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:35.4348549Z V1204 11:21:22.414000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.4349416Z V1204 11:21:22.416000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:35.4349854Z V1204 11:21:22.416000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.4350752Z V1204 11:21:22.418000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %add_tensor : [num_users=1] = call_function[target=torch.ops.aten.add.Tensor](args = (%sum_dim_int_list, %convert_element_type), kwargs = {}) 2025-12-04T11:24:35.4351302Z V1204 11:21:22.418000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via .inner at 0x7f11500feb00> 2025-12-04T11:24:35.4352140Z V1204 11:21:22.420000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.4352641Z V1204 11:21:22.421000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.4353195Z V1204 11:21:22.421000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.4353783Z V1204 11:21:22.424000 100010 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:35.4354525Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:35.4354628Z warnings.warn( 2025-12-04T11:24:35.4355119Z V1204 11:21:22.424000 100010 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:35.4355275Z =================================== FAILURES =================================== 2025-12-04T11:24:35.4355742Z _ TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True _ 2025-12-04T11:24:35.4355876Z Traceback (most recent call last): 2025-12-04T11:24:35.4356465Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 275, in test_decompose_linear_mixed_precision 2025-12-04T11:24:35.4356585Z self.assertEqual( 2025-12-04T11:24:35.4357046Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:35.4357215Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:35.4357766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:35.4357974Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:35.4358122Z AssertionError: Scalars are not equal! 2025-12-04T11:24:35.4358131Z 2025-12-04T11:24:35.4358237Z Expected 1 but got 0. 2025-12-04T11:24:35.4358346Z Absolute difference: 1 2025-12-04T11:24:35.4358503Z Relative difference: 1.0 2025-12-04T11:24:35.4358509Z 2025-12-04T11:24:35.4358726Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:35.4359604Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:35.4359620Z 2025-12-04T11:24:35.4359948Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:35.4360171Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:35.4360298Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:35.4360412Z stats [('calls_captured', 1)] 2025-12-04T11:24:35.4361227Z inductor [('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_count', 4), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_addmm', 1)] 2025-12-04T11:24:35.4361460Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:35.4361554Z graph_break [] 2025-12-04T11:24:35.4361785Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:35.4362283Z I1204 11:21:18.753000 100010 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:35.4362881Z I1204 11:21:19.000000 100010 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:35.4363329Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.4363819Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:35.4364811Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.4365891Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_self_modules_linear_parameters_bias_: "f32[2][1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.4366650Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:35.4367377Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_bias_ = L_self_modules_linear_parameters_bias_ 2025-12-04T11:24:35.4367828Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:35.4368258Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.4369232Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.4370621Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, l_self_modules_linear_parameters_bias_); l_x_ = l_self_modules_linear_parameters_weight_ = l_self_modules_linear_parameters_bias_ = None 2025-12-04T11:24:35.4371094Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:35.4371565Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.4371950Z V1204 11:21:19.003000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.4372504Z I1204 11:21:21.335000 100010 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:35.4373252Z V1204 11:21:21.714000 100010 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:35.4373955Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw: 2025-12-04T11:24:35.4374730Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4ral453goioq2ytnrbxj5hhqjyfnxjkats7ca6dgxms6usqpldw] gm: GraphModule() 2025-12-04T11:24:35.4375011Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.4375285Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.4375568Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.4376076Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2, primals_3): 2025-12-04T11:24:35.4377010Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.4377823Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.4378680Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_2 = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.4379456Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.4380304Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] addmm = torch.ops.aten.addmm.default(convert_element_type, convert_element_type_2, permute); convert_element_type = None 2025-12-04T11:24:35.4380947Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.4381456Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] return (addmm, convert_element_type_2, permute_1) 2025-12-04T11:24:35.4381764Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.4382332Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:35.4384633Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.4386938Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [di4vpbifhxsheleamv47gds75iwa6xnkz5malaf2qsqkrhojljw] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2]), stride=(1,), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=8, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.4389295Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[2]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.4389970Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:35.4390893Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:35.4391619Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:35.4392432Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:35.4393231Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:35.4393976Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:35.4394701Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:35.4395423Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:35.4396352Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:35.4397108Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [nnqi2dye4qdwjdgeujlxh7d67y2n67kat5ehf3472pn5avlyqbk] fx_kwargs[static_input_idxs]: [0, 1] 2025-12-04T11:24:35.4397785Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inputs_to_check[0]: 2 2025-12-04T11:24:35.4398673Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:35.4399459Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:35.4400652Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:35.4401429Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:35.4430450Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:35.4431559Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:35.4432327Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:35.4433113Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:35.4433954Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:35.4434814Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:35.4435532Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:35.4436319Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:35.4437099Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:35.4437873Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:35.4438739Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:35.4439629Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:35.4440500Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:35.4441379Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:35.4442243Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:35.4443172Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:35.4444032Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:35.4444859Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:35.4445862Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:35.4446633Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:35.4447507Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:35.4448263Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:35.4449098Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:35.4449858Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:35.4450578Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:35.4451375Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:35.4452133Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:35.4452882Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:35.4453725Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:35.4454475Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:35.4455286Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:35.4456070Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:35.4456924Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:35.4457715Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:35.4458525Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:35.4459296Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:35.4460088Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:35.4460870Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:35.4461645Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:35.4462411Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:35.4463277Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:35.4464043Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:35.4464819Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:35.4465628Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:35.4466425Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:35.4467264Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:35.4468047Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:35.4468926Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:35.4469696Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:35.4470468Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:35.4471243Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:35.4472050Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:35.4473013Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:35.4473806Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:35.4474643Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:35.4475471Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:35.4476229Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:35.4477417Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:35.4478222Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:35.4479126Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:35.4480009Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:35.4480827Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:35.4481635Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:35.4482491Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:35.4483367Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:35.4484179Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:35.4485113Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.4485938Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:35.4486932Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.4487732Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:35.4488736Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.4489555Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:35.4490466Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:35.4491230Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:35.4491990Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:35.4492825Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:35.4493676Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:35.4494493Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:35.4495305Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:35.4496270Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:35.4497165Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:35.4498040Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:35.4498973Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:35.4499828Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:35.4500611Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:35.4501437Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:35.4502326Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:35.4503195Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:35.4503990Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:35.4504928Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:35.4505792Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:35.4506721Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:35.4507588Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:35.4508406Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:35.4509206Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:35.4509941Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:35.4510779Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:35.4511584Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:35.4512391Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:35.4513223Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:35.4514069Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:35.4514885Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:35.4515707Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:35.4516445Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:35.4517234Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:35.4518161Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:35.4519121Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:35.4520067Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:35.4520899Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:35.4521731Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:35.4522601Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:35.4523453Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:35.4524350Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:35.4525206Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:35.4526002Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:35.4526843Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:35.4527684Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:35.4528518Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:35.4529305Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:35.4530094Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:35.4530934Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:35.4531726Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:35.4532495Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:35.4533312Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:35.4534123Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:35.4534997Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:35.4535853Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:35.4536628Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:35.4537611Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:35.4538393Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:35.4539304Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:35.4540114Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:35.4540907Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:35.4541661Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:35.4542436Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:35.4543255Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:35.4544084Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:35.4544935Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:35.4545757Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:35.4546587Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:35.4547441Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:35.4548198Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:35.4549105Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:35.4549947Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:35.4550810Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:35.4551608Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:35.4552448Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:35.4553247Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:35.4554012Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:35.4554849Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:35.4555612Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:35.4556366Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:35.4557148Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:35.4558005Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:35.4558880Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:35.4559707Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:35.4560489Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:35.4561255Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:35.4562077Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:35.4562883Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:35.4563739Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:35.4564616Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:35.4565440Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:35.4566276Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:35.4567103Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:35.4567935Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:35.4568791Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:35.4569597Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:35.4570411Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:35.4571289Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:35.4572083Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:35.4572892Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:35.4573764Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:35.4574601Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:35.4575368Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:35.4576184Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:35.4577100Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:35.4577890Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:35.4578722Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:35.4579684Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:35.4580689Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:35.4581678Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:35.4582681Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:35.4583663Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:35.4584661Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:35.4585699Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:35.4586701Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:35.4587692Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:35.4588537Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:35.4589356Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:35.4590186Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:35.4591135Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:35.4591979Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:35.4592758Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:35.4593536Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:35.4594302Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:35.4595109Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:35.4595906Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:35.4596843Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:35.4597648Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:35.4598529Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:35.4599425Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:35.4600257Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:35.4601056Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:35.4601883Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:35.4602701Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:35.4603486Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:35.4604317Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:35.4605159Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:35.4605954Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:35.4606811Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:35.4607550Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:35.4608330Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:35.4609138Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:35.4609966Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:35.4610930Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:35.4611723Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:35.4612460Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:35.4613311Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:35.4614121Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:35.4615026Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:35.4615834Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:35.4616649Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:35.4617708Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:35.4618583Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:35.4619477Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:35.4620263Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:35.4621181Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:35.4621978Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:35.4622803Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:35.4623623Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:35.4624443Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:35.4625324Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:35.4626143Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:35.4626900Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:35.4627706Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:35.4628510Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:35.4629267Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:35.4630052Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:35.4630877Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:35.4631710Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:35.4632505Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:35.4633399Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.4634281Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.4635090Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:35.4635941Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:35.4636787Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:35.4637638Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:35.4638500Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:35.4639424Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:35.4640269Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:35.4641136Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:35.4641945Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:35.4642766Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:35.4643600Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:35.4644429Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:35.4645255Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:35.4646088Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:35.4646964Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:35.4647798Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:35.4648646Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:35.4649550Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:35.4650371Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:35.4651301Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:35.4652168Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:35.4653041Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:35.4653968Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:35.4654900Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:35.4655818Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:35.4656737Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:35.4657648Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:35.4658517Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:35.4659354Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:35.4660296Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:35.4661180Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:35.4662042Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:35.4662866Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:35.4663729Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:35.4664559Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:35.4665445Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:35.4666246Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:35.4667097Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:35.4667941Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:35.4668755Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:35.4669614Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:35.4670506Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:35.4671359Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:35.4672175Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:35.4673075Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:35.4673981Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:35.4674809Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:35.4675670Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:35.4676559Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:35.4677399Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:35.4678261Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:35.4679142Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:35.4679984Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:35.4680786Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:35.4681655Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:35.4682461Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:35.4683287Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:35.4684104Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:35.4684944Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:35.4685908Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:35.4686787Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.4687669Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:35.4688547Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:35.4689412Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:35.4690277Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:35.4691087Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:35.4692074Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:35.4692902Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:35.4693743Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:35.4694531Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:35.4695388Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:35.4696467Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:35.4697405Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:35.4698427Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:35.4699274Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:35.4700128Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:35.4700960Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:35.4701843Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:35.4702766Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:35.4703687Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:35.4704534Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:35.4705371Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:35.4706294Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:35.4707146Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:35.4708027Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:35.4708835Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:35.4709686Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:35.4710531Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:35.4711326Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:35.4712324Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:35.4713181Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:35.4714023Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:35.4714814Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:35.4715720Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:35.4716657Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:35.4717564Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:35.4718457Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:35.4719392Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:35.4720291Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:35.4721158Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:35.4722049Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:35.4722972Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:35.4723839Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:35.4724691Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:35.4725513Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:35.4726354Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:35.4727226Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:35.4728131Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:35.4729012Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:35.4729922Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:35.4730682Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:35.4731468Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:35.4732293Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:35.4733091Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:35.4733913Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:35.4734725Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:35.4735563Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:35.4736457Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:35.4737458Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:35.4738362Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:35.4739169Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:35.4739954Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:35.4740812Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:35.4741650Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:35.4742511Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:35.4743363Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:35.4744258Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:35.4745133Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:35.4746020Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:35.4746844Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:35.4747721Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:35.4748595Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:35.4749480Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:35.4750322Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:35.4751051Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:35.4752035Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:35.4752847Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:35.4753625Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:35.4754401Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:35.4755183Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:35.4755993Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:35.4756857Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:35.4757637Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:35.4758390Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:35.4759229Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:35.4760109Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:35.4760951Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:35.4761863Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:35.4762717Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:35.4763549Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:35.4764346Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:35.4765177Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:35.4765920Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:35.4766680Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:35.4767483Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:35.4768278Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:35.4769100Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:35.4769955Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:35.4770794Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:35.4771572Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:35.4772335Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:35.4773142Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:35.4775099Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:35.4777631Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:35.4778444Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:35.4779357Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:35.4780192Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:35.4781030Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:35.4781994Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:35.4782846Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:35.4783748Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:35.4784695Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:35.4785579Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:35.4786482Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:35.4787392Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:35.4788378Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:35.4789277Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:35.4790111Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:35.4791050Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:35.4791970Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:35.4792896Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:35.4793767Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:35.4794763Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:35.4795514Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:35.4796442Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:35.4797202Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:35.4797913Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:35.4798651Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:35.4799375Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:35.4800314Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:35.4801167Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:35.4802023Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:35.4802761Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:35.4803408Z V1204 11:21:21.724000 100010 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.4804097Z I1204 11:21:21.725000 100010 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.4804585Z V1204 11:21:21.725000 100010 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:35.4805070Z V1204 11:21:21.725000 100010 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:35.4805585Z I1204 11:21:21.726000 100010 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:35.4806448Z V1204 11:21:21.768000 100010 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.addmm.default with input shape: torch.Size([2]), torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:35.4806908Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.4807462Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:35.4808415Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.4809288Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[2][1]cuda:0", primals_3: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.4810266Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.4811278Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2][1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.4812240Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.4813239Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_2: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.4814143Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.4815038Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_2, 2) 2025-12-04T11:24:35.4815799Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:35.4816911Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:35.4817738Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:35.4818765Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] add_tensor: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.add.Tensor(sum_dim_int_list, convert_element_type); sum_dim_int_list = convert_element_type = None 2025-12-04T11:24:35.4819546Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.4820169Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.4820611Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.4821004Z V1204 11:21:21.774000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.4821628Z V1204 11:21:21.778000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:35.4822194Z V1204 11:21:21.778000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:35.4822759Z V1204 11:21:21.779000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_3 : [num_users=1] = placeholder[target=primals_3] 2025-12-04T11:24:35.4823817Z V1204 11:21:21.779000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.4824326Z V1204 11:21:21.780000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.4825385Z V1204 11:21:21.781000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.4825886Z V1204 11:21:21.782000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.4826911Z V1204 11:21:21.813000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_2 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_3, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.4827407Z V1204 11:21:21.814000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.4828336Z V1204 11:21:21.815000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.4828807Z V1204 11:21:21.816000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.4829746Z V1204 11:21:21.821000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_2, 2), kwargs = {}) 2025-12-04T11:24:35.4830219Z V1204 11:21:21.822000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.4831092Z V1204 11:21:21.823000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:35.4831562Z V1204 11:21:21.823000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.4832471Z V1204 11:21:21.824000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:35.4832896Z V1204 11:21:21.825000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.4833759Z V1204 11:21:21.827000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:35.4834192Z V1204 11:21:21.827000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.4835102Z V1204 11:21:21.830000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %add_tensor : [num_users=1] = call_function[target=torch.ops.aten.add.Tensor](args = (%sum_dim_int_list, %convert_element_type), kwargs = {}) 2025-12-04T11:24:35.4835669Z V1204 11:21:21.830000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via .inner at 0x7f11500feb00> 2025-12-04T11:24:35.4836515Z V1204 11:21:21.832000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.4836992Z V1204 11:21:21.833000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.4837541Z V1204 11:21:21.833000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.4838128Z V1204 11:21:21.838000 100010 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:35.4838900Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:35.4839017Z warnings.warn( 2025-12-04T11:24:35.4839495Z V1204 11:21:21.839000 100010 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:35.4839721Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:35.4839844Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:35.4839959Z stats [('calls_captured', 1)] 2025-12-04T11:24:35.4840192Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:35.4840981Z inductor [('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_count', 4), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_addmm', 1)] 2025-12-04T11:24:35.4841113Z graph_break [] 2025-12-04T11:24:35.4841344Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:35.4841786Z V1204 11:21:22.073000 100010 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpq92dyh2h 2025-12-04T11:24:35.4842223Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.4842726Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:35.4843668Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.4844760Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_self_modules_linear_parameters_bias_: "f32[2][1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.4845510Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:35.4846247Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_bias_ = L_self_modules_linear_parameters_bias_ 2025-12-04T11:24:35.4846706Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:35.4847130Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.4848124Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.4849529Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, l_self_modules_linear_parameters_bias_); l_x_ = l_self_modules_linear_parameters_weight_ = l_self_modules_linear_parameters_bias_ = None 2025-12-04T11:24:35.4850018Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:35.4850478Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.4850877Z V1204 11:21:22.101000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.4851626Z V1204 11:21:22.368000 100010 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:35.4852352Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw: 2025-12-04T11:24:35.4853008Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4ral453goioq2ytnrbxj5hhqjyfnxjkats7ca6dgxms6usqpldw] gm: GraphModule() 2025-12-04T11:24:35.4853291Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.4853581Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.4853855Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.4854362Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2, primals_3): 2025-12-04T11:24:35.4855229Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.4856044Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.4856928Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_2 = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.4857692Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.4858555Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] addmm = torch.ops.aten.addmm.default(convert_element_type, convert_element_type_2, permute); convert_element_type = None 2025-12-04T11:24:35.4859190Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.4859708Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] return (addmm, convert_element_type_2, permute_1) 2025-12-04T11:24:35.4860005Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.4860581Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:35.4862937Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.4865254Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [di4vpbifhxsheleamv47gds75iwa6xnkz5malaf2qsqkrhojljw] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2]), stride=(1,), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=8, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.4867593Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[2]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.4868290Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:35.4869204Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:35.4869928Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:35.4870791Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:35.4871573Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:35.4872299Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:35.4873013Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:35.4873745Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:35.4874459Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:35.4875210Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [nnqi2dye4qdwjdgeujlxh7d67y2n67kat5ehf3472pn5avlyqbk] fx_kwargs[static_input_idxs]: [0, 1] 2025-12-04T11:24:35.4875898Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inputs_to_check[0]: 2 2025-12-04T11:24:35.4876776Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:35.4877575Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:35.4878786Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:35.4879581Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:35.4908746Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:35.4909859Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:35.4910639Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:35.4911404Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:35.4912264Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:35.4913110Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:35.4913830Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:35.4914616Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:35.4915400Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:35.4916304Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:35.4917158Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:35.4918036Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:35.4918890Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:35.4919776Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:35.4920675Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:35.4921568Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:35.4922440Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:35.4923253Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:35.4924311Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:35.4925069Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:35.4925919Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:35.4926662Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:35.4927494Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:35.4928270Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:35.4928976Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:35.4929783Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:35.4930530Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:35.4931289Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:35.4932176Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:35.4932942Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:35.4933763Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:35.4934540Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:35.4935355Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:35.4936138Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:35.4936977Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:35.4937746Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:35.4938543Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:35.4939337Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:35.4940141Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:35.4940924Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:35.4941736Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:35.4942512Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:35.4943279Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:35.4944095Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:35.4944913Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:35.4945741Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:35.4946540Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:35.4947403Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:35.4948223Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:35.4948977Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:35.4949753Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:35.4950575Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:35.4951532Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:35.4952334Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:35.4953120Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:35.4953966Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:35.4954710Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:35.4955931Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:35.4956740Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:35.4957608Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:35.4958501Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:35.4959323Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:35.4960141Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:35.4960983Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:35.4961857Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:35.4962680Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:35.4963644Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.4964488Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:35.4965447Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.4966295Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:35.4967227Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.4968087Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:35.4968958Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:35.4969718Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:35.4970484Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:35.4971349Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:35.4972215Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:35.4972977Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:35.4973806Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:35.4974591Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:35.4975429Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:35.4976305Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:35.4977284Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:35.4978166Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:35.4978935Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:35.4979819Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:35.4980694Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:35.4981497Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:35.4982313Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:35.4983197Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:35.4984103Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:35.4984970Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:35.4985856Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:35.4986650Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:35.4987498Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:35.4988233Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:35.4989023Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:35.4989839Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:35.4990636Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:35.4991480Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:35.4992304Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:35.4993132Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:35.4993942Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:35.4994678Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:35.4995516Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:35.4996556Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:35.4997493Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:35.4998484Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:35.4999342Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:35.5000166Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:35.5001000Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:35.5001844Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:35.5002750Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:35.5003660Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:35.5004444Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:35.5005265Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:35.5006087Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:35.5006916Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:35.5007703Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:35.5008494Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:35.5009338Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:35.5010124Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:35.5010904Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:35.5011760Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:35.5012580Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:35.5013423Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:35.5014332Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:35.5015124Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:35.5016031Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:35.5016823Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:35.5017746Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:35.5018557Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:35.5019359Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:35.5020152Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:35.5020947Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:35.5021731Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:35.5022569Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:35.5023411Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:35.5024252Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:35.5025071Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:35.5025919Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:35.5026690Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:35.5027620Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:35.5028482Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:35.5029292Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:35.5030122Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:35.5030955Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:35.5031764Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:35.5032542Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:35.5033313Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:35.5034092Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:35.5034840Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:35.5035667Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:35.5036517Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:35.5037359Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:35.5038190Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:35.5038981Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:35.5039758Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:35.5040567Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:35.5041384Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:35.5042226Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:35.5043099Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:35.5043956Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:35.5044796Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:35.5045631Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:35.5046449Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:35.5047333Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:35.5048123Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:35.5048906Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:35.5049795Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:35.5050590Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:35.5051453Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:35.5052314Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:35.5053134Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:35.5053896Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:35.5054720Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:35.5055531Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:35.5056323Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:35.5057209Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:35.5058171Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:35.5059182Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:35.5060194Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:35.5061180Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:35.5062186Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:35.5063181Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:35.5064229Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:35.5065208Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:35.5066212Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:35.5067050Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:35.5067918Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:35.5068732Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:35.5069646Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:35.5070499Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:35.5071286Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:35.5072088Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:35.5072840Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:35.5073658Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:35.5074443Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:35.5075241Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:35.5076057Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:35.5076942Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:35.5077803Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:35.5078623Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:35.5079417Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:35.5080231Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:35.5081019Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:35.5081790Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:35.5082631Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:35.5083517Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:35.5084309Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:35.5085143Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:35.5085873Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:35.5086673Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:35.5087469Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:35.5088292Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:35.5089250Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:35.5090051Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:35.5090797Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:35.5091670Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:35.5092500Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:35.5093353Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:35.5094189Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:35.5095018Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:35.5096108Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:35.5097051Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:35.5097900Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:35.5098699Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:35.5099667Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:35.5100468Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:35.5101260Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:35.5102074Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:35.5102905Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:35.5103770Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:35.5104606Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:35.5105344Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:35.5106147Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:35.5106954Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:35.5107752Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:35.5108556Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:35.5109331Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:35.5110228Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:35.5111012Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:35.5111927Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.5112809Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.5113577Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:35.5114445Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:35.5115281Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:35.5116185Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:35.5117041Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:35.5117947Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:35.5118785Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:35.5119658Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:35.5120474Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:35.5121295Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:35.5122132Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:35.5122952Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:35.5123833Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:35.5124653Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:35.5125499Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:35.5126356Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:35.5127205Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:35.5128046Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:35.5128848Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:35.5129751Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:35.5130611Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:35.5131478Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:35.5132438Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:35.5133338Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:35.5134279Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:35.5135183Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:35.5136043Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:35.5136958Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:35.5137802Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:35.5138729Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:35.5139605Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:35.5140503Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:35.5141330Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:35.5142200Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:35.5143011Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:35.5143916Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:35.5144697Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:35.5145516Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:35.5146354Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:35.5147176Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:35.5148091Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:35.5148963Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:35.5149789Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:35.5150589Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:35.5151485Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:35.5152394Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:35.5153216Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:35.5154085Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:35.5154949Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:35.5155801Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:35.5156682Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:35.5157575Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:35.5158386Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:35.5159191Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:35.5160075Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:35.5160864Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:35.5161671Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:35.5162473Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:35.5163328Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:35.5164785Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:35.5165662Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.5166522Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:35.5167395Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:35.5168273Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:35.5169132Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:35.5169946Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:35.5170926Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:35.5171758Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:35.5172650Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:35.5173441Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:35.5174282Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:35.5175220Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:35.5176129Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:35.5177212Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:35.5178025Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:35.5178863Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:35.5179695Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:35.5180628Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:35.5181542Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:35.5182430Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:35.5183271Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:35.5184125Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:35.5185039Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:35.5185894Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:35.5186775Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:35.5187581Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:35.5188444Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:35.5189310Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:35.5190123Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:35.5191085Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:35.5191945Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:35.5192795Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:35.5193585Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:35.5194479Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:35.5195400Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:35.5196634Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:35.5197605Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:35.5198562Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:35.5199425Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:35.5200285Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:35.5201190Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:35.5202111Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:35.5202983Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:35.5203827Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:35.5204665Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:35.5205546Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:35.5206426Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:35.5207291Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:35.5208206Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:35.5209107Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:35.5209897Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:35.5210667Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:35.5211482Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:35.5212281Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:35.5213111Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:35.5213959Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:35.5214760Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:35.5215642Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:35.5216609Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:35.5217565Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:35.5218389Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:35.5219152Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:35.5220016Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:35.5220864Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:35.5221761Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:35.5222629Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:35.5223503Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:35.5224390Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:35.5225240Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:35.5226068Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:35.5226916Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:35.5227792Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:35.5228684Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:35.5229564Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:35.5230295Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:35.5231228Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:35.5232042Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:35.5232817Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:35.5233596Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:35.5234388Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:35.5235175Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:35.5236051Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:35.5236821Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:35.5237600Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:35.5238445Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:35.5239286Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:35.5240181Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:35.5241053Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:35.5241956Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:35.5242741Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:35.5243552Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:35.5244366Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:35.5245148Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:35.5245934Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:35.5246690Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:35.5247500Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:35.5248312Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:35.5249181Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:35.5250005Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:35.5250774Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:35.5251555Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:35.5252348Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:35.5254359Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:35.5256813Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:35.5257696Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:35.5258611Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:35.5259419Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:35.5260253Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:35.5261228Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:35.5262115Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:35.5263018Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:35.5263935Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:35.5264817Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:35.5265735Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:35.5266640Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:35.5267633Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:35.5268517Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:35.5269375Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:35.5270333Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:35.5271255Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:35.5272199Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:35.5273072Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:35.5274095Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:35.5274837Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:35.5275549Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:35.5276297Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:35.5277011Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:35.5277773Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:35.5278500Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:35.5279387Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:35.5280230Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:35.5281087Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:35.5281815Z V1204 11:21:22.378000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:35.5282462Z V1204 11:21:22.379000 100010 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.5283145Z I1204 11:21:22.379000 100010 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.5283630Z V1204 11:21:22.379000 100010 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:35.5284126Z V1204 11:21:22.379000 100010 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:35.5284633Z I1204 11:21:22.380000 100010 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:35.5285545Z V1204 11:21:22.394000 100010 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.addmm.default with input shape: torch.Size([2]), torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:35.5285992Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.5286488Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:35.5287484Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.5288311Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[2][1]cuda:0", primals_3: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.5289352Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.5290307Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2][1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.5291283Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.5292272Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_2: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.5293222Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.5294060Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_2, 2) 2025-12-04T11:24:35.5294816Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:35.5295882Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:35.5296840Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:35.5297927Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] add_tensor: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.add.Tensor(sum_dim_int_list, convert_element_type); sum_dim_int_list = convert_element_type = None 2025-12-04T11:24:35.5298707Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.5299408Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.5299834Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.5300238Z V1204 11:21:22.400000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.5300809Z V1204 11:21:22.404000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:35.5301411Z V1204 11:21:22.405000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:35.5301986Z V1204 11:21:22.405000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_3 : [num_users=1] = placeholder[target=primals_3] 2025-12-04T11:24:35.5303039Z V1204 11:21:22.406000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.5303556Z V1204 11:21:22.406000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.5304570Z V1204 11:21:22.407000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.5305081Z V1204 11:21:22.407000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.5306088Z V1204 11:21:22.408000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_2 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_3, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.5306631Z V1204 11:21:22.408000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.5307536Z V1204 11:21:22.410000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.5307988Z V1204 11:21:22.410000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.5308933Z V1204 11:21:22.411000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_2, 2), kwargs = {}) 2025-12-04T11:24:35.5309387Z V1204 11:21:22.412000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.5310284Z V1204 11:21:22.413000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:35.5310735Z V1204 11:21:22.413000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.5311650Z V1204 11:21:22.414000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:35.5312090Z V1204 11:21:22.414000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.5312936Z V1204 11:21:22.416000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:35.5313413Z V1204 11:21:22.416000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.5314315Z V1204 11:21:22.418000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %add_tensor : [num_users=1] = call_function[target=torch.ops.aten.add.Tensor](args = (%sum_dim_int_list, %convert_element_type), kwargs = {}) 2025-12-04T11:24:35.5314900Z V1204 11:21:22.418000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via .inner at 0x7f11500feb00> 2025-12-04T11:24:35.5315739Z V1204 11:21:22.420000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.5316219Z V1204 11:21:22.421000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.5316791Z V1204 11:21:22.421000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.5317366Z V1204 11:21:22.424000 100010 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:35.5318117Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:35.5318224Z warnings.warn( 2025-12-04T11:24:35.5318707Z V1204 11:21:22.424000 100010 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:35.5318944Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:35.5319092Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:35.5319209Z stats [('calls_captured', 1)] 2025-12-04T11:24:35.5319444Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:35.5320236Z inductor [('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_count', 4), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_addmm', 1)] 2025-12-04T11:24:35.5320348Z graph_break [] 2025-12-04T11:24:35.5320568Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:35.5321008Z V1204 11:21:22.445000 100010 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmp7tpsu88v 2025-12-04T11:24:35.5321458Z V1204 11:21:22.472000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.5321948Z V1204 11:21:22.472000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:35.5322909Z V1204 11:21:22.472000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.5323990Z V1204 11:21:22.472000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_self_modules_linear_parameters_bias_: "f32[2][1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.5324755Z V1204 11:21:22.472000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:35.5325476Z V1204 11:21:22.472000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_bias_ = L_self_modules_linear_parameters_bias_ 2025-12-04T11:24:35.5325932Z V1204 11:21:22.472000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:35.5326390Z V1204 11:21:22.472000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.5327365Z V1204 11:21:22.472000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.5328776Z V1204 11:21:22.472000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, l_self_modules_linear_parameters_bias_); l_x_ = l_self_modules_linear_parameters_weight_ = l_self_modules_linear_parameters_bias_ = None 2025-12-04T11:24:35.5329279Z V1204 11:21:22.472000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:35.5329712Z V1204 11:21:22.472000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.5330096Z V1204 11:21:22.472000 100010 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.5330807Z V1204 11:21:22.730000 100010 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:35.5331531Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw: 2025-12-04T11:24:35.5332188Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4ral453goioq2ytnrbxj5hhqjyfnxjkats7ca6dgxms6usqpldw] gm: GraphModule() 2025-12-04T11:24:35.5332513Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.5332790Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.5333067Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.5333597Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2, primals_3): 2025-12-04T11:24:35.5334411Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.5335238Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.5336050Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_2 = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.5336831Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.5337764Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] addmm = torch.ops.aten.addmm.default(convert_element_type, convert_element_type_2, permute); convert_element_type = None 2025-12-04T11:24:35.5338592Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.5339117Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] return (addmm, convert_element_type_2, permute_1) 2025-12-04T11:24:35.5339414Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.5340004Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:35.5342398Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.5344706Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [di4vpbifhxsheleamv47gds75iwa6xnkz5malaf2qsqkrhojljw] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2]), stride=(1,), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=8, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.5347069Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[2]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.5347758Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:35.5348705Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:35.5349444Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:35.5350264Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:35.5351060Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:35.5351769Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:35.5352487Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:35.5353221Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:35.5353928Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:35.5354696Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [nnqi2dye4qdwjdgeujlxh7d67y2n67kat5ehf3472pn5avlyqbk] fx_kwargs[static_input_idxs]: [0, 1] 2025-12-04T11:24:35.5355369Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inputs_to_check[0]: 2 2025-12-04T11:24:35.5356258Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:35.5357075Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:35.5358266Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:35.5359097Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:35.5388081Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:35.5389185Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:35.5389958Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:35.5390726Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:35.5391585Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:35.5392430Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:35.5393166Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:35.5393968Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:35.5394762Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:35.5395565Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:35.5396579Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:35.5397491Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:35.5398349Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:35.5399212Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:35.5400060Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:35.5400970Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:35.5401878Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:35.5402701Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:35.5403708Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:35.5404463Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:35.5405308Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:35.5406053Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:35.5406903Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:35.5407663Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:35.5408381Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:35.5409173Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:35.5409964Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:35.5410734Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:35.5411609Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:35.5412386Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:35.5413199Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:35.5413988Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:35.5414761Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:35.5415542Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:35.5416333Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:35.5417166Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:35.5418015Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:35.5418792Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:35.5419578Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:35.5420343Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:35.5421156Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:35.5421937Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:35.5422694Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:35.5423519Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:35.5424317Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:35.5425162Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:35.5425995Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:35.5426854Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:35.5427660Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:35.5428421Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:35.5429287Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:35.5430221Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:35.5431162Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:35.5431954Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:35.5432745Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:35.5433636Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:35.5434390Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:35.5435596Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:35.5436395Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:35.5437278Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:35.5438174Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:35.5438988Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:35.5439815Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:35.5440661Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:35.5441600Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:35.5442404Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:35.5443355Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.5444206Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:35.5445181Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.5446024Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:35.5446967Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.5447802Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:35.5448680Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:35.5449486Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:35.5450241Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:35.5451094Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:35.5451942Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:35.5452701Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:35.5453540Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:35.5454324Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:35.5455180Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:35.5456043Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:35.5457036Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:35.5457937Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:35.5458708Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:35.5459547Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:35.5460457Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:35.5461276Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:35.5462089Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:35.5462989Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:35.5463851Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:35.5464735Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:35.5465638Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:35.5466432Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:35.5467245Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:35.5467967Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:35.5468773Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:35.5469580Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:35.5470386Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:35.5471215Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:35.5472039Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:35.5472869Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:35.5473708Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:35.5474459Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:35.5475251Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:35.5476202Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:35.5477120Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:35.5478082Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:35.5478901Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:35.5479726Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:35.5480574Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:35.5481448Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:35.5482367Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:35.5483222Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:35.5484004Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:35.5484821Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:35.5485659Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:35.5486497Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:35.5487264Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:35.5488063Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:35.5488890Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:35.5489704Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:35.5490492Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:35.5491297Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:35.5492163Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:35.5492998Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:35.5493911Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:35.5494688Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:35.5495574Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:35.5496510Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:35.5497438Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:35.5498333Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:35.5499123Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:35.5499894Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:35.5500676Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:35.5501474Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:35.5502303Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:35.5503146Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:35.5503991Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:35.5504815Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:35.5505731Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:35.5506490Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:35.5507399Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:35.5508280Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:35.5509100Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:35.5509955Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:35.5510787Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:35.5511571Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:35.5512341Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:35.5513129Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:35.5513952Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:35.5514694Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:35.5515488Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:35.5516338Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:35.5517199Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:35.5518029Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:35.5518834Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:35.5519595Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:35.5520453Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:35.5521266Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:35.5522138Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:35.5523026Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:35.5523864Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:35.5524708Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:35.5525532Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:35.5526351Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:35.5527187Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:35.5527983Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:35.5528779Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:35.5529680Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:35.5530489Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:35.5531300Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:35.5532163Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:35.5532985Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:35.5533746Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:35.5534577Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:35.5535368Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:35.5536177Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:35.5537044Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:35.5538041Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:35.5539049Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:35.5540063Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:35.5541040Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:35.5542062Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:35.5543071Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:35.5544067Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:35.5545055Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:35.5546086Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:35.5546935Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:35.5547759Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:35.5548574Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:35.5549495Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:35.5550335Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:35.5551125Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:35.5551901Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:35.5552667Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:35.5553470Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:35.5554292Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:35.5555093Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:35.5555909Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:35.5556803Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:35.5557656Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:35.5558489Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:35.5559238Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:35.5560051Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:35.5560839Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:35.5561638Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:35.5562487Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:35.5563327Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:35.5564133Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:35.5564951Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:35.5565688Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:35.5566487Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:35.5567277Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:35.5568120Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:35.5569060Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:35.5569902Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:35.5570641Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:35.5571484Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:35.5572332Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:35.5573183Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:35.5574041Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:35.5574844Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:35.5575795Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:35.5576666Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:35.5577564Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:35.5578402Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:35.5579303Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:35.5580113Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:35.5580891Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:35.5581719Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:35.5582542Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:35.5583425Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:35.5584257Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:35.5584996Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:35.5585856Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:35.5586650Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:35.5587419Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:35.5588239Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:35.5589033Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:35.5589903Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:35.5590686Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:35.5591562Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.5592432Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.5593219Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:35.5594103Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:35.5594952Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:35.5595803Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:35.5596782Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:35.5597694Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:35.5598536Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:35.5599423Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:35.5600224Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:35.5601070Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:35.5601961Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:35.5602783Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:35.5603627Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:35.5604496Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:35.5605357Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:35.5606223Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:35.5607087Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:35.5607875Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:35.5608698Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:35.5609587Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:35.5610487Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:35.5611371Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:35.5612279Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:35.5613192Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:35.5614120Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:35.5615037Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:35.5615877Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:35.5616732Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:35.5617623Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:35.5618581Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:35.5619470Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:35.5620346Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:35.5621182Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:35.5622041Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:35.5622855Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:35.5623723Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:35.5624507Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:35.5625333Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:35.5626187Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:35.5627012Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:35.5627870Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:35.5628758Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:35.5629571Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:35.5630378Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:35.5631288Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:35.5632188Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:35.5633021Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:35.5633872Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:35.5634779Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:35.5635617Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:35.5636490Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:35.5637381Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:35.5638200Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:35.5639011Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:35.5639845Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:35.5640654Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:35.5641450Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:35.5642286Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:35.5643143Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:35.5644089Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:35.5644986Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.5645821Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:35.5646716Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:35.5647580Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:35.5648450Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:35.5649252Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:35.5650268Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:35.5651111Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:35.5651935Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:35.5652768Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:35.5653594Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:35.5654546Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:35.5655437Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:35.5656368Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:35.5657239Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:35.5658074Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:35.5658961Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:35.5659840Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:35.5660771Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:35.5661648Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:35.5662495Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:35.5663348Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:35.5664261Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:35.5665137Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:35.5665996Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:35.5666869Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:35.5667728Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:35.5668607Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:35.5669406Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:35.5670370Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:35.5671242Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:35.5672045Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:35.5672846Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:35.5673731Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:35.5674703Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:35.5675607Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:35.5676487Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:35.5677433Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:35.5678291Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:35.5679169Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:35.5680048Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:35.5680979Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:35.5681839Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:35.5682719Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:35.5683536Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:35.5684357Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:35.5685352Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:35.5686206Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:35.5687134Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:35.5688018Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:35.5688779Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:35.5689540Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:35.5690349Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:35.5691190Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:35.5692006Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:35.5692829Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:35.5693617Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:35.5694516Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:35.5695462Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:35.5696472Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:35.5697352Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:35.5698116Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:35.5699051Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:35.5699878Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:35.5700744Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:35.5701633Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:35.5702470Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:35.5703400Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:35.5704224Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:35.5705060Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:35.5705890Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:35.5706826Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:35.5707712Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:35.5708549Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:35.5709271Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:35.5710202Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:35.5711036Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:35.5711802Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:35.5712595Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:35.5713380Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:35.5714184Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:35.5715089Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:35.5715860Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:35.5716619Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:35.5717480Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:35.5718337Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:35.5719223Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:35.5720115Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:35.5720971Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:35.5721760Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:35.5722611Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:35.5723425Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:35.5724180Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:35.5724945Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:35.5725714Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:35.5732090Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:35.5733001Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:35.5733875Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:35.5734707Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:35.5735496Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:35.5736262Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:35.5737238Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:35.5739269Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:35.5741702Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:35.5742512Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:35.5743438Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:35.5744233Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:35.5745107Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:35.5746082Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:35.5746926Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:35.5747843Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:35.5748758Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:35.5749664Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:35.5750562Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:35.5751464Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:35.5752463Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:35.5753382Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:35.5754228Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:35.5755180Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:35.5756113Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:35.5757026Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:35.5757907Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:35.5758871Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:35.5759611Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:35.5760319Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:35.5761099Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:35.5761822Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:35.5762535Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:35.5763270Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:35.5764139Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:35.5764990Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:35.5765848Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:35.5766563Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:35.5767231Z V1204 11:21:22.740000 100010 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.5767899Z I1204 11:21:22.741000 100010 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.5768400Z V1204 11:21:22.741000 100010 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:35.5768911Z V1204 11:21:22.741000 100010 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:35.5769418Z I1204 11:21:22.742000 100010 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 2 2025-12-04T11:24:35.5770316Z V1204 11:21:22.756000 100010 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.addmm.default with input shape: torch.Size([2]), torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:35.5770763Z V1204 11:21:22.762000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.5771272Z V1204 11:21:22.762000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:35.5772255Z V1204 11:21:22.762000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.5773080Z V1204 11:21:22.762000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[2][1]cuda:0", primals_3: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.5774071Z V1204 11:21:22.762000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.5775017Z V1204 11:21:22.762000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2][1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.5776021Z V1204 11:21:22.762000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.5777112Z V1204 11:21:22.762000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_2: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.5778044Z V1204 11:21:22.762000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.5778884Z V1204 11:21:22.762000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_2, 2) 2025-12-04T11:24:35.5779663Z V1204 11:21:22.762000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:35.5780703Z V1204 11:21:22.762000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:35.5781539Z V1204 11:21:22.762000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:35.5782586Z V1204 11:21:22.762000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] add_tensor: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.add.Tensor(sum_dim_int_list, convert_element_type); sum_dim_int_list = convert_element_type = None 2025-12-04T11:24:35.5783382Z V1204 11:21:22.762000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.5783998Z V1204 11:21:22.762000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.5784452Z V1204 11:21:22.762000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.5784855Z V1204 11:21:22.762000 100010 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.5785477Z V1204 11:21:22.766000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:35.5786052Z V1204 11:21:22.767000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:35.5786617Z V1204 11:21:22.767000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_3 : [num_users=1] = placeholder[target=primals_3] 2025-12-04T11:24:35.5787624Z V1204 11:21:22.768000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.5788136Z V1204 11:21:22.768000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.5789150Z V1204 11:21:22.769000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.5789697Z V1204 11:21:22.769000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.5790703Z V1204 11:21:22.770000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_2 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_3, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.5791217Z V1204 11:21:22.770000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.5792112Z V1204 11:21:22.772000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.5792567Z V1204 11:21:22.772000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.5793518Z V1204 11:21:22.773000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_2, 2), kwargs = {}) 2025-12-04T11:24:35.5793969Z V1204 11:21:22.774000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.5794859Z V1204 11:21:22.775000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:35.5795310Z V1204 11:21:22.775000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.5796383Z V1204 11:21:22.776000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:35.5796878Z V1204 11:21:22.776000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.5797727Z V1204 11:21:22.778000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:35.5798216Z V1204 11:21:22.778000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.5799123Z V1204 11:21:22.780000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %add_tensor : [num_users=1] = call_function[target=torch.ops.aten.add.Tensor](args = (%sum_dim_int_list, %convert_element_type), kwargs = {}) 2025-12-04T11:24:35.5799718Z V1204 11:21:22.780000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via .inner at 0x7f11500feb00> 2025-12-04T11:24:35.5800559Z V1204 11:21:22.782000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.5801016Z V1204 11:21:22.782000 100010 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.5801568Z V1204 11:21:22.783000 100010 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.5802145Z V1204 11:21:22.786000 100010 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 2 2025-12-04T11:24:35.5802891Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:35.5803035Z warnings.warn( 2025-12-04T11:24:35.5803530Z V1204 11:21:22.786000 100010 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:35.5804366Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-fd020a6d441a5738.xml - 2025-12-04T11:24:35.5804540Z =========================== short test summary info ============================ 2025-12-04T11:24:35.5805474Z FAILED [0.3597s] inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True - AssertionError: Scalars are not equal! 2025-12-04T11:24:35.5805484Z 2025-12-04T11:24:35.5805590Z Expected 1 but got 0. 2025-12-04T11:24:35.5805711Z Absolute difference: 1 2025-12-04T11:24:35.5805826Z Relative difference: 1.0 2025-12-04T11:24:35.5805832Z 2025-12-04T11:24:35.5806049Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:35.5806930Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:35.5806936Z 2025-12-04T11:24:35.5807205Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:35.5807395Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:24:35.5807592Z ================== 1 failed, 15 deselected, 2 rerun in 6.73s =================== 2025-12-04T11:24:35.5807701Z --- Logging error --- 2025-12-04T11:24:35.5807831Z Traceback (most recent call last): 2025-12-04T11:24:35.5808139Z File "/opt/conda/envs/py_3.10/lib/python3.10/logging/__init__.py", line 1103, in emit 2025-12-04T11:24:35.5808269Z stream.write(msg + self.terminator) 2025-12-04T11:24:35.5808417Z ValueError: I/O operation on closed file. 2025-12-04T11:24:35.5808512Z Call stack: 2025-12-04T11:24:35.5809073Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/remote_cache.py", line 432, in dump_cache_stats 2025-12-04T11:24:35.5809220Z log.info("Cache Metrics:%s", out.getvalue()) 2025-12-04T11:24:35.5809332Z Message: 'Cache Metrics:%s' 2025-12-04T11:24:35.5809446Z Arguments: (' None\n',) 2025-12-04T11:24:35.5809541Z Got exit code 1 2025-12-04T11:24:35.5809645Z Retrying single test... 2025-12-04T11:24:35.5810322Z Test results will be stored in test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-b7f552c9fcec58a4.xml 2025-12-04T11:24:35.5810515Z ============================= test session starts ============================== 2025-12-04T11:24:35.5810878Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:24:35.5811016Z cachedir: .pytest_cache 2025-12-04T11:24:35.5811537Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:24:35.5811670Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:24:35.5811779Z configfile: pytest.ini 2025-12-04T11:24:35.5812373Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:24:35.5813578Z collecting ... /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:52: PytestCollectionWarning: cannot collect test class 'TestDecomposeAddMM' because it has a __init__ constructor (from: test/inductor/test_decompose_mem_bound_mm.py) 2025-12-04T11:24:35.5813728Z class TestDecomposeAddMM(torch.nn.Module): 2025-12-04T11:24:35.5813886Z collected 37 items / 36 deselected / 1 selected 2025-12-04T11:24:35.5814827Z stepcurrent: skipping 15 already run items. Running only test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:35.5814970Z Running 1 items in this shard 2025-12-04T11:24:35.5814988Z 2025-12-04T11:24:35.5815808Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True ('RERUN', {'yellow': True}) [6.0142s] [100%] 2025-12-04T11:24:35.5816625Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True ('RERUN', {'yellow': True}) [0.3928s] [100%] 2025-12-04T11:24:35.5817435Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True FAILED [0.3689s] [100%] 2025-12-04T11:24:35.5817443Z 2025-12-04T11:24:35.5817586Z ==================================== RERUNS ==================================== 2025-12-04T11:24:35.5818064Z _ TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True _ 2025-12-04T11:24:35.5818184Z Traceback (most recent call last): 2025-12-04T11:24:35.5818769Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 275, in test_decompose_linear_mixed_precision 2025-12-04T11:24:35.5818887Z self.assertEqual( 2025-12-04T11:24:35.5819346Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:35.5819512Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:35.5820059Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:35.5820265Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:35.5820407Z AssertionError: Scalars are not equal! 2025-12-04T11:24:35.5820413Z 2025-12-04T11:24:35.5820519Z Expected 1 but got 0. 2025-12-04T11:24:35.5820627Z Absolute difference: 1 2025-12-04T11:24:35.5820744Z Relative difference: 1.0 2025-12-04T11:24:35.5820749Z 2025-12-04T11:24:35.5820965Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:35.5821876Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:35.5821882Z 2025-12-04T11:24:35.5822151Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:35.5822365Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:35.5822517Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:35.5822631Z stats [('calls_captured', 1)] 2025-12-04T11:24:35.5823425Z inductor [('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_count', 4), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_addmm', 1)] 2025-12-04T11:24:35.5823678Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:35.5823777Z graph_break [] 2025-12-04T11:24:35.5824010Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:35.5824508Z I1204 11:21:38.249000 100207 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:35.5825108Z I1204 11:21:38.495000 100207 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:35.5825563Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.5826053Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:35.5827003Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.5828114Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_self_modules_linear_parameters_bias_: "f32[2][1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.5828875Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:35.5829599Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_bias_ = L_self_modules_linear_parameters_bias_ 2025-12-04T11:24:35.5830053Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:35.5830490Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.5831465Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.5832856Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, l_self_modules_linear_parameters_bias_); l_x_ = l_self_modules_linear_parameters_weight_ = l_self_modules_linear_parameters_bias_ = None 2025-12-04T11:24:35.5833324Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:35.5833759Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.5834166Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.5834727Z I1204 11:21:40.836000 100207 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:35.5835453Z V1204 11:21:41.215000 100207 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:35.5836205Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw: 2025-12-04T11:24:35.5836876Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4ral453goioq2ytnrbxj5hhqjyfnxjkats7ca6dgxms6usqpldw] gm: GraphModule() 2025-12-04T11:24:35.5837190Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.5837467Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.5837750Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.5838259Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2, primals_3): 2025-12-04T11:24:35.5839083Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.5839890Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.5840715Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_2 = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.5841505Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.5842353Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] addmm = torch.ops.aten.addmm.default(convert_element_type, convert_element_type_2, permute); convert_element_type = None 2025-12-04T11:24:35.5843000Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.5843502Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] return (addmm, convert_element_type_2, permute_1) 2025-12-04T11:24:35.5843817Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.5844387Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:35.5846715Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.5849031Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [di4vpbifhxsheleamv47gds75iwa6xnkz5malaf2qsqkrhojljw] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2]), stride=(1,), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=8, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.5851412Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[2]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.5852055Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:35.5853010Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:35.5853735Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:35.5854554Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:35.5855335Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:35.5856050Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:35.5856813Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:35.5857595Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:35.5858322Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:35.5859072Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [nnqi2dye4qdwjdgeujlxh7d67y2n67kat5ehf3472pn5avlyqbk] fx_kwargs[static_input_idxs]: [0, 1] 2025-12-04T11:24:35.5859758Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inputs_to_check[0]: 2 2025-12-04T11:24:35.5860635Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:35.5861422Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:35.5862619Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:35.5863403Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:35.5892320Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:35.5893375Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:35.5894193Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:35.5894961Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:35.5895797Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:35.5896777Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:35.5897568Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:35.5898357Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:35.5899141Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:35.5899930Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:35.5900780Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:35.5901618Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:35.5902555Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:35.5903409Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:35.5904316Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:35.5905211Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:35.5906121Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:35.5906938Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:35.5907942Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:35.5908714Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:35.5909550Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:35.5910352Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:35.5911185Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:35.5911952Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:35.5912650Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:35.5913451Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:35.5914209Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:35.5914957Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:35.5915805Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:35.5916558Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:35.5917347Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:35.5918165Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:35.5918938Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:35.5919754Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:35.5920530Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:35.5921333Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:35.5922125Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:35.5922912Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:35.5923674Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:35.5924440Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:35.5925295Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:35.5926072Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:35.5926844Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:35.5927655Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:35.5928463Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:35.5929287Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:35.5930073Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:35.5930949Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:35.5931727Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:35.5932499Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:35.5933287Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:35.5934106Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:35.5935027Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:35.5935852Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:35.5936655Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:35.5937606Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:35.5938367Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:35.5939556Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:35.5940364Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:35.5941269Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:35.5942166Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:35.5942976Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:35.5943785Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:35.5944644Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:35.5945529Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:35.5946344Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:35.5947285Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.5948123Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:35.5949074Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.5949982Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:35.5950934Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.5951782Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:35.5952676Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:35.5953468Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:35.5954232Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:35.5955075Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:35.5955923Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:35.5956701Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:35.5957554Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:35.5958347Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:35.5959181Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:35.5960052Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:35.5960972Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:35.5961850Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:35.5962621Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:35.5963451Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:35.5964337Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:35.5965140Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:35.5965965Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:35.5966854Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:35.5967767Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:35.5968629Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:35.5969531Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:35.5970333Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:35.5971127Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:35.5971858Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:35.5972642Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:35.5973514Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:35.5974303Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:35.5975130Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:35.5975967Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:35.5976787Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:35.5977676Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:35.5978415Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:35.5979221Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:35.5980137Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:35.5981065Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:35.5982062Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:35.5982871Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:35.5983735Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:35.5984574Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:35.5985444Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:35.5986349Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:35.5987212Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:35.5987997Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:35.5988809Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:35.5989691Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:35.5990518Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:35.5991296Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:35.5992081Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:35.5992924Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:35.5993708Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:35.5994482Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:35.5995298Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:35.5996301Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:35.5997140Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:35.5998081Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:35.5998870Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:35.5999791Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:35.6000582Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:35.6001511Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:35.6002325Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:35.6003120Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:35.6003877Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:35.6004669Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:35.6005489Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:35.6006318Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:35.6007169Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:35.6008003Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:35.6008842Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:35.6009705Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:35.6010475Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:35.6011367Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:35.6012210Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:35.6013033Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:35.6013860Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:35.6014700Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:35.6015468Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:35.6016273Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:35.6017102Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:35.6017908Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:35.6018663Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:35.6019440Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:35.6020299Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:35.6021147Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:35.6022021Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:35.6022813Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:35.6023576Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:35.6024402Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:35.6025204Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:35.6026064Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:35.6026935Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:35.6027765Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:35.6028597Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:35.6029429Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:35.6030258Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:35.6031081Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:35.6031917Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:35.6032705Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:35.6033634Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:35.6034435Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:35.6035270Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:35.6036130Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:35.6036946Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:35.6037756Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:35.6038577Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:35.6039386Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:35.6040172Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:35.6041009Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:35.6041968Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:35.6042975Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:35.6043967Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:35.6044933Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:35.6045953Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:35.6046953Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:35.6047992Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:35.6048966Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:35.6050002Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:35.6050844Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:35.6051666Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:35.6052493Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:35.6053404Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:35.6054280Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:35.6055058Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:35.6055850Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:35.6056608Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:35.6057474Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:35.6058279Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:35.6059065Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:35.6059863Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:35.6060743Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:35.6061575Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:35.6062429Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:35.6063186Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:35.6064036Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:35.6064811Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:35.6065622Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:35.6066455Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:35.6067305Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:35.6068090Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:35.6068914Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:35.6069698Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:35.6070483Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:35.6071289Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:35.6072113Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:35.6073072Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:35.6073879Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:35.6074610Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:35.6075461Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:35.6076275Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:35.6077142Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:35.6077987Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:35.6078810Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:35.6079787Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:35.6080673Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:35.6081551Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:35.6082340Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:35.6083260Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:35.6084055Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:35.6084850Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:35.6085702Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:35.6086535Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:35.6087396Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:35.6088217Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:35.6088968Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:35.6089777Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:35.6090587Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:35.6091343Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:35.6092146Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:35.6092918Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:35.6093779Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:35.6094576Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:35.6095489Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.6096527Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.6097421Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:35.6098300Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:35.6099139Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:35.6099991Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:35.6100865Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:35.6101811Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:35.6102668Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:35.6103538Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:35.6104347Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:35.6105166Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:35.6105989Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:35.6106814Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:35.6107632Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:35.6108461Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:35.6109302Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:35.6110185Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:35.6111038Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:35.6111881Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:35.6112690Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:35.6113607Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:35.6114487Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:35.6115354Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:35.6116280Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:35.6117183Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:35.6118151Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:35.6119051Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:35.6119893Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:35.6120761Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:35.6121593Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:35.6122530Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:35.6123407Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:35.6124272Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:35.6125086Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:35.6125914Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:35.6126765Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:35.6127611Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:35.6128436Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:35.6129258Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:35.6130124Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:35.6130939Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:35.6131810Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:35.6132684Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:35.6133498Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:35.6134342Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:35.6135238Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:35.6136212Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:35.6137094Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:35.6137964Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:35.6138835Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:35.6139673Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:35.6140539Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:35.6141415Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:35.6142250Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:35.6143059Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:35.6143908Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:35.6144730Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:35.6145526Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:35.6146369Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:35.6147214Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:35.6148174Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:35.6149056Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.6149945Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:35.6150823Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:35.6151706Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:35.6152561Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:35.6153365Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:35.6154362Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:35.6155192Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:35.6156036Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:35.6156826Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:35.6157663Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:35.6158620Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:35.6159518Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:35.6160493Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:35.6161292Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:35.6162177Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:35.6163007Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:35.6163901Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:35.6164822Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:35.6165699Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:35.6166587Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:35.6167424Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:35.6168350Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:35.6169212Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:35.6170086Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:35.6170895Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:35.6171759Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:35.6172595Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:35.6173396Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:35.6174379Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:35.6175240Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:35.6176058Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:35.6176950Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:35.6177851Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:35.6178808Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:35.6179725Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:35.6180614Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:35.6181560Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:35.6182472Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:35.6183337Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:35.6184239Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:35.6185156Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:35.6186033Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:35.6186877Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:35.6187695Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:35.6188537Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:35.6189413Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:35.6190279Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:35.6191194Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:35.6192096Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:35.6192874Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:35.6193642Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:35.6194502Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:35.6195310Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:35.6196275Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:35.6197095Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:35.6197899Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:35.6198947Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:35.6199897Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:35.6200801Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:35.6201604Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:35.6202383Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:35.6203251Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:35.6204089Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:35.6204949Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:35.6205804Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:35.6206674Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:35.6207597Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:35.6208441Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:35.6209301Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:35.6210149Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:35.6211066Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:35.6211966Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:35.6212791Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:35.6213521Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:35.6214466Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:35.6215332Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:35.6216106Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:35.6216953Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:35.6217755Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:35.6218543Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:35.6219417Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:35.6220199Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:35.6220946Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:35.6221792Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:35.6222642Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:35.6223553Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:35.6224430Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:35.6225314Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:35.6226113Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:35.6226942Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:35.6227771Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:35.6228515Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:35.6229295Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:35.6230056Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:35.6230876Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:35.6231703Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:35.6232553Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:35.6233390Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:35.6234164Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:35.6234948Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:35.6235743Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:35.6237729Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:35.6240175Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:35.6240956Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:35.6241899Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:35.6242690Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:35.6243576Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:35.6244536Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:35.6245392Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:35.6246296Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:35.6247256Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:35.6248149Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:35.6249066Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:35.6249963Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:35.6250945Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:35.6251846Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:35.6252671Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:35.6253610Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:35.6254578Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:35.6255510Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:35.6256385Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:35.6257448Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:35.6258225Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:35.6258929Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:35.6259721Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:35.6260440Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:35.6261172Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:35.6261896Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:35.6262769Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:35.6263655Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:35.6264498Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:35.6265227Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:35.6265876Z V1204 11:21:41.225000 100207 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.6266550Z I1204 11:21:41.225000 100207 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.6267039Z V1204 11:21:41.226000 100207 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:35.6267523Z V1204 11:21:41.226000 100207 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:35.6268038Z I1204 11:21:41.226000 100207 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:35.6268902Z V1204 11:21:41.270000 100207 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.addmm.default with input shape: torch.Size([2]), torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:35.6269360Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.6269857Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:35.6270856Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.6271678Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[2][1]cuda:0", primals_3: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.6272685Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.6273650Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2][1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.6274661Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.6275656Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_2: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.6276565Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.6277411Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_2, 2) 2025-12-04T11:24:35.6278203Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:35.6279262Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:35.6280086Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:35.6281114Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] add_tensor: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.add.Tensor(sum_dim_int_list, convert_element_type); sum_dim_int_list = convert_element_type = None 2025-12-04T11:24:35.6281898Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.6282531Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.6282957Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.6283349Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.6283937Z V1204 11:21:41.279000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:35.6284536Z V1204 11:21:41.280000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:35.6285108Z V1204 11:21:41.280000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_3 : [num_users=1] = placeholder[target=primals_3] 2025-12-04T11:24:35.6286149Z V1204 11:21:41.281000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.6286656Z V1204 11:21:41.281000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.6287711Z V1204 11:21:41.283000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.6288215Z V1204 11:21:41.283000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.6289238Z V1204 11:21:41.314000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_2 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_3, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.6289737Z V1204 11:21:41.315000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.6290645Z V1204 11:21:41.317000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.6291130Z V1204 11:21:41.317000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.6292071Z V1204 11:21:41.323000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_2, 2), kwargs = {}) 2025-12-04T11:24:35.6292534Z V1204 11:21:41.323000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.6293417Z V1204 11:21:41.324000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:35.6293876Z V1204 11:21:41.325000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.6294794Z V1204 11:21:41.326000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:35.6295236Z V1204 11:21:41.326000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.6296221Z V1204 11:21:41.329000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:35.6296661Z V1204 11:21:41.329000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.6297632Z V1204 11:21:41.331000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %add_tensor : [num_users=1] = call_function[target=torch.ops.aten.add.Tensor](args = (%sum_dim_int_list, %convert_element_type), kwargs = {}) 2025-12-04T11:24:35.6298172Z V1204 11:21:41.332000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via .inner at 0x7f1a5818ab00> 2025-12-04T11:24:35.6299092Z V1204 11:21:41.334000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.6299543Z V1204 11:21:41.334000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.6300152Z V1204 11:21:41.335000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.6300734Z V1204 11:21:41.340000 100207 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:35.6301546Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:35.6301669Z warnings.warn( 2025-12-04T11:24:35.6302151Z V1204 11:21:41.341000 100207 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:35.6302629Z _ TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True _ 2025-12-04T11:24:35.6302753Z Traceback (most recent call last): 2025-12-04T11:24:35.6303336Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 275, in test_decompose_linear_mixed_precision 2025-12-04T11:24:35.6303463Z self.assertEqual( 2025-12-04T11:24:35.6303925Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:35.6304089Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:35.6304685Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:35.6304893Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:35.6305041Z AssertionError: Scalars are not equal! 2025-12-04T11:24:35.6305048Z 2025-12-04T11:24:35.6305154Z Expected 1 but got 0. 2025-12-04T11:24:35.6305260Z Absolute difference: 1 2025-12-04T11:24:35.6305382Z Relative difference: 1.0 2025-12-04T11:24:35.6305387Z 2025-12-04T11:24:35.6305599Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:35.6306476Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:35.6306494Z 2025-12-04T11:24:35.6306767Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:35.6306992Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:35.6307120Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:35.6307234Z stats [('calls_captured', 1)] 2025-12-04T11:24:35.6308022Z inductor [('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_count', 4), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_addmm', 1)] 2025-12-04T11:24:35.6308255Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:35.6308354Z graph_break [] 2025-12-04T11:24:35.6308582Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:35.6309091Z I1204 11:21:38.249000 100207 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:35.6309688Z I1204 11:21:38.495000 100207 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:35.6310142Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.6310672Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:35.6311627Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.6312737Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_self_modules_linear_parameters_bias_: "f32[2][1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.6313497Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:35.6314256Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_bias_ = L_self_modules_linear_parameters_bias_ 2025-12-04T11:24:35.6314712Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:35.6315146Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.6316130Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.6317517Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, l_self_modules_linear_parameters_bias_); l_x_ = l_self_modules_linear_parameters_weight_ = l_self_modules_linear_parameters_bias_ = None 2025-12-04T11:24:35.6318032Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:35.6318470Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.6318855Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.6319419Z I1204 11:21:40.836000 100207 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:35.6320154Z V1204 11:21:41.215000 100207 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:35.6320870Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw: 2025-12-04T11:24:35.6321547Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4ral453goioq2ytnrbxj5hhqjyfnxjkats7ca6dgxms6usqpldw] gm: GraphModule() 2025-12-04T11:24:35.6321830Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.6322110Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.6322398Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.6322912Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2, primals_3): 2025-12-04T11:24:35.6323749Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.6324606Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.6325438Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_2 = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.6326252Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.6327097Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] addmm = torch.ops.aten.addmm.default(convert_element_type, convert_element_type_2, permute); convert_element_type = None 2025-12-04T11:24:35.6327780Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.6328288Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] return (addmm, convert_element_type_2, permute_1) 2025-12-04T11:24:35.6328599Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.6329171Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:35.6331480Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.6333813Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [di4vpbifhxsheleamv47gds75iwa6xnkz5malaf2qsqkrhojljw] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2]), stride=(1,), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=8, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.6336163Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[2]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.6336810Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:35.6337805Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:35.6338535Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:35.6339358Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:35.6340180Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:35.6340893Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:35.6341620Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:35.6342379Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:35.6343109Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:35.6343889Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [nnqi2dye4qdwjdgeujlxh7d67y2n67kat5ehf3472pn5avlyqbk] fx_kwargs[static_input_idxs]: [0, 1] 2025-12-04T11:24:35.6344579Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inputs_to_check[0]: 2 2025-12-04T11:24:35.6345450Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:35.6346236Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:35.6347393Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:35.6348218Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:35.6377202Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:35.6378238Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:35.6379038Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:35.6379815Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:35.6380688Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:35.6381549Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:35.6382266Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:35.6383049Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:35.6383829Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:35.6384606Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:35.6385509Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:35.6386345Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:35.6387212Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:35.6388055Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:35.6388925Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:35.6389821Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:35.6390688Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:35.6391502Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:35.6392511Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:35.6393318Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:35.6394160Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:35.6394949Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:35.6395786Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:35.6396742Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:35.6397452Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:35.6398249Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:35.6399017Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:35.6399771Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:35.6400614Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:35.6401444Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:35.6402231Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:35.6403005Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:35.6403767Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:35.6404562Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:35.6405340Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:35.6406118Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:35.6406917Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:35.6407709Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:35.6408476Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:35.6409308Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:35.6410135Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:35.6410941Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:35.6411718Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:35.6412576Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:35.6413385Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:35.6414208Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:35.6414995Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:35.6415869Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:35.6416679Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:35.6417504Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:35.6418262Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:35.6419085Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:35.6420006Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:35.6420803Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:35.6421610Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:35.6422441Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:35.6423203Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:35.6424388Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:35.6425238Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:35.6426102Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:35.6427014Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:35.6427842Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:35.6428689Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:35.6429551Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:35.6430429Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:35.6431245Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:35.6432178Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.6433064Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:35.6434021Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.6434822Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:35.6435767Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.6436585Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:35.6437474Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:35.6438231Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:35.6438992Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:35.6439831Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:35.6440678Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:35.6441488Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:35.6442303Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:35.6443622Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:35.6444469Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:35.6445382Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:35.6446305Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:35.6447163Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:35.6447945Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:35.6448769Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:35.6449706Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:35.6450502Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:35.6451287Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:35.6452177Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:35.6453039Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:35.6453918Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:35.6454788Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:35.6455595Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:35.6456390Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:35.6457263Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:35.6458096Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:35.6458915Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:35.6459735Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:35.6460570Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:35.6461532Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:35.6462352Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:35.6463170Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:35.6463906Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:35.6464708Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:35.6465671Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:35.6466590Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:35.6467535Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:35.6468343Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:35.6469174Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:35.6470012Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:35.6470853Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:35.6471757Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:35.6472610Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:35.6473442Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:35.6474252Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:35.6475095Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:35.6475949Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:35.6476738Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:35.6477559Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:35.6478387Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:35.6479180Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:35.6479953Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:35.6480771Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:35.6481622Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:35.6482462Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:35.6483325Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:35.6484098Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:35.6484989Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:35.6485787Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:35.6486670Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:35.6487486Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:35.6488287Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:35.6489054Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:35.6489899Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:35.6490685Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:35.6491545Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:35.6492407Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:35.6493267Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:35.6494107Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:35.6494962Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:35.6495738Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:35.6496777Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:35.6497779Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:35.6498606Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:35.6499399Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:35.6500245Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:35.6501020Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:35.6501805Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:35.6502582Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:35.6503351Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:35.6504116Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:35.6504898Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:35.6505822Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:35.6506666Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:35.6507494Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:35.6508317Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:35.6509080Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:35.6509957Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:35.6510757Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:35.6511623Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:35.6512491Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:35.6513355Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:35.6514196Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:35.6514991Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:35.6515819Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:35.6516644Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:35.6517453Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:35.6518238Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:35.6519125Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:35.6519928Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:35.6520738Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:35.6521645Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:35.6522457Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:35.6523227Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:35.6524083Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:35.6524886Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:35.6525727Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:35.6526546Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:35.6527510Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:35.6528511Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:35.6529536Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:35.6530503Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:35.6531487Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:35.6532491Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:35.6533496Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:35.6534477Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:35.6535466Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:35.6536317Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:35.6537212Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:35.6538082Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:35.6538993Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:35.6539838Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:35.6540647Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:35.6541443Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:35.6542239Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:35.6543046Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:35.6543844Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:35.6544633Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:35.6545433Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:35.6546350Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:35.6547179Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:35.6547996Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:35.6548752Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:35.6549579Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:35.6550355Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:35.6551137Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:35.6551976Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:35.6552829Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:35.6553622Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:35.6554470Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:35.6555215Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:35.6556030Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:35.6556838Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:35.6557704Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:35.6558660Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:35.6559463Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:35.6560198Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:35.6561046Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:35.6561904Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:35.6562767Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:35.6563571Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:35.6564394Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:35.6565332Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:35.6566212Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:35.6567062Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:35.6567848Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:35.6568761Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:35.6569560Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:35.6570386Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:35.6571204Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:35.6572057Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:35.6572934Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:35.6573800Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:35.6574554Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:35.6575357Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:35.6576162Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:35.6576987Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:35.6577823Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:35.6578615Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:35.6579442Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:35.6580242Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:35.6581111Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.6582001Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.6582770Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:35.6583639Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:35.6584475Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:35.6585326Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:35.6586237Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:35.6587134Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:35.6588027Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:35.6588892Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:35.6589730Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:35.6590549Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:35.6591363Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:35.6592196Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:35.6593014Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:35.6593876Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:35.6594721Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:35.6595547Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:35.6596550Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:35.6597336Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:35.6598163Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:35.6599047Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:35.6599918Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:35.6600784Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:35.6601711Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:35.6602670Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:35.6603598Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:35.6604541Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:35.6605375Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:35.6606294Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:35.6607123Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:35.6608058Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:35.6608934Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:35.6609794Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:35.6610651Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:35.6611480Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:35.6612300Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:35.6613151Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:35.6613959Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:35.6614781Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:35.6615621Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:35.6616441Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:35.6617369Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:35.6618305Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:35.6619122Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:35.6619934Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:35.6620860Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:35.6621775Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:35.6622636Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:35.6623492Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:35.6624376Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:35.6625217Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:35.6626115Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:35.6626996Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:35.6627787Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:35.6628588Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:35.6629439Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:35.6630233Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:35.6631025Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:35.6631838Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:35.6632681Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:35.6633644Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:35.6634558Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.6635425Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:35.6636329Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:35.6637191Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:35.6638094Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:35.6638895Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:35.6639885Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:35.6640728Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:35.6641575Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:35.6642397Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:35.6643233Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:35.6644148Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:35.6645041Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:35.6645989Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:35.6646796Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:35.6647652Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:35.6648492Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:35.6649388Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:35.6650348Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:35.6651239Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:35.6652094Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:35.6652968Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:35.6653896Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:35.6654790Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:35.6655663Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:35.6656471Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:35.6657397Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:35.6658253Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:35.6659085Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:35.6660039Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:35.6660896Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:35.6661712Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:35.6662512Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:35.6663405Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:35.6664327Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:35.6665235Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:35.6666131Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:35.6667099Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:35.6667972Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:35.6668863Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:35.6669755Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:35.6670709Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:35.6671568Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:35.6672423Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:35.6673244Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:35.6674083Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:35.6675004Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:35.6675870Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:35.6676746Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:35.6677632Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:35.6678388Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:35.6679148Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:35.6679971Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:35.6680776Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:35.6681602Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:35.6682417Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:35.6683248Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:35.6684129Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:35.6685105Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:35.6686004Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:35.6686850Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:35.6687627Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:35.6688485Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:35.6689323Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:35.6690181Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:35.6691070Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:35.6691926Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:35.6692802Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:35.6693639Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:35.6694460Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:35.6695305Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:35.6696306Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:35.6697266Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:35.6698112Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:35.6698906Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:35.6699854Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:35.6700665Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:35.6701477Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:35.6702251Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:35.6703079Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:35.6703883Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:35.6704752Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:35.6705535Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:35.6706281Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:35.6707184Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:35.6708029Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:35.6708888Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:35.6709760Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:35.6710610Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:35.6711417Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:35.6712212Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:35.6713039Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:35.6713785Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:35.6714565Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:35.6715366Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:35.6716165Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:35.6717113Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:35.6717974Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:35.6718844Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:35.6719618Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:35.6720396Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:35.6721193Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:35.6723161Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:35.6725647Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:35.6726425Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:35.6727343Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:35.6728136Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:35.6728980Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:35.6729935Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:35.6730786Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:35.6731728Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:35.6732643Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:35.6733553Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:35.6734465Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:35.6735399Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:35.6736384Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:35.6737355Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:35.6738190Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:35.6739130Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:35.6740090Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:35.6740988Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:35.6741863Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:35.6742833Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:35.6743593Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:35.6744291Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:35.6745044Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:35.6745758Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:35.6746487Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:35.6747211Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:35.6748108Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:35.6748963Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:35.6749837Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:35.6750572Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:35.6751260Z V1204 11:21:41.225000 100207 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.6751936Z I1204 11:21:41.225000 100207 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.6752419Z V1204 11:21:41.226000 100207 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:35.6752901Z V1204 11:21:41.226000 100207 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:35.6753414Z I1204 11:21:41.226000 100207 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:35.6754271Z V1204 11:21:41.270000 100207 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.addmm.default with input shape: torch.Size([2]), torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:35.6754763Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.6755258Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:35.6756216Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.6757052Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[2][1]cuda:0", primals_3: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.6758031Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.6758998Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2][1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.6759959Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.6760949Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_2: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.6761896Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.6762746Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_2, 2) 2025-12-04T11:24:35.6763538Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:35.6764590Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:35.6765463Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:35.6766485Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] add_tensor: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.add.Tensor(sum_dim_int_list, convert_element_type); sum_dim_int_list = convert_element_type = None 2025-12-04T11:24:35.6767265Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.6767885Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.6768356Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.6768748Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.6769335Z V1204 11:21:41.279000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:35.6769902Z V1204 11:21:41.280000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:35.6770466Z V1204 11:21:41.280000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_3 : [num_users=1] = placeholder[target=primals_3] 2025-12-04T11:24:35.6771485Z V1204 11:21:41.281000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.6771995Z V1204 11:21:41.281000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.6773018Z V1204 11:21:41.283000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.6773523Z V1204 11:21:41.283000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.6774533Z V1204 11:21:41.314000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_2 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_3, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.6775035Z V1204 11:21:41.315000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.6775969Z V1204 11:21:41.317000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.6776420Z V1204 11:21:41.317000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.6777462Z V1204 11:21:41.323000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_2, 2), kwargs = {}) 2025-12-04T11:24:35.6777929Z V1204 11:21:41.323000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.6778851Z V1204 11:21:41.324000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:35.6779318Z V1204 11:21:41.325000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.6780233Z V1204 11:21:41.326000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:35.6780677Z V1204 11:21:41.326000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.6781528Z V1204 11:21:41.329000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:35.6781994Z V1204 11:21:41.329000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.6782912Z V1204 11:21:41.331000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %add_tensor : [num_users=1] = call_function[target=torch.ops.aten.add.Tensor](args = (%sum_dim_int_list, %convert_element_type), kwargs = {}) 2025-12-04T11:24:35.6783450Z V1204 11:21:41.332000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via .inner at 0x7f1a5818ab00> 2025-12-04T11:24:35.6784302Z V1204 11:21:41.334000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.6784751Z V1204 11:21:41.334000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.6785310Z V1204 11:21:41.335000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.6785900Z V1204 11:21:41.340000 100207 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:35.6786643Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:35.6786763Z warnings.warn( 2025-12-04T11:24:35.6787245Z V1204 11:21:41.341000 100207 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:35.6787474Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:35.6787600Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:35.6787717Z stats [('calls_captured', 1)] 2025-12-04T11:24:35.6787958Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:35.6788776Z inductor [('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_count', 4), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_addmm', 1)] 2025-12-04T11:24:35.6788879Z graph_break [] 2025-12-04T11:24:35.6789112Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:35.6789555Z V1204 11:21:41.578000 100207 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpmu524c4r 2025-12-04T11:24:35.6790011Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.6790538Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:35.6791484Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.6792613Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_self_modules_linear_parameters_bias_: "f32[2][1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.6793358Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:35.6794098Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_bias_ = L_self_modules_linear_parameters_bias_ 2025-12-04T11:24:35.6794554Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:35.6795020Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.6796298Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.6797676Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, l_self_modules_linear_parameters_bias_); l_x_ = l_self_modules_linear_parameters_weight_ = l_self_modules_linear_parameters_bias_ = None 2025-12-04T11:24:35.6798166Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:35.6798584Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.6798987Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.6799709Z V1204 11:21:41.894000 100207 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:35.6800434Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw: 2025-12-04T11:24:35.6801098Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4ral453goioq2ytnrbxj5hhqjyfnxjkats7ca6dgxms6usqpldw] gm: GraphModule() 2025-12-04T11:24:35.6801378Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.6801673Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.6801953Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.6802559Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2, primals_3): 2025-12-04T11:24:35.6803378Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.6804192Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.6805066Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_2 = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.6805879Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.6806741Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] addmm = torch.ops.aten.addmm.default(convert_element_type, convert_element_type_2, permute); convert_element_type = None 2025-12-04T11:24:35.6807374Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.6807901Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] return (addmm, convert_element_type_2, permute_1) 2025-12-04T11:24:35.6808198Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.6808772Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:35.6811137Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.6813431Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [di4vpbifhxsheleamv47gds75iwa6xnkz5malaf2qsqkrhojljw] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2]), stride=(1,), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=8, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.6815782Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[2]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.6816438Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:35.6817410Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:35.6818173Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:35.6818999Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:35.6819780Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:35.6820551Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:35.6821269Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:35.6822043Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:35.6822754Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:35.6823503Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [nnqi2dye4qdwjdgeujlxh7d67y2n67kat5ehf3472pn5avlyqbk] fx_kwargs[static_input_idxs]: [0, 1] 2025-12-04T11:24:35.6824192Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inputs_to_check[0]: 2 2025-12-04T11:24:35.6825064Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:35.6825906Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:35.6827086Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:35.6827884Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:35.6856797Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:35.6857903Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:35.6858718Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:35.6859490Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:35.6860341Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:35.6861187Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:35.6861920Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:35.6862699Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:35.6863515Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:35.6864304Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:35.6865157Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:35.6866002Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:35.6866862Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:35.6867722Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:35.6868570Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:35.6869474Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:35.6870344Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:35.6871201Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:35.6872217Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:35.6873003Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:35.6873856Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:35.6874628Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:35.6875465Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:35.6876234Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:35.6876940Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:35.6877746Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:35.6878499Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:35.6879292Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:35.6880125Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:35.6880886Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:35.6881677Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:35.6882455Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:35.6883236Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:35.6884015Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:35.6884803Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:35.6885564Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:35.6886361Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:35.6887186Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:35.6887954Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:35.6888762Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:35.6889570Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:35.6890380Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:35.6891142Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:35.6891952Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:35.6892765Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:35.6893586Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:35.6894420Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:35.6895287Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:35.6896206Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:35.6897017Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:35.6897767Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:35.6898593Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:35.6899518Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:35.6900331Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:35.6901119Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:35.6901962Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:35.6902787Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:35.6903975Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:35.6904823Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:35.6905689Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:35.6906631Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:35.6907446Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:35.6908270Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:35.6909121Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:35.6910019Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:35.6910860Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:35.6911795Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.6912633Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:35.6913593Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.6914408Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:35.6915352Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.6916177Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:35.6917047Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:35.6917808Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:35.6918570Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:35.6919449Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:35.6920306Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:35.6921095Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:35.6921922Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:35.6922738Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:35.6923576Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:35.6924447Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:35.6925376Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:35.6926245Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:35.6927054Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:35.6927893Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:35.6928778Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:35.6929588Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:35.6930360Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:35.6931254Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:35.6932124Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:35.6932991Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:35.6933882Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:35.6934682Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:35.6935528Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:35.6936250Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:35.6937133Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:35.6937959Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:35.6938784Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:35.6939623Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:35.6940450Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:35.6941287Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:35.6942094Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:35.6942869Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:35.6943671Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:35.6944584Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:35.6945524Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:35.6946447Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:35.6947269Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:35.6948090Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:35.6948929Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:35.6949770Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:35.6950676Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:35.6951578Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:35.6952361Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:35.6953212Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:35.6954041Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:35.6954917Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:35.6955700Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:35.6956486Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:35.6957333Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:35.6958114Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:35.6958960Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:35.6959765Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:35.6960597Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:35.6961426Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:35.6962292Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:35.6963088Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:35.6963961Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:35.6964764Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:35.6965629Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:35.6966461Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:35.6967279Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:35.6968038Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:35.6968866Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:35.6969649Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:35.6970525Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:35.6971374Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:35.6972225Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:35.6973058Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:35.6973917Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:35.6974727Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:35.6975626Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:35.6976482Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:35.6977387Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:35.6978203Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:35.6979043Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:35.6979813Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:35.6980594Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:35.6981365Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:35.6982144Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:35.6983014Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:35.6983812Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:35.6984661Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:35.6985534Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:35.6986368Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:35.6987197Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:35.6987973Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:35.6988790Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:35.6989611Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:35.6990456Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:35.6991361Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:35.6992189Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:35.6993018Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:35.6993823Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:35.6994642Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:35.6995480Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:35.6996400Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:35.6997183Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:35.6998057Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:35.6998917Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:35.6999745Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:35.7000602Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:35.7001464Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:35.7002219Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:35.7003082Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:35.7003894Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:35.7004681Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:35.7005513Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:35.7006468Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:35.7007523Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:35.7008498Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:35.7009474Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:35.7010443Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:35.7011445Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:35.7012450Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:35.7013424Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:35.7014421Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:35.7015292Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:35.7016132Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:35.7017025Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:35.7017990Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:35.7018826Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:35.7019656Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:35.7020447Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:35.7021201Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:35.7022017Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:35.7022798Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:35.7023632Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:35.7024413Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:35.7025300Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:35.7026128Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:35.7026946Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:35.7027713Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:35.7028521Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:35.7029305Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:35.7030074Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:35.7030919Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:35.7031803Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:35.7032587Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:35.7033457Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:35.7034186Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:35.7035018Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:35.7035806Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:35.7036636Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:35.7037595Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:35.7038390Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:35.7039169Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:35.7040006Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:35.7040826Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:35.7041681Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:35.7042484Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:35.7043303Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:35.7044236Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:35.7045120Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:35.7045962Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:35.7046764Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:35.7047708Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:35.7048516Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:35.7049314Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:35.7050130Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:35.7050994Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:35.7051851Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:35.7052688Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:35.7053427Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:35.7054246Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:35.7055071Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:35.7055820Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:35.7056617Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:35.7057444Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:35.7058285Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:35.7059071Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:35.7059949Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.7060812Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.7061581Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:35.7062448Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:35.7063322Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:35.7064189Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:35.7065070Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:35.7065974Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:35.7066843Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:35.7067709Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:35.7068518Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:35.7069339Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:35.7070171Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:35.7071021Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:35.7071855Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:35.7072669Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:35.7073515Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:35.7074349Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:35.7075206Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:35.7076005Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:35.7076810Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:35.7077709Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:35.7078565Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:35.7079476Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:35.7080390Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:35.7081336Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:35.7082278Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:35.7083215Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:35.7084067Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:35.7084925Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:35.7085772Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:35.7086691Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:35.7087606Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:35.7088464Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:35.7089281Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:35.7090114Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:35.7090924Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:35.7091784Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:35.7092567Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:35.7093385Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:35.7094222Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:35.7095072Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:35.7102755Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:35.7103724Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:35.7104691Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:35.7105496Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:35.7106468Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:35.7107381Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:35.7108207Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:35.7109079Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:35.7109997Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:35.7110851Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:35.7111702Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:35.7112583Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:35.7113382Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:35.7114193Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:35.7115045Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:35.7115831Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:35.7116641Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:35.7117436Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:35.7118334Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:35.7119298Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:35.7120214Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.7121066Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:35.7121979Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:35.7122857Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:35.7123714Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:35.7124526Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:35.7125503Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:35.7126374Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:35.7127218Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:35.7128005Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:35.7128846Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:35.7129745Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:35.7130653Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:35.7131599Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:35.7132404Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:35.7133248Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:35.7134080Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:35.7134999Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:35.7135907Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:35.7136827Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:35.7137752Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:35.7138645Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:35.7139566Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:35.7140419Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:35.7141292Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:35.7142098Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:35.7142995Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:35.7143832Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:35.7144632Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:35.7145573Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:35.7146436Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:35.7147251Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:35.7148036Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:35.7148921Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:35.7149836Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:35.7150810Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:35.7151681Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:35.7152648Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:35.7153519Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:35.7154415Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:35.7155302Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:35.7156221Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:35.7157089Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:35.7157924Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:35.7158787Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:35.7159610Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:35.7160485Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:35.7161348Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:35.7162225Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:35.7163125Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:35.7163865Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:35.7164632Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:35.7165442Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:35.7166241Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:35.7167097Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:35.7167914Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:35.7168744Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:35.7169628Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:35.7170615Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:35.7171494Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:35.7172306Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:35.7173083Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:35.7173946Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:35.7174817Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:35.7175673Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:35.7176540Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:35.7177446Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:35.7178335Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:35.7179167Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:35.7179987Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:35.7180835Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:35.7181709Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:35.7182650Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:35.7183484Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:35.7184221Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:35.7185185Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:35.7186000Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:35.7186805Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:35.7187574Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:35.7188367Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:35.7189158Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:35.7190031Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:35.7190832Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:35.7191578Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:35.7192418Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:35.7193257Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:35.7194116Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:35.7194996Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:35.7195864Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:35.7196875Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:35.7197670Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:35.7198501Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:35.7199305Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:35.7200075Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:35.7200868Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:35.7201675Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:35.7202537Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:35.7203389Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:35.7204222Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:35.7204995Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:35.7205764Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:35.7206605Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:35.7208583Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:35.7211002Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:35.7211781Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:35.7212683Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:35.7213494Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:35.7214324Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:35.7215324Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:35.7216166Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:35.7217180Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:35.7218099Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:35.7219014Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:35.7219922Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:35.7220818Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:35.7221814Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:35.7222693Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:35.7223558Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:35.7224486Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:35.7225403Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:35.7226293Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:35.7227173Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:35.7228149Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:35.7228884Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:35.7229596Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:35.7230339Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:35.7231082Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:35.7231815Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:35.7232536Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:35.7233443Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:35.7234288Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:35.7235176Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:35.7235898Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:35.7236542Z V1204 11:21:41.904000 100207 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.7237221Z I1204 11:21:41.904000 100207 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.7237703Z V1204 11:21:41.905000 100207 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:35.7238300Z V1204 11:21:41.905000 100207 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:35.7238803Z I1204 11:21:41.905000 100207 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:35.7239667Z V1204 11:21:41.920000 100207 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.addmm.default with input shape: torch.Size([2]), torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:35.7240123Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.7240621Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:35.7241590Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.7242418Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[2][1]cuda:0", primals_3: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.7243407Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.7244359Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2][1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.7245327Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.7246352Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_2: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.7247256Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.7248127Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_2, 2) 2025-12-04T11:24:35.7248918Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:35.7249973Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:35.7250796Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:35.7251807Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] add_tensor: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.add.Tensor(sum_dim_int_list, convert_element_type); sum_dim_int_list = convert_element_type = None 2025-12-04T11:24:35.7252621Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.7253245Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.7253670Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.7254061Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.7254640Z V1204 11:21:41.930000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:35.7255206Z V1204 11:21:41.931000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:35.7255784Z V1204 11:21:41.931000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_3 : [num_users=1] = placeholder[target=primals_3] 2025-12-04T11:24:35.7256799Z V1204 11:21:41.932000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.7257370Z V1204 11:21:41.932000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.7258384Z V1204 11:21:41.933000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.7258891Z V1204 11:21:41.933000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.7259945Z V1204 11:21:41.934000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_2 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_3, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.7260446Z V1204 11:21:41.934000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.7261387Z V1204 11:21:41.936000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.7261834Z V1204 11:21:41.936000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.7262837Z V1204 11:21:41.937000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_2, 2), kwargs = {}) 2025-12-04T11:24:35.7263290Z V1204 11:21:41.938000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.7264170Z V1204 11:21:41.939000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:35.7264633Z V1204 11:21:41.939000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.7265543Z V1204 11:21:41.940000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:35.7266018Z V1204 11:21:41.940000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.7266864Z V1204 11:21:41.942000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:35.7267313Z V1204 11:21:41.942000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.7268211Z V1204 11:21:41.944000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %add_tensor : [num_users=1] = call_function[target=torch.ops.aten.add.Tensor](args = (%sum_dim_int_list, %convert_element_type), kwargs = {}) 2025-12-04T11:24:35.7268746Z V1204 11:21:41.944000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via .inner at 0x7f1a5818ab00> 2025-12-04T11:24:35.7269600Z V1204 11:21:41.946000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.7270047Z V1204 11:21:41.947000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.7270612Z V1204 11:21:41.947000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.7271192Z V1204 11:21:41.950000 100207 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:35.7271943Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:35.7272049Z warnings.warn( 2025-12-04T11:24:35.7272535Z V1204 11:21:41.950000 100207 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:35.7272702Z =================================== FAILURES =================================== 2025-12-04T11:24:35.7273199Z _ TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True _ 2025-12-04T11:24:35.7273325Z Traceback (most recent call last): 2025-12-04T11:24:35.7273912Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 275, in test_decompose_linear_mixed_precision 2025-12-04T11:24:35.7274019Z self.assertEqual( 2025-12-04T11:24:35.7274523Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:35.7274692Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:35.7275224Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:35.7275471Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:35.7275602Z AssertionError: Scalars are not equal! 2025-12-04T11:24:35.7275612Z 2025-12-04T11:24:35.7275733Z Expected 1 but got 0. 2025-12-04T11:24:35.7275841Z Absolute difference: 1 2025-12-04T11:24:35.7275953Z Relative difference: 1.0 2025-12-04T11:24:35.7275959Z 2025-12-04T11:24:35.7276188Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:35.7277061Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:35.7277067Z 2025-12-04T11:24:35.7277340Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:35.7277577Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:35.7277687Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:35.7277846Z stats [('calls_captured', 1)] 2025-12-04T11:24:35.7278629Z inductor [('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_count', 4), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_addmm', 1)] 2025-12-04T11:24:35.7278846Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:35.7278949Z graph_break [] 2025-12-04T11:24:35.7279163Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:35.7279669Z I1204 11:21:38.249000 100207 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:35.7280262Z I1204 11:21:38.495000 100207 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:35.7280697Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.7281201Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:35.7282135Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.7283230Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_self_modules_linear_parameters_bias_: "f32[2][1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.7283975Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:35.7284711Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_bias_ = L_self_modules_linear_parameters_bias_ 2025-12-04T11:24:35.7285192Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:35.7285613Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.7286598Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.7288002Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, l_self_modules_linear_parameters_bias_); l_x_ = l_self_modules_linear_parameters_weight_ = l_self_modules_linear_parameters_bias_ = None 2025-12-04T11:24:35.7288520Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:35.7288942Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.7289335Z V1204 11:21:38.497000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.7289892Z I1204 11:21:40.836000 100207 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:35.7290603Z V1204 11:21:41.215000 100207 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:35.7291323Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw: 2025-12-04T11:24:35.7292017Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4ral453goioq2ytnrbxj5hhqjyfnxjkats7ca6dgxms6usqpldw] gm: GraphModule() 2025-12-04T11:24:35.7292311Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.7292585Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.7292855Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.7293377Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2, primals_3): 2025-12-04T11:24:35.7294190Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.7295018Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.7295828Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_2 = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.7296724Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.7297645Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] addmm = torch.ops.aten.addmm.default(convert_element_type, convert_element_type_2, permute); convert_element_type = None 2025-12-04T11:24:35.7298282Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.7298806Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] return (addmm, convert_element_type_2, permute_1) 2025-12-04T11:24:35.7299166Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.7299750Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:35.7302087Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.7304436Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [di4vpbifhxsheleamv47gds75iwa6xnkz5malaf2qsqkrhojljw] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2]), stride=(1,), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=8, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.7306754Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[2]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.7307457Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:35.7308370Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:35.7309100Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:35.7309918Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:35.7310714Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:35.7311431Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:35.7312146Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:35.7312877Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:35.7313589Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:35.7314349Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [nnqi2dye4qdwjdgeujlxh7d67y2n67kat5ehf3472pn5avlyqbk] fx_kwargs[static_input_idxs]: [0, 1] 2025-12-04T11:24:35.7315056Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inputs_to_check[0]: 2 2025-12-04T11:24:35.7315929Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:35.7316723Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:35.7317943Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:35.7318740Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:35.7347750Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:35.7377987Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:35.7379885Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:35.7381563Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:35.7383313Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:35.7385175Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:35.7386889Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:35.7388527Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:35.7390258Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:35.7391960Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:35.7393761Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:35.7395590Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:35.7397559Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:35.7399429Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:35.7401278Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:35.7403222Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:35.7405123Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:35.7406939Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:35.7408912Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:35.7410835Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:35.7412571Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:35.7414305Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:35.7416037Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:35.7417839Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:35.7419441Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:35.7421151Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:35.7422857Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:35.7424548Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:35.7426293Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:35.7428102Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:35.7429788Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:35.7431483Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:35.7433175Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:35.7434888Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:35.7436613Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:35.7438305Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:35.7440010Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:35.7441723Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:35.7443422Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:35.7445099Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:35.7446815Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:35.7448542Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:35.7450230Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:35.7451958Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:35.7453752Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:35.7455530Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:35.7457343Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:35.7459196Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:35.7460991Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:35.7462711Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:35.7464363Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:35.7466071Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:35.7467940Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:35.7469795Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:35.7471560Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:35.7473320Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:35.7475041Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:35.7477124Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:35.7479287Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:35.7481083Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:35.7482976Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:35.7484832Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:35.7486614Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:35.7488500Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:35.7490357Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:35.7492179Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:35.7494095Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.7496199Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:35.7498201Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.7500096Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:35.7501981Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.7503880Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:35.7505810Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:35.7507584Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:35.7509221Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:35.7510960Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:35.7512791Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:35.7514552Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:35.7516269Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:35.7518022Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:35.7519802Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:35.7521646Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:35.7523636Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:35.7525560Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:35.7527368Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:35.7529111Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:35.7531013Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:35.7532839Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:35.7534570Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:35.7536365Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:35.7538329Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:35.7540260Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:35.7542148Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:35.7543968Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:35.7545693Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:35.7547363Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:35.7549018Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:35.7550745Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:35.7552488Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:35.7554254Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:35.7556060Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:35.7557919Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:35.7559697Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:35.7561429Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:35.7563090Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:35.7564948Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:35.7566965Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:35.7568967Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:35.7570851Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:35.7572608Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:35.7574448Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:35.7576256Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:35.7578219Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:35.7580119Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:35.7581890Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:35.7583641Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:35.7585430Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:35.7587241Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:35.7588990Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:35.7590692Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:35.7592506Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:35.7594275Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:35.7596167Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:35.7597902Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:35.7599701Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:35.7601508Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:35.7603345Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:35.7605125Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:35.7606915Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:35.7608771Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:35.7610558Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:35.7612377Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:35.7614126Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:35.7615806Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:35.7617541Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:35.7619253Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:35.7621001Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:35.7622810Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:35.7624623Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:35.7626486Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:35.7628314Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:35.7630063Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:35.7631894Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:35.7633764Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:35.7635605Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:35.7637361Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:35.7639128Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:35.7640860Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:35.7642527Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:35.7644248Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:35.7645929Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:35.7647587Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:35.7649262Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:35.7651020Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:35.7652860Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:35.7654672Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:35.7656418Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:35.7658183Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:35.7659897Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:35.7661702Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:35.7663495Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:35.7665393Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:35.7667223Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:35.7669039Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:35.7670808Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:35.7672552Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:35.7674343Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:35.7676106Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:35.7677858Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:35.7679658Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:35.7681461Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:35.7683220Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:35.7685038Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:35.7686837Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:35.7688551Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:35.7690268Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:35.7692023Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:35.7693754Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:35.7695543Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:35.7697690Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:35.7699849Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:35.7701975Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:35.7704179Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:35.7706247Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:35.7708359Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:35.7710507Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:35.7712672Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:35.7714779Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:35.7716756Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:35.7718551Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:35.7720330Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:35.7722200Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:35.7724093Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:35.7725846Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:35.7727540Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:35.7729236Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:35.7730998Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:35.7732732Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:35.7734465Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:35.7736197Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:35.7738120Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:35.7739973Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:35.7741758Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:35.7743461Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:35.7745171Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:35.7746950Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:35.7748645Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:35.7750399Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:35.7752204Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:35.7753981Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:35.7755743Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:35.7757442Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:35.7759101Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:35.7760810Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:35.7762570Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:35.7764534Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:35.7766421Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:35.7768127Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:35.7769829Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:35.7771650Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:35.7773472Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:35.7775282Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:35.7777084Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:35.7778961Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:35.7780966Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:35.7782824Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:35.7784597Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:35.7786446Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:35.7788288Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:35.7790014Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:35.7791746Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:35.7793529Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:35.7795355Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:35.7797336Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:35.7799153Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:35.7800843Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:35.7802635Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:35.7804337Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:35.7806010Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:35.7807772Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:35.7809520Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:35.7811296Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:35.7813092Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.7814959Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.7816814Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:35.7818654Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:35.7820489Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:35.7822319Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:35.7824164Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:35.7826056Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:35.7827929Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:35.7829783Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:35.7831595Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:35.7833390Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:35.7835180Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:35.7836966Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:35.7838793Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:35.7840577Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:35.7842404Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:35.7844213Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:35.7846025Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:35.7847812Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:35.7849591Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:35.7851416Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:35.7853306Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:35.7855184Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:35.7857175Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:35.7859129Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:35.7861075Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:35.7863034Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:35.7864918Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:35.7866755Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:35.7868627Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:35.7870515Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:35.7872513Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:35.7874386Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:35.7876230Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:35.7878016Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:35.7879784Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:35.7881585Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:35.7883419Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:35.7885321Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:35.7887111Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:35.7888890Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:35.7890714Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:35.7892597Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:35.7894454Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:35.7896358Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:35.7898252Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:35.7900192Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:35.7902063Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:35.7904007Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:35.7905887Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:35.7907770Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:35.7909613Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:35.7911553Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:35.7913365Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:35.7915093Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:35.7916863Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:35.7918636Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:35.7920417Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:35.7922152Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:35.7923955Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:35.7925880Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:35.7927862Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.7929732Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:35.7931607Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:35.7933490Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:35.7935332Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:35.7937233Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:35.7939161Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:35.7941143Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:35.7942956Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:35.7944734Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:35.7946498Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:35.7948361Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:35.7950298Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:35.7952272Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:35.7954217Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:35.7955992Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:35.7957819Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:35.7959692Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:35.7961639Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:35.7963575Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:35.7965428Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:35.7967254Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:35.7969141Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:35.7971046Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:35.7972934Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:35.7974723Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:35.7976558Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:35.7978451Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:35.7980268Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:35.7982145Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:35.7984073Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:35.7985880Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:35.7987621Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:35.7989469Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:35.7991417Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:35.7993381Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:35.7995307Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:35.7997384Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:35.8002195Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:35.8004058Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:35.8005959Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:35.8007904Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:35.8009913Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:35.8011798Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:35.8013666Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:35.8015463Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:35.8017453Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:35.8019331Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:35.8021200Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:35.8023116Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:35.8024887Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:35.8026532Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:35.8028248Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:35.8029993Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:35.8031746Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:35.8033514Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:35.8035262Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:35.8037160Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:35.8039129Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:35.8041109Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:35.8042949Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:35.8044710Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:35.8046485Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:35.8048347Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:35.8050180Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:35.8052077Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:35.8053930Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:35.8055797Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:35.8057695Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:35.8059488Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:35.8061297Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:35.8063156Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:35.8065082Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:35.8066944Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:35.8068635Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:35.8070448Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:35.8072406Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:35.8074131Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:35.8075801Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:35.8077505Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:35.8079256Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:35.8081071Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:35.8082884Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:35.8084534Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:35.8086288Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:35.8088108Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:35.8089959Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:35.8091835Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:35.8093694Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:35.8095490Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:35.8097495Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:35.8099255Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:35.8100963Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:35.8102602Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:35.8104277Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:35.8106043Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:35.8107776Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:35.8109589Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:35.8111395Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:35.8113177Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:35.8114856Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:35.8116559Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:35.8119496Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:35.8124093Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:35.8127430Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:35.8129226Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:35.8131082Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:35.8132852Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:35.8134788Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:35.8136725Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:35.8138666Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:35.8140666Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:35.8142588Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:35.8144511Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:35.8146448Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:35.8148513Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:35.8150515Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:35.8152398Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:35.8154296Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:35.8156928Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:35.8158877Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:35.8160764Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:35.8162753Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:35.8164612Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:35.8166195Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:35.8167780Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:35.8169366Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:35.8170939Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:35.8172524Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:35.8174270Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:35.8176187Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:35.8178101Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:35.8179810Z V1204 11:21:41.224000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:35.8181321Z V1204 11:21:41.225000 100207 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.8182822Z I1204 11:21:41.225000 100207 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.8184106Z V1204 11:21:41.226000 100207 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:35.8185216Z V1204 11:21:41.226000 100207 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:35.8186374Z I1204 11:21:41.226000 100207 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:35.8187885Z V1204 11:21:41.270000 100207 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.addmm.default with input shape: torch.Size([2]), torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:35.8189359Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.8190432Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:35.8192024Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.8193935Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[2][1]cuda:0", primals_3: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.8195873Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.8198097Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2][1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.8200140Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.8202230Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_2: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.8204257Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.8206151Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_2, 2) 2025-12-04T11:24:35.8207961Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:35.8209907Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:35.8211899Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:35.8213917Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] add_tensor: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.add.Tensor(sum_dim_int_list, convert_element_type); sum_dim_int_list = convert_element_type = None 2025-12-04T11:24:35.8215894Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.8217496Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.8218731Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.8219700Z V1204 11:21:41.276000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.8220793Z V1204 11:21:41.279000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:35.8222066Z V1204 11:21:41.280000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:35.8223349Z V1204 11:21:41.280000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_3 : [num_users=1] = placeholder[target=primals_3] 2025-12-04T11:24:35.8225073Z V1204 11:21:41.281000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.8226726Z V1204 11:21:41.281000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.8228390Z V1204 11:21:41.283000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.8230047Z V1204 11:21:41.283000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.8231706Z V1204 11:21:41.314000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_2 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_3, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.8233352Z V1204 11:21:41.315000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.8234889Z V1204 11:21:41.317000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.8236420Z V1204 11:21:41.317000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.8237964Z V1204 11:21:41.323000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_2, 2), kwargs = {}) 2025-12-04T11:24:35.8239504Z V1204 11:21:41.323000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.8240991Z V1204 11:21:41.324000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:35.8242563Z V1204 11:21:41.325000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.8244088Z V1204 11:21:41.326000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:35.8245584Z V1204 11:21:41.326000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.8247062Z V1204 11:21:41.329000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:35.8248497Z V1204 11:21:41.329000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.8250005Z V1204 11:21:41.331000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %add_tensor : [num_users=1] = call_function[target=torch.ops.aten.add.Tensor](args = (%sum_dim_int_list, %convert_element_type), kwargs = {}) 2025-12-04T11:24:35.8251594Z V1204 11:21:41.332000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via .inner at 0x7f1a5818ab00> 2025-12-04T11:24:35.8253125Z V1204 11:21:41.334000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.8254565Z V1204 11:21:41.334000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.8255710Z V1204 11:21:41.335000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.8257037Z V1204 11:21:41.340000 100207 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:35.8258502Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:35.8259486Z warnings.warn( 2025-12-04T11:24:35.8260131Z V1204 11:21:41.341000 100207 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:35.8260988Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:35.8261480Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:35.8261820Z stats [('calls_captured', 1)] 2025-12-04T11:24:35.8262247Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:35.8263400Z inductor [('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_count', 4), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_addmm', 1)] 2025-12-04T11:24:35.8264413Z graph_break [] 2025-12-04T11:24:35.8264784Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:35.8265640Z V1204 11:21:41.578000 100207 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpmu524c4r 2025-12-04T11:24:35.8266662Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.8267742Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:35.8269303Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.8271509Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_self_modules_linear_parameters_bias_: "f32[2][1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.8273478Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:35.8274214Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_bias_ = L_self_modules_linear_parameters_bias_ 2025-12-04T11:24:35.8274697Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:35.8275119Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.8276135Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.8277517Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, l_self_modules_linear_parameters_bias_); l_x_ = l_self_modules_linear_parameters_weight_ = l_self_modules_linear_parameters_bias_ = None 2025-12-04T11:24:35.8278005Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:35.8278425Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.8278823Z V1204 11:21:41.607000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.8279545Z V1204 11:21:41.894000 100207 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:35.8280255Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw: 2025-12-04T11:24:35.8280929Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4ral453goioq2ytnrbxj5hhqjyfnxjkats7ca6dgxms6usqpldw] gm: GraphModule() 2025-12-04T11:24:35.8281211Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.8281499Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.8281771Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.8282284Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2, primals_3): 2025-12-04T11:24:35.8283109Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.8283959Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.8284782Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_2 = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.8285544Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.8286430Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] addmm = torch.ops.aten.addmm.default(convert_element_type, convert_element_type_2, permute); convert_element_type = None 2025-12-04T11:24:35.8287069Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.8287573Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] return (addmm, convert_element_type_2, permute_1) 2025-12-04T11:24:35.8287909Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.8288482Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:35.8290790Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.8293112Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [di4vpbifhxsheleamv47gds75iwa6xnkz5malaf2qsqkrhojljw] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2]), stride=(1,), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=8, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.8295456Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[2]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.8296259Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:35.8297239Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:35.8297970Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:35.8298803Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:35.8299649Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:35.8300386Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:35.8301098Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:35.8301818Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:35.8302596Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:35.8303357Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [nnqi2dye4qdwjdgeujlxh7d67y2n67kat5ehf3472pn5avlyqbk] fx_kwargs[static_input_idxs]: [0, 1] 2025-12-04T11:24:35.8304048Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inputs_to_check[0]: 2 2025-12-04T11:24:35.8304956Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:35.8305752Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:35.8307003Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:35.8307783Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:35.8336700Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:35.8337828Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:35.8338610Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:35.8339380Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:35.8340281Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:35.8341125Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:35.8341878Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:35.8342663Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:35.8343477Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:35.8344261Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:35.8345118Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:35.8345966Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:35.8346822Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:35.8347673Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:35.8348529Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:35.8349425Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:35.8350292Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:35.8351107Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:35.8352132Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:35.8352925Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:35.8353768Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:35.8354526Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:35.8355364Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:35.8356168Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:35.8356875Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:35.8357705Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:35.8358456Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:35.8359242Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:35.8360083Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:35.8360849Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:35.8361637Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:35.8362412Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:35.8363194Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:35.8363974Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:35.8364747Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:35.8365516Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:35.8366310Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:35.8367098Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:35.8367905Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:35.8368682Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:35.8369494Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:35.8370260Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:35.8371070Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:35.8371879Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:35.8372695Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:35.8373560Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:35.8374361Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:35.8375260Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:35.8376032Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:35.8376798Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:35.8377608Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:35.8378427Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:35.8379350Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:35.8380157Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:35.8380945Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:35.8381772Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:35.8382530Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:35.8383717Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:35.8384567Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:35.8385435Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:35.8386333Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:35.8387185Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:35.8388013Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:35.8388858Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:35.8389761Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:35.8390569Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:35.8391539Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.8392380Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:35.8393336Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.8394170Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:35.8395110Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.8395928Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:35.8396966Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:35.8397730Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:35.8398494Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:35.8399333Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:35.8400271Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:35.8401032Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:35.8401865Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:35.8402654Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:35.8403540Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:35.8404418Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:35.8405390Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:35.8406265Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:35.8407077Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:35.8407916Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:35.8408799Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:35.8409606Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:35.8410388Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:35.8411282Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:35.8412160Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:35.8413026Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:35.8413904Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:35.8414698Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:35.8415496Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:35.8416268Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:35.8417109Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:35.8417934Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:35.8418726Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:35.8419611Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:35.8420440Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:35.8421290Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:35.8422115Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:35.8422884Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:35.8423693Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:35.8424611Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:35.8425547Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:35.8426474Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:35.8427299Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:35.8428120Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:35.8428953Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:35.8429801Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:35.8430704Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:35.8431572Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:35.8432409Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:35.8433231Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:35.8434060Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:35.8434887Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:35.8435701Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:35.8436492Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:35.8437367Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:35.8438147Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:35.8438968Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:35.8439777Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:35.8440593Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:35.8441435Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:35.8442291Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:35.8443081Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:35.8443954Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:35.8444747Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:35.8445614Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:35.8446421Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:35.8447222Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:35.8448010Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:35.8448799Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:35.8449580Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:35.8450413Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:35.8451282Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:35.8452117Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:35.8452948Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:35.8453833Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:35.8454598Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:35.8455542Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:35.8456391Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:35.8457255Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:35.8458050Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:35.8458895Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:35.8459666Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:35.8460445Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:35.8461215Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:35.8461990Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:35.8462735Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:35.8463519Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:35.8464420Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:35.8465301Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:35.8466263Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:35.8467056Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:35.8467872Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:35.8468693Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:35.8469535Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:35.8470386Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:35.8471298Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:35.8472129Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:35.8472962Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:35.8473772Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:35.8474590Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:35.8475426Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:35.8476215Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:35.8476995Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:35.8477882Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:35.8478680Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:35.8479510Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:35.8480412Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:35.8481236Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:35.8481992Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:35.8482811Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:35.8483645Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:35.8484440Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:35.8485299Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:35.8486257Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:35.8487307Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:35.8488284Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:35.8489246Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:35.8490234Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:35.8491235Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:35.8492248Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:35.8493224Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:35.8494235Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:35.8495067Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:35.8495900Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:35.8496947Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:35.8497865Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:35.8498715Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:35.8499497Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:35.8500361Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:35.8501120Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:35.8501979Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:35.8502767Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:35.8503600Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:35.8504400Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:35.8505285Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:35.8506115Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:35.8506934Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:35.8507707Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:35.8508515Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:35.8509289Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:35.8510072Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:35.8510906Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:35.8511757Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:35.8512587Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:35.8513417Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:35.8514150Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:35.8514933Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:35.8515764Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:35.8516592Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:35.8517588Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:35.8518380Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:35.8519158Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:35.8519994Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:35.8520816Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:35.8521671Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:35.8522473Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:35.8523293Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:35.8524241Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:35.8525133Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:35.8525973Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:35.8526779Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:35.8527682Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:35.8528573Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:35.8529362Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:35.8530178Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:35.8531012Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:35.8531903Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:35.8532745Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:35.8533485Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:35.8534320Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:35.8542588Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:35.8543569Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:35.8544385Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:35.8545170Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:35.8546005Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:35.8546810Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:35.8547684Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.8548573Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.8549343Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:35.8550214Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:35.8551056Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:35.8551910Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:35.8552827Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:35.8553722Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:35.8554572Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:35.8555480Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:35.8556296Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:35.8557121Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:35.8557976Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:35.8558807Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:35.8559668Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:35.8560499Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:35.8561346Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:35.8562182Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:35.8563035Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:35.8563844Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:35.8564652Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:35.8565541Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:35.8566410Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:35.8567284Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:35.8568245Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:35.8569147Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:35.8570082Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:35.8570985Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:35.8571852Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:35.8572725Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:35.8573580Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:35.8574518Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:35.8575426Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:35.8576283Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:35.8577190Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:35.8578020Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:35.8578837Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:35.8579698Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:35.8580503Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:35.8581327Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:35.8582166Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:35.8582980Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:35.8583854Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:35.8584765Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:35.8585578Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:35.8586388Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:35.8587437Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:35.8588402Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:35.8589232Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:35.8590134Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:35.8591002Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:35.8591872Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:35.8592739Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:35.8593620Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:35.8594418Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:35.8595221Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:35.8596239Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:35.8597036Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:35.8597832Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:35.8598645Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:35.8599487Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:35.8600443Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:35.8601396Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.8602256Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:35.8603133Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:35.8604056Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:35.8604918Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:35.8605720Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:35.8606748Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:35.8607579Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:35.8608464Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:35.8609255Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:35.8610097Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:35.8611000Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:35.8611893Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:35.8612842Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:35.8613645Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:35.8614494Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:35.8615332Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:35.8616227Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:35.8617232Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:35.8618114Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:35.8618967Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:35.8619809Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:35.8620767Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:35.8621636Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:35.8622539Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:35.8623340Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:35.8624243Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:35.8625077Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:35.8625869Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:35.8626825Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:35.8627676Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:35.8628498Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:35.8629290Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:35.8630177Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:35.8631094Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:35.8632000Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:35.8632889Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:35.8633849Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:35.8634720Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:35.8635583Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:35.8636503Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:35.8637422Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:35.8638324Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:35.8639165Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:35.8640014Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:35.8640854Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:35.8641740Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:35.8642602Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:35.8643483Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:35.8644382Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:35.8645127Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:35.8645884Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:35.8646704Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:35.8647501Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:35.8648332Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:35.8649190Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:35.8649992Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:35.8650871Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:35.8651819Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:35.8652755Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:35.8653563Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:35.8654364Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:35.8655226Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:35.8656096Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:35.8657027Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:35.8657886Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:35.8658743Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:35.8659621Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:35.8660460Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:35.8661280Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:35.8662121Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:35.8662990Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:35.8663883Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:35.8664717Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:35.8665470Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:35.8666421Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:35.8667238Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:35.8668016Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:35.8668819Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:35.8669613Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:35.8670426Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:35.8671290Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:35.8672100Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:35.8672855Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:35.8673698Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:35.8674542Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:35.8675402Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:35.8676283Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:35.8677137Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:35.8677938Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:35.8678725Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:35.8679547Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:35.8680293Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:35.8681100Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:35.8681854Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:35.8682642Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:35.8683462Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:35.8684340Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:35.8685176Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:35.8685947Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:35.8686747Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:35.8687546Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:35.8689556Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:35.8691984Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:35.8692767Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:35.8693679Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:35.8694484Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:35.8695312Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:35.8696419Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:35.8697338Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:35.8698334Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:35.8699253Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:35.8700134Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:35.8701088Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:35.8701993Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:35.8703021Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:35.8703917Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:35.8704809Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:35.8705748Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:35.8706674Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:35.8707567Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:35.8708440Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:35.8709428Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:35.8710169Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:35.8710870Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:35.8711628Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:35.8712343Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:35.8713076Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:35.8713830Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:35.8714718Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:35.8715560Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:35.8716412Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:35.8717176Z V1204 11:21:41.903000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:35.8717832Z V1204 11:21:41.904000 100207 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.8718516Z I1204 11:21:41.904000 100207 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.8719024Z V1204 11:21:41.905000 100207 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:35.8719508Z V1204 11:21:41.905000 100207 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:35.8720055Z I1204 11:21:41.905000 100207 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:35.8720914Z V1204 11:21:41.920000 100207 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.addmm.default with input shape: torch.Size([2]), torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:35.8721376Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.8721875Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:35.8722840Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.8723665Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[2][1]cuda:0", primals_3: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.8724647Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.8725625Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2][1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.8726587Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.8727590Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_2: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.8728522Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.8729382Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_2, 2) 2025-12-04T11:24:35.8730142Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:35.8731227Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:35.8732052Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:35.8733101Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] add_tensor: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.add.Tensor(sum_dim_int_list, convert_element_type); sum_dim_int_list = convert_element_type = None 2025-12-04T11:24:35.8733884Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.8734543Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.8734971Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.8735357Z V1204 11:21:41.926000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.8735941Z V1204 11:21:41.930000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:35.8736503Z V1204 11:21:41.931000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:35.8737136Z V1204 11:21:41.931000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_3 : [num_users=1] = placeholder[target=primals_3] 2025-12-04T11:24:35.8738152Z V1204 11:21:41.932000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.8738655Z V1204 11:21:41.932000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.8739682Z V1204 11:21:41.933000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.8740179Z V1204 11:21:41.933000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.8741200Z V1204 11:21:41.934000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_2 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_3, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.8741732Z V1204 11:21:41.934000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.8742638Z V1204 11:21:41.936000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.8743088Z V1204 11:21:41.936000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.8744029Z V1204 11:21:41.937000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_2, 2), kwargs = {}) 2025-12-04T11:24:35.8744524Z V1204 11:21:41.938000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.8745405Z V1204 11:21:41.939000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:35.8745867Z V1204 11:21:41.939000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.8746802Z V1204 11:21:41.940000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:35.8747246Z V1204 11:21:41.940000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.8748132Z V1204 11:21:41.942000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:35.8748570Z V1204 11:21:41.942000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.8749479Z V1204 11:21:41.944000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %add_tensor : [num_users=1] = call_function[target=torch.ops.aten.add.Tensor](args = (%sum_dim_int_list, %convert_element_type), kwargs = {}) 2025-12-04T11:24:35.8750020Z V1204 11:21:41.944000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via .inner at 0x7f1a5818ab00> 2025-12-04T11:24:35.8750874Z V1204 11:21:41.946000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.8751327Z V1204 11:21:41.947000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.8751893Z V1204 11:21:41.947000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.8752473Z V1204 11:21:41.950000 100207 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:35.8753213Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:35.8753327Z warnings.warn( 2025-12-04T11:24:35.8753810Z V1204 11:21:41.950000 100207 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:35.8754046Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:35.8754159Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:35.8754275Z stats [('calls_captured', 1)] 2025-12-04T11:24:35.8754512Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:35.8755329Z inductor [('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_count', 4), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_addmm', 1)] 2025-12-04T11:24:35.8755427Z graph_break [] 2025-12-04T11:24:35.8755660Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:35.8756101Z V1204 11:21:41.973000 100207 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpjw1kct_u 2025-12-04T11:24:35.8756550Z V1204 11:21:42.000000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.8757046Z V1204 11:21:42.000000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:35.8758019Z V1204 11:21:42.000000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.8759117Z V1204 11:21:42.000000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_self_modules_linear_parameters_bias_: "f32[2][1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.8759902Z V1204 11:21:42.000000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:35.8760638Z V1204 11:21:42.000000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_bias_ = L_self_modules_linear_parameters_bias_ 2025-12-04T11:24:35.8761126Z V1204 11:21:42.000000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:35.8761560Z V1204 11:21:42.000000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.8762537Z V1204 11:21:42.000000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.8763922Z V1204 11:21:42.000000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, l_self_modules_linear_parameters_bias_); l_x_ = l_self_modules_linear_parameters_weight_ = l_self_modules_linear_parameters_bias_ = None 2025-12-04T11:24:35.8764399Z V1204 11:21:42.000000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:35.8764821Z V1204 11:21:42.000000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.8765221Z V1204 11:21:42.000000 100207 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.8765931Z V1204 11:21:42.266000 100207 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:35.8766655Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw: 2025-12-04T11:24:35.8767313Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4ral453goioq2ytnrbxj5hhqjyfnxjkats7ca6dgxms6usqpldw] gm: GraphModule() 2025-12-04T11:24:35.8767593Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.8767883Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.8768188Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.8768707Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2, primals_3): 2025-12-04T11:24:35.8769522Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.8770331Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.8771155Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_2 = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.8771942Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.8772807Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] addmm = torch.ops.aten.addmm.default(convert_element_type, convert_element_type_2, permute); convert_element_type = None 2025-12-04T11:24:35.8773464Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.8773982Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] return (addmm, convert_element_type_2, permute_1) 2025-12-04T11:24:35.8774374Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.8774953Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:35.8777390Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.8779689Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [di4vpbifhxsheleamv47gds75iwa6xnkz5malaf2qsqkrhojljw] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2]), stride=(1,), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=8, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.8782025Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[2]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.8782683Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:35.8783597Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:35.8784371Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:35.8785184Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:35.8785967Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:35.8786695Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:35.8787438Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:35.8788178Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:35.8788891Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:35.8789672Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [nnqi2dye4qdwjdgeujlxh7d67y2n67kat5ehf3472pn5avlyqbk] fx_kwargs[static_input_idxs]: [0, 1] 2025-12-04T11:24:35.8790359Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inputs_to_check[0]: 2 2025-12-04T11:24:35.8791270Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:35.8792070Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:35.8793266Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:35.8794069Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:35.8823240Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:35.8824329Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:35.8825167Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:35.8825945Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:35.8826831Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:35.8827678Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:35.8828451Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:35.8829221Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:35.8830001Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:35.8830789Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:35.8831642Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:35.8832492Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:35.8833355Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:35.8834211Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:35.8835060Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:35.8835950Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:35.8836819Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:35.8837666Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:35.8838687Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:35.8839447Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:35.8840303Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:35.8841092Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:35.8841935Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:35.8842735Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:35.8843436Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:35.8844273Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:35.8845024Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:35.8845793Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:35.8846632Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:35.8847392Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:35.8848182Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:35.8848956Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:35.8849735Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:35.8850519Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:35.8851305Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:35.8852071Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:35.8852901Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:35.8853690Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:35.8854453Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:35.8855235Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:35.8856087Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:35.8856924Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:35.8857686Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:35.8858534Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:35.8859347Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:35.8860201Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:35.8861002Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:35.8861877Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:35.8862663Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:35.8863418Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:35.8864184Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:35.8864990Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:35.8865908Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:35.8866710Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:35.8867503Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:35.8868348Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:35.8869134Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:35.8870338Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:35.8871138Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:35.8872038Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:35.8872940Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:35.8873761Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:35.8874621Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:35.8875464Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:35.8876395Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:35.8877197Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:35.8878137Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.8878975Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:35.8879936Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.8880759Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:35.8881696Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.8882525Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:35.8883402Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:35.8884181Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:35.8884974Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:35.8885807Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:35.8886668Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:35.8887426Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:35.8888286Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:35.8889073Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:35.8889921Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:35.8890808Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:35.8891730Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:35.8892628Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:35.8893399Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:35.8894245Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:35.8895122Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:35.8895936Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:35.8896929Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:35.8897828Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:35.8898702Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:35.8899566Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:35.8900452Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:35.8901332Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:35.8902135Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:35.8902849Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:35.8903628Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:35.8904492Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:35.8905284Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:35.8906125Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:35.8906994Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:35.8907824Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:35.8908669Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:35.8909403Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:35.8910207Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:35.8911119Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:35.8912052Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:35.8912982Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:35.8913805Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:35.8914631Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:35.8915478Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:35.8916314Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:35.8917250Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:35.8918111Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:35.8918896Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:35.8919714Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:35.8920575Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:35.8921419Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:35.8922222Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:35.8923017Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:35.8923890Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:35.8924676Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:35.8925462Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:35.8926261Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:35.8927087Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:35.8927919Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:35.8928785Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:35.8929577Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:35.8930450Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:35.8931247Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:35.8932119Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:35.8932974Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:35.8933757Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:35.8934519Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:35.8935307Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:35.8936118Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:35.8937008Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:35.8937858Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:35.8938741Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:35.8939567Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:35.8940482Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:35.8941252Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:35.8942146Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:35.8943000Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:35.8943814Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:35.8944624Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:35.8945456Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:35.8946242Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:35.8947006Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:35.8947778Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:35.8948567Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:35.8949348Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:35.8950145Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:35.8950992Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:35.8951852Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:35.8952692Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:35.8953482Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:35.8954283Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:35.8955099Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:35.8955952Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:35.8956794Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:35.8957682Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:35.8958497Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:35.8959325Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:35.8960130Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:35.8960945Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:35.8961785Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:35.8962578Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:35.8963371Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:35.8964238Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:35.8965087Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:35.8965909Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:35.8966770Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:35.8967598Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:35.8968384Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:35.8969218Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:35.8970044Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:35.8970832Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:35.8971699Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:35.8972655Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:35.8973673Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:35.8974650Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:35.8975631Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:35.8976611Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:35.8977694Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:35.8978693Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:35.8979663Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:35.8980672Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:35.8981543Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:35.8982381Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:35.8983189Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:35.8984112Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:35.8984975Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:35.8985757Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:35.8986583Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:35.8987339Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:35.8988183Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:35.8988968Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:35.8989775Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:35.8990562Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:35.8991449Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:35.8992290Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:35.8993105Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:35.8993879Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:35.8994693Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:35.8995476Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:35.8996418Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:35.8997326Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:35.8998179Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:35.8998967Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:35.8999803Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:35.9000584Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:35.9001380Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:35.9002170Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:35.9003044Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:35.9003986Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:35.9004825Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:35.9005572Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:35.9006413Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:35.9007233Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:35.9008087Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:35.9008907Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:35.9009716Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:35.9010660Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:35.9011550Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:35.9012397Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:35.9013232Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:35.9014133Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:35.9014945Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:35.9015727Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:35.9016576Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:35.9017491Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:35.9018359Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:35.9019248Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:35.9019991Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:35.9020844Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:35.9021642Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:35.9022396Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:35.9023195Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:35.9023970Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:35.9024816Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:35.9025599Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:35.9026480Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.9027355Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.9028129Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:35.9028998Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:35.9029866Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:35.9030735Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:35.9031594Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:35.9032535Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:35.9033377Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:35.9034257Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:35.9035091Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:35.9035918Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:35.9036785Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:35.9037600Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:35.9038445Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:35.9039265Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:35.9040122Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:35.9040934Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:35.9041795Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:35.9042598Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:35.9043404Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:35.9044314Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:35.9045172Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:35.9046087Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:35.9046997Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:35.9047893Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:35.9048857Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:35.9049762Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:35.9050725Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:35.9051581Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:35.9052453Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:35.9053374Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:35.9054268Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:35.9055113Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:35.9055932Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:35.9056775Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:35.9057641Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:35.9058510Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:35.9059294Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:35.9060125Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:35.9060954Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:35.9061818Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:35.9062689Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:35.9063564Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:35.9064391Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:35.9065222Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:35.9066141Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:35.9067066Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:35.9067895Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:35.9068800Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:35.9069668Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:35.9070522Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:35.9071381Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:35.9072276Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:35.9073067Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:35.9073871Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:35.9074722Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:35.9075511Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:35.9076321Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:35.9077126Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:35.9078020Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:35.9078966Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:35.9079860Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.9080705Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:35.9081626Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:35.9082501Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:35.9083385Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:35.9084203Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:35.9085207Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:35.9086053Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:35.9086883Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:35.9087666Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:35.9088504Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:35.9089410Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:35.9090311Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:35.9091246Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:35.9092054Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:35.9092887Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:35.9093757Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:35.9094647Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:35.9095561Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:35.9096603Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:35.9097576Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:35.9098438Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:35.9099386Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:35.9100264Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:35.9101166Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:35.9101973Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:35.9102846Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:35.9103678Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:35.9104486Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:35.9105425Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:35.9106295Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:35.9107099Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:35.9107894Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:35.9108786Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:35.9109707Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:35.9110676Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:35.9111554Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:35.9112497Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:35.9113413Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:35.9114291Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:35.9115171Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:35.9116126Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:35.9117000Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:35.9117873Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:35.9118701Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:35.9119523Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:35.9120422Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:35.9121275Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:35.9122158Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:35.9123058Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:35.9123801Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:35.9124577Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:35.9125390Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:35.9126232Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:35.9127045Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:35.9127858Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:35.9128661Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:35.9129570Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:35.9130534Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:35.9131447Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:35.9132265Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:35.9133063Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:35.9133937Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:35.9134764Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:35.9135623Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:35.9136483Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:35.9137389Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:35.9138282Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:35.9139108Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:35.9139936Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:35.9140763Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:35.9141647Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:35.9142599Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:35.9143430Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:35.9144176Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:35.9145106Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:35.9145965Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:35.9146733Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:35.9147542Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:35.9148332Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:35.9149154Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:35.9150034Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:35.9150810Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:35.9151574Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:35.9152401Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:35.9153249Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:35.9154109Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:35.9154985Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:35.9155858Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:35.9156652Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:35.9157470Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:35.9158319Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:35.9159066Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:35.9159840Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:35.9160595Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:35.9161426Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:35.9162242Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:35.9163101Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:35.9163950Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:35.9164736Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:35.9165542Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:35.9166336Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:35.9168321Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:35.9170743Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:35.9171525Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:35.9172434Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:35.9173243Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:35.9174080Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:35.9175088Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:35.9175930Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:35.9176912Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:35.9177859Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:35.9178750Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:35.9179692Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:35.9180597Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:35.9181630Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:35.9182515Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:35.9183360Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:35.9184283Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:35.9185214Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:35.9186103Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:35.9186981Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:35.9187968Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:35.9188707Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:35.9189426Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:35.9190169Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:35.9190933Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:35.9191649Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:35.9192374Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:35.9193266Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:35.9194141Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:35.9195011Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:35.9195754Z V1204 11:21:42.275000 100207 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:35.9196572Z V1204 11:21:42.276000 100207 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.9197243Z I1204 11:21:42.276000 100207 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.9197806Z V1204 11:21:42.277000 100207 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:35.9198305Z V1204 11:21:42.277000 100207 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:35.9198809Z I1204 11:21:42.277000 100207 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 2 2025-12-04T11:24:35.9199689Z V1204 11:21:42.292000 100207 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.addmm.default with input shape: torch.Size([2]), torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:35.9200134Z V1204 11:21:42.298000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.9200629Z V1204 11:21:42.298000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:35.9201593Z V1204 11:21:42.298000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.9202415Z V1204 11:21:42.298000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[2][1]cuda:0", primals_3: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.9203412Z V1204 11:21:42.298000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.9204362Z V1204 11:21:42.298000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2][1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.9205337Z V1204 11:21:42.298000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.9206362Z V1204 11:21:42.298000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_2: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.9207275Z V1204 11:21:42.298000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.9208148Z V1204 11:21:42.298000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_2, 2) 2025-12-04T11:24:35.9208925Z V1204 11:21:42.298000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:35.9210019Z V1204 11:21:42.298000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:35.9210846Z V1204 11:21:42.298000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:35.9211889Z V1204 11:21:42.298000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] add_tensor: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.add.Tensor(sum_dim_int_list, convert_element_type); sum_dim_int_list = convert_element_type = None 2025-12-04T11:24:35.9212669Z V1204 11:21:42.298000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.9213300Z V1204 11:21:42.298000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.9213725Z V1204 11:21:42.298000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.9214123Z V1204 11:21:42.298000 100207 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.9214704Z V1204 11:21:42.302000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:35.9215271Z V1204 11:21:42.303000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:35.9215846Z V1204 11:21:42.303000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_3 : [num_users=1] = placeholder[target=primals_3] 2025-12-04T11:24:35.9216912Z V1204 11:21:42.304000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.9217432Z V1204 11:21:42.304000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.9218448Z V1204 11:21:42.305000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.9219032Z V1204 11:21:42.305000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.9220035Z V1204 11:21:42.306000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_2 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_3, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.9220541Z V1204 11:21:42.306000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.9221448Z V1204 11:21:42.308000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.9221934Z V1204 11:21:42.308000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.9222891Z V1204 11:21:42.309000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_2, 2), kwargs = {}) 2025-12-04T11:24:35.9223340Z V1204 11:21:42.310000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.9224259Z V1204 11:21:42.311000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:35.9224708Z V1204 11:21:42.311000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.9225652Z V1204 11:21:42.312000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:35.9226098Z V1204 11:21:42.312000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.9226948Z V1204 11:21:42.314000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:35.9227394Z V1204 11:21:42.314000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.9228294Z V1204 11:21:42.316000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %add_tensor : [num_users=1] = call_function[target=torch.ops.aten.add.Tensor](args = (%sum_dim_int_list, %convert_element_type), kwargs = {}) 2025-12-04T11:24:35.9228846Z V1204 11:21:42.316000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via .inner at 0x7f1a5818ab00> 2025-12-04T11:24:35.9229685Z V1204 11:21:42.318000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.9230133Z V1204 11:21:42.319000 100207 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.9230700Z V1204 11:21:42.319000 100207 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.9231275Z V1204 11:21:42.322000 100207 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 2 2025-12-04T11:24:35.9232026Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:35.9232130Z warnings.warn( 2025-12-04T11:24:35.9232646Z V1204 11:21:42.322000 100207 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:35.9233502Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-b7f552c9fcec58a4.xml - 2025-12-04T11:24:35.9233677Z =========================== short test summary info ============================ 2025-12-04T11:24:35.9234607Z FAILED [0.3689s] inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True - AssertionError: Scalars are not equal! 2025-12-04T11:24:35.9234618Z 2025-12-04T11:24:35.9234728Z Expected 1 but got 0. 2025-12-04T11:24:35.9234835Z Absolute difference: 1 2025-12-04T11:24:35.9234957Z Relative difference: 1.0 2025-12-04T11:24:35.9234963Z 2025-12-04T11:24:35.9235181Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:35.9236090Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:35.9236098Z 2025-12-04T11:24:35.9236369Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:35.9236550Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:24:35.9236792Z ================== 1 failed, 36 deselected, 2 rerun in 6.82s =================== 2025-12-04T11:24:35.9236900Z --- Logging error --- 2025-12-04T11:24:35.9237035Z Traceback (most recent call last): 2025-12-04T11:24:35.9237345Z File "/opt/conda/envs/py_3.10/lib/python3.10/logging/__init__.py", line 1103, in emit 2025-12-04T11:24:35.9237504Z stream.write(msg + self.terminator) 2025-12-04T11:24:35.9237650Z ValueError: I/O operation on closed file. 2025-12-04T11:24:35.9237747Z Call stack: 2025-12-04T11:24:35.9238266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/remote_cache.py", line 432, in dump_cache_stats 2025-12-04T11:24:35.9238430Z log.info("Cache Metrics:%s", out.getvalue()) 2025-12-04T11:24:35.9238542Z Message: 'Cache Metrics:%s' 2025-12-04T11:24:35.9238659Z Arguments: (' None\n',) 2025-12-04T11:24:35.9238754Z Got exit code 1 2025-12-04T11:24:35.9238859Z Retrying single test... 2025-12-04T11:24:35.9239544Z Test results will be stored in test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-ba19ebfa7960f6c4.xml 2025-12-04T11:24:35.9239706Z ============================= test session starts ============================== 2025-12-04T11:24:35.9240055Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:24:35.9240176Z cachedir: .pytest_cache 2025-12-04T11:24:35.9240701Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:24:35.9240838Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:24:35.9240946Z configfile: pytest.ini 2025-12-04T11:24:35.9241542Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:24:35.9242751Z collecting ... /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:52: PytestCollectionWarning: cannot collect test class 'TestDecomposeAddMM' because it has a __init__ constructor (from: test/inductor/test_decompose_mem_bound_mm.py) 2025-12-04T11:24:35.9242901Z class TestDecomposeAddMM(torch.nn.Module): 2025-12-04T11:24:35.9243052Z collected 37 items / 36 deselected / 1 selected 2025-12-04T11:24:35.9244018Z stepcurrent: skipping 15 already run items. Running only test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:35.9244133Z Running 1 items in this shard 2025-12-04T11:24:35.9244170Z 2025-12-04T11:24:35.9245008Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True ('RERUN', {'yellow': True}) [5.9772s] [100%] 2025-12-04T11:24:35.9245821Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True ('RERUN', {'yellow': True}) [0.3938s] [100%] 2025-12-04T11:24:35.9246571Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True FAILED [0.3889s] [100%] 2025-12-04T11:24:35.9246577Z 2025-12-04T11:24:35.9246721Z ==================================== RERUNS ==================================== 2025-12-04T11:24:35.9247182Z _ TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True _ 2025-12-04T11:24:35.9247346Z Traceback (most recent call last): 2025-12-04T11:24:35.9247928Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 275, in test_decompose_linear_mixed_precision 2025-12-04T11:24:35.9248049Z self.assertEqual( 2025-12-04T11:24:35.9248505Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:35.9248672Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:35.9249274Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:35.9249481Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:35.9249645Z AssertionError: Scalars are not equal! 2025-12-04T11:24:35.9249663Z 2025-12-04T11:24:35.9249767Z Expected 1 but got 0. 2025-12-04T11:24:35.9249876Z Absolute difference: 1 2025-12-04T11:24:35.9249999Z Relative difference: 1.0 2025-12-04T11:24:35.9250004Z 2025-12-04T11:24:35.9250222Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:35.9251088Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:35.9251094Z 2025-12-04T11:24:35.9251374Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:35.9251594Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:35.9251717Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:35.9251831Z stats [('calls_captured', 1)] 2025-12-04T11:24:35.9252614Z inductor [('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_count', 4), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_addmm', 1)] 2025-12-04T11:24:35.9252849Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:35.9252952Z graph_break [] 2025-12-04T11:24:35.9253171Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:35.9253681Z I1204 11:21:58.159000 100404 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:35.9254276Z I1204 11:21:58.404000 100404 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:35.9254730Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.9255218Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:35.9256177Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.9257356Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_self_modules_linear_parameters_bias_: "f32[2][1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.9258111Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:35.9258845Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_bias_ = L_self_modules_linear_parameters_bias_ 2025-12-04T11:24:35.9259303Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:35.9259769Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.9260748Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.9262166Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, l_self_modules_linear_parameters_bias_); l_x_ = l_self_modules_linear_parameters_weight_ = l_self_modules_linear_parameters_bias_ = None 2025-12-04T11:24:35.9262643Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:35.9263094Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.9263491Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.9264048Z I1204 11:22:00.743000 100404 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:35.9264777Z V1204 11:22:01.120000 100404 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:35.9265484Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw: 2025-12-04T11:24:35.9266153Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4ral453goioq2ytnrbxj5hhqjyfnxjkats7ca6dgxms6usqpldw] gm: GraphModule() 2025-12-04T11:24:35.9266437Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.9266712Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.9267001Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.9267509Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2, primals_3): 2025-12-04T11:24:35.9268339Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.9269157Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.9269970Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_2 = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.9270777Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.9271619Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] addmm = torch.ops.aten.addmm.default(convert_element_type, convert_element_type_2, permute); convert_element_type = None 2025-12-04T11:24:35.9272262Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.9272768Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] return (addmm, convert_element_type_2, permute_1) 2025-12-04T11:24:35.9273078Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.9273674Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:35.9276003Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.9278316Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [di4vpbifhxsheleamv47gds75iwa6xnkz5malaf2qsqkrhojljw] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2]), stride=(1,), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=8, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.9280684Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[2]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.9281334Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:35.9282264Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:35.9282990Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:35.9283804Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:35.9284599Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:35.9285314Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:35.9286046Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:35.9286802Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:35.9287536Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:35.9288292Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [nnqi2dye4qdwjdgeujlxh7d67y2n67kat5ehf3472pn5avlyqbk] fx_kwargs[static_input_idxs]: [0, 1] 2025-12-04T11:24:35.9288970Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inputs_to_check[0]: 2 2025-12-04T11:24:35.9289962Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:35.9290753Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:35.9291969Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:35.9292747Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:35.9321915Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:35.9322943Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:35.9323788Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:35.9324581Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:35.9325421Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:35.9326277Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:35.9327058Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:35.9327849Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:35.9328637Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:35.9329457Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:35.9330329Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:35.9331210Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:35.9332089Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:35.9332941Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:35.9333800Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:35.9334708Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:35.9335563Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:35.9336394Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:35.9337456Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:35.9338232Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:35.9339077Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:35.9339880Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:35.9340722Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:35.9341488Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:35.9342207Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:35.9343037Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:35.9343797Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:35.9344546Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:35.9345415Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:35.9346173Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:35.9346983Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:35.9347766Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:35.9348527Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:35.9349319Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:35.9350091Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:35.9350865Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:35.9351661Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:35.9352431Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:35.9353204Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:35.9353965Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:35.9354789Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:35.9355588Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:35.9356360Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:35.9357170Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:35.9357973Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:35.9358844Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:35.9359631Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:35.9360538Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:35.9361309Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:35.9362077Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:35.9362860Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:35.9363662Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:35.9364599Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:35.9365389Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:35.9366194Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:35.9367029Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:35.9367794Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:35.9368974Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:35.9369785Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:35.9370654Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:35.9371562Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:35.9372385Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:35.9373191Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:35.9374048Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:35.9374959Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:35.9375783Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:35.9376742Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.9377649Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:35.9378671Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.9379474Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:35.9380431Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.9381251Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:35.9382139Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:35.9382909Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:35.9383659Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:35.9384513Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:35.9385365Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:35.9386139Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:35.9386955Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:35.9387797Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:35.9388634Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:35.9389499Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:35.9390438Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:35.9391326Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:35.9392106Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:35.9392956Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:35.9393844Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:35.9394671Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:35.9395461Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:35.9396496Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:35.9397366Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:35.9398248Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:35.9399117Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:35.9399926Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:35.9400721Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:35.9401449Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:35.9402235Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:35.9403043Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:35.9403896Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:35.9404728Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:35.9405565Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:35.9406386Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:35.9407243Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:35.9407983Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:35.9408815Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:35.9409737Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:35.9410707Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:35.9411645Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:35.9412449Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:35.9413277Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:35.9414109Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:35.9414943Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:35.9415859Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:35.9416711Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:35.9417569Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:35.9418378Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:35.9419222Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:35.9420084Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:35.9420871Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:35.9421656Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:35.9422489Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:35.9423316Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:35.9424092Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:35.9424937Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:35.9425755Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:35.9426628Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:35.9427518Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:35.9428322Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:35.9429246Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:35.9430056Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:35.9430942Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:35.9431810Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:35.9432624Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:35.9433385Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:35.9434168Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:35.9434964Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:35.9435830Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:35.9436690Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:35.9437523Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:35.9438358Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:35.9439247Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:35.9440005Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:35.9440941Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:35.9441781Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:35.9442653Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:35.9443454Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:35.9444305Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:35.9445084Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:35.9445852Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:35.9446636Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:35.9447405Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:35.9448188Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:35.9448967Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:35.9449838Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:35.9450693Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:35.9451546Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:35.9452354Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:35.9453121Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:35.9453945Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:35.9454779Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:35.9455639Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:35.9456513Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:35.9457423Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:35.9458279Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:35.9459110Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:35.9459947Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:35.9460774Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:35.9461586Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:35.9462375Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:35.9463268Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:35.9464073Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:35.9464891Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:35.9465765Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:35.9466582Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:35.9467352Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:35.9468205Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:35.9469015Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:35.9469809Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:35.9470634Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:35.9471640Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:35.9472653Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:35.9473677Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:35.9474645Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:35.9475669Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:35.9476674Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:35.9477694Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:35.9478669Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:35.9479665Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:35.9480523Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:35.9481351Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:35.9482182Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:35.9483096Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:35.9483949Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:35.9484791Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:35.9485570Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:35.9486347Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:35.9487186Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:35.9487993Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:35.9488780Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:35.9489621Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:35.9490509Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:35.9491364Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:35.9492190Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:35.9492951Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:35.9493772Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:35.9494552Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:35.9495338Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:35.9496322Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:35.9497215Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:35.9498025Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:35.9498853Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:35.9499602Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:35.9500486Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:35.9501297Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:35.9502125Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:35.9503079Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:35.9504495Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:35.9505247Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:35.9506165Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:35.9506985Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:35.9507909Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:35.9508722Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:35.9509533Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:35.9510491Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:35.9511365Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:35.9512246Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:35.9513041Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:35.9513966Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:35.9514763Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:35.9515556Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:35.9516376Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:35.9517243Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:35.9518132Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:35.9518955Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:35.9519712Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:35.9520559Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:35.9521370Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:35.9522154Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:35.9522951Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:35.9523773Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:35.9524608Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:35.9525406Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:35.9526280Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.9527166Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.9527946Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:35.9528807Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:35.9529662Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:35.9530517Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:35.9531388Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:35.9532294Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:35.9533177Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:35.9534045Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:35.9534840Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:35.9535674Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:35.9536524Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:35.9537440Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:35.9538303Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:35.9539128Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:35.9540004Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:35.9540840Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:35.9541728Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:35.9542555Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:35.9543366Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:35.9544279Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:35.9545143Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:35.9546011Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:35.9546936Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:35.9547841Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:35.9548777Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:35.9549725Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:35.9550573Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:35.9551422Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:35.9552303Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:35.9553227Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:35.9554098Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:35.9554988Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:35.9555814Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:35.9556685Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:35.9557499Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:35.9558365Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:35.9559146Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:35.9559968Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:35.9560805Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:35.9561626Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:35.9562502Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:35.9563379Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:35.9564207Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:35.9565013Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:35.9566030Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:35.9566927Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:35.9567761Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:35.9568679Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:35.9569554Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:35.9570412Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:35.9571293Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:35.9572192Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:35.9573007Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:35.9573814Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:35.9574668Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:35.9575459Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:35.9576275Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:35.9577133Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:35.9578006Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:35.9578963Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:35.9579842Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:35.9580700Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:35.9581621Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:35.9582497Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:35.9583355Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:35.9584173Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:35.9585183Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:35.9586037Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:35.9586897Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:35.9587686Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:35.9588556Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:35.9589465Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:35.9590371Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:35.9591314Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:35.9592125Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:35.9592966Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:35.9593799Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:35.9594693Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:35.9595610Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:35.9596672Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:35.9597515Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:35.9598436Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:35.9599353Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:35.9600214Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:35.9601145Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:35.9601963Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:35.9602828Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:35.9603701Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:35.9604518Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:35.9605501Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:35.9606376Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:35.9607178Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:35.9607974Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:35.9608875Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:35.9609801Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:35.9610731Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:35.9611610Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:35.9612563Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:35.9613431Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:35.9614325Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:35.9615225Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:35.9616149Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:35.9617075Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:35.9617957Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:35.9618795Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:35.9619650Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:35.9620533Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:35.9621419Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:35.9622304Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:35.9623208Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:35.9623956Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:35.9624728Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:35.9625547Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:35.9626369Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:35.9627186Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:35.9628009Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:35.9628810Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:35.9629695Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:35.9630703Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:35.9631593Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:35.9632409Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:35.9633172Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:35.9634060Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:35.9634908Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:35.9635796Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:35.9636664Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:35.9637536Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:35.9638426Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:35.9639245Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:35.9640059Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:35.9640906Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:35.9641778Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:35.9642667Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:35.9643493Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:35.9644230Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:35.9645162Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:35.9645994Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:35.9646785Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:35.9647560Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:35.9648354Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:35.9649151Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:35.9650057Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:35.9650830Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:35.9651628Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:35.9652460Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:35.9653332Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:35.9654197Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:35.9655075Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:35.9655945Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:35.9656739Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:35.9657621Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:35.9658442Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:35.9659189Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:35.9659964Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:35.9660722Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:35.9661531Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:35.9662381Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:35.9663249Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:35.9664069Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:35.9664839Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:35.9665642Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:35.9666433Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:35.9668449Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:35.9670902Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:35.9671685Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:35.9672599Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:35.9673404Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:35.9674230Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:35.9675201Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:35.9676043Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:35.9676963Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:35.9677868Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:35.9678785Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:35.9679708Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:35.9680615Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:35.9681654Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:35.9682539Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:35.9683382Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:35.9684331Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:35.9685252Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:35.9686195Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:35.9687068Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:35.9688048Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:35.9688788Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:35.9689504Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:35.9690254Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:35.9690970Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:35.9691703Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:35.9692420Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:35.9693306Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:35.9694146Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:35.9695032Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:35.9695755Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:35.9696532Z V1204 11:22:01.131000 100404 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.9697291Z I1204 11:22:01.131000 100404 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:35.9697850Z V1204 11:22:01.131000 100404 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:35.9698352Z V1204 11:22:01.131000 100404 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:35.9698858Z I1204 11:22:01.132000 100404 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:35.9699777Z V1204 11:22:01.176000 100404 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.addmm.default with input shape: torch.Size([2]), torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:35.9700229Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.9700771Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:35.9701744Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.9702564Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[2][1]cuda:0", primals_3: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.9703561Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.9704510Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2][1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.9705488Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.9706479Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_2: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.9707403Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.9708248Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_2, 2) 2025-12-04T11:24:35.9709053Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:35.9710105Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:35.9710937Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:35.9711991Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] add_tensor: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.add.Tensor(sum_dim_int_list, convert_element_type); sum_dim_int_list = convert_element_type = None 2025-12-04T11:24:35.9712777Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.9713438Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.9713864Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.9714299Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:35.9714871Z V1204 11:22:01.186000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:35.9715440Z V1204 11:22:01.186000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:35.9716019Z V1204 11:22:01.186000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_3 : [num_users=1] = placeholder[target=primals_3] 2025-12-04T11:24:35.9717032Z V1204 11:22:01.187000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.9717552Z V1204 11:22:01.187000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.9718564Z V1204 11:22:01.189000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.9719082Z V1204 11:22:01.189000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.9720097Z V1204 11:22:01.221000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_2 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_3, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:35.9720601Z V1204 11:22:01.221000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.9721511Z V1204 11:22:01.223000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.9721972Z V1204 11:22:01.223000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.9722958Z V1204 11:22:01.229000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_2, 2), kwargs = {}) 2025-12-04T11:24:35.9723414Z V1204 11:22:01.230000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.9724309Z V1204 11:22:01.231000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:35.9724763Z V1204 11:22:01.231000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.9725705Z V1204 11:22:01.232000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:35.9726152Z V1204 11:22:01.233000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.9727026Z V1204 11:22:01.235000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:35.9727475Z V1204 11:22:01.236000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.9728373Z V1204 11:22:01.238000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %add_tensor : [num_users=1] = call_function[target=torch.ops.aten.add.Tensor](args = (%sum_dim_int_list, %convert_element_type), kwargs = {}) 2025-12-04T11:24:35.9728962Z V1204 11:22:01.238000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via .inner at 0x7f2e9019ab00> 2025-12-04T11:24:35.9729803Z V1204 11:22:01.240000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:35.9730254Z V1204 11:22:01.241000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:35.9730825Z V1204 11:22:01.241000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:35.9731403Z V1204 11:22:01.246000 100404 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:35.9732155Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:35.9732260Z warnings.warn( 2025-12-04T11:24:35.9732745Z V1204 11:22:01.247000 100404 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:35.9733224Z _ TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True _ 2025-12-04T11:24:35.9733351Z Traceback (most recent call last): 2025-12-04T11:24:35.9733947Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 275, in test_decompose_linear_mixed_precision 2025-12-04T11:24:35.9734055Z self.assertEqual( 2025-12-04T11:24:35.9734513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:35.9734693Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:35.9735230Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:35.9735451Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:35.9735643Z AssertionError: Scalars are not equal! 2025-12-04T11:24:35.9735650Z 2025-12-04T11:24:35.9735758Z Expected 1 but got 0. 2025-12-04T11:24:35.9735883Z Absolute difference: 1 2025-12-04T11:24:35.9735996Z Relative difference: 1.0 2025-12-04T11:24:35.9736001Z 2025-12-04T11:24:35.9736219Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:35.9737174Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:35.9737182Z 2025-12-04T11:24:35.9737456Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:35.9737697Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:35.9737811Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:35.9737964Z stats [('calls_captured', 1)] 2025-12-04T11:24:35.9738776Z inductor [('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_count', 4), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_addmm', 1)] 2025-12-04T11:24:35.9738999Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:35.9739100Z graph_break [] 2025-12-04T11:24:35.9739336Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:35.9739882Z I1204 11:21:58.159000 100404 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:35.9740491Z I1204 11:21:58.404000 100404 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:35.9740964Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:35.9741459Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:35.9742425Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:35.9743510Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_self_modules_linear_parameters_bias_: "f32[2][1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:35.9744275Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:35.9745006Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_bias_ = L_self_modules_linear_parameters_bias_ 2025-12-04T11:24:35.9745479Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:35.9745900Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.9746896Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:35.9748271Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, l_self_modules_linear_parameters_bias_); l_x_ = l_self_modules_linear_parameters_weight_ = l_self_modules_linear_parameters_bias_ = None 2025-12-04T11:24:35.9748784Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:35.9749223Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.9749609Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:35.9750183Z I1204 11:22:00.743000 100404 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:35.9750900Z V1204 11:22:01.120000 100404 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:35.9751657Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw: 2025-12-04T11:24:35.9752324Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4ral453goioq2ytnrbxj5hhqjyfnxjkats7ca6dgxms6usqpldw] gm: GraphModule() 2025-12-04T11:24:35.9752607Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.9752895Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.9753197Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.9753712Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2, primals_3): 2025-12-04T11:24:35.9754575Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:35.9755394Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:35.9756227Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_2 = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:35.9756996Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:35.9757852Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] addmm = torch.ops.aten.addmm.default(convert_element_type, convert_element_type_2, permute); convert_element_type = None 2025-12-04T11:24:35.9758488Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:35.9759012Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] return (addmm, convert_element_type_2, permute_1) 2025-12-04T11:24:35.9759307Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:35.9759877Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:35.9762211Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.9764536Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [di4vpbifhxsheleamv47gds75iwa6xnkz5malaf2qsqkrhojljw] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2]), stride=(1,), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=8, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.9766903Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[2]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:35.9767563Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:35.9768511Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:35.9769233Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:35.9770059Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:35.9770877Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:35.9771648Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:35.9772364Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:35.9773100Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:35.9773813Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:35.9774569Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [nnqi2dye4qdwjdgeujlxh7d67y2n67kat5ehf3472pn5avlyqbk] fx_kwargs[static_input_idxs]: [0, 1] 2025-12-04T11:24:35.9775258Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inputs_to_check[0]: 2 2025-12-04T11:24:35.9776125Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:35.9776978Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:35.9778138Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:35.9778951Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:35.9808139Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:35.9809216Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:35.9810002Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:35.9810769Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:35.9811626Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:35.9812468Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:35.9813182Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:35.9813973Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:35.9814758Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:35.9815554Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:35.9816450Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:35.9817358Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:35.9818226Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:35.9819073Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:35.9819991Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:35.9820891Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:35.9821789Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:35.9822606Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:35.9823728Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:35.9824493Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:35.9825348Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:35.9826096Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:35.9826939Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:35.9827716Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:35.9828428Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:35.9829244Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:35.9830002Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:35.9830765Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:35.9831606Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:35.9832374Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:35.9833197Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:35.9833973Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:35.9834752Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:35.9835567Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:35.9836361Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:35.9837123Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:35.9837945Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:35.9838744Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:35.9839543Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:35.9840331Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:35.9841150Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:35.9841925Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:35.9842683Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:35.9843494Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:35.9844310Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:35.9845132Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:35.9845933Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:35.9846795Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:35.9847575Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:35.9848364Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:35.9849114Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:35.9849936Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:35.9850855Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:35.9851685Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:35.9852478Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:35.9853351Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:35.9854100Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:35.9855322Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:35.9856138Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:35.9857054Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:35.9857958Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:35.9858770Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:35.9859591Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:35.9860439Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:35.9861314Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:35.9862130Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:35.9863065Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.9863908Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:35.9864912Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.9865728Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:35.9866669Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:35.9867537Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:35.9868413Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:35.9869177Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:35.9869969Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:35.9870809Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:35.9871706Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:35.9872466Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:35.9873292Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:35.9874085Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:35.9874917Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:35.9875788Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:35.9876713Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:35.9877578Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:35.9878348Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:35.9879187Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:35.9880061Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:35.9880912Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:35.9881706Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:35.9882597Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:35.9883505Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:35.9884373Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:35.9885254Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:35.9886079Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:35.9886889Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:35.9887644Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:35.9888432Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:35.9889256Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:35.9890044Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:35.9890889Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:35.9891716Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:35.9892534Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:35.9900401Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:35.9901172Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:35.9901990Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:35.9902910Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:35.9903989Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:35.9904924Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:35.9905733Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:35.9906627Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:35.9907471Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:35.9908313Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:35.9909259Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:35.9910127Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:35.9910967Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:35.9911780Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:35.9912627Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:35.9913458Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:35.9914247Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:35.9915043Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:35.9915885Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:35.9916667Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:35.9917456Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:35.9918262Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:35.9919085Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:35.9919969Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:35.9920833Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:35.9921617Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:35.9922498Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:35.9923334Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:35.9924206Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:35.9925043Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:35.9925843Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:35.9926635Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:35.9927428Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:35.9928205Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:35.9929046Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:35.9929893Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:35.9930724Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:35.9931569Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:35.9932425Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:35.9933191Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:35.9934083Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:35.9934946Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:35.9935795Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:35.9936597Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:35.9937520Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:35.9938290Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:35.9939104Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:35.9939882Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:35.9940692Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:35.9941442Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:35.9942272Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:35.9943135Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:35.9943977Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:35.9944803Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:35.9945592Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:35.9946367Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:35.9947181Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:35.9947980Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:35.9948835Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:35.9949703Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:35.9950529Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:35.9951390Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:35.9952194Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:35.9953014Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:35.9953836Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:35.9954669Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:35.9955452Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:35.9956333Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:35.9957159Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:35.9957988Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:35.9958878Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:35.9959690Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:35.9960453Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:35.9961274Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:35.9962083Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:35.9962874Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:35.9963708Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:35.9964666Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:35.9965683Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:35.9966661Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:35.9967657Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:35.9968645Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:35.9969645Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:35.9970681Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:35.9971661Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:35.9972657Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:35.9973523Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:35.9974385Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:35.9975198Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:35.9976110Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:35.9977062Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:35.9977851Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:35.9978637Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:35.9979398Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:35.9980218Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:35.9981005Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:35.9981798Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:35.9982610Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:35.9983497Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:35.9984366Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:35.9985187Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:35.9985954Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:35.9986793Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:35.9987567Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:35.9988342Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:35.9989207Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:35.9990054Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:35.9990910Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:35.9991741Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:35.9992467Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:35.9993257Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:35.9994063Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:35.9994903Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:35.9995860Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:35.9996794Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:35.9997539Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:35.9998382Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:35.9999196Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:36.0000122Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:36.0000929Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:36.0001746Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:36.0002684Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:36.0003611Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:36.0004459Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:36.0005301Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:36.0006222Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:36.0007060Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:36.0007848Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:36.0008664Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:36.0009496Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:36.0010361Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:36.0011202Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:36.0011944Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:36.0012752Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:36.0013559Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:36.0014311Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:36.0015119Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:36.0015926Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:36.0016776Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:36.0017625Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:36.0018488Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.0019412Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.0020185Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:36.0021081Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:36.0021919Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:36.0022813Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:36.0023667Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:36.0024560Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:36.0025408Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:36.0026279Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:36.0027090Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:36.0027917Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:36.0028757Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:36.0029570Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:36.0030393Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:36.0031221Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:36.0032103Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:36.0032933Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:36.0033778Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:36.0034575Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:36.0035415Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:36.0036311Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:36.0037211Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:36.0038081Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:36.0039040Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:36.0039934Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:36.0040865Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:36.0041761Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:36.0042609Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:36.0043463Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:36.0044291Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:36.0045223Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:36.0046101Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:36.0046960Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:36.0047813Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:36.0048646Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:36.0049452Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:36.0050305Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:36.0051127Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:36.0051951Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:36.0052789Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:36.0053635Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:36.0054504Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:36.0055415Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:36.0056231Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:36.0057107Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:36.0058016Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:36.0058931Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:36.0059760Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:36.0060635Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:36.0061505Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:36.0062354Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:36.0063214Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:36.0064140Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:36.0064930Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:36.0065727Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:36.0066576Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:36.0067412Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:36.0068230Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:36.0069061Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:36.0069908Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:36.0070896Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:36.0071775Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.0072630Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:36.0073511Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:36.0074386Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:36.0075243Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:36.0076050Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:36.0077050Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:36.0077890Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:36.0078738Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:36.0079529Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:36.0080398Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:36.0081301Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:36.0082204Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:36.0083165Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:36.0083972Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:36.0084818Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:36.0085752Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:36.0086644Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:36.0087587Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:36.0088481Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:36.0089321Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:36.0090153Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:36.0091072Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:36.0091928Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:36.0092793Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:36.0093599Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:36.0094468Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:36.0095304Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:36.0096290Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:36.0097283Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:36.0098144Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:36.0098957Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:36.0099819Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:36.0100711Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:36.0101666Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:36.0102588Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:36.0103504Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:36.0104441Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:36.0105308Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:36.0106173Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:36.0107064Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:36.0107988Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:36.0108857Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:36.0109698Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:36.0110515Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:36.0111356Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:36.0112233Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:36.0113144Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:36.0114025Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:36.0114917Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:36.0115690Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:36.0116468Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:36.0117278Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:36.0118115Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:36.0118947Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:36.0119788Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:36.0120595Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:36.0121471Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:36.0122433Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:36.0123324Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:36.0124136Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:36.0124919Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:36.0125780Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:36.0126621Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:36.0127484Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:36.0128349Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:36.0129242Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:36.0130116Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:36.0130951Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:36.0131800Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:36.0132649Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:36.0133521Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:36.0134440Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:36.0135269Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:36.0136021Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:36.0137020Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:36.0137839Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:36.0138612Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:36.0139385Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:36.0140173Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:36.0140963Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:36.0141840Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:36.0142612Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:36.0143362Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:36.0144207Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:36.0145085Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:36.0145946Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:36.0146826Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:36.0147721Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:36.0148511Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:36.0149306Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:36.0150157Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:36.0150904Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:36.0151713Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:36.0152471Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:36.0153272Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:36.0154084Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:36.0154932Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:36.0155768Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:36.0156543Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:36.0157312Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:36.0158107Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:36.0160086Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:36.0162536Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:36.0163313Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:36.0164280Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:36.0165094Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:36.0165945Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:36.0166906Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:36.0167786Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:36.0168699Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:36.0169617Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:36.0170510Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:36.0171422Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:36.0172325Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:36.0173322Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:36.0174203Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:36.0175031Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:36.0175965Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:36.0176987Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:36.0177878Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:36.0178748Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:36.0179733Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:36.0180506Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:36.0181208Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:36.0181990Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:36.0182704Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:36.0183429Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:36.0184197Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:36.0185086Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:36.0185924Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:36.0186771Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:36.0187501Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:36.0188154Z V1204 11:22:01.131000 100404 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:36.0188833Z I1204 11:22:01.131000 100404 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:36.0189319Z V1204 11:22:01.131000 100404 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:36.0189814Z V1204 11:22:01.131000 100404 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:36.0190312Z I1204 11:22:01.132000 100404 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:36.0191174Z V1204 11:22:01.176000 100404 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.addmm.default with input shape: torch.Size([2]), torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:36.0191634Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.0192159Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:36.0193123Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.0193940Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[2][1]cuda:0", primals_3: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:36.0194964Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:36.0195920Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2][1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.0197081Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.0198079Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_2: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:36.0199025Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.0199883Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_2, 2) 2025-12-04T11:24:36.0200642Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:36.0201696Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:36.0202525Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:36.0203552Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] add_tensor: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.add.Tensor(sum_dim_int_list, convert_element_type); sum_dim_int_list = convert_element_type = None 2025-12-04T11:24:36.0204344Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:36.0204972Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:36.0205397Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.0205828Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.0206414Z V1204 11:22:01.186000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:36.0206984Z V1204 11:22:01.186000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:36.0207566Z V1204 11:22:01.186000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_3 : [num_users=1] = placeholder[target=primals_3] 2025-12-04T11:24:36.0208616Z V1204 11:22:01.187000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.0209137Z V1204 11:22:01.187000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.0210153Z V1204 11:22:01.189000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.0210687Z V1204 11:22:01.189000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.0211713Z V1204 11:22:01.221000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_2 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_3, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.0212247Z V1204 11:22:01.221000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.0213159Z V1204 11:22:01.223000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.0213611Z V1204 11:22:01.223000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.0214571Z V1204 11:22:01.229000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_2, 2), kwargs = {}) 2025-12-04T11:24:36.0215028Z V1204 11:22:01.230000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.0215911Z V1204 11:22:01.231000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:36.0216381Z V1204 11:22:01.231000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.0217352Z V1204 11:22:01.232000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:36.0217804Z V1204 11:22:01.233000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.0218654Z V1204 11:22:01.235000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:36.0219106Z V1204 11:22:01.236000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.0220043Z V1204 11:22:01.238000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %add_tensor : [num_users=1] = call_function[target=torch.ops.aten.add.Tensor](args = (%sum_dim_int_list, %convert_element_type), kwargs = {}) 2025-12-04T11:24:36.0220581Z V1204 11:22:01.238000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via .inner at 0x7f2e9019ab00> 2025-12-04T11:24:36.0221444Z V1204 11:22:01.240000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.0221894Z V1204 11:22:01.241000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.0222466Z V1204 11:22:01.241000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:36.0223079Z V1204 11:22:01.246000 100404 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:36.0223819Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:36.0223939Z warnings.warn( 2025-12-04T11:24:36.0224448Z V1204 11:22:01.247000 100404 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:36.0224693Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:36.0224807Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:36.0224954Z stats [('calls_captured', 1)] 2025-12-04T11:24:36.0225196Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:36.0225990Z inductor [('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_count', 4), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_addmm', 1)] 2025-12-04T11:24:36.0226107Z graph_break [] 2025-12-04T11:24:36.0226329Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:36.0226770Z V1204 11:22:01.491000 100404 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpvdhcq94w 2025-12-04T11:24:36.0227223Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.0227718Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:36.0228674Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.0229754Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_self_modules_linear_parameters_bias_: "f32[2][1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:36.0230506Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:36.0231243Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_bias_ = L_self_modules_linear_parameters_bias_ 2025-12-04T11:24:36.0231697Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:36.0232129Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.0233104Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:36.0234520Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, l_self_modules_linear_parameters_bias_); l_x_ = l_self_modules_linear_parameters_weight_ = l_self_modules_linear_parameters_bias_ = None 2025-12-04T11:24:36.0234992Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:36.0235408Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.0235828Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.0236539Z V1204 11:22:01.804000 100404 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:36.0237259Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw: 2025-12-04T11:24:36.0237960Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4ral453goioq2ytnrbxj5hhqjyfnxjkats7ca6dgxms6usqpldw] gm: GraphModule() 2025-12-04T11:24:36.0238252Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.0238557Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.0238830Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.0239352Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2, primals_3): 2025-12-04T11:24:36.0240173Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.0240997Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.0241804Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_2 = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:36.0242576Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.0243435Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] addmm = torch.ops.aten.addmm.default(convert_element_type, convert_element_type_2, permute); convert_element_type = None 2025-12-04T11:24:36.0244068Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:36.0244587Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] return (addmm, convert_element_type_2, permute_1) 2025-12-04T11:24:36.0244883Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.0245473Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:36.0247764Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.0250094Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [di4vpbifhxsheleamv47gds75iwa6xnkz5malaf2qsqkrhojljw] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2]), stride=(1,), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=8, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.0252502Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[2]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.0253167Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:36.0254082Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:36.0254848Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:36.0255662Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:36.0256455Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:36.0257237Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:36.0257956Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:36.0258688Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:36.0259401Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:36.0260162Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [nnqi2dye4qdwjdgeujlxh7d67y2n67kat5ehf3472pn5avlyqbk] fx_kwargs[static_input_idxs]: [0, 1] 2025-12-04T11:24:36.0260838Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inputs_to_check[0]: 2 2025-12-04T11:24:36.0261710Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:36.0262511Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:36.0263741Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:36.0264541Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:36.0293460Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:36.0294520Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:36.0295290Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:36.0296193Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:36.0297105Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:36.0297959Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:36.0298694Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:36.0299472Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:36.0300334Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:36.0301110Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:36.0301972Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:36.0302825Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:36.0303737Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:36.0304608Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:36.0305455Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:36.0306403Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:36.0307302Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:36.0308124Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:36.0309148Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:36.0309907Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:36.0310760Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:36.0311512Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:36.0312363Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:36.0313121Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:36.0313823Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:36.0314620Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:36.0315368Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:36.0316141Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:36.0317003Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:36.0317780Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:36.0318555Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:36.0319376Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:36.0320157Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:36.0320940Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:36.0321757Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:36.0322524Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:36.0323365Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:36.0324143Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:36.0324906Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:36.0325682Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:36.0326488Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:36.0327273Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:36.0328033Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:36.0328852Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:36.0329661Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:36.0330480Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:36.0331281Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:36.0332170Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:36.0332957Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:36.0333714Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:36.0334467Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:36.0335306Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:36.0336233Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:36.0337124Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:36.0337919Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:36.0338763Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:36.0339675Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:36.0340883Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:36.0341680Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:36.0342555Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:36.0343443Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:36.0344261Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:36.0345076Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:36.0345924Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:36.0346816Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:36.0347620Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:36.0348624Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.0349451Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:36.0350407Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.0351233Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:36.0352290Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.0353128Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:36.0354023Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:36.0354799Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:36.0355586Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:36.0356426Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:36.0357283Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:36.0358042Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:36.0358868Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:36.0359649Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:36.0360495Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:36.0361355Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:36.0362292Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:36.0363147Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:36.0363919Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:36.0364791Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:36.0365667Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:36.0366482Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:36.0367259Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:36.0368185Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:36.0369057Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:36.0369950Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:36.0370833Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:36.0371663Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:36.0372475Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:36.0373195Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:36.0373997Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:36.0374803Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:36.0375598Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:36.0376444Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:36.0377331Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:36.0378168Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:36.0378984Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:36.0379740Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:36.0380592Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:36.0381509Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:36.0382451Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:36.0383413Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:36.0384244Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:36.0385072Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:36.0385957Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:36.0386787Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:36.0387739Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:36.0388598Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:36.0389378Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:36.0390206Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:36.0391033Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:36.0391886Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:36.0392662Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:36.0393472Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:36.0394296Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:36.0395086Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:36.0395879Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:36.0396846Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:36.0397678Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:36.0398505Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:36.0399390Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:36.0400230Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:36.0401110Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:36.0401945Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:36.0402814Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:36.0403688Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:36.0404474Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:36.0405246Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:36.0406029Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:36.0406805Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:36.0407650Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:36.0408492Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:36.0409329Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:36.0410156Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:36.0411026Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:36.0411793Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:36.0412726Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:36.0413582Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:36.0414392Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:36.0415206Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:36.0416068Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:36.0416901Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:36.0417708Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:36.0418484Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:36.0419295Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:36.0420038Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:36.0420831Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:36.0421686Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:36.0422538Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:36.0423356Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:36.0424143Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:36.0424913Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:36.0425724Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:36.0426537Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:36.0427383Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:36.0428300Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:36.0429115Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:36.0429943Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:36.0430744Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:36.0431602Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:36.0432443Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:36.0433232Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:36.0434055Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:36.0434929Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:36.0435783Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:36.0436596Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:36.0437459Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:36.0438280Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:36.0439040Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:36.0439869Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:36.0440668Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:36.0441470Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:36.0442289Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:36.0443244Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:36.0444267Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:36.0445272Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:36.0446257Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:36.0447233Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:36.0448278Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:36.0449277Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:36.0450297Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:36.0451287Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:36.0452152Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:36.0452987Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:36.0453801Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:36.0454727Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:36.0455562Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:36.0456357Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:36.0457190Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:36.0457954Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:36.0458774Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:36.0459563Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:36.0460379Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:36.0461206Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:36.0462105Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:36.0462924Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:36.0463767Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:36.0464540Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:36.0465353Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:36.0466170Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:36.0466940Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:36.0467822Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:36.0468664Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:36.0469457Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:36.0470295Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:36.0471027Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:36.0471844Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:36.0472677Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:36.0473560Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:36.0474557Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:36.0475392Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:36.0476133Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:36.0477028Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:36.0477843Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:36.0478690Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:36.0479509Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:36.0480343Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:36.0481293Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:36.0482198Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:36.0483053Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:36.0483871Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:36.0484794Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:36.0485586Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:36.0486370Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:36.0487200Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:36.0488024Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:36.0488897Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:36.0489720Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:36.0490474Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:36.0491282Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:36.0492080Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:36.0492886Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:36.0493682Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:36.0494466Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:36.0495299Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:36.0496301Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:36.0497223Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.0498166Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.0498960Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:36.0499858Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:36.0500713Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:36.0501570Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:36.0502442Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:36.0503340Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:36.0504174Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:36.0505055Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:36.0505856Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:36.0506697Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:36.0507523Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:36.0508360Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:36.0509227Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:36.0510047Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:36.0510907Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:36.0511727Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:36.0512623Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:36.0513415Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:36.0514277Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:36.0515166Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:36.0516072Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:36.0516943Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:36.0517854Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:36.0518775Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:36.0519701Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:36.0520622Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:36.0521466Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:36.0522343Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:36.0523179Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:36.0524109Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:36.0525007Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:36.0525885Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:36.0526720Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:36.0527553Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:36.0528411Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:36.0529267Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:36.0530066Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:36.0530917Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:36.0531748Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:36.0532622Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:36.0533487Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:36.0534383Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:36.0535206Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:36.0536029Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:36.0536993Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:36.0537902Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:36.0538759Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:36.0539617Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:36.0540508Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:36.0541421Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:36.0542292Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:36.0543180Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:36.0543967Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:36.0544819Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:36.0545667Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:36.0546498Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:36.0547299Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:36.0548148Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:36.0549003Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:36.0549950Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:36.0550846Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.0551695Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:36.0552589Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:36.0553455Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:36.0554319Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:36.0555126Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:36.0556119Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:36.0556954Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:36.0557824Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:36.0558625Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:36.0559449Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:36.0560363Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:36.0561298Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:36.0562253Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:36.0563085Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:36.0563934Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:36.0564819Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:36.0565704Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:36.0566633Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:36.0567517Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:36.0568370Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:36.0569216Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:36.0570135Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:36.0570994Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:36.0571855Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:36.0572676Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:36.0573566Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:36.0574415Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:36.0575214Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:36.0576173Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:36.0577181Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:36.0578004Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:36.0578903Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:36.0579816Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:36.0580806Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:36.0581798Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:36.0582692Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:36.0583633Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:36.0584492Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:36.0585383Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:36.0586290Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:36.0587229Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:36.0588096Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:36.0588954Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:36.0589774Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:36.0590657Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:36.0591539Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:36.0592391Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:36.0593325Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:36.0594212Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:36.0594966Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:36.0595757Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:36.0596709Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:36.0597572Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:36.0598384Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:36.0599211Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:36.0600001Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:36.0600894Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:36.0601844Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:36.0602748Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:36.0603558Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:36.0604330Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:36.0605200Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:36.0606031Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:36.0607021Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:36.0607882Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:36.0608741Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:36.0609664Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:36.0610508Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:36.0611328Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:36.0612203Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:36.0613092Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:36.0614006Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:36.0614852Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:36.0615581Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:36.0616531Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:36.0617408Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:36.0618177Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:36.0618965Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:36.0619748Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:36.0620550Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:36.0621419Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:36.0622208Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:36.0623011Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:36.0623844Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:36.0624696Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:36.0625543Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:36.0626464Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:36.0627318Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:36.0628159Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:36.0628961Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:36.0629807Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:36.0630568Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:36.0631332Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:36.0632106Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:36.0632895Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:36.0633725Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:36.0634579Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:36.0635401Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:36.0636195Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:36.0636953Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:36.0637764Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:36.0639764Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:36.0642234Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:36.0643009Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:36.0643967Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:36.0644768Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:36.0645644Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:36.0646604Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:36.0647461Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:36.0648370Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:36.0649274Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:36.0650176Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:36.0651080Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:36.0651995Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:36.0652984Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:36.0653883Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:36.0654718Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:36.0655672Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:36.0656609Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:36.0657545Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:36.0658475Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:36.0659445Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:36.0660225Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:36.0660931Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:36.0661718Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:36.0662435Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:36.0663161Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:36.0663897Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:36.0664770Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:36.0665629Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:36.0666481Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:36.0667221Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:36.0667873Z V1204 11:22:01.815000 100404 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:36.0668545Z I1204 11:22:01.815000 100404 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:36.0669045Z V1204 11:22:01.815000 100404 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:36.0669534Z V1204 11:22:01.815000 100404 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:36.0670079Z I1204 11:22:01.816000 100404 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:36.0670938Z V1204 11:22:01.831000 100404 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.addmm.default with input shape: torch.Size([2]), torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:36.0671389Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.0671904Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:36.0672854Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.0673727Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[2][1]cuda:0", primals_3: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:36.0674709Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:36.0675709Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2][1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.0676714Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.0677710Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_2: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:36.0678625Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.0679477Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_2, 2) 2025-12-04T11:24:36.0680244Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:36.0681287Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:36.0682123Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:36.0683135Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] add_tensor: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.add.Tensor(sum_dim_int_list, convert_element_type); sum_dim_int_list = convert_element_type = None 2025-12-04T11:24:36.0683933Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:36.0684581Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:36.0685018Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.0685413Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.0685989Z V1204 11:22:01.841000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:36.0686569Z V1204 11:22:01.841000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:36.0687172Z V1204 11:22:01.842000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_3 : [num_users=1] = placeholder[target=primals_3] 2025-12-04T11:24:36.0688196Z V1204 11:22:01.842000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.0688730Z V1204 11:22:01.843000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.0689756Z V1204 11:22:01.844000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.0690293Z V1204 11:22:01.844000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.0691298Z V1204 11:22:01.845000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_2 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_3, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.0691815Z V1204 11:22:01.845000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.0692712Z V1204 11:22:01.846000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.0693176Z V1204 11:22:01.847000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.0694120Z V1204 11:22:01.848000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_2, 2), kwargs = {}) 2025-12-04T11:24:36.0694584Z V1204 11:22:01.848000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.0695470Z V1204 11:22:01.849000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:36.0695922Z V1204 11:22:01.850000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.0697055Z V1204 11:22:01.851000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:36.0697493Z V1204 11:22:01.851000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.0698425Z V1204 11:22:01.853000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:36.0698860Z V1204 11:22:01.853000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.0699772Z V1204 11:22:01.855000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %add_tensor : [num_users=1] = call_function[target=torch.ops.aten.add.Tensor](args = (%sum_dim_int_list, %convert_element_type), kwargs = {}) 2025-12-04T11:24:36.0700309Z V1204 11:22:01.855000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via .inner at 0x7f2e9019ab00> 2025-12-04T11:24:36.0701194Z V1204 11:22:01.857000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.0701660Z V1204 11:22:01.857000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.0702212Z V1204 11:22:01.858000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:36.0702843Z V1204 11:22:01.860000 100404 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:36.0703586Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:36.0703747Z warnings.warn( 2025-12-04T11:24:36.0704228Z V1204 11:22:01.861000 100404 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:36.0704382Z =================================== FAILURES =================================== 2025-12-04T11:24:36.0704857Z _ TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True _ 2025-12-04T11:24:36.0704982Z Traceback (most recent call last): 2025-12-04T11:24:36.0705564Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 275, in test_decompose_linear_mixed_precision 2025-12-04T11:24:36.0705703Z self.assertEqual( 2025-12-04T11:24:36.0706165Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:36.0706341Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:36.0706876Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:36.0707084Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:36.0707231Z AssertionError: Scalars are not equal! 2025-12-04T11:24:36.0707245Z 2025-12-04T11:24:36.0707351Z Expected 1 but got 0. 2025-12-04T11:24:36.0707461Z Absolute difference: 1 2025-12-04T11:24:36.0707583Z Relative difference: 1.0 2025-12-04T11:24:36.0707589Z 2025-12-04T11:24:36.0707805Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:36.0708682Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:36.0708689Z 2025-12-04T11:24:36.0708958Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:36.0709181Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:36.0709307Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:36.0709421Z stats [('calls_captured', 1)] 2025-12-04T11:24:36.0710218Z inductor [('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_count', 4), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_addmm', 1)] 2025-12-04T11:24:36.0710472Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:36.0710570Z graph_break [] 2025-12-04T11:24:36.0710799Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:36.0711301Z I1204 11:21:58.159000 100404 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:36.0711904Z I1204 11:21:58.404000 100404 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:36.0712342Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.0712836Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:36.0713825Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.0714935Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_self_modules_linear_parameters_bias_: "f32[2][1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:36.0715700Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:36.0716455Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_bias_ = L_self_modules_linear_parameters_bias_ 2025-12-04T11:24:36.0716928Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:36.0717351Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.0718333Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:36.0719720Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, l_self_modules_linear_parameters_bias_); l_x_ = l_self_modules_linear_parameters_weight_ = l_self_modules_linear_parameters_bias_ = None 2025-12-04T11:24:36.0720203Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:36.0720638Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.0721024Z V1204 11:21:58.406000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.0721597Z I1204 11:22:00.743000 100404 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:36.0722312Z V1204 11:22:01.120000 100404 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:36.0723022Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw: 2025-12-04T11:24:36.0723697Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4ral453goioq2ytnrbxj5hhqjyfnxjkats7ca6dgxms6usqpldw] gm: GraphModule() 2025-12-04T11:24:36.0724008Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.0724298Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.0724571Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.0725082Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2, primals_3): 2025-12-04T11:24:36.0725914Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.0726760Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.0727591Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_2 = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:36.0728350Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.0729252Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] addmm = torch.ops.aten.addmm.default(convert_element_type, convert_element_type_2, permute); convert_element_type = None 2025-12-04T11:24:36.0729913Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:36.0730417Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] return (addmm, convert_element_type_2, permute_1) 2025-12-04T11:24:36.0730729Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.0731303Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:36.0733616Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.0735906Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [di4vpbifhxsheleamv47gds75iwa6xnkz5malaf2qsqkrhojljw] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2]), stride=(1,), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=8, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.0738327Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[2]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.0738980Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:36.0739938Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:36.0740669Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:36.0741492Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:36.0742280Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:36.0743038Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:36.0743758Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:36.0744511Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:36.0745243Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:36.0746027Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [nnqi2dye4qdwjdgeujlxh7d67y2n67kat5ehf3472pn5avlyqbk] fx_kwargs[static_input_idxs]: [0, 1] 2025-12-04T11:24:36.0746718Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inputs_to_check[0]: 2 2025-12-04T11:24:36.0747594Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:36.0748383Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:36.0749542Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:36.0750326Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:36.0779307Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:36.0780372Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:36.0781150Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:36.0781947Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:36.0782791Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:36.0784191Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:36.0784912Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:36.0785697Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:36.0786479Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:36.0787263Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:36.0788131Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:36.0788974Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:36.0789851Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:36.0790697Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:36.0791561Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:36.0792459Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:36.0793369Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:36.0794183Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:36.0795201Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:36.0796312Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:36.0797198Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:36.0797962Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:36.0798863Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:36.0799645Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:36.0800398Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:36.0801216Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:36.0801959Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:36.0802711Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:36.0803559Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:36.0804320Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:36.0805108Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:36.0805880Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:36.0806662Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:36.0807441Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:36.0808223Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:36.0809041Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:36.0809835Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:36.0810626Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:36.0811394Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:36.0812233Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:36.0813048Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:36.0813812Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:36.0814635Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:36.0815450Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:36.0816306Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:36.0817198Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:36.0818016Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:36.0818885Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:36.0819662Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:36.0820438Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:36.0821190Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:36.0822014Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:36.0822935Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:36.0823745Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:36.0824540Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:36.0825412Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:36.0826181Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:36.0827368Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:36.0828212Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:36.0829086Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:36.0830014Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:36.0830833Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:36.0831677Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:36.0832541Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:36.0833426Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:36.0834243Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:36.0835188Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.0836035Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:36.0836995Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.0837816Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:36.0838751Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.0839570Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:36.0840456Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:36.0841255Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:36.0842021Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:36.0842859Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:36.0843724Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:36.0844510Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:36.0845328Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:36.0846160Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:36.0847001Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:36.0847904Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:36.0848837Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:36.0849717Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:36.0850495Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:36.0851317Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:36.0852209Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:36.0853016Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:36.0853804Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:36.0854693Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:36.0855567Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:36.0856439Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:36.0857390Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:36.0858206Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:36.0859005Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:36.0859743Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:36.0860568Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:36.0861388Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:36.0862204Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:36.0863051Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:36.0863980Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:36.0864800Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:36.0865617Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:36.0866354Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:36.0867162Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:36.0868077Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:36.0869017Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:36.0869944Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:36.0870759Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:36.0871592Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:36.0872434Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:36.0873322Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:36.0874231Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:36.0875102Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:36.0875893Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:36.0876731Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:36.0877581Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:36.0878446Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:36.0879235Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:36.0880056Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:36.0880901Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:36.0881683Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:36.0882461Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:36.0883288Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:36.0884114Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:36.0884961Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:36.0885826Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:36.0886618Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:36.0887492Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:36.0888278Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:36.0889181Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:36.0889998Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:36.0890800Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:36.0891557Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:36.0892388Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:36.0893172Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:36.0894024Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:36.0894878Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:36.0895738Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:36.0896725Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:36.0897638Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:36.0898421Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:36.0899324Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:36.0900188Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:36.0901008Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:36.0901815Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:36.0902673Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:36.0903450Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:36.0904228Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:36.0905077Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:36.0905856Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:36.0906610Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:36.0907387Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:36.0908294Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:36.0909146Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:36.0909979Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:36.0910814Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:36.0911588Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:36.0912443Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:36.0913251Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:36.0914112Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:36.0914985Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:36.0915813Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:36.0916648Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:36.0917463Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:36.0918287Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:36.0919117Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:36.0919924Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:36.0920709Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:36.0921622Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:36.0922427Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:36.0923253Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:36.0924119Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:36.0924966Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:36.0925732Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:36.0926586Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:36.0927392Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:36.0928230Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:36.0929071Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:36.0930022Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:36.0931029Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:36.0932011Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:36.0932979Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:36.0933966Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:36.0934964Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:36.0935971Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:36.0937007Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:36.0938053Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:36.0938892Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:36.0939721Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:36.0940583Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:36.0941495Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:36.0942343Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:36.0943150Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:36.0943946Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:36.0944743Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:36.0945565Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:36.0946355Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:36.0947141Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:36.0947940Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:36.0948831Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:36.0949666Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:36.0950481Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:36.0951256Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:36.0952067Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:36.0952844Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:36.0953668Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:36.0954508Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:36.0955353Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:36.0956144Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:36.0957012Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:36.0957744Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:36.0958566Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:36.0959377Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:36.0960237Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:36.0961193Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:36.0961992Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:36.0962738Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:36.0963576Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:36.0964392Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:36.0965262Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:36.0966068Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:36.0966888Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:36.0967831Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:36.0968720Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:36.0969598Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:36.0970388Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:36.0971311Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:36.0972106Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:36.0972937Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:36.0973762Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:36.0974631Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:36.0975498Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:36.0976356Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:36.0977176Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:36.0977986Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:36.0978803Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:36.0979562Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:36.0980371Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:36.0981150Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:36.0981991Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:36.0982795Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:36.0983661Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.0984565Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.0985376Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:36.0986249Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:36.0987091Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:36.0987958Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:36.0988846Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:36.0989750Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:36.0990644Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:36.0991517Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:36.0992357Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:36.0993180Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:36.0994015Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:36.0994836Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:36.0995661Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:36.0996646Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:36.0997489Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:36.0998325Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:36.0999179Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:36.0999986Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:36.1000804Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:36.1001748Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:36.1002617Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:36.1003488Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:36.1004421Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:36.1005364Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:36.1006300Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:36.1007242Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:36.1008105Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:36.1009001Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:36.1009839Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:36.1010778Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:36.1011654Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:36.1012510Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:36.1013335Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:36.1014181Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:36.1014994Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:36.1015860Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:36.1016662Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:36.1017547Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:36.1018426Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:36.1019244Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:36.1020116Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:36.1021023Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:36.1021846Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:36.1022668Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:36.1023610Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:36.1024523Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:36.1025385Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:36.1026261Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:36.1027142Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:36.1027982Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:36.1028843Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:36.1029725Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:36.1030523Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:36.1031326Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:36.1032183Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:36.1032979Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:36.1033791Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:36.1034635Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:36.1035482Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:36.1036476Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:36.1037450Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.1038306Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:36.1039245Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:36.1040145Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:36.1041061Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:36.1041868Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:36.1042851Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:36.1043702Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:36.1044533Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:36.1045335Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:36.1046168Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:36.1047086Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:36.1047979Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:36.1048914Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:36.1049733Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:36.1050605Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:36.1051452Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:36.1052334Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:36.1053256Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:36.1054185Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:36.1055038Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:36.1055903Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:36.1056813Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:36.1057767Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:36.1058633Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:36.1059456Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:36.1060313Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:36.1061161Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:36.1061962Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:36.1062906Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:36.1063789Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:36.1064594Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:36.1065406Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:36.1066288Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:36.1067259Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:36.1068173Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:36.1069052Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:36.1070036Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:36.1070907Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:36.1071814Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:36.1072694Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:36.1073659Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:36.1074527Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:36.1075385Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:36.1076211Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:36.1077040Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:36.1077934Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:36.1078795Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:36.1079695Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:36.1080588Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:36.1081344Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:36.1082117Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:36.1082968Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:36.1083782Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:36.1084599Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:36.1085426Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:36.1086248Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:36.1087144Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:36.1088126Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:36.1089013Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:36.1089863Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:36.1090636Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:36.1091524Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:36.1092366Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:36.1093243Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:36.1094100Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:36.1094963Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:36.1095839Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:36.1096814Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:36.1097714Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:36.1098555Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:36.1099518Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:36.1100408Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:36.1101256Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:36.1101985Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:36.1102970Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:36.1103802Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:36.1104603Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:36.1105394Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:36.1106229Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:36.1107038Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:36.1107902Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:36.1108673Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:36.1109438Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:36.1110272Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:36.1111127Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:36.1111978Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:36.1112870Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:36.1113728Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:36.1114529Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:36.1115389Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:36.1116212Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:36.1116976Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:36.1117741Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:36.1118554Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:36.1119352Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:36.1120277Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:36.1121147Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:36.1121999Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:36.1122794Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:36.1123557Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:36.1124375Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:36.1126347Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:36.1128791Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:36.1129560Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:36.1130494Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:36.1131334Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:36.1132185Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:36.1133145Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:36.1133995Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:36.1134949Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:36.1135865Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:36.1136797Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:36.1137751Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:36.1138706Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:36.1139697Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:36.1140593Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:36.1141449Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:36.1142368Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:36.1143315Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:36.1144198Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:36.1145088Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:36.1146052Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:36.1146805Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:36.1147510Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:36.1148295Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:36.1149022Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:36.1149741Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:36.1150485Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:36.1151388Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:36.1152249Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:36.1153131Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:36.1153852Z V1204 11:22:01.130000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:36.1154549Z V1204 11:22:01.131000 100404 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:36.1155219Z I1204 11:22:01.131000 100404 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:36.1155721Z V1204 11:22:01.131000 100404 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:36.1156205Z V1204 11:22:01.131000 100404 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:36.1156711Z I1204 11:22:01.132000 100404 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:36.1157581Z V1204 11:22:01.176000 100404 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.addmm.default with input shape: torch.Size([2]), torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:36.1158034Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.1158546Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:36.1159503Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.1160340Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[2][1]cuda:0", primals_3: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:36.1161322Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:36.1162290Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2][1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.1163293Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.1164314Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_2: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:36.1165279Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.1166123Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_2, 2) 2025-12-04T11:24:36.1166936Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:36.1167986Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:36.1168884Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:36.1169904Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] add_tensor: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.add.Tensor(sum_dim_int_list, convert_element_type); sum_dim_int_list = convert_element_type = None 2025-12-04T11:24:36.1170701Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:36.1171316Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:36.1171758Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.1172152Z V1204 11:22:01.182000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.1172726Z V1204 11:22:01.186000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:36.1173301Z V1204 11:22:01.186000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:36.1173865Z V1204 11:22:01.186000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_3 : [num_users=1] = placeholder[target=primals_3] 2025-12-04T11:24:36.1174889Z V1204 11:22:01.187000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.1175400Z V1204 11:22:01.187000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.1176466Z V1204 11:22:01.189000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.1177057Z V1204 11:22:01.189000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.1178075Z V1204 11:22:01.221000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_2 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_3, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.1178591Z V1204 11:22:01.221000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.1179530Z V1204 11:22:01.223000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.1180004Z V1204 11:22:01.223000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.1180978Z V1204 11:22:01.229000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_2, 2), kwargs = {}) 2025-12-04T11:24:36.1181449Z V1204 11:22:01.230000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.1182331Z V1204 11:22:01.231000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:36.1182818Z V1204 11:22:01.231000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.1183749Z V1204 11:22:01.232000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:36.1184181Z V1204 11:22:01.233000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.1185045Z V1204 11:22:01.235000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:36.1185485Z V1204 11:22:01.236000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.1186400Z V1204 11:22:01.238000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %add_tensor : [num_users=1] = call_function[target=torch.ops.aten.add.Tensor](args = (%sum_dim_int_list, %convert_element_type), kwargs = {}) 2025-12-04T11:24:36.1186947Z V1204 11:22:01.238000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via .inner at 0x7f2e9019ab00> 2025-12-04T11:24:36.1187793Z V1204 11:22:01.240000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.1188259Z V1204 11:22:01.241000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.1188812Z V1204 11:22:01.241000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:36.1189403Z V1204 11:22:01.246000 100404 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:36.1190175Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:36.1190279Z warnings.warn( 2025-12-04T11:24:36.1190771Z V1204 11:22:01.247000 100404 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:36.1191001Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:36.1191132Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:36.1191248Z stats [('calls_captured', 1)] 2025-12-04T11:24:36.1191473Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:36.1192276Z inductor [('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_count', 4), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_addmm', 1)] 2025-12-04T11:24:36.1192407Z graph_break [] 2025-12-04T11:24:36.1192630Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:36.1193085Z V1204 11:22:01.491000 100404 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpvdhcq94w 2025-12-04T11:24:36.1193523Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.1194062Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:36.1195009Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.1196275Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_self_modules_linear_parameters_bias_: "f32[2][1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:36.1197033Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:36.1197761Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_bias_ = L_self_modules_linear_parameters_bias_ 2025-12-04T11:24:36.1198232Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:36.1198659Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.1199659Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:36.1201043Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, l_self_modules_linear_parameters_bias_); l_x_ = l_self_modules_linear_parameters_weight_ = l_self_modules_linear_parameters_bias_ = None 2025-12-04T11:24:36.1201537Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:36.1201958Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.1202346Z V1204 11:22:01.519000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.1203075Z V1204 11:22:01.804000 100404 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:36.1203844Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw: 2025-12-04T11:24:36.1204523Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4ral453goioq2ytnrbxj5hhqjyfnxjkats7ca6dgxms6usqpldw] gm: GraphModule() 2025-12-04T11:24:36.1204804Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.1205084Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.1205377Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.1205897Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2, primals_3): 2025-12-04T11:24:36.1206783Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.1207601Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.1208464Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_2 = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:36.1209232Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.1210115Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] addmm = torch.ops.aten.addmm.default(convert_element_type, convert_element_type_2, permute); convert_element_type = None 2025-12-04T11:24:36.1210765Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:36.1211270Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] return (addmm, convert_element_type_2, permute_1) 2025-12-04T11:24:36.1211581Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.1212158Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:36.1214471Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.1216761Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [di4vpbifhxsheleamv47gds75iwa6xnkz5malaf2qsqkrhojljw] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2]), stride=(1,), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=8, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.1219179Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[2]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.1219867Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:36.1220790Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:36.1221513Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:36.1222371Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:36.1223160Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:36.1223914Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:36.1224645Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:36.1225402Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:36.1226129Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:36.1226886Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [nnqi2dye4qdwjdgeujlxh7d67y2n67kat5ehf3472pn5avlyqbk] fx_kwargs[static_input_idxs]: [0, 1] 2025-12-04T11:24:36.1227589Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inputs_to_check[0]: 2 2025-12-04T11:24:36.1228466Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:36.1229256Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:36.1230431Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:36.1231213Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:36.1260199Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:36.1261290Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:36.1262074Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:36.1262881Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:36.1263721Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:36.1271958Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:36.1272782Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:36.1273586Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:36.1274378Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:36.1275158Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:36.1276026Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:36.1276864Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:36.1277738Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:36.1278596Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:36.1279558Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:36.1280461Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:36.1281323Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:36.1282194Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:36.1283208Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:36.1283979Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:36.1284853Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:36.1285611Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:36.1286490Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:36.1287258Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:36.1287979Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:36.1288770Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:36.1289533Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:36.1290289Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:36.1291140Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:36.1291904Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:36.1292683Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:36.1293464Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:36.1294232Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:36.1295062Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:36.1295833Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:36.1296793Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:36.1297657Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:36.1298516Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:36.1299304Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:36.1300078Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:36.1300944Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:36.1301712Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:36.1302530Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:36.1303351Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:36.1304155Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:36.1304991Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:36.1305782Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:36.1306657Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:36.1307429Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:36.1308201Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:36.1308951Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:36.1309755Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:36.1310687Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:36.1311523Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:36.1312322Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:36.1313151Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:36.1313907Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:36.1315136Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:36.1315948Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:36.1316837Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:36.1317728Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:36.1318588Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:36.1319401Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:36.1320259Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:36.1321139Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:36.1321948Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:36.1322882Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.1323712Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:36.1324684Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.1325478Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:36.1326431Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.1327279Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:36.1328160Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:36.1328918Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:36.1329676Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:36.1330545Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:36.1331394Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:36.1332206Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:36.1333019Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:36.1333843Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:36.1334678Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:36.1335547Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:36.1336477Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:36.1337388Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:36.1338178Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:36.1339009Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:36.1339903Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:36.1340706Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:36.1341495Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:36.1342386Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:36.1343283Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:36.1344159Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:36.1345023Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:36.1345829Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:36.1346651Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:36.1347381Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:36.1348201Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:36.1349007Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:36.1349838Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:36.1350675Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:36.1351518Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:36.1352341Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:36.1353150Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:36.1353890Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:36.1354687Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:36.1355607Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:36.1356526Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:36.1357469Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:36.1358278Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:36.1359141Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:36.1359975Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:36.1360817Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:36.1361723Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:36.1362609Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:36.1363409Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:36.1364249Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:36.1365085Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:36.1365945Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:36.1366730Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:36.1367519Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:36.1368351Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:36.1369142Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:36.1369917Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:36.1370738Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:36.1371554Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:36.1372396Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:36.1373259Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:36.1374035Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:36.1374962Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:36.1375741Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:36.1376618Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:36.1377489Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:36.1378328Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:36.1379088Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:36.1379902Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:36.1380694Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:36.1381522Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:36.1382411Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:36.1383247Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:36.1384078Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:36.1384930Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:36.1385687Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:36.1386589Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:36.1387430Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:36.1388254Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:36.1389046Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:36.1389888Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:36.1390687Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:36.1391455Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:36.1392235Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:36.1393002Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:36.1393860Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:36.1394642Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:36.1395505Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:36.1396558Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:36.1397376Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:36.1398224Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:36.1398984Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:36.1399804Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:36.1400607Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:36.1401459Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:36.1402339Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:36.1403161Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:36.1403994Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:36.1404786Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:36.1405619Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:36.1406450Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:36.1407295Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:36.1408084Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:36.1408967Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:36.1409770Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:36.1410628Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:36.1411511Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:36.1412349Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:36.1413121Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:36.1413980Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:36.1414791Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:36.1415585Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:36.1416413Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:36.1417446Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:36.1418461Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:36.1419459Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:36.1420429Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:36.1421426Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:36.1422426Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:36.1423491Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:36.1424478Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:36.1425474Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:36.1426332Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:36.1427215Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:36.1428059Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:36.1429017Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:36.1429880Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:36.1430711Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:36.1431499Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:36.1432280Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:36.1433090Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:36.1433907Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:36.1434709Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:36.1435518Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:36.1436412Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:36.1437241Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:36.1438086Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:36.1438850Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:36.1439725Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:36.1440504Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:36.1441293Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:36.1442138Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:36.1443027Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:36.1443836Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:36.1444693Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:36.1445448Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:36.1446273Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:36.1447084Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:36.1447917Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:36.1448871Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:36.1449690Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:36.1450476Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:36.1451333Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:36.1452153Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:36.1453029Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:36.1453845Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:36.1454674Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:36.1455679Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:36.1456564Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:36.1457482Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:36.1458274Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:36.1459254Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:36.1460053Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:36.1460877Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:36.1461698Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:36.1462561Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:36.1463431Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:36.1464255Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:36.1465012Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:36.1465819Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:36.1466631Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:36.1467384Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:36.1468174Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:36.1468959Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:36.1469797Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:36.1470606Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:36.1471518Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.1472401Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.1473171Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:36.1474030Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:36.1474919Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:36.1475769Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:36.1476670Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:36.1477569Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:36.1478453Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:36.1479326Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:36.1480134Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:36.1480972Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:36.1481789Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:36.1482622Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:36.1483449Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:36.1484280Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:36.1485124Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:36.1485951Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:36.1486804Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:36.1487683Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:36.1488500Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:36.1489397Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:36.1490264Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:36.1491168Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:36.1492093Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:36.1493040Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:36.1493962Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:36.1494915Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:36.1495756Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:36.1496821Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:36.1497703Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:36.1498652Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:36.1499533Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:36.1500398Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:36.1501226Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:36.1502049Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:36.1502877Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:36.1503736Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:36.1504646Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:36.1505464Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:36.1506306Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:36.1507123Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:36.1508035Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:36.1508933Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:36.1509787Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:36.1510600Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:36.1511550Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:36.1512461Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:36.1513289Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:36.1514141Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:36.1515025Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:36.1515870Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:36.1516738Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:36.1517619Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:36.1518415Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:36.1519218Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:36.1520056Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:36.1520898Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:36.1521695Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:36.1522515Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:36.1523395Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:36.1524357Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:36.1525243Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.1526127Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:36.1527009Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:36.1527903Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:36.1528771Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:36.1529571Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:36.1530568Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:36.1531400Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:36.1532246Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:36.1533035Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:36.1533862Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:36.1534777Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:36.1535673Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:36.1536652Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:36.1537508Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:36.1538357Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:36.1539189Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:36.1540105Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:36.1541031Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:36.1541940Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:36.1542787Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:36.1543655Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:36.1544578Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:36.1545433Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:36.1546309Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:36.1547112Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:36.1547965Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:36.1548814Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:36.1549613Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:36.1550564Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:36.1551421Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:36.1552239Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:36.1553064Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:36.1553943Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:36.1554878Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:36.1555835Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:36.1556731Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:36.1557676Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:36.1558578Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:36.1559443Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:36.1560368Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:36.1561291Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:36.1562150Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:36.1563006Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:36.1563829Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:36.1564669Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:36.1565543Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:36.1566412Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:36.1567290Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:36.1568181Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:36.1568970Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:36.1569732Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:36.1570559Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:36.1571358Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:36.1572211Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:36.1573027Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:36.1573850Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:36.1574757Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:36.1575740Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:36.1576644Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:36.1577521Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:36.1578303Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:36.1579164Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:36.1579999Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:36.1580880Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:36.1581736Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:36.1582593Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:36.1583470Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:36.1584312Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:36.1585169Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:36.1586015Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:36.1586892Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:36.1587778Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:36.1588656Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:36.1589381Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:36.1590360Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:36.1591178Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:36.1591994Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:36.1592774Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:36.1593554Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:36.1594362Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:36.1595228Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:36.1596220Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:36.1597019Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:36.1597867Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:36.1598708Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:36.1599556Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:36.1600454Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:36.1601409Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:36.1602213Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:36.1603008Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:36.1603837Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:36.1604644Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:36.1605409Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:36.1606219Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:36.1607034Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:36.1607905Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:36.1608782Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:36.1609645Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:36.1610422Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:36.1611179Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:36.1611987Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:36.1613953Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:36.1616392Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:36.1617266Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:36.1618239Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:36.1619033Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:36.1619870Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:36.1620865Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:36.1621724Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:36.1622662Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:36.1623586Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:36.1624510Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:36.1625415Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:36.1626347Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:36.1627351Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:36.1628262Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:36.1629107Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:36.1630053Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:36.1630984Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:36.1631880Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:36.1632778Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:36.1633761Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:36.1634570Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:36.1635279Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:36.1636040Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:36.1636765Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:36.1637547Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:36.1638288Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:36.1639199Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:36.1640070Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:36.1640965Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:36.1641706Z V1204 11:22:01.814000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:36.1642362Z V1204 11:22:01.815000 100404 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:36.1643050Z I1204 11:22:01.815000 100404 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:36.1643533Z V1204 11:22:01.815000 100404 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:36.1644027Z V1204 11:22:01.815000 100404 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:36.1644546Z I1204 11:22:01.816000 100404 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:36.1645407Z V1204 11:22:01.831000 100404 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.addmm.default with input shape: torch.Size([2]), torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:36.1645870Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.1646368Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:36.1647314Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.1648169Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[2][1]cuda:0", primals_3: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:36.1649299Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:36.1650275Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2][1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.1651238Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.1652282Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_2: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:36.1653197Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.1654079Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_2, 2) 2025-12-04T11:24:36.1654846Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:36.1655940Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:36.1656760Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:36.1657857Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] add_tensor: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.add.Tensor(sum_dim_int_list, convert_element_type); sum_dim_int_list = convert_element_type = None 2025-12-04T11:24:36.1658650Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:36.1659274Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:36.1659717Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.1660113Z V1204 11:22:01.837000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.1660709Z V1204 11:22:01.841000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:36.1661278Z V1204 11:22:01.841000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:36.1661843Z V1204 11:22:01.842000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_3 : [num_users=1] = placeholder[target=primals_3] 2025-12-04T11:24:36.1662911Z V1204 11:22:01.842000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.1663416Z V1204 11:22:01.843000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.1664444Z V1204 11:22:01.844000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.1664951Z V1204 11:22:01.844000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.1666013Z V1204 11:22:01.845000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_2 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_3, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.1666515Z V1204 11:22:01.845000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.1667439Z V1204 11:22:01.846000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.1667900Z V1204 11:22:01.847000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.1668840Z V1204 11:22:01.848000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_2, 2), kwargs = {}) 2025-12-04T11:24:36.1669353Z V1204 11:22:01.848000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.1670235Z V1204 11:22:01.849000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:36.1670701Z V1204 11:22:01.850000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.1671611Z V1204 11:22:01.851000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:36.1672045Z V1204 11:22:01.851000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.1672909Z V1204 11:22:01.853000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:36.1673343Z V1204 11:22:01.853000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.1674251Z V1204 11:22:01.855000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %add_tensor : [num_users=1] = call_function[target=torch.ops.aten.add.Tensor](args = (%sum_dim_int_list, %convert_element_type), kwargs = {}) 2025-12-04T11:24:36.1674790Z V1204 11:22:01.855000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via .inner at 0x7f2e9019ab00> 2025-12-04T11:24:36.1675646Z V1204 11:22:01.857000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.1676097Z V1204 11:22:01.857000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.1676684Z V1204 11:22:01.858000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:36.1677272Z V1204 11:22:01.860000 100404 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:36.1678012Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:36.1678129Z warnings.warn( 2025-12-04T11:24:36.1678613Z V1204 11:22:01.861000 100404 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:36.1678849Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:36.1679005Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:36.1679121Z stats [('calls_captured', 1)] 2025-12-04T11:24:36.1679359Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:36.1680152Z inductor [('pattern_matcher_nodes', 6), ('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pattern_matcher_count', 4), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_addmm', 1)] 2025-12-04T11:24:36.1680250Z graph_break [] 2025-12-04T11:24:36.1680510Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:36.1680955Z V1204 11:22:01.886000 100404 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmp2l0z8yco 2025-12-04T11:24:36.1681394Z V1204 11:22:01.914000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.1681930Z V1204 11:22:01.914000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:36.1682873Z V1204 11:22:01.914000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.1683979Z V1204 11:22:01.914000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_self_modules_linear_parameters_weight_: "f32[2, 5][5, 1]cuda:0", L_self_modules_linear_parameters_bias_: "f32[2][1]cuda:0", L_x_: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:36.1684723Z V1204 11:22:01.914000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_weight_ = L_self_modules_linear_parameters_weight_ 2025-12-04T11:24:36.1685462Z V1204 11:22:01.914000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_self_modules_linear_parameters_bias_ = L_self_modules_linear_parameters_bias_ 2025-12-04T11:24:36.1685918Z V1204 11:22:01.914000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_x_ = L_x_ 2025-12-04T11:24:36.1686341Z V1204 11:22:01.914000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.1687334Z V1204 11:22:01.914000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:36.1688706Z V1204 11:22:01.914000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] linear: "bf16[20480, 2][2, 1]cuda:0" = torch._C._nn.linear(l_x_, l_self_modules_linear_parameters_weight_, l_self_modules_linear_parameters_bias_); l_x_ = l_self_modules_linear_parameters_weight_ = l_self_modules_linear_parameters_bias_ = None 2025-12-04T11:24:36.1689194Z V1204 11:22:01.914000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (linear,) 2025-12-04T11:24:36.1689654Z V1204 11:22:01.914000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.1690049Z V1204 11:22:01.914000 100404 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.1690776Z V1204 11:22:02.196000 100404 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:36.1691501Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw: 2025-12-04T11:24:36.1692168Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4ral453goioq2ytnrbxj5hhqjyfnxjkats7ca6dgxms6usqpldw] gm: GraphModule() 2025-12-04T11:24:36.1692485Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.1692787Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.1693062Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.1693578Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2, primals_3): 2025-12-04T11:24:36.1694457Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.1695272Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.1696343Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_2 = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:36.1697188Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.1698054Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] addmm = torch.ops.aten.addmm.default(convert_element_type, convert_element_type_2, permute); convert_element_type = None 2025-12-04T11:24:36.1698687Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:36.1699209Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] return (addmm, convert_element_type_2, permute_1) 2025-12-04T11:24:36.1699503Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.1700074Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:36.1702392Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [oshwlakhjhdhy5kxz5sjtm7blvef7odydhb66euas3oclhi7u5y] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=40, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.1704669Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [di4vpbifhxsheleamv47gds75iwa6xnkz5malaf2qsqkrhojljw] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([2]), stride=(1,), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=8, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.1707115Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[2]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.1707774Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:36.1708731Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:36.1709462Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:36.1710330Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:36.1711119Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:36.1711884Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:36.1712605Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:36.1713352Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:36.1714063Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:36.1714816Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [nnqi2dye4qdwjdgeujlxh7d67y2n67kat5ehf3472pn5avlyqbk] fx_kwargs[static_input_idxs]: [0, 1] 2025-12-04T11:24:36.1715504Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inputs_to_check[0]: 2 2025-12-04T11:24:36.1716375Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:36.1717177Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:36.1718364Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:36.1719161Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:36.1748240Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:36.1749332Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:36.1750117Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:36.1750896Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:36.1751748Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:36.1752588Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:36.1753306Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:36.1754099Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:36.1754887Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:36.1755683Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:36.1756536Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:36.1757386Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:36.1758278Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:36.1759130Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:36.1759991Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:36.1760895Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:36.1761803Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:36.1762628Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:36.1763679Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:36.1764435Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:36.1765323Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:36.1766080Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:36.1766921Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:36.1767700Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:36.1768409Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:36.1769225Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:36.1769977Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:36.1770743Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:36.1771576Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:36.1772335Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:36.1773133Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:36.1773944Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:36.1774724Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:36.1775504Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:36.1776294Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:36.1777194Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:36.1777998Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:36.1778792Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:36.1779588Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:36.1780369Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:36.1781215Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:36.1781994Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:36.1782756Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:36.1783563Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:36.1784378Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:36.1785202Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:36.1785997Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:36.1786862Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:36.1787646Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:36.1788405Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:36.1789153Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:36.1790001Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:36.1790928Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:36.1791732Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:36.1792527Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:36.1793402Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:36.1794155Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:36.1795365Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:36.1796374Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:36.1797322Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:36.1798223Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:36.1799041Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:36.1799862Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:36.1800708Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:36.1801588Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:36.1802403Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:36.1803335Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.1804181Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:36.1805149Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.1806019Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:36.1806958Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.1807792Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:36.1808666Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:36.1809486Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:36.1810253Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:36.1811129Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:36.1811990Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:36.1812780Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:36.1813613Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:36.1814400Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:36.1815243Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:36.1816120Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:36.1817139Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:36.1818027Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:36.1818798Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:36.1819643Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:36.1820521Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:36.1821329Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:36.1822160Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:36.1823050Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:36.1823934Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:36.1824811Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:36.1825725Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:36.1826527Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:36.1827349Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:36.1828080Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:36.1828895Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:36.1829717Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:36.1830514Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:36.1831355Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:36.1832181Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:36.1833009Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:36.1833814Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:36.1834554Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:36.1835359Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:36.1836269Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:36.1837206Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:36.1838165Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:36.1838981Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:36.1839809Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:36.1840645Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:36.1841523Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:36.1842422Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:36.1843317Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:36.1844095Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:36.1844944Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:36.1845773Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:36.1846604Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:36.1847391Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:36.1848183Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:36.1849039Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:36.1849821Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:36.1850608Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:36.1851413Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:36.1852232Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:36.1853083Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:36.1853982Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:36.1854766Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:36.1855644Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:36.1856441Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:36.1857410Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:36.1858227Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:36.1859054Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:36.1859811Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:36.1860604Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:36.1861422Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:36.1862260Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:36.1863105Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:36.1863935Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:36.1864771Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:36.1865629Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:36.1866405Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:36.1867300Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:36.1868153Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:36.1868973Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:36.1869834Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:36.1870667Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:36.1871442Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:36.1872223Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:36.1873027Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:36.1873807Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:36.1874553Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:36.1875381Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:36.1876229Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:36.1877106Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:36.1877937Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:36.1878729Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:36.1879505Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:36.1880312Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:36.1881129Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:36.1881977Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:36.1882847Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:36.1883672Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:36.1884509Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:36.1885313Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:36.1886179Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:36.1887016Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:36.1887807Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:36.1888600Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:36.1889518Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:36.1890325Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:36.1891194Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:36.1892067Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:36.1892923Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:36.1893679Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:36.1894499Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:36.1895311Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:36.1896301Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:36.1897202Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:36.1898168Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:36.1899197Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:36.1900174Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:36.1901155Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:36.1902131Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:36.1903229Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:36.1904244Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:36.1905222Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:36.1906285Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:36.1907132Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:36.1908008Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:36.1908822Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:36.1909852Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:36.1910707Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:36.1911491Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:36.1912298Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:36.1913055Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:36.1913879Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:36.1914671Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:36.1915483Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:36.1916274Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:36.1917163Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:36.1918009Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:36.1918867Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:36.1919643Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:36.1920455Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:36.1921252Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:36.1922058Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:36.1922903Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:36.1923796Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:36.1924586Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:36.1925470Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:36.1926200Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:36.1927005Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:36.1927804Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:36.1928638Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:36.1929602Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:36.1930399Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:36.1931154Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:36.1931995Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:36.1932823Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:36.1933680Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:36.1934531Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:36.1935351Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:36.1936290Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:36.1937224Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:36.1938108Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:36.1938909Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:36.1939848Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:36.1940645Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:36.1941469Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:36.1942287Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:36.1943121Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:36.1943978Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:36.1944809Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:36.1945550Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:36.1946360Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:36.1947173Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:36.1947933Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:36.1948731Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:36.1949508Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:36.1950356Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:36.1951183Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:36.1952047Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.1952935Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.1953741Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:36.1954607Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:36.1955444Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:36.1956342Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:36.1957197Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:36.1958138Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:36.1958970Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:36.1959834Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:36.1960640Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:36.1961462Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:36.1962295Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:36.1963112Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:36.1963947Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:36.1964757Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:36.1965603Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:36.1966466Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:36.1967318Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:36.1968127Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:36.1968932Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:36.1969869Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:36.1970725Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:36.1971622Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:36.1972553Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:36.1973485Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:36.1974420Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:36.1975332Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:36.1976181Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:36.1977095Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:36.1977946Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:36.1978867Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:36.1979744Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:36.1980605Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:36.1981424Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:36.1982265Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:36.1983121Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:36.1983991Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:36.1984772Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:36.1985593Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:36.1986459Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:36.1987276Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:36.1988202Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:36.1989075Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:36.1989941Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:36.1990744Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:36.1991644Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:36.1992554Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:36.1993384Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:36.1994258Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:36.1995138Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:36.1996210Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:36.1997066Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:36.1997966Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:36.1998753Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:36.1999634Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:36.2000485Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:36.2001277Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:36.2002090Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:36.2002940Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:36.2003810Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:36.2004804Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:36.2005693Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.2006588Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:36.2007462Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:36.2008343Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:36.2009198Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:36.2010013Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:36.2010999Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:36.2011841Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:36.2012688Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:36.2013476Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:36.2014314Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:36.2015257Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:36.2016170Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:36.2017182Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:36.2017995Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:36.2018874Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:36.2019707Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:36.2020631Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:36.2021542Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:36.2022471Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:36.2023313Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:36.2024167Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:36.2025080Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:36.2025942Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:36.2026819Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:36.2027625Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:36.2028498Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:36.2029332Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:36.2030138Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:36.2031080Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:36.2031977Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:36.2032791Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:36.2033579Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:36.2034468Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:36.2035424Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:36.2036345Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:36.2037252Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:36.2038203Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:36.2039091Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:36.2039955Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:36.2040849Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:36.2041770Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:36.2042645Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:36.2043488Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:36.2044320Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:36.2045143Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:36.2046021Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:36.2046894Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:36.2047811Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:36.2048712Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:36.2049455Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:36.2050236Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:36.2051099Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:36.2051906Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:36.2052764Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:36.2053576Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:36.2054413Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:36.2055296Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:36.2056262Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:36.2057222Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:36.2058051Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:36.2058821Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:36.2059686Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:36.2060530Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:36.2061397Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:36.2062271Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:36.2063116Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:36.2064049Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:36.2064871Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:36.2065690Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:36.2066538Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:36.2067757Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:36.2068663Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:36.2069526Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:36.2070271Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:36.2071236Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:36.2072054Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:36.2072830Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:36.2073611Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:36.2074405Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:36.2075203Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:36.2076090Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:36.2076857Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:36.2077607Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:36.2078447Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:36.2079298Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:36.2080199Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:36.2081078Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:36.2081954Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:36.2082741Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:36.2083583Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:36.2084402Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:36.2085176Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:36.2085957Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:36.2086745Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:36.2087545Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:36.2088359Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:36.2089227Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:36.2090056Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:36.2090828Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:36.2091598Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:36.2092392Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:36.2094360Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:36.2097037Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:36.2097916Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:36.2098826Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:36.2099687Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:36.2100525Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:36.2101494Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:36.2102372Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:36.2103279Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:36.2104267Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:36.2105153Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:36.2106069Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:36.2106964Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:36.2107979Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:36.2108862Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:36.2109705Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:36.2110633Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:36.2111551Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:36.2112447Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:36.2113371Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:36.2114351Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:36.2115089Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:36.2115802Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:36.2116577Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:36.2117292Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:36.2118048Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:36.2118772Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:36.2119689Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:36.2120532Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:36.2121392Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:36.2122112Z V1204 11:22:02.205000 100404 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:36.2122760Z V1204 11:22:02.206000 100404 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:36.2123439Z I1204 11:22:02.206000 100404 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key f4d3easobhsg2334tv7nx5vkeodlymlhlpwzuo7sqybz7w7yh3iw 2025-12-04T11:24:36.2123925Z V1204 11:22:02.207000 100404 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:36.2124426Z V1204 11:22:02.207000 100404 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:36.2124927Z I1204 11:22:02.207000 100404 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 2 2025-12-04T11:24:36.2125800Z V1204 11:22:02.222000 100404 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.addmm.default with input shape: torch.Size([2]), torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:36.2126246Z V1204 11:22:02.228000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.2126745Z V1204 11:22:02.228000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:36.2127711Z V1204 11:22:02.228000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.2128572Z V1204 11:22:02.228000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[2, 5][5, 1]cuda:0", primals_2: "f32[2][1]cuda:0", primals_3: "f32[20480, 5][5, 1]cuda:0"): 2025-12-04T11:24:36.2129561Z V1204 11:22:02.228000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:31 in forward, code: return self.linear(x) 2025-12-04T11:24:36.2130545Z V1204 11:22:02.228000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[2][1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.2131520Z V1204 11:22:02.228000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.2132526Z V1204 11:22:02.228000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_2: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_3, torch.bfloat16); primals_3 = None 2025-12-04T11:24:36.2133440Z V1204 11:22:02.228000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 2][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.2134312Z V1204 11:22:02.228000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_2, 2) 2025-12-04T11:24:36.2135077Z V1204 11:22:02.228000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][5, 1, 5]cuda:0" = torch.ops.aten.unsqueeze.default(permute, 0) 2025-12-04T11:24:36.2136134Z V1204 11:22:02.228000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 1, 5]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:36.2137030Z V1204 11:22:02.228000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:36.2138054Z V1204 11:22:02.228000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] add_tensor: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.add.Tensor(sum_dim_int_list, convert_element_type); sum_dim_int_list = convert_element_type = None 2025-12-04T11:24:36.2138833Z V1204 11:22:02.228000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][5, 1]cuda:0" = torch.ops.aten.permute.default(permute, [1, 0]); permute = None 2025-12-04T11:24:36.2139469Z V1204 11:22:02.228000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:36.2139895Z V1204 11:22:02.228000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.2140300Z V1204 11:22:02.228000 100404 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.2140874Z V1204 11:22:02.233000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:36.2141477Z V1204 11:22:02.233000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:36.2142056Z V1204 11:22:02.234000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_3 : [num_users=1] = placeholder[target=primals_3] 2025-12-04T11:24:36.2143066Z V1204 11:22:02.234000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.2143583Z V1204 11:22:02.234000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.2144624Z V1204 11:22:02.235000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=1] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.2145142Z V1204 11:22:02.236000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.2146175Z V1204 11:22:02.236000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_2 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_3, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.2146673Z V1204 11:22:02.237000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.2147611Z V1204 11:22:02.238000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=2] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.2148064Z V1204 11:22:02.238000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.2149023Z V1204 11:22:02.240000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_2, 2), kwargs = {}) 2025-12-04T11:24:36.2149478Z V1204 11:22:02.240000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.2150373Z V1204 11:22:02.241000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%permute, 0), kwargs = {}) 2025-12-04T11:24:36.2150824Z V1204 11:22:02.241000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.2151737Z V1204 11:22:02.242000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:36.2152184Z V1204 11:22:02.242000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.2153029Z V1204 11:22:02.244000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:36.2153476Z V1204 11:22:02.245000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.2154376Z V1204 11:22:02.246000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %add_tensor : [num_users=1] = call_function[target=torch.ops.aten.add.Tensor](args = (%sum_dim_int_list, %convert_element_type), kwargs = {}) 2025-12-04T11:24:36.2154920Z V1204 11:22:02.247000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via .inner at 0x7f2e9019ab00> 2025-12-04T11:24:36.2155809Z V1204 11:22:02.249000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%permute, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.2156259Z V1204 11:22:02.249000 100404 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.2156825Z V1204 11:22:02.250000 100404 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (add_tensor, convert_element_type_2, permute_1) 2025-12-04T11:24:36.2157404Z V1204 11:22:02.252000 100404 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 2 2025-12-04T11:24:36.2158186Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:36.2158293Z warnings.warn( 2025-12-04T11:24:36.2158772Z V1204 11:22:02.253000 100404 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:36.2159631Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-ba19ebfa7960f6c4.xml - 2025-12-04T11:24:36.2159903Z =========================== short test summary info ============================ 2025-12-04T11:24:36.2160830Z FAILED [0.3889s] inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True - AssertionError: Scalars are not equal! 2025-12-04T11:24:36.2160884Z 2025-12-04T11:24:36.2160993Z Expected 1 but got 0. 2025-12-04T11:24:36.2161101Z Absolute difference: 1 2025-12-04T11:24:36.2161224Z Relative difference: 1.0 2025-12-04T11:24:36.2161230Z 2025-12-04T11:24:36.2161451Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:36.2162334Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:36.2162340Z 2025-12-04T11:24:36.2162614Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:36.2162799Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:24:36.2163011Z ================== 1 failed, 36 deselected, 2 rerun in 6.80s =================== 2025-12-04T11:24:36.2163119Z --- Logging error --- 2025-12-04T11:24:36.2163244Z Traceback (most recent call last): 2025-12-04T11:24:36.2163569Z File "/opt/conda/envs/py_3.10/lib/python3.10/logging/__init__.py", line 1103, in emit 2025-12-04T11:24:36.2163700Z stream.write(msg + self.terminator) 2025-12-04T11:24:36.2163855Z ValueError: I/O operation on closed file. 2025-12-04T11:24:36.2163953Z Call stack: 2025-12-04T11:24:36.2164475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/remote_cache.py", line 432, in dump_cache_stats 2025-12-04T11:24:36.2164640Z log.info("Cache Metrics:%s", out.getvalue()) 2025-12-04T11:24:36.2164752Z Message: 'Cache Metrics:%s' 2025-12-04T11:24:36.2164856Z Arguments: (' None\n',) 2025-12-04T11:24:36.2164964Z Got exit code 1 2025-12-04T11:24:36.2165758Z FAILED CONSISTENTLY: test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:36.2166182Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:24:36.2166856Z Test results will be stored in test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-6aac4defd9d3a3cc.xml 2025-12-04T11:24:36.2167026Z ============================= test session starts ============================== 2025-12-04T11:24:36.2167422Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:24:36.2167530Z cachedir: .pytest_cache 2025-12-04T11:24:36.2168052Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:24:36.2168189Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:24:36.2168297Z configfile: pytest.ini 2025-12-04T11:24:36.2168906Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:24:36.2170112Z collecting ... /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:52: PytestCollectionWarning: cannot collect test class 'TestDecomposeAddMM' because it has a __init__ constructor (from: test/inductor/test_decompose_mem_bound_mm.py) 2025-12-04T11:24:36.2170295Z class TestDecomposeAddMM(torch.nn.Module): 2025-12-04T11:24:36.2170463Z collected 37 items / 16 deselected / 21 selected 2025-12-04T11:24:36.2170605Z stepcurrent: skipping 16 already run items. 2025-12-04T11:24:36.2170731Z Running 21 items in this shard 2025-12-04T11:24:36.2170737Z 2025-12-04T11:24:36.2171481Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_2048_k_2_n_2_should_decompose_False_has_bias_False PASSED [5.8290s] [ 4%] 2025-12-04T11:24:36.2172244Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_2048_k_2_n_2_should_decompose_False_has_bias_True PASSED [0.3469s] [ 9%] 2025-12-04T11:24:36.2172855Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_cpu_m_1_k_64_n_16_should_decompose_True PASSED [1.1792s] [ 14%] 2025-12-04T11:24:36.2173460Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_cpu_m_1_k_64_n_32_should_decompose_True PASSED [0.7919s] [ 19%] 2025-12-04T11:24:36.2174052Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_cpu_m_2_k_64_n_16_should_decompose_False PASSED [0.1310s] [ 23%] 2025-12-04T11:24:36.2174697Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_m_20480_k_32_n_2_should_decompose_False_has_bias_False PASSED [0.3467s] [ 28%] 2025-12-04T11:24:36.2175346Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_m_20480_k_32_n_2_should_decompose_False_has_bias_True PASSED [0.3335s] [ 33%] 2025-12-04T11:24:36.2175981Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_m_20480_k_5_n_2_should_decompose_True_has_bias_False PASSED [1.2161s] [ 38%] 2025-12-04T11:24:36.2176609Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_m_20480_k_5_n_2_should_decompose_True_has_bias_True PASSED [1.1608s] [ 42%] 2025-12-04T11:24:36.2177341Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_m_2048_k_2_n_2_should_decompose_False_has_bias_False PASSED [0.3579s] [ 47%] 2025-12-04T11:24:36.2177974Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_m_2048_k_2_n_2_should_decompose_False_has_bias_True PASSED [0.3514s] [ 52%] 2025-12-04T11:24:36.2178711Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_32_n_2_should_decompose_False_has_bias_False PASSED [0.3324s] [ 57%] 2025-12-04T11:24:36.2179426Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_32_n_2_should_decompose_False_has_bias_True PASSED [0.2916s] [ 61%] 2025-12-04T11:24:36.2180227Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False ('RERUN', {'yellow': True}) [0.3532s] [ 66%] 2025-12-04T11:24:36.2181041Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False ('RERUN', {'yellow': True}) [0.3092s] [ 66%] 2025-12-04T11:24:36.2181794Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False FAILED [0.3087s] [ 66%] 2025-12-04T11:24:36.2181800Z 2025-12-04T11:24:36.2181959Z ==================================== RERUNS ==================================== 2025-12-04T11:24:36.2182414Z _ TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False _ 2025-12-04T11:24:36.2182553Z Traceback (most recent call last): 2025-12-04T11:24:36.2183124Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 384, in test_decompose_mm_mixed_precision 2025-12-04T11:24:36.2183233Z self.assertEqual( 2025-12-04T11:24:36.2183714Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:36.2183915Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:36.2184460Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:36.2184681Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:36.2184812Z AssertionError: Scalars are not equal! 2025-12-04T11:24:36.2184818Z 2025-12-04T11:24:36.2184938Z Expected 1 but got 0. 2025-12-04T11:24:36.2185046Z Absolute difference: 1 2025-12-04T11:24:36.2185186Z Relative difference: 1.0 2025-12-04T11:24:36.2185191Z 2025-12-04T11:24:36.2185423Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:36.2186274Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:36.2186309Z 2025-12-04T11:24:36.2186599Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:36.2186825Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:36.2186941Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:36.2187072Z stats [('calls_captured', 1)] 2025-12-04T11:24:36.2187296Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:36.2188053Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:36.2188168Z graph_break [] 2025-12-04T11:24:36.2188388Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:36.2188852Z V1204 11:22:27.905000 100601 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpf4czj7g6 2025-12-04T11:24:36.2189297Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.2189794Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:36.2190759Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.2191499Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.2192009Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:36.2192500Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:36.2192939Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.2193990Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.2194734Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:36.2195221Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:36.2195644Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.2196314Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.2197041Z V1204 11:22:28.167000 100601 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:36.2197755Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:36.2198472Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:36.2198753Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.2199043Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.2199358Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.2199832Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:36.2200667Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.2201483Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.2202119Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:36.2202887Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.2203652Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.2204105Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:36.2204399Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.2204987Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:36.2207317Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.2209688Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.2210337Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:36.2211304Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:36.2212032Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:36.2212867Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:36.2213696Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:36.2214423Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:36.2215171Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:36.2215900Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:36.2216627Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:36.2217428Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:36.2218130Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:36.2218818Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:36.2219705Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:36.2220497Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:36.2221673Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:36.2222466Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:36.2251392Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:36.2252498Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:36.2253272Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:36.2254042Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:36.2254896Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:36.2255741Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:36.2256476Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:36.2257323Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:36.2258125Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:36.2258901Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:36.2259760Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:36.2260604Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:36.2261498Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:36.2262363Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:36.2263211Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:36.2264164Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:36.2265026Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:36.2265856Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:36.2266885Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:36.2267670Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:36.2268524Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:36.2269274Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:36.2270124Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:36.2270885Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:36.2271606Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:36.2272406Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:36.2273151Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:36.2273918Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:36.2274752Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:36.2275529Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:36.2276314Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:36.2277134Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:36.2277902Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:36.2278683Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:36.2279469Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:36.2280283Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:36.2281096Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:36.2281905Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:36.2282684Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:36.2283482Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:36.2284292Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:36.2285073Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:36.2285833Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:36.2286660Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:36.2287461Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:36.2288306Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:36.2289093Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:36.2289961Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:36.2290748Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:36.2291515Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:36.2292274Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:36.2293112Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:36.2294046Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:36.2294840Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:36.2295676Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:36.2296731Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:36.2297558Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:36.2298827Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:36.2299668Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:36.2300547Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:36.2301437Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:36.2302255Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:36.2303075Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:36.2303925Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:36.2304822Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:36.2305624Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:36.2306574Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.2307399Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:36.2308371Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.2309225Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:36.2310162Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.2310996Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:36.2311870Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:36.2313745Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:36.2315419Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:36.2317196Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:36.2319031Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:36.2320832Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:36.2322554Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:36.2324307Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:36.2326074Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:36.2327940Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:36.2329875Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:36.2331856Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:36.2333629Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:36.2335384Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:36.2337310Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:36.2339147Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:36.2340904Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:36.2342726Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:36.2344620Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:36.2346504Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:36.2348420Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:36.2350218Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:36.2351994Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:36.2353666Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:36.2355363Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:36.2357107Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:36.2358829Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:36.2360603Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:36.2362408Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:36.2364202Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:36.2365973Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:36.2367651Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:36.2369330Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:36.2371191Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:36.2373179Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:36.2375240Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:36.2377202Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:36.2378992Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:36.2380817Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:36.2382687Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:36.2384575Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:36.2386510Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:36.2388309Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:36.2390092Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:36.2391892Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:36.2393702Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:36.2395441Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:36.2397318Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:36.2399097Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:36.2400869Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:36.2402572Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:36.2404279Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:36.2406043Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:36.2407835Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:36.2409792Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:36.2411592Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:36.2413396Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:36.2415214Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:36.2417153Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:36.2418994Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:36.2420805Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:36.2422492Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:36.2424244Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:36.2425958Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:36.2427719Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:36.2429520Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:36.2431339Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:36.2433141Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:36.2434962Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:36.2436713Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:36.2438496Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:36.2440380Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:36.2442195Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:36.2443988Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:36.2445757Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:36.2447536Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:36.2449236Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:36.2450971Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:36.2452670Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:36.2454320Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:36.2456053Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:36.2457883Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:36.2459779Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:36.2461597Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:36.2463346Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:36.2465049Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:36.2466780Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:36.2468556Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:36.2470358Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:36.2472211Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:36.2474046Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:36.2475853Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:36.2477633Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:36.2479448Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:36.2481235Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:36.2483003Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:36.2484765Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:36.2486579Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:36.2488386Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:36.2490157Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:36.2491977Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:36.2493841Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:36.2495565Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:36.2497581Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:36.2502435Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:36.2504226Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:36.2506001Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:36.2507938Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:36.2510029Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:36.2512178Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:36.2514272Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:36.2516449Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:36.2518560Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:36.2520712Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:36.2522827Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:36.2525005Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:36.2526991Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:36.2528782Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:36.2530567Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:36.2532528Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:36.2534422Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:36.2536187Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:36.2538057Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:36.2539752Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:36.2541458Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:36.2543204Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:36.2544934Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:36.2546652Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:36.2548482Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:36.2550339Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:36.2552240Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:36.2553965Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:36.2555666Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:36.2557410Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:36.2559148Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:36.2560911Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:36.2562734Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:36.2564496Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:36.2566282Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:36.2567997Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:36.2569665Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:36.2571397Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:36.2573205Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:36.2575129Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:36.2577081Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:36.2578757Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:36.2580464Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:36.2582266Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:36.2584091Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:36.2585956Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:36.2587722Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:36.2589601Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:36.2591564Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:36.2593462Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:36.2595247Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:36.2597281Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:36.2599135Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:36.2600915Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:36.2602670Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:36.2611783Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:36.2613660Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:36.2615650Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:36.2617460Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:36.2619170Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:36.2620924Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:36.2622628Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:36.2624306Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:36.2626030Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:36.2627858Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:36.2629627Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:36.2631419Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.2633294Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.2635143Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:36.2636919Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:36.2638755Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:36.2640593Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:36.2642478Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:36.2644376Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:36.2646250Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:36.2648102Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:36.2649961Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:36.2651728Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:36.2653509Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:36.2655289Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:36.2657146Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:36.2658931Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:36.2660726Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:36.2662570Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:36.2664384Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:36.2666162Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:36.2667904Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:36.2669795Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:36.2671690Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:36.2673560Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:36.2675490Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:36.2677484Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:36.2679448Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:36.2681408Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:36.2683294Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:36.2685179Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:36.2687023Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:36.2688922Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:36.2690848Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:36.2692720Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:36.2694532Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:36.2696483Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:36.2698393Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:36.2700187Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:36.2701974Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:36.2703776Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:36.2705571Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:36.2707364Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:36.2709175Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:36.2711058Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:36.2712948Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:36.2714709Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:36.2716548Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:36.2718519Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:36.2720392Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:36.2722228Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:36.2724109Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:36.2725958Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:36.2727777Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:36.2729657Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:36.2731507Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:36.2733238Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:36.2735028Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:36.2736794Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:36.2738620Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:36.2740362Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:36.2742150Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:36.2744081Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:36.2746041Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.2747942Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:36.2749809Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:36.2751696Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:36.2753601Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:36.2755396Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:36.2757325Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:36.2759278Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:36.2761088Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:36.2762852Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:36.2764598Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:36.2766505Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:36.2768453Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:36.2770438Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:36.2772322Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:36.2774124Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:36.2775935Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:36.2777845Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:36.2779791Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:36.2781774Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:36.2783622Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:36.2785448Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:36.2787381Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:36.2789296Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:36.2791174Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:36.2793078Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:36.2794916Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:36.2796887Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:36.2798671Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:36.2800555Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:36.2802596Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:36.2804406Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:36.2806153Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:36.2808029Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:36.2809983Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:36.2811947Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:36.2813889Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:36.2815903Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:36.2817898Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:36.2819774Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:36.2821671Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:36.2823664Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:36.2825599Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:36.2827447Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:36.2829256Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:36.2831049Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:36.2832882Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:36.2834753Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:36.2836672Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:36.2838577Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:36.2840355Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:36.2841999Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:36.2843754Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:36.2845525Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:36.2847283Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:36.2849057Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:36.2850825Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:36.2852635Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:36.2854609Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:36.2856586Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:36.2858537Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:36.2860254Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:36.2862033Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:36.2863866Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:36.2865706Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:36.2867565Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:36.2869402Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:36.2871299Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:36.2873144Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:36.2874933Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:36.2876732Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:36.2878610Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:36.2880527Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:36.2882384Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:36.2884093Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:36.2885930Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:36.2887810Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:36.2889529Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:36.2891207Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:36.2892941Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:36.2894668Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:36.2896609Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:36.2898451Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:36.2900125Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:36.2901858Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:36.2903673Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:36.2905588Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:36.2907463Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:36.2909347Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:36.2911144Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:36.2912906Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:36.2914674Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:36.2916384Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:36.2918046Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:36.2919757Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:36.2921438Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:36.2923192Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:36.2924999Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:36.2926911Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:36.2928664Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:36.2930321Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:36.2932022Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:36.2934924Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:36.2939519Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:36.2942891Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:36.2944714Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:36.2946604Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:36.2948361Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:36.2950295Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:36.2952257Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:36.2954180Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:36.2956144Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:36.2958085Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:36.2960010Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:36.2961999Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:36.2964041Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:36.2966059Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:36.2967926Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:36.2969824Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:36.2971821Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:36.2973772Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:36.2975713Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:36.2977793Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:36.2979633Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:36.2981270Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:36.2982866Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:36.2984471Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:36.2986054Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:36.2987628Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:36.2989411Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:36.2991279Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:36.2993113Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:36.2994815Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:36.2996537Z V1204 11:22:28.177000 100601 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.2998010Z I1204 11:22:28.177000 100601 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.2999307Z V1204 11:22:28.178000 100601 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:36.3000411Z V1204 11:22:28.178000 100601 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:36.3001530Z I1204 11:22:28.178000 100601 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 13 2025-12-04T11:24:36.3002958Z V1204 11:22:28.195000 100601 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:36.3004311Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.3005399Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:36.3007032Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.3008868Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.3010781Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.3012965Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.3015054Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.3017081Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:36.3018880Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:36.3020931Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:36.3022932Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:36.3024832Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.3026826Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.3028458Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.3029602Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.3030567Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.3031679Z V1204 11:22:28.202000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:36.3032961Z V1204 11:22:28.204000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:36.3034690Z V1204 11:22:28.205000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.3036332Z V1204 11:22:28.205000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.3038026Z V1204 11:22:28.206000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.3039678Z V1204 11:22:28.206000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.3041270Z V1204 11:22:28.207000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:36.3042857Z V1204 11:22:28.208000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.3044380Z V1204 11:22:28.209000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:36.3045915Z V1204 11:22:28.209000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.3047419Z V1204 11:22:28.211000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:36.3048906Z V1204 11:22:28.211000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.3050407Z V1204 11:22:28.213000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:36.3051829Z V1204 11:22:28.213000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.3053305Z V1204 11:22:28.215000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.3054793Z V1204 11:22:28.215000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.3056315Z V1204 11:22:28.216000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.3057888Z V1204 11:22:28.216000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.3059004Z V1204 11:22:28.217000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.3060245Z V1204 11:22:28.217000 100601 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 13 2025-12-04T11:24:36.3061704Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:36.3062682Z warnings.warn( 2025-12-04T11:24:36.3063329Z V1204 11:22:28.218000 100601 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:36.3064405Z _ TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False _ 2025-12-04T11:24:36.3065120Z Traceback (most recent call last): 2025-12-04T11:24:36.3065925Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 384, in test_decompose_mm_mixed_precision 2025-12-04T11:24:36.3066785Z self.assertEqual( 2025-12-04T11:24:36.3067431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:36.3068201Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:36.3069028Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:36.3069917Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:36.3070402Z AssertionError: Scalars are not equal! 2025-12-04T11:24:36.3070660Z 2025-12-04T11:24:36.3070770Z Expected 1 but got 0. 2025-12-04T11:24:36.3071064Z Absolute difference: 1 2025-12-04T11:24:36.3071368Z Relative difference: 1.0 2025-12-04T11:24:36.3071561Z 2025-12-04T11:24:36.3071794Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:36.3073012Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:36.3074003Z 2025-12-04T11:24:36.3074276Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:36.3074911Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:36.3075388Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:36.3075713Z stats [('calls_captured', 1)] 2025-12-04T11:24:36.3076154Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:36.3077282Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:36.3078300Z graph_break [] 2025-12-04T11:24:36.3078686Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:36.3079492Z V1204 11:22:27.905000 100601 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpf4czj7g6 2025-12-04T11:24:36.3080527Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.3081598Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:36.3083329Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.3085157Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.3086534Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:36.3087673Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:36.3088713Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.3090309Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.3092222Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:36.3093587Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:36.3094660Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.3095615Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.3097071Z V1204 11:22:28.167000 100601 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:36.3098659Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:36.3100188Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:36.3101344Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.3102061Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.3102767Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.3103652Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:36.3105090Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.3106872Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.3108514Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:36.3110057Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.3111700Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.3113042Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:36.3113984Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.3115004Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:36.3118057Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.3122850Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.3125942Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:36.3127666Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:36.3129440Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:36.3131117Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:36.3132856Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:36.3134529Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:36.3136095Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:36.3137736Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:36.3139316Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:36.3140952Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:36.3142499Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:36.3144014Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:36.3145713Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:36.3147557Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:36.3149693Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:36.3151436Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:36.3181449Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:36.3212122Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:36.3214033Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:36.3215728Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:36.3217534Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:36.3219364Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:36.3221070Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:36.3222707Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:36.3224487Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:36.3226200Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:36.3227968Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:36.3229805Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:36.3231650Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:36.3233501Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:36.3235418Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:36.3237346Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:36.3239255Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:36.3241073Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:36.3243087Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:36.3244998Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:36.3246724Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:36.3248447Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:36.3250171Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:36.3251931Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:36.3252642Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:36.3253450Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:36.3254196Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:36.3255013Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:36.3255852Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:36.3256615Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:36.3257466Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:36.3258239Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:36.3259020Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:36.3259835Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:36.3260660Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:36.3261418Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:36.3262215Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:36.3263009Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:36.3263824Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:36.3264610Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:36.3265416Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:36.3266198Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:36.3266960Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:36.3267777Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:36.3268594Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:36.3269425Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:36.3270253Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:36.3271120Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:36.3271912Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:36.3272674Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:36.3273423Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:36.3274247Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:36.3275165Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:36.3276009Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:36.3276825Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:36.3277673Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:36.3278422Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:36.3279650Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:36.3280460Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:36.3281321Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:36.3282226Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:36.3283039Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:36.3283868Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:36.3284714Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:36.3285591Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:36.3286447Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:36.3287388Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.3288225Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:36.3289188Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.3290002Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:36.3290940Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.3291830Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:36.3292732Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:36.3293492Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:36.3294256Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:36.3295100Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:36.3296171Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:36.3296991Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:36.3297830Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:36.3298620Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:36.3299459Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:36.3300340Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:36.3301266Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:36.3302140Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:36.3302984Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:36.3303841Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:36.3304725Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:36.3305534Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:36.3306324Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:36.3307216Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:36.3308149Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:36.3309054Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:36.3309931Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:36.3310732Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:36.3311572Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:36.3312306Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:36.3313097Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:36.3313911Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:36.3314706Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:36.3315550Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:36.3316382Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:36.3317202Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:36.3318017Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:36.3318782Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:36.3319598Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:36.3320509Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:36.3321439Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:36.3322365Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:36.3323188Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:36.3324041Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:36.3324914Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:36.3325757Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:36.3326657Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:36.3327555Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:36.3328338Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:36.3329162Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:36.3329991Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:36.3330825Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:36.3331611Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:36.3332406Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:36.3333252Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:36.3334037Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:36.3334860Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:36.3335676Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:36.3336493Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:36.3337403Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:36.3338266Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:36.3339060Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:36.3339980Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:36.3340812Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:36.3341685Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:36.3342496Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:36.3343296Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:36.3344085Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:36.3344884Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:36.3345663Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:36.3346497Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:36.3347350Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:36.3348182Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:36.3349014Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:36.3349865Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:36.3350701Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:36.3351604Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:36.3352459Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:36.3353272Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:36.3354085Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:36.3354917Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:36.3355694Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:36.3356506Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:36.3357312Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:36.3358087Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:36.3358838Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:36.3359659Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:36.3360513Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:36.3361361Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:36.3362194Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:36.3362991Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:36.3363771Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:36.3364591Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:36.3365409Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:36.3366252Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:36.3367164Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:36.3367996Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:36.3368831Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:36.3369631Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:36.3370449Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:36.3371285Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:36.3372110Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:36.3372923Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:36.3373804Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:36.3374602Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:36.3375465Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:36.3376327Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:36.3377221Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:36.3377978Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:36.3378807Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:36.3379615Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:36.3380403Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:36.3381243Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:36.3382200Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:36.3383251Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:36.3384237Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:36.3385217Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:36.3386189Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:36.3387192Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:36.3388239Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:36.3389247Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:36.3390256Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:36.3391096Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:36.3391969Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:36.3392793Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:36.3393704Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:36.3394554Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:36.3395337Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:36.3396286Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:36.3397046Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:36.3397863Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:36.3398648Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:36.3399512Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:36.3400315Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:36.3401201Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:36.3402027Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:36.3402841Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:36.3403618Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:36.3404480Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:36.3405311Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:36.3406078Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:36.3406913Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:36.3407767Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:36.3408625Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:36.3409463Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:36.3410191Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:36.3410989Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:36.3411779Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:36.3412615Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:36.3413579Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:36.3414374Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:36.3415155Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:36.3415989Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:36.3416819Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:36.3417726Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:36.3418538Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:36.3419363Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:36.3420338Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:36.3421255Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:36.3422094Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:36.3422893Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:36.3423799Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:36.3424633Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:36.3425424Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:36.3426244Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:36.3427079Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:36.3427939Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:36.3428779Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:36.3429521Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:36.3430328Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:36.3431174Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:36.3431933Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:36.3432736Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:36.3433509Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:36.3434354Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:36.3435144Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:36.3436009Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.3436918Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.3437722Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:36.3438591Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:36.3439432Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:36.3440429Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:36.3441286Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:36.3442198Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:36.3443038Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:36.3443911Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:36.3444730Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:36.3445552Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:36.3446389Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:36.3447257Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:36.3448100Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:36.3448919Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:36.3449764Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:36.3450602Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:36.3451454Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:36.3452287Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:36.3453101Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:36.3454038Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:36.3454895Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:36.3455762Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:36.3456716Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:36.3457684Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:36.3458629Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:36.3459538Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:36.3460391Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:36.3461255Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:36.3462102Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:36.3463021Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:36.3463937Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:36.3464810Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:36.3465628Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:36.3466465Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:36.3467270Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:36.3468134Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:36.3468951Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:36.3469815Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:36.3470657Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:36.3471477Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:36.3472381Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:36.3473260Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:36.3474089Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:36.3474886Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:36.3475793Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:36.3476704Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:36.3477534Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:36.3478407Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:36.3479279Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:36.3480163Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:36.3481016Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:36.3481900Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:36.3482694Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:36.3483504Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:36.3484358Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:36.3485182Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:36.3486015Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:36.3486817Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:36.3487680Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:36.3488659Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:36.3489539Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.3490391Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:36.3491272Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:36.3492153Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:36.3493016Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:36.3493828Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:36.3494800Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:36.3495671Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:36.3496673Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:36.3497526Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:36.3498370Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:36.3499271Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:36.3500178Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:36.3501192Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:36.3502034Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:36.3502876Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:36.3503711Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:36.3504605Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:36.3505562Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:36.3506457Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:36.3507296Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:36.3508153Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:36.3509070Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:36.3509925Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:36.3510805Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:36.3511607Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:36.3512530Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:36.3513369Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:36.3514180Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:36.3515119Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:36.3515982Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:36.3516803Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:36.3517633Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:36.3518560Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:36.3519488Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:36.3520422Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:36.3521344Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:36.3522300Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:36.3523156Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:36.3524027Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:36.3524924Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:36.3525850Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:36.3526720Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:36.3527564Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:36.3528436Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:36.3529269Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:36.3530150Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:36.3531017Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:36.3531902Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:36.3532795Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:36.3533591Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:36.3534397Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:36.3535210Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:36.3536016Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:36.3536906Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:36.3537765Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:36.3538570Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:36.3539446Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:36.3540418Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:36.3541308Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:36.3542140Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:36.3542905Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:36.3543768Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:36.3544647Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:36.3545517Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:36.3546389Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:36.3547230Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:36.3548116Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:36.3548942Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:36.3549801Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:36.3550679Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:36.3551556Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:36.3552451Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:36.3553311Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:36.3554048Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:36.3554982Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:36.3555798Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:36.3556577Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:36.3557356Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:36.3558148Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:36.3558944Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:36.3559821Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:36.3560629Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:36.3561387Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:36.3562229Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:36.3563074Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:36.3563939Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:36.3564817Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:36.3565717Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:36.3566540Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:36.3567338Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:36.3568167Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:36.3568918Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:36.3569732Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:36.3570498Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:36.3571303Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:36.3572122Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:36.3572976Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:36.3573816Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:36.3574591Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:36.3575363Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:36.3576191Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:36.3578261Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:36.3580693Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:36.3581479Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:36.3582425Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:36.3583266Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:36.3584103Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:36.3585076Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:36.3585954Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:36.3586865Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:36.3587778Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:36.3588668Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:36.3589583Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:36.3590484Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:36.3591490Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:36.3592372Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:36.3593289Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:36.3594223Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:36.3595146Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:36.3596186Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:36.3597065Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:36.3598054Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:36.3598853Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:36.3599616Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:36.3600359Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:36.3601074Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:36.3601854Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:36.3602580Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:36.3603471Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:36.3604318Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:36.3605184Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:36.3605906Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:36.3606561Z V1204 11:22:28.177000 100601 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.3607248Z I1204 11:22:28.177000 100601 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.3607735Z V1204 11:22:28.178000 100601 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:36.3608236Z V1204 11:22:28.178000 100601 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:36.3608789Z I1204 11:22:28.178000 100601 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 13 2025-12-04T11:24:36.3609559Z V1204 11:22:28.195000 100601 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:36.3610021Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.3610522Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:36.3611495Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.3612243Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.3613325Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.3614323Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.3615316Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.3616152Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:36.3617060Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:36.3618128Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:36.3618954Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:36.3619897Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.3620795Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.3621407Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.3621834Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.3622238Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.3622852Z V1204 11:22:28.202000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:36.3623427Z V1204 11:22:28.204000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:36.3624458Z V1204 11:22:28.205000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.3624959Z V1204 11:22:28.205000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.3625986Z V1204 11:22:28.206000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.3626492Z V1204 11:22:28.206000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.3627478Z V1204 11:22:28.207000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:36.3627959Z V1204 11:22:28.208000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.3628892Z V1204 11:22:28.209000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:36.3629359Z V1204 11:22:28.209000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.3630271Z V1204 11:22:28.211000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:36.3630748Z V1204 11:22:28.211000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.3631600Z V1204 11:22:28.213000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:36.3632045Z V1204 11:22:28.213000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.3632942Z V1204 11:22:28.215000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.3633392Z V1204 11:22:28.215000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.3634300Z V1204 11:22:28.216000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.3634752Z V1204 11:22:28.216000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.3635276Z V1204 11:22:28.217000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.3635854Z V1204 11:22:28.217000 100601 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 13 2025-12-04T11:24:36.3636625Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:36.3636742Z warnings.warn( 2025-12-04T11:24:36.3637225Z V1204 11:22:28.218000 100601 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:36.3637471Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:36.3637583Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:36.3637698Z stats [('calls_captured', 1)] 2025-12-04T11:24:36.3637935Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:36.3638688Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:36.3638798Z graph_break [] 2025-12-04T11:24:36.3639019Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:36.3639465Z V1204 11:22:28.260000 100601 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpw391jgfi 2025-12-04T11:24:36.3639919Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.3640454Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:36.3641452Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.3642178Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.3642675Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:36.3643246Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:36.3643669Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.3644712Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.3645449Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:36.3645941Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:36.3646361Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.3646746Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.3647479Z V1204 11:22:28.509000 100601 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:36.3648190Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:36.3648862Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:36.3649173Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.3649451Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.3649744Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.3650213Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:36.3651045Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.3651860Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.3652490Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:36.3653272Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.3654056Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.3654549Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:36.3654845Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.3655437Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:36.3657865Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.3660229Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.3660880Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:36.3661813Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:36.3662537Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:36.3663354Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:36.3664186Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:36.3664899Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:36.3665632Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:36.3666356Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:36.3667083Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:36.3667817Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:36.3668497Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:36.3669231Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:36.3670132Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:36.3670929Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:36.3672085Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:36.3672884Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:36.3702137Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:36.3703177Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:36.3703965Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:36.3704734Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:36.3705591Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:36.3706434Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:36.3707197Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:36.3708125Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:36.3708910Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:36.3709700Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:36.3710606Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:36.3711464Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:36.3712324Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:36.3713169Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:36.3714029Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:36.3714931Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:36.3715803Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:36.3716615Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:36.3717689Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:36.3718451Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:36.3719301Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:36.3720045Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:36.3720876Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:36.3721647Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:36.3722360Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:36.3723192Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:36.3723989Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:36.3724752Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:36.3725588Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:36.3726376Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:36.3727171Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:36.3727942Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:36.3728720Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:36.3729501Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:36.3730292Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:36.3731052Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:36.3731855Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:36.3732640Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:36.3733439Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:36.3734226Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:36.3735034Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:36.3735819Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:36.3736576Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:36.3737457Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:36.3738274Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:36.3739133Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:36.3739960Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:36.3740825Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:36.3741614Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:36.3742407Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:36.3743159Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:36.3743974Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:36.3744898Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:36.3745710Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:36.3746504Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:36.3747351Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:36.3748102Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:36.3749316Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:36.3750126Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:36.3750998Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:36.3751906Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:36.3752719Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:36.3753543Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:36.3754423Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:36.3755337Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:36.3756150Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:36.3757090Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.3757930Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:36.3758926Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.3759741Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:36.3760682Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.3761516Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:36.3762397Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:36.3763157Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:36.3763921Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:36.3764761Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:36.3765650Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:36.3766415Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:36.3767249Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:36.3768034Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:36.3768870Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:36.3769753Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:36.3770719Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:36.3771621Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:36.3772397Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:36.3773242Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:36.3774150Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:36.3774951Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:36.3775747Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:36.3776637Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:36.3777585Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:36.3778458Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:36.3779344Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:36.3780148Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:36.3780967Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:36.3781738Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:36.3782539Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:36.3783364Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:36.3784152Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:36.3784993Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:36.3785827Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:36.3786710Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:36.3787552Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:36.3788295Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:36.3789108Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:36.3790025Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:36.3791006Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:36.3791937Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:36.3792754Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:36.3793577Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:36.3794417Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:36.3795261Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:36.3796326Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:36.3797196Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:36.3798045Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:36.3798872Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:36.3799697Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:36.3800531Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:36.3801313Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:36.3802101Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:36.3803005Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:36.3803829Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:36.3804614Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:36.3805420Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:36.3806233Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:36.3807126Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:36.3807990Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:36.3808782Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:36.3809660Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:36.3810460Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:36.3811324Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:36.3812156Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:36.3812940Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:36.3813736Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:36.3814537Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:36.3815327Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:36.3816165Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:36.3817081Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:36.3817929Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:36.3818788Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:36.3819673Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:36.3820436Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:36.3821330Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:36.3822183Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:36.3823029Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:36.3823839Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:36.3824671Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:36.3825445Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:36.3826225Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:36.3826999Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:36.3827783Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:36.3828529Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:36.3829354Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:36.3830202Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:36.3831048Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:36.3831882Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:36.3832665Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:36.3833446Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:36.3834263Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:36.3835128Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:36.3835999Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:36.3836876Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:36.3837703Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:36.3839131Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:36.3839938Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:36.3840763Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:36.3841605Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:36.3842401Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:36.3843189Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:36.3844074Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:36.3844868Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:36.3845693Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:36.3846610Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:36.3847437Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:36.3848190Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:36.3849019Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:36.3849822Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:36.3850613Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:36.3851479Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:36.3852468Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:36.3853479Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:36.3854461Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:36.3855467Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:36.3856443Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:36.3857516Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:36.3858526Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:36.3859510Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:36.3860521Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:36.3861357Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:36.3862195Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:36.3863047Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:36.3863981Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:36.3864820Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:36.3865603Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:36.3866397Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:36.3867151Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:36.3867998Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:36.3868812Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:36.3869616Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:36.3870403Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:36.3871321Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:36.3872150Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:36.3872968Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:36.3873737Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:36.3874547Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:36.3875329Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:36.3876095Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:36.3876931Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:36.3877778Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:36.3878599Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:36.3879436Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:36.3880169Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:36.3880973Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:36.3881765Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:36.3882596Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:36.3883584Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:36.3884415Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:36.3885162Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:36.3886003Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:36.3886829Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:36.3887715Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:36.3888526Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:36.3889347Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:36.3890295Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:36.3891184Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:36.3892029Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:36.3892831Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:36.3893736Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:36.3894566Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:36.3895365Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:36.3896343Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:36.3897256Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:36.3898119Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:36.3898950Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:36.3899756Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:36.3900610Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:36.3901422Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:36.3902178Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:36.3902985Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:36.3903829Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:36.3904681Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:36.3905466Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:36.3906359Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.3907227Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.3908000Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:36.3908877Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:36.3909715Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:36.3910629Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:36.3911495Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:36.3912403Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:36.3913243Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:36.3914113Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:36.3914924Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:36.3915784Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:36.3916652Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:36.3917466Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:36.3918308Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:36.3919132Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:36.3920014Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:36.3920852Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:36.3921706Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:36.3922516Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:36.3923323Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:36.3924229Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:36.3925086Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:36.3925957Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:36.3926919Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:36.3927825Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:36.3928763Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:36.3929673Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:36.3930531Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:36.3931390Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:36.3932266Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:36.3933218Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:36.3934100Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:36.3934965Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:36.3935810Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:36.3936645Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:36.3937514Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:36.3938384Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:36.3939170Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:36.3939995Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:36.3940841Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:36.3941657Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:36.3942530Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:36.3943438Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:36.3944271Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:36.3945075Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:36.3945970Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:36.3946876Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:36.3947702Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:36.3948605Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:36.3949511Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:36.3950360Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:36.3951209Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:36.3952140Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:36.3952921Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:36.3953726Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:36.3954575Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:36.3955368Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:36.3956176Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:36.3956982Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:36.3957836Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:36.3958816Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:36.3959701Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.3960552Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:36.3961431Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:36.3962306Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:36.3963168Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:36.3964088Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:36.3965103Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:36.3965937Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:36.3966776Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:36.3967565Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:36.3968433Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:36.3969333Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:36.3970238Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:36.3971180Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:36.3971997Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:36.3972831Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:36.3973664Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:36.3974560Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:36.3975532Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:36.3976424Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:36.3977320Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:36.3978178Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:36.3979090Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:36.3979950Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:36.3980862Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:36.3981696Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:36.3982561Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:36.3983402Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:36.3984244Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:36.3985182Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:36.3986041Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:36.3986858Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:36.3987655Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:36.3988552Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:36.3989475Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:36.3990392Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:36.3991305Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:36.3992256Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:36.3993113Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:36.3993976Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:36.3994870Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:36.3995793Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:36.3997020Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:36.3997913Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:36.3998745Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:36.3999577Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:36.4000451Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:36.4001366Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:36.4002246Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:36.4003152Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:36.4003894Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:36.4004666Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:36.4005479Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:36.4006288Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:36.4007115Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:36.4007971Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:36.4008776Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:36.4009653Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:36.4010619Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:36.4011502Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:36.4012325Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:36.4013122Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:36.4014016Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:36.4014859Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:36.4015717Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:36.4016621Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:36.4017531Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:36.4018423Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:36.4019243Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:36.4020064Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:36.4020914Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:36.4021790Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:36.4022681Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:36.4023513Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:36.4024289Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:36.4025323Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:36.4026145Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:36.4026918Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:36.4027697Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:36.4028491Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:36.4029332Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:36.4030240Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:36.4031009Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:36.4031761Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:36.4032604Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:36.4033490Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:36.4034356Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:36.4035228Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:36.4036099Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:36.4036894Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:36.4037700Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:36.4038517Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:36.4039262Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:36.4040072Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:36.4040834Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:36.4041645Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:36.4042462Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:36.4043326Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:36.4044151Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:36.4044952Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:36.4045752Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:36.4046545Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:36.4048534Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:36.4050978Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:36.4058974Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:36.4060018Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:36.4060839Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:36.4061680Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:36.4062640Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:36.4063611Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:36.4064528Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:36.4065450Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:36.4066341Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:36.4067263Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:36.4068163Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:36.4069209Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:36.4070130Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:36.4070959Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:36.4071908Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:36.4072870Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:36.4073768Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:36.4074648Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:36.4075629Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:36.4076371Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:36.4077081Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:36.4077837Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:36.4078551Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:36.4079280Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:36.4080031Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:36.4080924Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:36.4081767Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:36.4082611Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:36.4083348Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:36.4084004Z V1204 11:22:28.519000 100601 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.4084720Z I1204 11:22:28.519000 100601 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.4085235Z V1204 11:22:28.519000 100601 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:36.4085729Z V1204 11:22:28.520000 100601 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:36.4086240Z I1204 11:22:28.520000 100601 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 14 2025-12-04T11:24:36.4087006Z V1204 11:22:28.533000 100601 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:36.4087499Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.4088001Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:36.4088970Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.4089716Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.4090765Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.4091727Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.4092712Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.4093566Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:36.4094425Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:36.4095482Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:36.4096530Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:36.4097544Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.4098443Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.4099123Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.4099605Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.4100010Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.4100583Z V1204 11:22:28.540000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:36.4101148Z V1204 11:22:28.540000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:36.4102223Z V1204 11:22:28.541000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.4102731Z V1204 11:22:28.541000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.4103749Z V1204 11:22:28.542000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.4104251Z V1204 11:22:28.542000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.4105189Z V1204 11:22:28.543000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:36.4105655Z V1204 11:22:28.544000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.4106589Z V1204 11:22:28.545000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:36.4107051Z V1204 11:22:28.545000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.4107959Z V1204 11:22:28.547000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:36.4108443Z V1204 11:22:28.547000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.4109298Z V1204 11:22:28.549000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:36.4109736Z V1204 11:22:28.549000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.4110640Z V1204 11:22:28.551000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.4111089Z V1204 11:22:28.552000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.4111991Z V1204 11:22:28.552000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.4112471Z V1204 11:22:28.553000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.4113001Z V1204 11:22:28.553000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.4113617Z V1204 11:22:28.554000 100601 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 14 2025-12-04T11:24:36.4114362Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:36.4114478Z warnings.warn( 2025-12-04T11:24:36.4114962Z V1204 11:22:28.554000 100601 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:36.4115155Z =================================== FAILURES =================================== 2025-12-04T11:24:36.4115604Z _ TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False _ 2025-12-04T11:24:36.4115730Z Traceback (most recent call last): 2025-12-04T11:24:36.4116314Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 384, in test_decompose_mm_mixed_precision 2025-12-04T11:24:36.4116419Z self.assertEqual( 2025-12-04T11:24:36.4116878Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:36.4117054Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:36.4117588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:36.4117807Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:36.4117938Z AssertionError: Scalars are not equal! 2025-12-04T11:24:36.4117950Z 2025-12-04T11:24:36.4118058Z Expected 1 but got 0. 2025-12-04T11:24:36.4118176Z Absolute difference: 1 2025-12-04T11:24:36.4118283Z Relative difference: 1.0 2025-12-04T11:24:36.4118289Z 2025-12-04T11:24:36.4118518Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:36.4119365Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:36.4119371Z 2025-12-04T11:24:36.4119640Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:36.4119875Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:36.4119986Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:36.4120101Z stats [('calls_captured', 1)] 2025-12-04T11:24:36.4120370Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:36.4121133Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:36.4121248Z graph_break [] 2025-12-04T11:24:36.4121472Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:36.4121914Z V1204 11:22:27.905000 100601 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpf4czj7g6 2025-12-04T11:24:36.4122363Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.4122858Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:36.4123815Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.4124578Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.4125291Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:36.4125778Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:36.4126197Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.4127244Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.4128030Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:36.4128516Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:36.4128935Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.4129319Z V1204 11:22:27.928000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.4130043Z V1204 11:22:28.167000 100601 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:36.4130756Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:36.4131436Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:36.4131719Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.4131997Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.4132280Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.4132746Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:36.4133607Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.4134427Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.4135062Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:36.4135825Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.4136570Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.4137093Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:36.4137393Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.4138022Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:36.4140356Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.4142726Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.4143420Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:36.4144343Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:36.4145067Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:36.4145887Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:36.4146671Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:36.4147380Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:36.4148104Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:36.4148817Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:36.4149577Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:36.4150311Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:36.4150993Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:36.4151690Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:36.4152561Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:36.4153360Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:36.4154595Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:36.4155426Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:36.4184418Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:36.4185479Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:36.4186296Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:36.4187074Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:36.4187926Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:36.4188771Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:36.4189502Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:36.4190279Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:36.4191067Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:36.4191892Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:36.4192783Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:36.4193631Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:36.4194489Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:36.4195388Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:36.4196417Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:36.4197316Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:36.4198187Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:36.4199007Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:36.4200037Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:36.4200795Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:36.4201644Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:36.4202453Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:36.4203309Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:36.4204068Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:36.4204774Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:36.4205574Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:36.4206324Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:36.4207091Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:36.4207967Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:36.4208769Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:36.4209556Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:36.4210331Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:36.4211153Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:36.4211935Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:36.4212726Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:36.4213495Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:36.4214293Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:36.4215084Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:36.4215855Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:36.4216638Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:36.4217507Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:36.4218328Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:36.4219091Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:36.4219907Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:36.4220727Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:36.4221547Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:36.4222351Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:36.4223250Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:36.4224087Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:36.4224845Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:36.4225597Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:36.4226424Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:36.4227383Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:36.4228195Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:36.4228984Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:36.4229833Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:36.4230584Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:36.4231790Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:36.4232591Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:36.4233464Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:36.4234473Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:36.4235295Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:36.4236125Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:36.4236971Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:36.4237896Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:36.4238772Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:36.4239738Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.4240610Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:36.4241568Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.4242381Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:36.4243353Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.4244181Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:36.4245051Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:36.4245823Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:36.4246577Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:36.4247419Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:36.4248282Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:36.4249039Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:36.4249857Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:36.4250690Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:36.4251551Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:36.4252416Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:36.4253337Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:36.4254208Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:36.4254978Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:36.4255842Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:36.4256754Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:36.4257653Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:36.4258432Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:36.4259363Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:36.4260237Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:36.4261102Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:36.4261982Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:36.4262778Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:36.4263596Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:36.4264319Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:36.4265100Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:36.4265912Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:36.4266731Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:36.4267584Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:36.4268468Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:36.4269304Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:36.4270112Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:36.4270856Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:36.4271697Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:36.4272639Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:36.4273576Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:36.4274508Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:36.4275361Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:36.4276179Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:36.4277025Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:36.4277850Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:36.4278757Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:36.4279632Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:36.4280416Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:36.4281235Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:36.4282061Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:36.4282929Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:36.4283701Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:36.4284490Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:36.4285331Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:36.4286116Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:36.4286902Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:36.4287758Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:36.4288622Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:36.4289455Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:36.4290318Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:36.4291136Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:36.4292017Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:36.4292821Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:36.4293688Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:36.4294522Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:36.4295316Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:36.4296234Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:36.4297096Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:36.4297880Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:36.4298787Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:36.4299630Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:36.4300478Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:36.4301308Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:36.4302167Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:36.4302940Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:36.4303888Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:36.4304782Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:36.4305593Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:36.4306397Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:36.4307230Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:36.4308066Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:36.4308837Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:36.4309605Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:36.4310385Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:36.4311130Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:36.4311931Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:36.4312782Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:36.4313635Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:36.4314483Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:36.4315276Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:36.4316055Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:36.4316867Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:36.4317677Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:36.4318525Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:36.4319416Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:36.4320253Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:36.4321114Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:36.4321918Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:36.4322739Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:36.4323609Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:36.4324403Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:36.4325197Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:36.4326069Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:36.4326871Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:36.4327697Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:36.4328559Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:36.4329373Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:36.4330160Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:36.4330992Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:36.4331786Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:36.4332582Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:36.4333414Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:36.4334375Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:36.4335389Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:36.4336394Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:36.4337459Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:36.4338439Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:36.4339448Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:36.4340486Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:36.4341462Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:36.4342466Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:36.4343305Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:36.4344152Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:36.4344968Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:36.4345893Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:36.4346770Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:36.4347557Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:36.4348347Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:36.4349111Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:36.4349927Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:36.4350717Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:36.4351520Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:36.4352336Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:36.4353252Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:36.4354079Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:36.4354898Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:36.4355699Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:36.4356508Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:36.4357296Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:36.4358066Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:36.4358901Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:36.4359757Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:36.4360547Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:36.4361380Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:36.4362111Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:36.4362937Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:36.4363735Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:36.4364594Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:36.4365532Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:36.4366334Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:36.4367085Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:36.4367956Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:36.4368814Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:36.4369669Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:36.4370495Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:36.4371333Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:36.4372276Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:36.4373164Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:36.4374004Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:36.4374806Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:36.4375714Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:36.4376529Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:36.4377372Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:36.4378193Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:36.4379070Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:36.4379942Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:36.4380787Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:36.4381530Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:36.4382349Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:36.4383152Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:36.4383944Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:36.4384778Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:36.4385552Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:36.4386401Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:36.4387199Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:36.4388107Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.4388978Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.4389752Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:36.4390619Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:36.4391455Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:36.4392315Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:36.4393174Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:36.4394081Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:36.4394952Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:36.4395834Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:36.4396790Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:36.4397608Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:36.4398440Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:36.4399261Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:36.4400163Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:36.4401028Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:36.4401885Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:36.4402703Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:36.4403553Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:36.4404413Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:36.4405221Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:36.4406121Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:36.4406982Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:36.4407865Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:36.4408783Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:36.4409685Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:36.4410613Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:36.4411566Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:36.4412427Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:36.4413280Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:36.4414124Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:36.4415045Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:36.4415931Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:36.4416805Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:36.4417704Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:36.4418542Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:36.4419351Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:36.4420256Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:36.4421039Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:36.4421870Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:36.4422694Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:36.4423520Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:36.4424402Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:36.4425278Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:36.4426108Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:36.4426941Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:36.4427855Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:36.4428757Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:36.4429581Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:36.4430446Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:36.4431319Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:36.4432205Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:36.4433087Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:36.4433977Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:36.4434756Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:36.4435559Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:36.4436438Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:36.4437231Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:36.4438037Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:36.4438839Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:36.4439700Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:36.4440648Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:36.4441544Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.4442389Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:36.4443298Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:36.4444181Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:36.4445037Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:36.4445855Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:36.4446835Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:36.4447684Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:36.4448538Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:36.4449353Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:36.4450190Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:36.4451095Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:36.4452033Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:36.4452970Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:36.4453789Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:36.4454626Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:36.4455456Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:36.4456352Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:36.4457332Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:36.4458225Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:36.4459063Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:36.4459950Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:36.4460862Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:36.4461735Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:36.4462599Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:36.4463405Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:36.4464278Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:36.4465141Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:36.4465997Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:36.4466929Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:36.4467807Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:36.4468642Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:36.4469433Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:36.4470325Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:36.4471244Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:36.4472170Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:36.4473055Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:36.4474001Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:36.4474858Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:36.4475768Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:36.4476649Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:36.4477573Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:36.4478444Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:36.4479285Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:36.4480117Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:36.4480972Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:36.4481886Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:36.4482740Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:36.4483623Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:36.4484613Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:36.4485352Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:36.4486126Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:36.4486935Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:36.4487750Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:36.4488570Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:36.4489380Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:36.4490183Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:36.4491051Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:36.4492039Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:36.4492933Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:36.4493757Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:36.4494527Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:36.4495407Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:36.4496380Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:36.4497357Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:36.4498270Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:36.4499116Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:36.4500006Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:36.4500875Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:36.4501706Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:36.4502541Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:36.4503425Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:36.4504324Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:36.4505165Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:36.4505903Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:36.4506833Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:36.4507716Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:36.4508484Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:36.4509263Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:36.4510059Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:36.4510851Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:36.4511731Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:36.4512503Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:36.4513290Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:36.4514145Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:36.4514989Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:36.4515855Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:36.4516765Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:36.4517632Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:36.4518424Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:36.4519234Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:36.4520052Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:36.4520795Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:36.4521581Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:36.4522335Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:36.4523140Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:36.4523987Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:36.4524862Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:36.4525695Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:36.4526473Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:36.4527245Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:36.4528042Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:36.4530078Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:36.4532532Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:36.4533349Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:36.4534261Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:36.4535079Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:36.4535911Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:36.4536935Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:36.4537784Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:36.4538712Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:36.4539614Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:36.4540532Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:36.4541455Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:36.4542362Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:36.4543372Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:36.4544264Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:36.4545112Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:36.4546064Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:36.4547037Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:36.4547917Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:36.4548789Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:36.4549801Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:36.4550543Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:36.4551255Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:36.4551994Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:36.4552721Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:36.4553445Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:36.4554165Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:36.4555049Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:36.4555888Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:36.4556775Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:36.4557496Z V1204 11:22:28.176000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:36.4558159Z V1204 11:22:28.177000 100601 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.4558827Z I1204 11:22:28.177000 100601 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.4559312Z V1204 11:22:28.178000 100601 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:36.4559807Z V1204 11:22:28.178000 100601 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:36.4560313Z I1204 11:22:28.178000 100601 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 13 2025-12-04T11:24:36.4561121Z V1204 11:22:28.195000 100601 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:36.4561595Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.4562090Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:36.4563055Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.4563797Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.4564882Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.4565844Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.4566841Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.4567681Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:36.4568508Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:36.4569551Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:36.4570430Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:36.4571364Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.4572259Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.4572862Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.4573290Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.4573695Z V1204 11:22:28.200000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.4574267Z V1204 11:22:28.202000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:36.4574876Z V1204 11:22:28.204000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:36.4575913Z V1204 11:22:28.205000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.4576417Z V1204 11:22:28.205000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.4577491Z V1204 11:22:28.206000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.4578034Z V1204 11:22:28.206000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.4578995Z V1204 11:22:28.207000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:36.4579450Z V1204 11:22:28.208000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.4580402Z V1204 11:22:28.209000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:36.4580856Z V1204 11:22:28.209000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.4581773Z V1204 11:22:28.211000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:36.4582217Z V1204 11:22:28.211000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.4583063Z V1204 11:22:28.213000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:36.4583510Z V1204 11:22:28.213000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.4584448Z V1204 11:22:28.215000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.4584916Z V1204 11:22:28.215000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.4585806Z V1204 11:22:28.216000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.4586253Z V1204 11:22:28.216000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.4586776Z V1204 11:22:28.217000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.4587358Z V1204 11:22:28.217000 100601 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 13 2025-12-04T11:24:36.4588106Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:36.4588209Z warnings.warn( 2025-12-04T11:24:36.4588723Z V1204 11:22:28.218000 100601 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:36.4588992Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:36.4589108Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:36.4589239Z stats [('calls_captured', 1)] 2025-12-04T11:24:36.4589463Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:36.4590219Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:36.4590333Z graph_break [] 2025-12-04T11:24:36.4590585Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:36.4591023Z V1204 11:22:28.260000 100601 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpw391jgfi 2025-12-04T11:24:36.4591482Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.4591983Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:36.4592935Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.4593664Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.4594170Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:36.4594660Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:36.4595080Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.4596289Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.4597092Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:36.4597586Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:36.4598007Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.4598395Z V1204 11:22:28.280000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.4599123Z V1204 11:22:28.509000 100601 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:36.4599835Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:36.4600512Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:36.4600794Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.4601086Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.4601405Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.4601911Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:36.4602743Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.4603558Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.4604192Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:36.4605000Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.4605753Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.4606215Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:36.4606509Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.4607096Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:36.4609449Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.4611805Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.4612452Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:36.4613375Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:36.4614097Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:36.4614924Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:36.4615709Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:36.4616453Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:36.4617235Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:36.4617995Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:36.4618720Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:36.4619453Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:36.4620174Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:36.4620859Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:36.4621732Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:36.4622534Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:36.4623686Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:36.4624484Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:36.4653362Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:36.4654411Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:36.4655192Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:36.4655958Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:36.4656812Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:36.4657763Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:36.4658500Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:36.4659274Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:36.4660059Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:36.4660850Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:36.4661708Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:36.4662559Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:36.4663420Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:36.4664316Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:36.4665174Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:36.4666087Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:36.4666949Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:36.4667761Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:36.4668786Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:36.4669577Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:36.4670457Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:36.4671198Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:36.4672048Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:36.4672809Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:36.4673548Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:36.4674356Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:36.4675104Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:36.4675865Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:36.4676699Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:36.4677476Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:36.4678249Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:36.4679020Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:36.4679798Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:36.4680603Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:36.4681407Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:36.4682169Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:36.4682982Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:36.4683763Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:36.4684530Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:36.4685346Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:36.4686187Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:36.4686971Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:36.4687734Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:36.4688562Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:36.4689400Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:36.4690223Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:36.4691020Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:36.4691889Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:36.4692679Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:36.4693440Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:36.4694210Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:36.4695017Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:36.4696127Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:36.4697018Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:36.4697814Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:36.4698661Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:36.4699411Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:36.4700624Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:36.4701488Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:36.4702392Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:36.4703288Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:36.4704103Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:36.4704966Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:36.4705810Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:36.4706703Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:36.4707500Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:36.4708457Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.4709289Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:36.4710249Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.4711067Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:36.4712042Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.4712873Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:36.4713745Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:36.4714518Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:36.4715267Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:36.4716105Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:36.4716967Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:36.4717751Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:36.4718629Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:36.4719414Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:36.4720262Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:36.4721168Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:36.4722094Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:36.4722960Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:36.4723730Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:36.4724565Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:36.4725446Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:36.4726263Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:36.4727032Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:36.4727920Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:36.4728824Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:36.4729695Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:36.4730579Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:36.4731370Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:36.4732176Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:36.4732897Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:36.4733724Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:36.4734555Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:36.4735345Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:36.4736184Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:36.4737102Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:36.4737936Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:36.4738743Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:36.4739496Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:36.4740298Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:36.4741212Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:36.4742149Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:36.4743079Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:36.4743934Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:36.4744757Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:36.4745609Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:36.4746438Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:36.4747337Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:36.4748207Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:36.4748991Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:36.4749840Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:36.4750770Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:36.4751618Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:36.4752389Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:36.4753214Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:36.4754064Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:36.4754848Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:36.4755632Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:36.4756442Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:36.4757280Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:36.4758119Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:36.4758988Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:36.4759780Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:36.4760684Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:36.4761491Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:36.4762356Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:36.4763178Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:36.4763969Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:36.4764727Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:36.4765548Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:36.4766357Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:36.4767195Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:36.4768038Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:36.4768907Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:36.4769734Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:36.4770600Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:36.4771359Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:36.4772252Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:36.4773110Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:36.4773924Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:36.4774730Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:36.4775560Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:36.4776375Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:36.4777233Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:36.4778011Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:36.4778789Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:36.4779535Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:36.4780330Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:36.4781228Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:36.4782113Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:36.4782929Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:36.4783720Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:36.4784497Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:36.4785338Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:36.4786154Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:36.4786997Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:36.4787878Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:36.4788687Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:36.4789520Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:36.4790332Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:36.4791153Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:36.4792014Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:36.4792817Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:36.4793614Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:36.4794478Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:36.4795275Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:36.4796249Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:36.4797169Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:36.4797996Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:36.4798791Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:36.4799621Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:36.4800414Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:36.4801247Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:36.4802081Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:36.4803039Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:36.4804064Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:36.4805042Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:36.4806015Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:36.4806990Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:36.4808001Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:36.4809033Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:36.4810017Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:36.4811025Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:36.4811866Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:36.4812704Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:36.4813558Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:36.4814484Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:36.4815351Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:36.4816148Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:36.4816974Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:36.4817769Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:36.4818587Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:36.4819372Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:36.4820179Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:36.4820971Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:36.4821878Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:36.4822701Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:36.4823512Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:36.4824315Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:36.4825125Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:36.4825915Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:36.4826683Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:36.4827535Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:36.4828375Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:36.4829164Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:36.4830026Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:36.4830788Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:36.4831586Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:36.4832377Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:36.4833250Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:36.4834192Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:36.4834997Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:36.4835749Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:36.4836588Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:36.4837415Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:36.4838279Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:36.4839106Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:36.4839910Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:36.4840906Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:36.4841801Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:36.4842641Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:36.4843439Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:36.4844346Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:36.4845159Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:36.4845970Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:36.4846816Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:36.4847661Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:36.4848525Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:36.4849399Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:36.4850146Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:36.4850967Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:36.4851762Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:36.4852517Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:36.4853323Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:36.4854095Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:36.4854946Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:36.4855727Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:36.4856638Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.4857594Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.4858382Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:36.4859231Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:36.4860064Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:36.4860932Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:36.4861827Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:36.4862767Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:36.4863600Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:36.4864485Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:36.4865316Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:36.4866142Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:36.4866975Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:36.4867795Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:36.4868640Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:36.4869455Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:36.4870309Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:36.4871130Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:36.4871980Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:36.4872807Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:36.4873620Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:36.4874525Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:36.4875379Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:36.4876254Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:36.4877168Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:36.4878096Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:36.4879058Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:36.4879964Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:36.4880820Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:36.4881702Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:36.4882545Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:36.4883466Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:36.4884351Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:36.4885200Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:36.4886022Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:36.4886857Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:36.4887666Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:36.4888557Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:36.4889346Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:36.4890178Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:36.4891008Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:36.4891833Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:36.4892707Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:36.4893636Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:36.4894500Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:36.4895303Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:36.4896351Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:36.4897301Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:36.4898204Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:36.4899070Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:36.4899939Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:36.4900798Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:36.4901648Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:36.4902546Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:36.4903332Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:36.4904152Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:36.4905038Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:36.4905834Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:36.4906644Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:36.4907450Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:36.4908304Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:36.4909254Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:36.4910196Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.4911081Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:36.4911962Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:36.4912840Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:36.4913727Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:36.4914543Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:36.4915525Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:36.4916373Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:36.4917206Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:36.4917998Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:36.4918840Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:36.4919749Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:36.4920656Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:36.4921630Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:36.4922451Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:36.4923284Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:36.4924129Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:36.4925009Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:36.4925922Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:36.4926844Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:36.4927711Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:36.4928556Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:36.4929467Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:36.4930368Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:36.4931234Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:36.4932036Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:36.4932906Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:36.4933741Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:36.4934551Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:36.4935486Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:36.4936355Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:36.4937256Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:36.4938054Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:36.4938944Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:36.4939868Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:36.4940789Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:36.4941664Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:36.4942644Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:36.4943524Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:36.4944399Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:36.4945279Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:36.4946229Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:36.4947103Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:36.4947945Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:36.4948776Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:36.4949608Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:36.4950500Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:36.4951358Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:36.4952237Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:36.4953173Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:36.4953917Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:36.4954693Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:36.4955500Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:36.4956316Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:36.4957135Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:36.4957952Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:36.4958789Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:36.4959711Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:36.4960673Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:36.4961568Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:36.4962418Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:36.4963181Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:36.4964059Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:36.4964888Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:36.4965749Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:36.4966618Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:36.4967467Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:36.4968358Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:36.4969220Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:36.4970055Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:36.4970895Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:36.4971770Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:36.4972672Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:36.4973503Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:36.4974241Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:36.4975204Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:36.4976061Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:36.4976825Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:36.4977659Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:36.4978499Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:36.4979291Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:36.4980171Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:36.4980940Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:36.4981706Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:36.4982537Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:36.4983383Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:36.4984257Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:36.4985132Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:36.4986033Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:36.4986832Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:36.4987643Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:36.4988457Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:36.4989221Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:36.4989987Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:36.4990773Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:36.4991601Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:36.4992415Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:36.4993284Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:36.4994144Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:36.4994934Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:36.4995702Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:36.4996657Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:36.4998646Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:36.5001065Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:36.5001897Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:36.5002818Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:36.5003632Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:36.5004461Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:36.5005435Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:36.5006277Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:36.5007328Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:36.5008304Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:36.5009191Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:36.5010110Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:36.5011056Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:36.5012066Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:36.5012956Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:36.5013806Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:36.5014741Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:36.5015684Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:36.5016565Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:36.5017487Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:36.5018514Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:36.5019257Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:36.5019973Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:36.5020716Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:36.5021441Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:36.5022162Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:36.5022888Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:36.5023822Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:36.5024700Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:36.5025560Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:36.5026280Z V1204 11:22:28.518000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:36.5026971Z V1204 11:22:28.519000 100601 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.5027638Z I1204 11:22:28.519000 100601 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.5028124Z V1204 11:22:28.519000 100601 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:36.5028616Z V1204 11:22:28.520000 100601 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:36.5029120Z I1204 11:22:28.520000 100601 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 14 2025-12-04T11:24:36.5029901Z V1204 11:22:28.533000 100601 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:36.5030347Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.5030844Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:36.5031805Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.5032548Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.5033632Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.5034597Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.5035592Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.5036431Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:36.5037258Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:36.5038335Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:36.5039200Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:36.5040138Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.5041074Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.5041667Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.5042092Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.5042497Z V1204 11:22:28.537000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.5043069Z V1204 11:22:28.540000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:36.5043650Z V1204 11:22:28.540000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:36.5044657Z V1204 11:22:28.541000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.5045163Z V1204 11:22:28.541000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.5046187Z V1204 11:22:28.542000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.5046723Z V1204 11:22:28.542000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.5047682Z V1204 11:22:28.543000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:36.5048137Z V1204 11:22:28.544000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.5049084Z V1204 11:22:28.545000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:36.5049536Z V1204 11:22:28.545000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.5050446Z V1204 11:22:28.547000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:36.5050890Z V1204 11:22:28.547000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.5051767Z V1204 11:22:28.549000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:36.5052244Z V1204 11:22:28.549000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.5053135Z V1204 11:22:28.551000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.5053598Z V1204 11:22:28.552000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.5054516Z V1204 11:22:28.552000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.5054964Z V1204 11:22:28.553000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.5055493Z V1204 11:22:28.553000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.5056071Z V1204 11:22:28.554000 100601 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 14 2025-12-04T11:24:36.5056822Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:36.5056980Z warnings.warn( 2025-12-04T11:24:36.5057466Z V1204 11:22:28.554000 100601 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:36.5057712Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:36.5057829Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:36.5057969Z stats [('calls_captured', 1)] 2025-12-04T11:24:36.5058193Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:36.5058950Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:36.5059071Z graph_break [] 2025-12-04T11:24:36.5059292Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:36.5059732Z V1204 11:22:28.571000 100601 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmp0qko7afv 2025-12-04T11:24:36.5060220Z V1204 11:22:28.591000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.5060719Z V1204 11:22:28.591000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:36.5061677Z V1204 11:22:28.591000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.5062412Z V1204 11:22:28.591000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.5062915Z V1204 11:22:28.591000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:36.5063400Z V1204 11:22:28.591000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:36.5063823Z V1204 11:22:28.591000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.5064896Z V1204 11:22:28.591000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.5065661Z V1204 11:22:28.591000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:36.5066153Z V1204 11:22:28.591000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:36.5066571Z V1204 11:22:28.591000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.5066985Z V1204 11:22:28.591000 100601 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.5067713Z V1204 11:22:28.819000 100601 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:36.5068430Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:36.5069106Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:36.5069384Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.5069673Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.5069949Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.5070420Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:36.5071253Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.5072071Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.5072702Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:36.5073513Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.5074260Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.5074724Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:36.5075021Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.5075610Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:36.5077941Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.5080306Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.5080979Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:36.5081904Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:36.5082669Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:36.5083495Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:36.5084279Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:36.5084991Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:36.5085723Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:36.5086450Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:36.5087174Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:36.5087899Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:36.5088589Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:36.5089302Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:36.5090177Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:36.5090986Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:36.5092145Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:36.5092939Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:36.5122149Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:36.5123255Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:36.5124042Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:36.5124817Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:36.5125669Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:36.5126554Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:36.5127296Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:36.5128070Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:36.5128869Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:36.5129645Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:36.5130501Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:36.5131859Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:36.5132758Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:36.5133627Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:36.5134479Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:36.5135396Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:36.5136285Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:36.5137158Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:36.5138183Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:36.5138946Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:36.5139793Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:36.5140544Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:36.5141390Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:36.5142151Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:36.5142897Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:36.5143707Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:36.5144460Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:36.5145221Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:36.5146051Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:36.5146822Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:36.5147597Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:36.5148402Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:36.5149211Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:36.5149991Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:36.5150786Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:36.5151599Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:36.5152409Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:36.5153192Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:36.5153958Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:36.5154743Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:36.5155555Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:36.5156341Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:36.5157109Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:36.5157936Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:36.5158768Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:36.5159602Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:36.5160403Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:36.5161264Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:36.5162054Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:36.5162813Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:36.5163619Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:36.5164448Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:36.5165373Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:36.5166183Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:36.5166975Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:36.5167853Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:36.5168606Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:36.5169807Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:36.5170609Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:36.5171478Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:36.5172384Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:36.5173201Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:36.5174023Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:36.5174900Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:36.5175798Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:36.5176598Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:36.5177606Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.5178440Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:36.5179403Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.5180256Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:36.5181219Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.5182052Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:36.5182927Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:36.5183734Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:36.5184487Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:36.5185323Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:36.5186187Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:36.5186952Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:36.5187782Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:36.5188566Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:36.5189415Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:36.5190306Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:36.5191230Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:36.5192103Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:36.5192876Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:36.5193714Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:36.5194592Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:36.5195406Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:36.5196392Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:36.5197339Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:36.5198199Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:36.5199067Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:36.5199992Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:36.5200788Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:36.5201593Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:36.5202313Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:36.5203113Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:36.5203920Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:36.5204713Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:36.5205562Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:36.5206427Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:36.5207258Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:36.5208069Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:36.5208823Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:36.5209621Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:36.5210535Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:36.5211474Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:36.5212432Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:36.5213293Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:36.5214114Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:36.5214962Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:36.5215826Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:36.5216733Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:36.5217653Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:36.5218438Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:36.5219268Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:36.5220103Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:36.5220948Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:36.5221720Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:36.5222542Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:36.5223387Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:36.5224174Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:36.5224969Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:36.5225777Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:36.5226619Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:36.5227454Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:36.5228359Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:36.5229163Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:36.5230038Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:36.5230835Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:36.5231734Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:36.5232563Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:36.5233351Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:36.5234123Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:36.5234902Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:36.5235685Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:36.5236530Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:36.5237376Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:36.5238216Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:36.5239067Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:36.5239938Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:36.5240697Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:36.5241590Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:36.5242445Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:36.5243256Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:36.5244094Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:36.5244955Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:36.5245737Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:36.5246637Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:36.5247451Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:36.5248232Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:36.5248973Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:36.5249764Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:36.5250609Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:36.5251463Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:36.5252280Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:36.5253070Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:36.5253847Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:36.5254687Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:36.5255501Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:36.5256347Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:36.5257289Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:36.5258104Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:36.5258941Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:36.5259786Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:36.5260633Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:36.5261474Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:36.5262265Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:36.5263063Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:36.5263966Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:36.5264762Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:36.5265583Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:36.5266448Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:36.5267275Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:36.5268027Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:36.5268861Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:36.5269657Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:36.5270543Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:36.5271365Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:36.5272320Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:36.5273345Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:36.5274317Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:36.5275301Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:36.5276316Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:36.5277357Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:36.5278354Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:36.5279335Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:36.5280374Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:36.5281216Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:36.5282056Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:36.5282875Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:36.5283799Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:36.5284636Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:36.5285434Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:36.5286213Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:36.5287000Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:36.5287819Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:36.5288604Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:36.5289409Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:36.5290194Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:36.5291093Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:36.5291942Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:36.5292761Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:36.5293564Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:36.5294373Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:36.5295161Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:36.5295963Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:36.5297010Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:36.5297850Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:36.5298639Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:36.5299568Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:36.5300676Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:36.5301953Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:36.5302749Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:36.5303595Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:36.5304658Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:36.5305467Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:36.5306220Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:36.5307059Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:36.5307894Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:36.5308752Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:36.5309629Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:36.5310469Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:36.5311408Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:36.5312301Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:36.5313196Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:36.5314004Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:36.5314911Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:36.5315721Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:36.5316505Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:36.5317345Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:36.5318172Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:36.5319032Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:36.5319901Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:36.5320644Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:36.5321467Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:36.5322271Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:36.5323048Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:36.5323836Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:36.5324614Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:36.5325488Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:36.5326309Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:36.5327194Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.5328067Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.5328884Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:36.5329733Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:36.5330569Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:36.5331433Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:36.5332293Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:36.5333204Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:36.5334042Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:36.5334922Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:36.5335725Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:36.5336598Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:36.5337510Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:36.5338329Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:36.5339165Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:36.5339977Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:36.5340831Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:36.5341686Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:36.5342570Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:36.5343374Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:36.5344182Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:36.5345113Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:36.5345971Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:36.5346855Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:36.5347767Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:36.5348682Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:36.5349606Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:36.5350511Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:36.5351366Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:36.5352249Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:36.5353089Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:36.5354010Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:36.5354900Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:36.5355741Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:36.5356566Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:36.5357436Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:36.5358271Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:36.5359138Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:36.5359925Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:36.5360757Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:36.5361627Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:36.5362442Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:36.5363308Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:36.5364184Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:36.5365014Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:36.5365816Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:36.5366725Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:36.5367619Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:36.5368483Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:36.5369348Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:36.5370218Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:36.5371066Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:36.5371914Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:36.5372814Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:36.5373632Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:36.5374480Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:36.5375318Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:36.5376110Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:36.5376998Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:36.5377846Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:36.5378717Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:36.5379662Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:36.5380557Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.5381403Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:36.5382286Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:36.5383165Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:36.5384024Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:36.5384878Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:36.5385865Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:36.5386713Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:36.5387542Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:36.5388344Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:36.5389171Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:36.5390108Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:36.5391038Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:36.5391974Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:36.5392785Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:36.5393664Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:36.5394515Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:36.5395399Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:36.5396517Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:36.5397412Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:36.5398252Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:36.5399113Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:36.5400026Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:36.5400968Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:36.5401830Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:36.5402637Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:36.5403509Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:36.5404344Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:36.5405158Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:36.5406141Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:36.5407053Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:36.5407867Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:36.5408668Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:36.5409549Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:36.5410517Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:36.5411443Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:36.5412323Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:36.5413276Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:36.5414133Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:36.5415008Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:36.5415888Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:36.5416812Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:36.5417767Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:36.5418613Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:36.5419449Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:36.5420274Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:36.5421162Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:36.5422015Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:36.5422934Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:36.5423847Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:36.5424593Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:36.5425366Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:36.5426203Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:36.5427015Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:36.5427832Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:36.5428665Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:36.5429461Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:36.5430343Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:36.5431306Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:36.5432190Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:36.5433043Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:36.5433814Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:36.5434693Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:36.5435524Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:36.5436386Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:36.5437255Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:36.5438099Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:36.5439022Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:36.5439872Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:36.5440704Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:36.5441533Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:36.5442440Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:36.5443340Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:36.5449931Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:36.5450748Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:36.5451710Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:36.5452552Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:36.5453322Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:36.5454100Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:36.5454975Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:36.5455775Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:36.5456660Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:36.5457540Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:36.5458308Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:36.5459142Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:36.5459988Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:36.5460899Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:36.5461832Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:36.5462705Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:36.5463498Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:36.5464349Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:36.5465167Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:36.5465913Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:36.5466695Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:36.5467455Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:36.5468263Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:36.5469079Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:36.5469948Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:36.5470773Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:36.5471578Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:36.5472350Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:36.5473147Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:36.5475132Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:36.5477586Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:36.5478401Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:36.5479311Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:36.5480147Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:36.5480980Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:36.5481958Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:36.5482809Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:36.5483729Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:36.5484636Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:36.5485530Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:36.5486441Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:36.5487366Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:36.5488382Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:36.5489268Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:36.5490117Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:36.5491040Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:36.5491969Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:36.5492881Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:36.5493786Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:36.5494761Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:36.5495501Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:36.5496452Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:36.5497251Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:36.5497979Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:36.5498695Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:36.5499415Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:36.5500297Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:36.5501141Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:36.5502001Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:36.5502719Z V1204 11:22:28.828000 100601 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:36.5503453Z V1204 11:22:28.829000 100601 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.5504122Z I1204 11:22:28.829000 100601 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.5504610Z V1204 11:22:28.829000 100601 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:36.5505104Z V1204 11:22:28.830000 100601 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:36.5505609Z I1204 11:22:28.830000 100601 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 15 2025-12-04T11:24:36.5506386Z V1204 11:22:28.843000 100601 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:36.5506834Z V1204 11:22:28.847000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.5507333Z V1204 11:22:28.847000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:36.5508350Z V1204 11:22:28.847000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.5509141Z V1204 11:22:28.847000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.5510187Z V1204 11:22:28.847000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.5511146Z V1204 11:22:28.847000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.5512181Z V1204 11:22:28.847000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.5513021Z V1204 11:22:28.847000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:36.5513850Z V1204 11:22:28.847000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:36.5514887Z V1204 11:22:28.847000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:36.5515732Z V1204 11:22:28.847000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:36.5516667Z V1204 11:22:28.847000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.5517593Z V1204 11:22:28.847000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.5518196Z V1204 11:22:28.847000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.5518622Z V1204 11:22:28.847000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.5519026Z V1204 11:22:28.847000 100601 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.5519602Z V1204 11:22:28.850000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:36.5520181Z V1204 11:22:28.850000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:36.5521189Z V1204 11:22:28.851000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.5521738Z V1204 11:22:28.851000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.5522780Z V1204 11:22:28.852000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.5523281Z V1204 11:22:28.852000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.5524238Z V1204 11:22:28.853000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:36.5524693Z V1204 11:22:28.854000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.5525746Z V1204 11:22:28.855000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:36.5526198Z V1204 11:22:28.855000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.5527103Z V1204 11:22:28.857000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:36.5527552Z V1204 11:22:28.857000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.5528404Z V1204 11:22:28.859000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:36.5528858Z V1204 11:22:28.859000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.5529753Z V1204 11:22:28.861000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.5530216Z V1204 11:22:28.862000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.5531134Z V1204 11:22:28.862000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.5531582Z V1204 11:22:28.863000 100601 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.5532111Z V1204 11:22:28.863000 100601 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.5532691Z V1204 11:22:28.864000 100601 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 15 2025-12-04T11:24:36.5533448Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:36.5533551Z warnings.warn( 2025-12-04T11:24:36.5534028Z V1204 11:22:28.864000 100601 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:36.5534893Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-6aac4defd9d3a3cc.xml - 2025-12-04T11:24:36.5535071Z =========================== short test summary info ============================ 2025-12-04T11:24:36.5536022Z FAILED [0.3087s] inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False - AssertionError: Scalars are not equal! 2025-12-04T11:24:36.5536061Z 2025-12-04T11:24:36.5536169Z Expected 1 but got 0. 2025-12-04T11:24:36.5536278Z Absolute difference: 1 2025-12-04T11:24:36.5536398Z Relative difference: 1.0 2025-12-04T11:24:36.5536404Z 2025-12-04T11:24:36.5536623Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:36.5537541Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:36.5537548Z 2025-12-04T11:24:36.5537822Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:36.5538042Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:24:36.5538271Z ============ 1 failed, 13 passed, 16 deselected, 2 rerun in 13.73s ============= 2025-12-04T11:24:36.5538380Z --- Logging error --- 2025-12-04T11:24:36.5538509Z Traceback (most recent call last): 2025-12-04T11:24:36.5538837Z File "/opt/conda/envs/py_3.10/lib/python3.10/logging/__init__.py", line 1103, in emit 2025-12-04T11:24:36.5538964Z stream.write(msg + self.terminator) 2025-12-04T11:24:36.5539112Z ValueError: I/O operation on closed file. 2025-12-04T11:24:36.5539209Z Call stack: 2025-12-04T11:24:36.5539728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/remote_cache.py", line 432, in dump_cache_stats 2025-12-04T11:24:36.5539887Z log.info("Cache Metrics:%s", out.getvalue()) 2025-12-04T11:24:36.5539997Z Message: 'Cache Metrics:%s' 2025-12-04T11:24:36.5540618Z Arguments: ('\n LocalAutotuneCache: {hit: 0, miss: 0, put: 4, exception: 0}\n backend:_LocalAutotuneCacheBackend: {hit: 0, miss: 0, put: 4, exception: 0}\n',) 2025-12-04T11:24:36.5540730Z Got exit code 1 2025-12-04T11:24:36.5540835Z Retrying single test... 2025-12-04T11:24:36.5541515Z Test results will be stored in test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-041aad81ad32b1c4.xml 2025-12-04T11:24:36.5541685Z ============================= test session starts ============================== 2025-12-04T11:24:36.5542036Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:24:36.5542154Z cachedir: .pytest_cache 2025-12-04T11:24:36.5542676Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:24:36.5542799Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:24:36.5542919Z configfile: pytest.ini 2025-12-04T11:24:36.5543541Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:24:36.5544756Z collecting ... /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:52: PytestCollectionWarning: cannot collect test class 'TestDecomposeAddMM' because it has a __init__ constructor (from: test/inductor/test_decompose_mem_bound_mm.py) 2025-12-04T11:24:36.5544911Z class TestDecomposeAddMM(torch.nn.Module): 2025-12-04T11:24:36.5545059Z collected 37 items / 36 deselected / 1 selected 2025-12-04T11:24:36.5545996Z stepcurrent: skipping 29 already run items. Running only test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:36.5546110Z Running 1 items in this shard 2025-12-04T11:24:36.5546116Z 2025-12-04T11:24:36.5546937Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False ('RERUN', {'yellow': True}) [5.9447s] [100%] 2025-12-04T11:24:36.5547778Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False ('RERUN', {'yellow': True}) [0.3417s] [100%] 2025-12-04T11:24:36.5548537Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False FAILED [0.3294s] [100%] 2025-12-04T11:24:36.5548544Z 2025-12-04T11:24:36.5548689Z ==================================== RERUNS ==================================== 2025-12-04T11:24:36.5549137Z _ TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False _ 2025-12-04T11:24:36.5549268Z Traceback (most recent call last): 2025-12-04T11:24:36.5549835Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 384, in test_decompose_mm_mixed_precision 2025-12-04T11:24:36.5549953Z self.assertEqual( 2025-12-04T11:24:36.5550445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:36.5550610Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:36.5551164Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:36.5551372Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:36.5551502Z AssertionError: Scalars are not equal! 2025-12-04T11:24:36.5551507Z 2025-12-04T11:24:36.5551622Z Expected 1 but got 0. 2025-12-04T11:24:36.5551729Z Absolute difference: 1 2025-12-04T11:24:36.5551848Z Relative difference: 1.0 2025-12-04T11:24:36.5551853Z 2025-12-04T11:24:36.5552069Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:36.5552912Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:36.5552919Z 2025-12-04T11:24:36.5553200Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:36.5553428Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:36.5553554Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:36.5553667Z stats [('calls_captured', 1)] 2025-12-04T11:24:36.5554423Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:36.5554655Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:36.5554750Z graph_break [] 2025-12-04T11:24:36.5554971Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:36.5555510Z I1204 11:22:43.005000 100983 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:36.5556107Z I1204 11:22:43.253000 100983 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:36.5556560Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.5557054Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:36.5557992Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.5558733Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.5559225Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:36.5559754Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:36.5560200Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.5561233Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.5561979Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:36.5562483Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:36.5562912Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.5563297Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.5563866Z I1204 11:22:45.464000 100983 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:36.5564583Z V1204 11:22:45.952000 100983 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:36.5565312Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:36.5565976Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:36.5566260Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.5566555Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.5566830Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.5567316Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:36.5568130Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.5568980Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.5569619Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:36.5570386Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.5571149Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.5571600Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:36.5571901Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.5572485Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:36.5574851Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.5577264Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.5577947Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:36.5578870Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:36.5579589Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:36.5580418Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:36.5581204Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:36.5581933Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:36.5582651Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:36.5583373Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:36.5584128Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:36.5584861Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:36.5585554Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:36.5586236Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:36.5587123Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:36.5587916Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:36.5589097Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:36.5589931Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:36.5619110Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:36.5620222Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:36.5620991Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:36.5621823Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:36.5622670Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:36.5623517Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:36.5624246Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:36.5625021Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:36.5625812Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:36.5626639Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:36.5627545Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:36.5628375Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:36.5629233Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:36.5630087Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:36.5630969Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:36.5631882Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:36.5632736Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:36.5633556Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:36.5634562Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:36.5635323Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:36.5636170Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:36.5636917Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:36.5637798Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:36.5638569Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:36.5639284Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:36.5640077Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:36.5640826Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:36.5641591Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:36.5642460Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:36.5643261Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:36.5644039Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:36.5644823Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:36.5645590Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:36.5646422Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:36.5647210Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:36.5647973Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:36.5648781Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:36.5649561Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:36.5650340Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:36.5651108Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:36.5651916Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:36.5652693Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:36.5653481Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:36.5654306Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:36.5655103Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:36.5655933Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:36.5656715Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:36.5657645Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:36.5658468Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:36.5659260Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:36.5660019Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:36.5660824Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:36.5661754Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:36.5662580Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:36.5663371Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:36.5664215Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:36.5664969Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:36.5666177Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:36.5666973Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:36.5667852Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:36.5668774Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:36.5669601Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:36.5670414Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:36.5671263Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:36.5672156Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:36.5672960Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:36.5673911Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.5674763Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:36.5675764Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.5676564Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:36.5677499Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.5678358Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:36.5679231Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:36.5680001Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:36.5680755Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:36.5681599Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:36.5682449Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:36.5683216Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:36.5684027Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:36.5684844Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:36.5685696Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:36.5686559Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:36.5687498Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:36.5688358Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:36.5689138Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:36.5689970Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:36.5690879Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:36.5691720Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:36.5692491Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:36.5693395Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:36.5694283Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:36.5695161Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:36.5696193Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:36.5697052Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:36.5697856Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:36.5698582Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:36.5699382Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:36.5700182Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:36.5701049Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:36.5701885Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:36.5702711Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:36.5703546Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:36.5704354Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:36.5705108Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:36.5705904Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:36.5706871Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:36.5707843Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:36.5708766Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:36.5709588Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:36.5710455Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:36.5711305Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:36.5712128Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:36.5713042Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:36.5713895Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:36.5714695Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:36.5715501Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:36.5716332Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:36.5717203Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:36.5717973Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:36.5718779Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:36.5719610Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:36.5720399Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:36.5721175Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:36.5721984Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:36.5722847Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:36.5723709Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:36.5724579Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:36.5725356Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:36.5726275Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:36.5727059Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:36.5727923Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:36.5728745Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:36.5729529Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:36.5730307Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:36.5731087Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:36.5731872Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:36.5732697Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:36.5733568Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:36.5734415Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:36.5735240Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:36.5736109Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:36.5736920Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:36.5737850Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:36.5738727Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:36.5739570Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:36.5740379Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:36.5741211Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:36.5742031Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:36.5742800Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:36.5743594Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:36.5744360Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:36.5745107Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:36.5745905Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:36.5746753Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:36.5747613Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:36.5748429Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:36.5749265Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:36.5750036Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:36.5750870Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:36.5751670Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:36.5752509Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:36.5753405Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:36.5754252Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:36.5755132Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:36.5755926Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:36.5756765Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:36.5757589Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:36.5758421Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:36.5759223Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:36.5760094Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:36.5760909Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:36.5761725Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:36.5762603Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:36.5763416Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:36.5764168Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:36.5765030Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:36.5765833Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:36.5766643Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:36.5767465Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:36.5768436Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:36.5769442Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:36.5770464Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:36.5771473Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:36.5772445Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:36.5773456Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:36.5774507Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:36.5775499Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:36.5776489Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:36.5777398Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:36.5778225Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:36.5779041Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:36.5779965Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:36.5780800Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:36.5781630Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:36.5782421Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:36.5783193Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:36.5783991Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:36.5784779Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:36.5785585Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:36.5786404Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:36.5787399Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:36.5788212Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:36.5789040Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:36.5789798Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:36.5790640Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:36.5791431Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:36.5792202Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:36.5793048Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:36.5793887Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:36.5794695Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:36.5795524Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:36.5796434Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:36.5797301Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:36.5798095Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:36.5798943Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:36.5799890Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:36.5800696Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:36.5801432Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:36.5802282Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:36.5803138Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:36.5804029Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:36.5804846Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:36.5805653Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:36.5806655Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:36.5807530Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:36.5808389Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:36.5809176Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:36.5810082Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:36.5810897Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:36.5811680Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:36.5812504Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:36.5813359Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:36.5814235Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:36.5815057Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:36.5815799Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:36.5816616Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:36.5817488Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:36.5818257Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:36.5819085Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:36.5819905Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:36.5820744Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:36.5821534Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:36.5822445Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.5823324Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.5824111Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:36.5824961Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:36.5825806Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:36.5826660Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:36.5827519Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:36.5828425Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:36.5829317Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:36.5830202Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:36.5830994Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:36.5831826Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:36.5832648Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:36.5833473Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:36.5834297Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:36.5835137Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:36.5836019Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:36.5836836Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:36.5837702Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:36.5838524Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:36.5839342Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:36.5840231Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:36.5841084Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:36.5841968Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:36.5842884Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:36.5843798Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:36.5844714Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:36.5845661Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:36.5846503Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:36.5847360Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:36.5848209Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:36.5849125Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:36.5850014Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:36.5850891Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:36.5851751Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:36.5852577Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:36.5853400Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:36.5854254Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:36.5855070Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:36.5855904Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:36.5856726Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:36.5857620Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:36.5858482Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:36.5859374Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:36.5860191Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:36.5860996Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:36.5861941Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:36.5862846Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:36.5863686Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:36.5864543Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:36.5865427Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:36.5866266Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:36.5867147Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:36.5868068Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:36.5868852Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:36.5869665Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:36.5870541Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:36.5871344Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:36.5872140Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:36.5872940Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:36.5873796Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:36.5874745Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:36.5875638Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.5876477Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:36.5877363Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:36.5878252Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:36.5879127Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:36.5879927Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:36.5880904Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:36.5881750Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:36.5882582Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:36.5883418Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:36.5884274Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:36.5885183Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:36.5886074Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:36.5887063Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:36.5887886Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:36.5888719Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:36.5889569Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:36.5890452Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:36.5891383Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:36.5892263Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:36.5893113Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:36.5893983Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:36.5894902Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:36.5895773Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:36.5896808Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:36.5897701Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:36.5898565Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:36.5899487Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:36.5900327Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:36.5901268Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:36.5902145Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:36.5902999Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:36.5903804Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:36.5904681Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:36.5905620Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:36.5906536Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:36.5907423Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:36.5908380Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:36.5909236Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:36.5910171Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:36.5911058Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:36.5911998Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:36.5912864Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:36.5913719Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:36.5914542Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:36.5915376Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:36.5916310Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:36.5917199Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:36.5918092Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:36.5918976Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:36.5919769Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:36.5920529Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:36.5921341Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:36.5922164Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:36.5922983Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:36.5923816Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:36.5924606Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:36.5925502Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:36.5926487Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:36.5927378Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:36.5928199Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:36.5928967Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:36.5929840Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:36.5930671Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:36.5931577Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:36.5932462Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:36.5933322Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:36.5934198Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:36.5935023Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:36.5935888Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:36.5936723Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:36.5937673Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:36.5938562Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:36.5939409Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:36.5940135Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:36.5941072Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:36.5941889Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:36.5942687Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:36.5943479Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:36.5944263Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:36.5945069Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:36.5945937Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:36.5946711Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:36.5947475Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:36.5948349Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:36.5949251Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:36.5950096Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:36.5950990Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:36.5951877Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:36.5952668Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:36.5953479Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:36.5954294Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:36.5955056Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:36.5955825Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:36.5956595Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:36.5957382Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:36.5958192Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:36.5959091Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:36.5959917Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:36.5960704Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:36.5961464Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:36.5962278Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:36.5964268Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:36.5966726Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:36.5967525Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:36.5968458Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:36.5969256Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:36.5970103Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:36.5971065Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:36.5971906Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:36.5972825Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:36.5973726Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:36.5974652Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:36.5975557Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:36.5976474Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:36.5977543Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:36.5978426Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:36.5979274Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:36.5980240Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:36.5981206Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:36.5982087Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:36.5982978Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:36.5983979Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:36.5984725Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:36.5985431Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:36.5986177Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:36.5986905Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:36.5987626Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:36.5988360Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:36.5989227Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:36.5990076Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:36.5990948Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:36.5991667Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:36.5992332Z V1204 11:22:45.962000 100983 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.5993005Z I1204 11:22:45.963000 100983 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.5993501Z V1204 11:22:45.963000 100983 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:36.5993985Z V1204 11:22:45.963000 100983 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:36.5994490Z I1204 11:22:45.963000 100983 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:36.5995299Z V1204 11:22:46.004000 100983 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:36.5995774Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.5996502Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:36.5997459Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.5998219Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.5999322Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.6000294Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.6001280Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.6002121Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:36.6002955Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:36.6003998Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:36.6004876Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:36.6005807Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.6006716Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.6007305Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.6007744Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.6008134Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.6008711Z V1204 11:22:46.012000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:36.6009350Z V1204 11:22:46.013000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:36.6010395Z V1204 11:22:46.014000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.6010916Z V1204 11:22:46.014000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6011924Z V1204 11:22:46.046000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.6012468Z V1204 11:22:46.047000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6013404Z V1204 11:22:46.048000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:36.6013859Z V1204 11:22:46.049000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6014806Z V1204 11:22:46.050000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:36.6015259Z V1204 11:22:46.051000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6016185Z V1204 11:22:46.052000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:36.6016621Z V1204 11:22:46.052000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6017540Z V1204 11:22:46.055000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:36.6017980Z V1204 11:22:46.055000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6018905Z V1204 11:22:46.058000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.6019371Z V1204 11:22:46.058000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6020265Z V1204 11:22:46.059000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.6020729Z V1204 11:22:46.060000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6021242Z V1204 11:22:46.060000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.6021819Z V1204 11:22:46.061000 100983 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:36.6022570Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:36.6022677Z warnings.warn( 2025-12-04T11:24:36.6023206Z V1204 11:22:46.061000 100983 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:36.6023690Z _ TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False _ 2025-12-04T11:24:36.6023812Z Traceback (most recent call last): 2025-12-04T11:24:36.6024394Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 384, in test_decompose_mm_mixed_precision 2025-12-04T11:24:36.6024502Z self.assertEqual( 2025-12-04T11:24:36.6024974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:36.6025139Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:36.6025675Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:36.6025926Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:36.6026057Z AssertionError: Scalars are not equal! 2025-12-04T11:24:36.6026065Z 2025-12-04T11:24:36.6026174Z Expected 1 but got 0. 2025-12-04T11:24:36.6026296Z Absolute difference: 1 2025-12-04T11:24:36.6026405Z Relative difference: 1.0 2025-12-04T11:24:36.6026411Z 2025-12-04T11:24:36.6026638Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:36.6027491Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:36.6027496Z 2025-12-04T11:24:36.6027768Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:36.6028005Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:36.6028119Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:36.6028260Z stats [('calls_captured', 1)] 2025-12-04T11:24:36.6029026Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:36.6029250Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:36.6029364Z graph_break [] 2025-12-04T11:24:36.6029588Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:36.6030090Z I1204 11:22:43.005000 100983 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:36.6030702Z I1204 11:22:43.253000 100983 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:36.6031177Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.6031687Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:36.6032632Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.6033378Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.6033868Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:36.6034356Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:36.6034797Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.6035926Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.6036709Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:36.6037182Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:36.6037618Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.6038007Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.6038593Z I1204 11:22:45.464000 100983 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:36.6039329Z V1204 11:22:45.952000 100983 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:36.6040050Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:36.6040730Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:36.6041013Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.6041293Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.6041588Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.6042058Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:36.6042896Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.6043721Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.6044345Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:36.6045156Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.6045917Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.6046382Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:36.6046677Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.6047265Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:36.6049638Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.6052004Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.6052651Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:36.6053608Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:36.6054331Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:36.6055140Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:36.6055934Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:36.6056650Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:36.6057468Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:36.6058192Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:36.6058912Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:36.6059636Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:36.6060362Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:36.6061062Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:36.6061931Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:36.6062729Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:36.6063897Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:36.6064702Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:36.6093665Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:36.6094752Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:36.6095534Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:36.6096461Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:36.6097455Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:36.6098303Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:36.6099044Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:36.6099823Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:36.6100610Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:36.6101396Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:36.6102256Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:36.6103145Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:36.6104049Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:36.6104912Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:36.6105765Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:36.6106715Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:36.6107591Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:36.6108409Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:36.6109428Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:36.6110189Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:36.6111046Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:36.6111793Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:36.6112627Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:36.6113431Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:36.6114142Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:36.6114955Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:36.6115711Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:36.6116473Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:36.6117306Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:36.6118064Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:36.6118893Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:36.6119694Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:36.6120478Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:36.6121258Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:36.6122047Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:36.6122838Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:36.6123632Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:36.6124426Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:36.6125191Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:36.6125974Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:36.6126794Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:36.6127576Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:36.6128339Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:36.6129183Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:36.6130006Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:36.6130832Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:36.6131641Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:36.6132509Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:36.6133298Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:36.6134059Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:36.6134837Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:36.6135681Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:36.6136603Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:36.6137491Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:36.6138346Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:36.6139190Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:36.6139942Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:36.6141133Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:36.6141939Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:36.6142807Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:36.6143706Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:36.6144517Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:36.6145367Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:36.6146219Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:36.6147113Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:36.6147918Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:36.6148856Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.6149697Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:36.6150701Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.6151551Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:36.6152490Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.6153325Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:36.6154221Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:36.6154992Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:36.6155756Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:36.6156589Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:36.6157448Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:36.6158208Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:36.6159037Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:36.6159822Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:36.6160656Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:36.6161561Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:36.6162489Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:36.6163358Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:36.6164130Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:36.6164966Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:36.6165846Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:36.6166695Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:36.6167497Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:36.6168388Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:36.6169267Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:36.6170162Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:36.6171039Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:36.6171835Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:36.6172647Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:36.6173374Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:36.6174166Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:36.6174985Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:36.6175779Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:36.6176624Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:36.6177554Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:36.6178398Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:36.6179203Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:36.6179939Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:36.6180746Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:36.6181661Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:36.6182635Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:36.6183594Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:36.6184413Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:36.6185238Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:36.6186104Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:36.6186948Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:36.6187850Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:36.6188717Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:36.6189501Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:36.6190327Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:36.6191157Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:36.6191990Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:36.6192772Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:36.6193587Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:36.6194435Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:36.6195222Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:36.6196174Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:36.6196981Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:36.6197821Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:36.6198707Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:36.6199640Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:36.6200437Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:36.6201313Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:36.6202111Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:36.6203026Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:36.6203858Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:36.6204646Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:36.6205411Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:36.6206212Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:36.6207000Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:36.6207845Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:36.6208689Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:36.6209575Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:36.6210407Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:36.6211265Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:36.6212040Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:36.6212940Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:36.6213798Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:36.6214655Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:36.6215496Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:36.6216333Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:36.6217160Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:36.6217954Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:36.6218764Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:36.6219547Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:36.6220297Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:36.6221091Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:36.6221941Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:36.6222795Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:36.6223630Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:36.6224427Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:36.6225203Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:36.6226047Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:36.6226863Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:36.6227712Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:36.6228582Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:36.6229410Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:36.6230250Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:36.6231087Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:36.6231940Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:36.6232781Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:36.6233580Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:36.6234391Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:36.6235277Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:36.6236076Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:36.6236901Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:36.6237760Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:36.6238590Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:36.6239344Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:36.6240159Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:36.6240958Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:36.6241785Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:36.6242631Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:36.6243589Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:36.6244613Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:36.6245581Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:36.6246572Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:36.6247581Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:36.6248607Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:36.6249627Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:36.6250634Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:36.6251644Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:36.6252484Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:36.6253318Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:36.6254135Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:36.6255075Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:36.6255914Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:36.6256695Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:36.6257539Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:36.6258339Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:36.6259159Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:36.6259952Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:36.6260754Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:36.6261544Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:36.6262431Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:36.6263312Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:36.6264155Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:36.6264924Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:36.6265736Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:36.6266549Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:36.6267320Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:36.6268162Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:36.6269013Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:36.6269805Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:36.6270645Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:36.6271373Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:36.6272179Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:36.6272971Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:36.6273830Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:36.6274798Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:36.6275602Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:36.6276350Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:36.6277187Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:36.6278011Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:36.6278895Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:36.6279731Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:36.6280545Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:36.6281486Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:36.6282369Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:36.6283246Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:36.6284046Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:36.6284950Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:36.6285765Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:36.6286544Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:36.6287362Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:36.6288197Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:36.6289061Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:36.6289918Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:36.6290660Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:36.6291479Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:36.6292278Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:36.6293030Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:36.6293834Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:36.6294641Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:36.6295516Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:36.6296426Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:36.6297353Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.6298230Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.6299066Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:36.6299932Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:36.6300768Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:36.6301634Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:36.6302487Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:36.6303401Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:36.6304244Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:36.6305109Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:36.6306026Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:36.6306849Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:36.6307691Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:36.6308512Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:36.6309356Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:36.6310174Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:36.6311056Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:36.6311927Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:36.6312783Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:36.6313583Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:36.6314400Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:36.6315334Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:36.6316193Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:36.6317079Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:36.6317993Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:36.6318898Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:36.6319832Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:36.6320735Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:36.6321586Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:36.6322486Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:36.6323332Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:36.6324263Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:36.6325138Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:36.6325993Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:36.6326813Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:36.6327680Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:36.6328520Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:36.6329386Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:36.6330180Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:36.6331034Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:36.6331875Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:36.6332692Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:36.6333569Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:36.6334454Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:36.6335286Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:36.6336091Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:36.6337050Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:36.6337984Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:36.6338816Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:36.6339690Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:36.6340561Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:36.6341413Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:36.6342265Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:36.6343158Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:36.6343973Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:36.6344805Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:36.6345661Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:36.6346459Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:36.6347301Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:36.6348102Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:36.6348974Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:36.6349916Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:36.6350789Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.6351651Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:36.6352533Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:36.6353412Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:36.6354300Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:36.6355120Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:36.6356102Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:36.6356954Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:36.6357784Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:36.6358577Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:36.6359452Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:36.6360379Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:36.6361286Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:36.6362225Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:36.6363042Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:36.6363916Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:36.6364751Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:36.6365649Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:36.6366562Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:36.6367462Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:36.6368306Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:36.6369160Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:36.6370070Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:36.6370963Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:36.6371843Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:36.6372650Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:36.6373519Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:36.6374360Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:36.6375170Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:36.6376139Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:36.6377099Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:36.6377911Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:36.6378708Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:36.6379662Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:36.6380590Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:36.6381520Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:36.6382394Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:36.6383348Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:36.6384213Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:36.6385081Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:36.6385975Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:36.6386926Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:36.6387806Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:36.6388647Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:36.6389478Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:36.6390304Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:36.6391181Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:36.6392073Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:36.6392981Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:36.6393877Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:36.6394620Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:36.6395393Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:36.6397018Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:36.6397840Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:36.6398649Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:36.6399460Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:36.6400257Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:36.6401141Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:36.6402111Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:36.6402998Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:36.6403878Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:36.6404649Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:36.6405514Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:36.6406358Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:36.6407220Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:36.6408086Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:36.6408978Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:36.6409911Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:36.6410738Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:36.6411562Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:36.6412462Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:36.6413338Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:36.6414236Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:36.6415074Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:36.6415813Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:36.6416751Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:36.6417644Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:36.6418412Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:36.6419188Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:36.6420015Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:36.6420814Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:36.6421695Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:36.6422460Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:36.6423223Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:36.6424056Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:36.6424932Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:36.6425826Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:36.6426702Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:36.6427575Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:36.6428389Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:36.6429201Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:36.6430017Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:36.6430764Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:36.6431544Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:36.6432307Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:36.6433107Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:36.6433922Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:36.6434788Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:36.6435644Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:36.6436417Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:36.6437185Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:36.6437978Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:36.6439961Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:36.6442404Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:36.6443225Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:36.6444137Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:36.6444972Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:36.6445804Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:36.6446779Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:36.6447622Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:36.6448530Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:36.6449450Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:36.6450337Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:36.6451254Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:36.6452185Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:36.6453193Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:36.6454076Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:36.6454918Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:36.6455853Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:36.6456773Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:36.6457752Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:36.6458661Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:36.6459647Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:36.6460386Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:36.6461135Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:36.6461881Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:36.6462591Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:36.6463322Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:36.6464046Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:36.6464929Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:36.6465774Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:36.6466632Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:36.6467355Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:36.6468033Z V1204 11:22:45.962000 100983 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.6468721Z I1204 11:22:45.963000 100983 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.6469205Z V1204 11:22:45.963000 100983 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:36.6469697Z V1204 11:22:45.963000 100983 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:36.6470199Z I1204 11:22:45.963000 100983 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:36.6470975Z V1204 11:22:46.004000 100983 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:36.6471423Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.6471920Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:36.6472913Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.6473684Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.6474731Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.6475724Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.6476719Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.6477564Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:36.6478392Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:36.6479438Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:36.6480263Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:36.6481209Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.6482131Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.6482736Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.6483167Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.6483572Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.6484146Z V1204 11:22:46.012000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:36.6484712Z V1204 11:22:46.013000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:36.6485739Z V1204 11:22:46.014000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.6486274Z V1204 11:22:46.014000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6487332Z V1204 11:22:46.046000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.6487833Z V1204 11:22:46.047000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6488791Z V1204 11:22:46.048000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:36.6489278Z V1204 11:22:46.049000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6490273Z V1204 11:22:46.050000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:36.6490739Z V1204 11:22:46.051000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6491650Z V1204 11:22:46.052000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:36.6492092Z V1204 11:22:46.052000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6492947Z V1204 11:22:46.055000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:36.6493395Z V1204 11:22:46.055000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6494284Z V1204 11:22:46.058000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.6494736Z V1204 11:22:46.058000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6495674Z V1204 11:22:46.059000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.6496277Z V1204 11:22:46.060000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6496808Z V1204 11:22:46.060000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.6497439Z V1204 11:22:46.061000 100983 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:36.6498200Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:36.6498306Z warnings.warn( 2025-12-04T11:24:36.6498787Z V1204 11:22:46.061000 100983 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:36.6499030Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:36.6499146Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:36.6499268Z stats [('calls_captured', 1)] 2025-12-04T11:24:36.6499512Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:36.6500353Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:36.6500508Z graph_break [] 2025-12-04T11:24:36.6500733Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:36.6501179Z V1204 11:22:46.298000 100983 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmp5h3w0da2 2025-12-04T11:24:36.6501637Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.6502135Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:36.6503142Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.6503878Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.6504373Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:36.6504875Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:36.6505296Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.6506339Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.6507086Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:36.6507575Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:36.6507993Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.6508374Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.6509145Z V1204 11:22:46.577000 100983 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:36.6509866Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:36.6510550Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:36.6510832Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.6511112Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.6511402Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.6511876Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:36.6512711Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.6513555Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.6514214Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:36.6514989Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.6515746Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.6516245Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:36.6516543Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.6517134Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:36.6519468Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.6521796Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.6522446Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:36.6523455Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:36.6524185Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:36.6525001Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:36.6525795Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:36.6526510Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:36.6527242Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:36.6527966Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:36.6528722Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:36.6529490Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:36.6530168Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:36.6530865Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:36.6531742Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:36.6532576Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:36.6533734Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:36.6534527Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:36.6563567Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:36.6564602Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:36.6565417Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:36.6566188Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:36.6567071Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:36.6567912Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:36.6568650Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:36.6569528Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:36.6570314Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:36.6571110Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:36.6571966Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:36.6572812Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:36.6573679Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:36.6574539Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:36.6575388Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:36.6576327Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:36.6577282Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:36.6578106Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:36.6579134Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:36.6579895Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:36.6580742Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:36.6581490Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:36.6582361Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:36.6583164Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:36.6583872Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:36.6584676Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:36.6585453Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:36.6586217Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:36.6587055Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:36.6587816Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:36.6588609Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:36.6589384Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:36.6590162Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:36.6590947Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:36.6591732Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:36.6592526Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:36.6593325Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:36.6594121Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:36.6594889Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:36.6595666Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:36.6596629Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:36.6597478Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:36.6598274Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:36.6599077Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:36.6599898Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:36.6600728Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:36.6601581Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:36.6602448Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:36.6603235Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:36.6603992Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:36.6604745Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:36.6605571Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:36.6606486Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:36.6607292Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:36.6608121Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:36.6608968Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:36.6609723Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:36.6610907Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:36.6611710Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:36.6612574Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:36.6613500Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:36.6614342Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:36.6615168Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:36.6616009Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:36.6616951Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:36.6617788Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:36.6618726Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.6619563Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:36.6620521Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.6621340Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:36.6622276Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.6623108Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:36.6623981Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:36.6624771Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:36.6625541Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:36.6626381Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:36.6627243Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:36.6628000Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:36.6628825Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:36.6629636Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:36.6630501Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:36.6631376Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:36.6632297Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:36.6633213Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:36.6633986Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:36.6634828Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:36.6635701Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:36.6636511Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:36.6637290Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:36.6638180Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:36.6639053Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:36.6639919Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:36.6640828Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:36.6641627Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:36.6642439Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:36.6643162Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:36.6643943Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:36.6644759Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:36.6645578Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:36.6646457Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:36.6647286Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:36.6648123Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:36.6648958Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:36.6649700Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:36.6650509Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:36.6651417Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:36.6652357Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:36.6653286Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:36.6654108Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:36.6654929Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:36.6655762Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:36.6656636Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:36.6657629Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:36.6658500Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:36.6659283Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:36.6660111Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:36.6660940Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:36.6661825Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:36.6662631Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:36.6663419Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:36.6664265Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:36.6665076Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:36.6665872Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:36.6666683Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:36.6667512Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:36.6668351Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:36.6669218Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:36.6670014Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:36.6670892Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:36.6671693Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:36.6672590Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:36.6673420Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:36.6674214Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:36.6674971Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:36.6675772Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:36.6676559Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:36.6677431Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:36.6678307Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:36.6679155Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:36.6679987Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:36.6680846Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:36.6681657Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:36.6682557Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:36.6683416Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:36.6684231Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:36.6685044Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:36.6685876Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:36.6686659Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:36.6687439Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:36.6688245Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:36.6689024Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:36.6689779Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:36.6690577Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:36.6691431Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:36.6692278Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:36.6693105Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:36.6693918Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:36.6694740Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:36.6695554Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:36.6696519Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:36.6697498Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:36.6698380Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:36.6699207Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:36.6700050Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:36.6700866Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:36.6701692Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:36.6702537Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:36.6703327Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:36.6704123Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:36.6705042Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:36.6705850Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:36.6706677Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:36.6707543Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:36.6708367Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:36.6709127Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:36.6710001Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:36.6710837Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:36.6711629Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:36.6712466Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:36.6713450Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:36.6714470Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:36.6715442Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:36.6716419Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:36.6717386Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:36.6718388Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:36.6719399Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:36.6720377Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:36.6721424Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:36.6722264Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:36.6723103Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:36.6723915Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:36.6724842Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:36.6725684Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:36.6726502Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:36.6727323Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:36.6728076Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:36.6728889Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:36.6729713Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:36.6730520Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:36.6731304Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:36.6732189Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:36.6733020Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:36.6733836Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:36.6734605Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:36.6735416Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:36.6736206Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:36.6737074Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:36.6737920Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:36.6738772Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:36.6739560Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:36.6740396Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:36.6741130Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:36.6741961Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:36.6742784Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:36.6743613Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:36.6744570Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:36.6745400Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:36.6746151Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:36.6746997Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:36.6747821Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:36.6748678Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:36.6749491Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:36.6750317Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:36.6751268Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:36.6752155Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:36.6753037Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:36.6753846Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:36.6754761Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:36.6755568Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:36.6756347Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:36.6757167Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:36.6758047Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:36.6758932Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:36.6759765Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:36.6760514Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:36.6761331Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:36.6762156Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:36.6762909Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:36.6763704Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:36.6764483Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:36.6765324Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:36.6766108Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:36.6766983Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.6767851Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.6768654Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:36.6769522Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:36.6770360Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:36.6771224Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:36.6772086Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:36.6772991Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:36.6773857Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:36.6774749Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:36.6775559Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:36.6776381Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:36.6777279Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:36.6778135Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:36.6778980Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:36.6779795Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:36.6780648Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:36.6781471Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:36.6782320Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:36.6783128Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:36.6783934Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:36.6784864Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:36.6785728Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:36.6786603Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:36.6787515Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:36.6788413Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:36.6789341Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:36.6790275Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:36.6791163Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:36.6792018Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:36.6792868Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:36.6793828Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:36.6794710Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:36.6795574Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:36.6796570Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:36.6797422Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:36.6798232Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:36.6799103Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:36.6799885Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:36.6800715Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:36.6801594Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:36.6802421Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:36.6803299Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:36.6804174Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:36.6805010Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:36.6805816Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:36.6806776Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:36.6814595Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:36.6815512Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:36.6816407Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:36.6817508Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:36.6818366Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:36.6819216Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:36.6820102Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:36.6820905Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:36.6821716Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:36.6822575Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:36.6823367Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:36.6824250Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:36.6825059Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:36.6825918Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:36.6826867Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:36.6827745Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.6828603Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:36.6829520Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:36.6830549Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:36.6831413Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:36.6832224Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:36.6833205Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:36.6834090Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:36.6834945Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:36.6835735Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:36.6836581Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:36.6837486Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:36.6838390Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:36.6839328Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:36.6840129Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:36.6841004Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:36.6841844Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:36.6842747Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:36.6843657Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:36.6844549Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:36.6845386Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:36.6846268Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:36.6847211Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:36.6848073Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:36.6848951Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:36.6849791Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:36.6850660Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:36.6851497Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:36.6852310Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:36.6853243Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:36.6854110Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:36.6854927Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:36.6855716Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:36.6856638Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:36.6857620Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:36.6858545Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:36.6859421Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:36.6860365Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:36.6861224Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:36.6862120Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:36.6863043Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:36.6863964Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:36.6864832Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:36.6865698Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:36.6866530Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:36.6867357Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:36.6868230Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:36.6869101Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:36.6869982Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:36.6870877Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:36.6871619Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:36.6872391Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:36.6873236Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:36.6874043Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:36.6874874Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:36.6875682Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:36.6876482Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:36.6877364Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:36.6878354Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:36.6879269Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:36.6880086Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:36.6880849Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:36.6881755Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:36.6882598Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:36.6883457Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:36.6884321Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:36.6885158Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:36.6886047Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:36.6886871Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:36.6887689Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:36.6888524Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:36.6889426Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:36.6890325Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:36.6891154Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:36.6891890Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:36.6892819Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:36.6893638Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:36.6894446Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:36.6895249Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:36.6896204Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:36.6897053Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:36.6898008Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:36.6898782Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:36.6899533Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:36.6900380Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:36.6901220Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:36.6902085Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:36.6902960Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:36.6903829Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:36.6904615Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:36.6905455Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:36.6906287Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:36.6907038Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:36.6907815Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:36.6908574Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:36.6909372Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:36.6910234Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:36.6911125Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:36.6911961Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:36.6912735Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:36.6913508Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:36.6914335Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:36.6916313Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:36.6918739Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:36.6919518Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:36.6920428Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:36.6921269Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:36.6922099Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:36.6923072Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:36.6923908Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:36.6924815Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:36.6925727Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:36.6926645Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:36.6927584Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:36.6928483Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:36.6929481Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:36.6930385Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:36.6931227Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:36.6932156Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:36.6933076Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:36.6933973Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:36.6934852Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:36.6935832Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:36.6936566Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:36.6937381Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:36.6938124Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:36.6938838Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:36.6939568Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:36.6940286Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:36.6941166Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:36.6942007Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:36.6942894Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:36.6943664Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:36.6944316Z V1204 11:22:46.587000 100983 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.6944999Z I1204 11:22:46.587000 100983 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.6945484Z V1204 11:22:46.588000 100983 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:36.6946009Z V1204 11:22:46.588000 100983 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:36.6946513Z I1204 11:22:46.588000 100983 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:36.6947282Z V1204 11:22:46.602000 100983 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:36.6947741Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.6948236Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:36.6949209Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.6949956Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.6951001Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.6951993Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.6952995Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.6953836Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:36.6954647Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:36.6955708Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:36.6956567Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:36.6957534Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.6958424Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.6959027Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.6959484Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.6959886Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.6960457Z V1204 11:22:46.609000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:36.6961021Z V1204 11:22:46.610000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:36.6962036Z V1204 11:22:46.610000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.6962538Z V1204 11:22:46.610000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6963561Z V1204 11:22:46.612000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.6964063Z V1204 11:22:46.612000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6965018Z V1204 11:22:46.613000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:36.6965473Z V1204 11:22:46.613000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6966439Z V1204 11:22:46.615000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:36.6966908Z V1204 11:22:46.615000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6967819Z V1204 11:22:46.616000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:36.6968265Z V1204 11:22:46.616000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6969116Z V1204 11:22:46.618000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:36.6969567Z V1204 11:22:46.619000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6970489Z V1204 11:22:46.621000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.6970968Z V1204 11:22:46.621000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6971872Z V1204 11:22:46.622000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.6972320Z V1204 11:22:46.622000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.6972849Z V1204 11:22:46.623000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.6973457Z V1204 11:22:46.623000 100983 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:36.6974196Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:36.6974317Z warnings.warn( 2025-12-04T11:24:36.6974801Z V1204 11:22:46.624000 100983 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:36.6974967Z =================================== FAILURES =================================== 2025-12-04T11:24:36.6975415Z _ TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False _ 2025-12-04T11:24:36.6975541Z Traceback (most recent call last): 2025-12-04T11:24:36.6976126Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 384, in test_decompose_mm_mixed_precision 2025-12-04T11:24:36.6976236Z self.assertEqual( 2025-12-04T11:24:36.6976710Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:36.6976939Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:36.6977478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:36.6977702Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:36.6977835Z AssertionError: Scalars are not equal! 2025-12-04T11:24:36.6977842Z 2025-12-04T11:24:36.6977949Z Expected 1 but got 0. 2025-12-04T11:24:36.6978072Z Absolute difference: 1 2025-12-04T11:24:36.6978183Z Relative difference: 1.0 2025-12-04T11:24:36.6978189Z 2025-12-04T11:24:36.6978465Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:36.6979322Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:36.6979329Z 2025-12-04T11:24:36.6979605Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:36.6979850Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:36.6979962Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:36.6980092Z stats [('calls_captured', 1)] 2025-12-04T11:24:36.6980855Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:36.6981079Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:36.6981194Z graph_break [] 2025-12-04T11:24:36.6981419Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:36.6981925Z I1204 11:22:43.005000 100983 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:36.6982567Z I1204 11:22:43.253000 100983 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:36.6983039Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.6983547Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:36.6984489Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.6985240Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.6985765Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:36.6986257Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:36.6986691Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.6987720Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.6988483Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:36.6988956Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:36.6989386Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.6989768Z V1204 11:22:43.255000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.6990318Z I1204 11:22:45.464000 100983 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:36.6991041Z V1204 11:22:45.952000 100983 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:36.6991790Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:36.6992469Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:36.6992752Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.6993031Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.6993315Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.6993783Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:36.6994618Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.6995434Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.6996289Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:36.6997109Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.6997856Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.6998320Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:36.6998655Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.6999243Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:36.7001583Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.7006677Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.7007339Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:36.7008250Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:36.7009054Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:36.7009903Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:36.7010692Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:36.7011420Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:36.7012130Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:36.7012861Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:36.7013582Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:36.7014358Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:36.7015073Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:36.7015756Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:36.7016642Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:36.7017575Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:36.7018793Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:36.7019576Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:36.7048549Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:36.7049656Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:36.7050426Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:36.7051205Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:36.7052072Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:36.7052959Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:36.7053676Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:36.7054470Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:36.7055251Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:36.7056026Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:36.7056948Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:36.7057789Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:36.7058721Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:36.7059569Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:36.7060428Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:36.7061325Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:36.7062213Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:36.7063043Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:36.7064053Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:36.7064827Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:36.7065665Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:36.7066426Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:36.7067257Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:36.7068047Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:36.7068794Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:36.7069585Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:36.7070347Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:36.7071097Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:36.7071945Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:36.7072702Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:36.7073478Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:36.7074315Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:36.7075084Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:36.7075878Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:36.7076658Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:36.7077432Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:36.7078255Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:36.7079032Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:36.7079817Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:36.7080588Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:36.7081410Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:36.7082179Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:36.7082950Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:36.7083789Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:36.7084615Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:36.7085448Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:36.7086242Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:36.7087120Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:36.7087891Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:36.7088653Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:36.7089396Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:36.7090233Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:36.7091163Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:36.7091960Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:36.7092759Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:36.7093587Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:36.7094439Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:36.7095631Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:36.7096622Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:36.7097562Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:36.7098457Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:36.7099357Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:36.7100209Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:36.7101062Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:36.7101938Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:36.7102758Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:36.7103698Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.7104520Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:36.7105487Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.7106291Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:36.7107291Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.7108111Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:36.7108993Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:36.7109755Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:36.7110553Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:36.7111400Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:36.7112254Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:36.7113027Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:36.7113840Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:36.7114632Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:36.7115501Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:36.7116403Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:36.7117322Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:36.7118181Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:36.7118966Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:36.7119802Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:36.7120694Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:36.7121499Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:36.7122291Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:36.7123227Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:36.7124086Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:36.7124979Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:36.7125845Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:36.7126693Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:36.7127493Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:36.7128231Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:36.7129014Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:36.7129816Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:36.7130623Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:36.7131494Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:36.7132396Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:36.7133214Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:36.7134024Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:36.7134770Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:36.7135568Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:36.7136496Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:36.7137477Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:36.7138424Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:36.7139271Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:36.7140110Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:36.7140949Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:36.7141795Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:36.7142731Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:36.7143595Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:36.7144397Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:36.7145211Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:36.7146054Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:36.7146893Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:36.7147714Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:36.7148535Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:36.7149363Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:36.7150157Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:36.7150928Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:36.7151745Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:36.7152560Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:36.7155895Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:36.7156782Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:36.7157608Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:36.7158494Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:36.7159278Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:36.7160177Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:36.7161059Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:36.7161844Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:36.7162614Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:36.7163395Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:36.7164176Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:36.7165021Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:36.7165901Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:36.7166748Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:36.7167572Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:36.7168447Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:36.7169210Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:36.7170108Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:36.7170961Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:36.7171777Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:36.7172657Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:36.7173518Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:36.7174297Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:36.7175061Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:36.7175835Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:36.7176645Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:36.7177509Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:36.7178311Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:36.7179156Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:36.7180014Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:36.7180836Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:36.7181627Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:36.7182448Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:36.7183263Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:36.7184081Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:36.7184925Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:36.7185806Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:36.7186621Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:36.7187449Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:36.7188301Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:36.7189154Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:36.7189989Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:36.7190781Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:36.7191576Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:36.7192444Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:36.7193287Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:36.7194117Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:36.7194975Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:36.7195796Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:36.7196731Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:36.7197567Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:36.7198422Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:36.7199228Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:36.7200048Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:36.7201005Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:36.7202020Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:36.7203000Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:36.7203979Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:36.7205005Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:36.7206067Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:36.7207061Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:36.7208046Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:36.7209037Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:36.7209927Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:36.7210773Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:36.7211588Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:36.7212516Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:36.7213351Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:36.7214144Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:36.7214953Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:36.7215708Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:36.7216521Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:36.7217359Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:36.7218167Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:36.7218954Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:36.7219851Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:36.7220706Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:36.7221559Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:36.7222327Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:36.7223132Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:36.7223917Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:36.7224690Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:36.7225568Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:36.7226406Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:36.7227192Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:36.7228021Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:36.7228750Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:36.7229553Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:36.7230374Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:36.7231217Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:36.7232156Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:36.7232956Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:36.7233701Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:36.7234537Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:36.7235360Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:36.7236208Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:36.7237062Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:36.7237895Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:36.7238839Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:36.7239718Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:36.7240559Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:36.7241392Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:36.7242299Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:36.7243104Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:36.7243882Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:36.7244716Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:36.7245541Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:36.7246429Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:36.7247264Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:36.7248009Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:36.7248827Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:36.7249627Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:36.7250396Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:36.7251179Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:36.7251955Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:36.7252834Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:36.7253665Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:36.7254542Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.7255409Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.7256197Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:36.7257161Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:36.7258004Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:36.7258874Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:36.7259728Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:36.7260638Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:36.7261480Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:36.7262394Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:36.7263191Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:36.7264014Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:36.7264849Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:36.7265664Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:36.7266503Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:36.7267314Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:36.7268167Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:36.7269020Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:36.7269898Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:36.7270696Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:36.7271501Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:36.7272402Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:36.7273296Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:36.7274176Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:36.7275090Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:36.7275999Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:36.7276923Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:36.7277827Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:36.7278706Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:36.7279565Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:36.7280409Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:36.7281331Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:36.7282220Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:36.7283069Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:36.7283885Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:36.7284762Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:36.7285602Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:36.7286473Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:36.7287256Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:36.7288089Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:36.7288918Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:36.7289765Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:36.7290641Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:36.7291515Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:36.7292345Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:36.7293147Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:36.7294059Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:36.7294987Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:36.7295822Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:36.7296840Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:36.7297781Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:36.7298637Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:36.7300508Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:36.7302378Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:36.7304280Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:36.7306089Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:36.7307885Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:36.7309663Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:36.7311442Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:36.7313207Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:36.7315004Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:36.7316951Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:36.7318935Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.7320811Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:36.7322668Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:36.7324614Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:36.7326492Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:36.7328292Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:36.7330226Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:36.7332186Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:36.7334010Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:36.7335779Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:36.7337592Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:36.7339542Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:36.7341500Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:36.7343473Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:36.7345364Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:36.7347150Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:36.7349023Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:36.7350878Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:36.7352808Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:36.7354743Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:36.7356623Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:36.7358473Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:36.7360416Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:36.7362338Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:36.7364203Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:36.7366021Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:36.7367831Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:36.7369659Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:36.7371449Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:36.7373376Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:36.7375363Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:36.7377229Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:36.7378967Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:36.7380788Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:36.7382786Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:36.7384778Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:36.7386716Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:36.7388681Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:36.7390637Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:36.7392520Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:36.7394446Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:36.7396665Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:36.7398608Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:36.7400475Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:36.7402281Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:36.7404092Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:36.7405937Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:36.7407994Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:36.7409885Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:36.7411860Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:36.7413646Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:36.7415295Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:36.7417074Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:36.7418905Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:36.7420683Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:36.7422467Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:36.7424218Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:36.7426034Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:36.7428015Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:36.7430045Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:36.7431891Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:36.7433602Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:36.7435384Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:36.7437224Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:36.7439060Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:36.7440923Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:36.7442788Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:36.7444689Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:36.7446545Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:36.7448343Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:36.7450138Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:36.7451984Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:36.7453926Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:36.7455786Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:36.7457555Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:36.7459371Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:36.7461255Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:36.7462988Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:36.7464729Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:36.7466435Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:36.7468154Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:36.7469946Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:36.7471731Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:36.7473410Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:36.7475145Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:36.7477005Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:36.7478835Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:36.7480738Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:36.7482616Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:36.7484404Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:36.7486140Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:36.7487924Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:36.7489629Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:36.7491293Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:36.7492959Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:36.7494637Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:36.7496552Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:36.7498496Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:36.7500333Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:36.7502080Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:36.7503753Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:36.7505454Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:36.7508368Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:36.7512912Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:36.7516306Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:36.7518126Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:36.7519967Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:36.7521787Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:36.7523711Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:36.7525655Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:36.7527541Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:36.7529497Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:36.7531435Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:36.7533389Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:36.7535330Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:36.7537423Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:36.7539458Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:36.7541317Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:36.7543216Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:36.7545208Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:36.7547196Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:36.7549128Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:36.7551127Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:36.7552964Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:36.7554548Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:36.7556131Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:36.7557753Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:36.7559331Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:36.7560901Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:36.7562640Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:36.7564496Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:36.7566337Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:36.7568084Z V1204 11:22:45.961000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:36.7569587Z V1204 11:22:45.962000 100983 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.7571048Z I1204 11:22:45.963000 100983 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.7572346Z V1204 11:22:45.963000 100983 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:36.7573466Z V1204 11:22:45.963000 100983 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:36.7574589Z I1204 11:22:45.963000 100983 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:36.7576011Z V1204 11:22:46.004000 100983 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:36.7577416Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.7578506Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:36.7580171Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.7582033Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.7583950Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.7586089Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.7588217Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.7590191Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:36.7591997Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:36.7593981Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:36.7596166Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:36.7598141Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.7600109Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.7601723Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.7602868Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.7603832Z V1204 11:22:46.009000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.7604938Z V1204 11:22:46.012000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:36.7606218Z V1204 11:22:46.013000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:36.7607944Z V1204 11:22:46.014000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.7609625Z V1204 11:22:46.014000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.7611324Z V1204 11:22:46.046000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.7612983Z V1204 11:22:46.047000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.7614581Z V1204 11:22:46.048000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:36.7616125Z V1204 11:22:46.049000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.7617717Z V1204 11:22:46.050000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:36.7619308Z V1204 11:22:46.051000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.7620825Z V1204 11:22:46.052000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:36.7622309Z V1204 11:22:46.052000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.7623736Z V1204 11:22:46.055000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:36.7625152Z V1204 11:22:46.055000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.7626625Z V1204 11:22:46.058000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.7628138Z V1204 11:22:46.058000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.7629628Z V1204 11:22:46.059000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.7631095Z V1204 11:22:46.060000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.7632207Z V1204 11:22:46.060000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.7633441Z V1204 11:22:46.061000 100983 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:36.7634891Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:36.7635869Z warnings.warn( 2025-12-04T11:24:36.7636511Z V1204 11:22:46.061000 100983 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:36.7637361Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:36.7637845Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:36.7638174Z stats [('calls_captured', 1)] 2025-12-04T11:24:36.7638655Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:36.7639780Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:36.7640816Z graph_break [] 2025-12-04T11:24:36.7641182Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:36.7641981Z V1204 11:22:46.298000 100983 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmp5h3w0da2 2025-12-04T11:24:36.7643008Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.7644073Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:36.7645649Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.7647492Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.7648850Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:36.7649974Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:36.7651016Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.7652609Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.7654519Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:36.7655883Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:36.7657031Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.7657973Z V1204 11:22:46.322000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.7659225Z V1204 11:22:46.577000 100983 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:36.7660798Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:36.7662310Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:36.7663391Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.7664097Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.7664800Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.7665687Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:36.7667111Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.7668913Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.7670531Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:36.7672064Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.7673724Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.7675054Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:36.7675952Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.7677007Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:36.7680059Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.7684840Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.7687927Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:36.7689654Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:36.7691435Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:36.7693117Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:36.7694856Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:36.7696671Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:36.7698294Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:36.7700044Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:36.7701721Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:36.7703383Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:36.7704948Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:36.7706453Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:36.7708159Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:36.7709970Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:36.7712153Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:36.7713898Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:36.7743850Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:36.7774078Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:36.7775981Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:36.7777751Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:36.7779545Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:36.7781384Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:36.7783079Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:36.7784719Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:36.7786461Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:36.7788162Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:36.7789947Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:36.7791762Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:36.7793594Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:36.7795448Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:36.7797429Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:36.7799380Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:36.7801258Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:36.7803071Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:36.7805045Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:36.7806957Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:36.7808689Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:36.7810402Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:36.7812180Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:36.7813973Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:36.7815592Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:36.7817315Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:36.7818996Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:36.7820645Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:36.7822435Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:36.7824173Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:36.7825838Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:36.7827530Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:36.7829210Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:36.7830901Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:36.7832639Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:36.7834314Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:36.7836013Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:36.7837731Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:36.7839420Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:36.7841098Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:36.7842807Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:36.7844531Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:36.7846240Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:36.7847989Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:36.7849745Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:36.7851498Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:36.7853260Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:36.7855097Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:36.7856949Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:36.7858670Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:36.7860304Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:36.7862005Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:36.7863874Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:36.7865732Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:36.7867498Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:36.7869249Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:36.7870969Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:36.7873054Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:36.7875196Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:36.7877003Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:36.7878888Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:36.7880784Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:36.7882613Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:36.7884424Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:36.7886290Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:36.7888106Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:36.7890026Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.7891942Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:36.7893867Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.7895778Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:36.7897843Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.7899748Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:36.7901668Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:36.7903446Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:36.7905102Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:36.7906825Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:36.7908667Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:36.7910421Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:36.7912152Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:36.7913950Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:36.7915705Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:36.7917598Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:36.7919528Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:36.7921458Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:36.7923237Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:36.7925017Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:36.7926878Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:36.7928709Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:36.7930431Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:36.7932249Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:36.7934133Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:36.7936043Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:36.7937993Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:36.7939807Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:36.7941552Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:36.7943208Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:36.7944861Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:36.7946601Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:36.7948392Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:36.7950195Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:36.7951985Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:36.7953792Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:36.7955569Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:36.7957254Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:36.7958961Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:36.7960805Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:36.7962781Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:36.7964771Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:36.7966657Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:36.7968435Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:36.7970333Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:36.7972145Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:36.7974023Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:36.7975927Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:36.7977764Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:36.7979493Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:36.7981275Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:36.7983114Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:36.7984892Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:36.7986591Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:36.7988347Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:36.7990104Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:36.7991809Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:36.7993570Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:36.7995338Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:36.7997278Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:36.7999129Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:36.8000919Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:36.8002721Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:36.8004621Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:36.8006408Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:36.8008227Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:36.8009978Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:36.8011671Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:36.8013350Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:36.8015039Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:36.8016795Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:36.8018723Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:36.8020594Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:36.8022393Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:36.8024206Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:36.8025962Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:36.8027800Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:36.8029685Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:36.8031485Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:36.8033224Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:36.8034995Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:36.8036748Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:36.8038461Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:36.8040132Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:36.8041821Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:36.8043473Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:36.8045145Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:36.8046921Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:36.8048744Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:36.8050550Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:36.8052336Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:36.8054064Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:36.8055790Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:36.8057603Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:36.8059397Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:36.8061292Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:36.8063127Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:36.8064912Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:36.8066671Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:36.8068428Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:36.8070217Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:36.8071986Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:36.8073723Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:36.8075507Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:36.8077331Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:36.8079096Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:36.8080934Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:36.8082754Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:36.8084451Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:36.8086209Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:36.8088009Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:36.8089748Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:36.8091509Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:36.8093418Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:36.8095555Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:36.8097898Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:36.8099998Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:36.8102080Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:36.8104190Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:36.8106331Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:36.8108527Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:36.8110649Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:36.8112632Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:36.8114442Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:36.8116221Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:36.8118094Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:36.8119986Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:36.8121791Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:36.8124045Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:36.8125714Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:36.8127409Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:36.8129143Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:36.8130864Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:36.8132637Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:36.8134445Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:36.8136295Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:36.8138144Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:36.8139861Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:36.8141561Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:36.8143324Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:36.8145025Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:36.8146782Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:36.8148605Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:36.8150366Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:36.8152123Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:36.8153820Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:36.8155529Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:36.8157258Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:36.8159045Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:36.8160959Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:36.8162843Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:36.8164535Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:36.8166296Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:36.8168082Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:36.8169892Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:36.8171703Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:36.8173459Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:36.8175347Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:36.8177384Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:36.8179249Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:36.8181023Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:36.8182878Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:36.8184726Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:36.8186432Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:36.8188166Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:36.8189979Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:36.8191830Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:36.8193657Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:36.8195350Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:36.8197184Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:36.8198945Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:36.8200703Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:36.8202390Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:36.8204080Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:36.8205832Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:36.8207603Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:36.8209399Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.8211326Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.8213102Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:36.8214870Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:36.8216703Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:36.8218593Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:36.8220450Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:36.8222334Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:36.8224258Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:36.8226114Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:36.8227970Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:36.8229732Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:36.8231508Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:36.8233289Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:36.8242789Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:36.8244609Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:36.8246405Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:36.8248217Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:36.8249084Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:36.8249878Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:36.8250748Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:36.8251644Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:36.8252503Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:36.8253391Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:36.8254305Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:36.8255214Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:36.8256132Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:36.8257191Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:36.8258098Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:36.8258957Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:36.8259803Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:36.8260725Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:36.8261650Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:36.8262502Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:36.8263335Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:36.8264160Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:36.8264982Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:36.8265836Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:36.8266620Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:36.8267487Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:36.8268316Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:36.8269144Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:36.8270006Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:36.8270896Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:36.8271709Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:36.8272508Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:36.8273450Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:36.8274372Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:36.8275212Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:36.8276073Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:36.8276958Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:36.8277822Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:36.8278676Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:36.8279570Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:36.8280352Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:36.8281168Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:36.8282008Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:36.8282806Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:36.8283626Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:36.8284434Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:36.8285276Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:36.8286224Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:36.8287109Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.8287944Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:36.8288833Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:36.8289727Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:36.8290621Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:36.8291424Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:36.8292405Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:36.8293248Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:36.8294106Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:36.8294906Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:36.8295728Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:36.8296788Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:36.8297757Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:36.8298702Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:36.8299586Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:36.8300425Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:36.8301270Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:36.8302147Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:36.8303072Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:36.8303950Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:36.8304799Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:36.8305693Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:36.8306652Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:36.8307523Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:36.8308385Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:36.8309204Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:36.8310097Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:36.8310948Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:36.8311742Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:36.8312679Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:36.8313553Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:36.8314357Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:36.8315159Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:36.8316063Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:36.8316992Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:36.8317908Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:36.8318806Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:36.8319749Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:36.8320605Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:36.8321477Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:36.8322406Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:36.8323361Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:36.8324219Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:36.8325067Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:36.8325890Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:36.8326748Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:36.8327640Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:36.8328489Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:36.8329379Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:36.8330263Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:36.8331012Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:36.8331797Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:36.8332605Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:36.8333415Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:36.8334228Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:36.8335050Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:36.8335840Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:36.8336727Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:36.8337757Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:36.8338691Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:36.8339512Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:36.8340277Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:36.8341155Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:36.8341989Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:36.8342895Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:36.8343756Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:36.8344608Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:36.8345479Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:36.8346299Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:36.8347137Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:36.8347993Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:36.8348879Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:36.8349767Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:36.8350611Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:36.8351327Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:36.8352258Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:36.8353083Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:36.8353877Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:36.8354698Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:36.8355476Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:36.8356275Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:36.8357134Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:36.8357907Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:36.8358699Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:36.8359532Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:36.8360382Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:36.8361225Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:36.8362105Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:36.8362963Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:36.8363783Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:36.8364586Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:36.8365402Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:36.8366162Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:36.8366924Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:36.8367699Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:36.8368490Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:36.8369305Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:36.8370207Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:36.8371061Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:36.8371849Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:36.8372603Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:36.8373417Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:36.8375479Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:36.8377982Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:36.8378747Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:36.8379719Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:36.8380519Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:36.8381360Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:36.8382319Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:36.8383164Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:36.8384079Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:36.8384984Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:36.8385907Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:36.8386844Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:36.8387757Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:36.8388747Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:36.8389639Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:36.8390469Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:36.8391432Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:36.8392361Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:36.8393243Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:36.8394130Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:36.8395101Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:36.8395882Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:36.8396703Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:36.8397442Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:36.8398168Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:36.8398884Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:36.8399620Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:36.8400495Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:36.8401356Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:36.8402252Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:36.8403017Z V1204 11:22:46.586000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:36.8403681Z V1204 11:22:46.587000 100983 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.8404353Z I1204 11:22:46.587000 100983 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.8404851Z V1204 11:22:46.588000 100983 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:36.8405340Z V1204 11:22:46.588000 100983 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:36.8405847Z I1204 11:22:46.588000 100983 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:36.8406666Z V1204 11:22:46.602000 100983 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:36.8407109Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.8407622Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:36.8408572Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.8409336Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.8410375Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.8411403Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.8412392Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.8413238Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:36.8414065Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:36.8415108Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:36.8415951Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:36.8416990Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.8417934Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.8418527Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.8418967Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.8419358Z V1204 11:22:46.606000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.8419964Z V1204 11:22:46.609000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:36.8420550Z V1204 11:22:46.610000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:36.8421568Z V1204 11:22:46.610000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.8422085Z V1204 11:22:46.610000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.8423103Z V1204 11:22:46.612000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.8423623Z V1204 11:22:46.612000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.8424562Z V1204 11:22:46.613000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:36.8425048Z V1204 11:22:46.613000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.8425996Z V1204 11:22:46.615000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:36.8426450Z V1204 11:22:46.615000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.8427369Z V1204 11:22:46.616000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:36.8427806Z V1204 11:22:46.616000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.8428667Z V1204 11:22:46.618000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:36.8429097Z V1204 11:22:46.619000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.8430020Z V1204 11:22:46.621000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.8430511Z V1204 11:22:46.621000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.8431397Z V1204 11:22:46.622000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.8431859Z V1204 11:22:46.622000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.8432367Z V1204 11:22:46.623000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.8432955Z V1204 11:22:46.623000 100983 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:36.8433693Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:36.8433843Z warnings.warn( 2025-12-04T11:24:36.8434334Z V1204 11:22:46.624000 100983 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:36.8434559Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:36.8434672Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:36.8434799Z stats [('calls_captured', 1)] 2025-12-04T11:24:36.8435019Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:36.8435789Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:36.8435887Z graph_break [] 2025-12-04T11:24:36.8436109Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:36.8436562Z V1204 11:22:46.642000 100983 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpd_4s21sj 2025-12-04T11:24:36.8437003Z V1204 11:22:46.665000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.8437510Z V1204 11:22:46.665000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:36.8438482Z V1204 11:22:46.665000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.8439212Z V1204 11:22:46.665000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.8439717Z V1204 11:22:46.665000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:36.8440209Z V1204 11:22:46.665000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:36.8440638Z V1204 11:22:46.665000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.8441663Z V1204 11:22:46.665000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.8442414Z V1204 11:22:46.665000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:36.8442920Z V1204 11:22:46.665000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:36.8443369Z V1204 11:22:46.665000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.8443761Z V1204 11:22:46.665000 100983 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.8444476Z V1204 11:22:46.910000 100983 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:36.8445203Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:36.8445862Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:36.8446142Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.8446429Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.8446735Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.8447219Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:36.8448036Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.8448844Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.8449479Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:36.8450247Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.8450999Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.8451479Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:36.8451788Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.8452362Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:36.8454701Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.8457093Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.8457809Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:36.8458763Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:36.8459489Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:36.8460319Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:36.8461184Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:36.8462047Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:36.8462767Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:36.8463504Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:36.8464213Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:36.8464944Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:36.8465634Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:36.8466317Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:36.8467234Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:36.8468022Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:36.8469234Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:36.8470018Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:36.8499203Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:36.8500317Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:36.8501082Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:36.8501859Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:36.8502704Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:36.8503545Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:36.8504317Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:36.8505086Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:36.8505877Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:36.8506659Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:36.8507523Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:36.8508354Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:36.8509217Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:36.8510074Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:36.8510950Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:36.8511915Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:36.8512775Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:36.8513597Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:36.8514603Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:36.8515397Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:36.8516233Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:36.8516984Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:36.8517829Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:36.8518585Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:36.8519303Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:36.8520094Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:36.8520885Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:36.8521638Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:36.8522471Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:36.8523248Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:36.8524023Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:36.8524811Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:36.8525572Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:36.8526392Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:36.8527199Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:36.8527955Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:36.8528764Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:36.8529539Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:36.8530326Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:36.8531129Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:36.8531948Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:36.8532714Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:36.8533478Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:36.8534295Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:36.8535105Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:36.8535970Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:36.8536756Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:36.8537688Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:36.8538466Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:36.8539232Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:36.8539994Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:36.8540793Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:36.8541731Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:36.8542563Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:36.8543393Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:36.8544224Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:36.8544975Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:36.8546173Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:36.8547001Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:36.8547880Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:36.8548765Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:36.8549591Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:36.8550398Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:36.8551245Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:36.8552164Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:36.8552967Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:36.8553914Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.8554743Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:36.8555712Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.8556510Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:36.8557454Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.8558300Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:36.8559200Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:36.8559988Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:36.8560741Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:36.8561594Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:36.8562449Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:36.8563251Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:36.8564077Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:36.8564859Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:36.8565711Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:36.8566583Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:36.8567523Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:36.8568409Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:36.8569195Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:36.8570024Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:36.8570901Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:36.8571716Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:36.8572491Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:36.8573399Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:36.8574301Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:36.8575210Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:36.8576079Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:36.8576947Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:36.8577748Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:36.8578471Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:36.8579313Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:36.8580120Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:36.8580925Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:36.8581755Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:36.8582599Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:36.8583427Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:36.8584279Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:36.8585032Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:36.8585829Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:36.8586760Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:36.8587690Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:36.8588634Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:36.8589439Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:36.8590296Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:36.8591185Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:36.8592019Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:36.8593053Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:36.8593909Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:36.8594709Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:36.8595562Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:36.8596565Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:36.8597415Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:36.8598182Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:36.8598988Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:36.8599822Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:36.8600680Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:36.8601458Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:36.8602270Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:36.8603102Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:36.8603931Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:36.8604805Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:36.8605584Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:36.8606512Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:36.8607371Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:36.8608234Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:36.8609058Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:36.8609841Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:36.8610611Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:36.8611438Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:36.8612235Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:36.8613060Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:36.8613918Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:36.8614748Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:36.8615577Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:36.8616478Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:36.8617285Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:36.8618191Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:36.8619038Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:36.8619874Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:36.8620673Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:36.8621496Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:36.8622313Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:36.8623082Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:36.8623903Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:36.8624666Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:36.8625422Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:36.8626200Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:36.8627078Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:36.8627935Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:36.8628750Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:36.8629549Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:36.8630307Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:36.8631129Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:36.8631928Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:36.8632814Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:36.8633703Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:36.8634524Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:36.8635373Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:36.8636167Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:36.8636996Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:36.8637823Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:36.8638647Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:36.8639559Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:36.8640433Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:36.8641252Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:36.8642072Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:36.8642979Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:36.8643799Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:36.8644552Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:36.8645386Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:36.8646183Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:36.8646991Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:36.8647810Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:36.8648803Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:36.8649804Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:36.8650796Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:36.8651762Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:36.8652738Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:36.8653749Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:36.8654779Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:36.8655797Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:36.8656797Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:36.8657713Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:36.8658534Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:36.8659380Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:36.8660310Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:36.8661152Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:36.8661942Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:36.8662720Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:36.8663491Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:36.8664295Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:36.8665127Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:36.8665918Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:36.8666710Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:36.8667611Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:36.8668431Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:36.8669255Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:36.8670009Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:36.8670864Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:36.8671670Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:36.8672446Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:36.8673298Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:36.8674132Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:36.8674933Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:36.8675790Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:36.8676536Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:36.8677324Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:36.8678114Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:36.8678951Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:36.8679903Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:36.8680739Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:36.8681473Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:36.8682323Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:36.8683137Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:36.8683996Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:36.8684818Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:36.8685625Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:36.8686609Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:36.8687509Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:36.8688365Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:36.8689148Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:36.8690052Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:36.8690860Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:36.8691684Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:36.8692516Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:36.8693340Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:36.8694212Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:36.8695035Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:36.8695782Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:36.8696810Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:36.8697681Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:36.8698451Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:36.8699241Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:36.8700028Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:36.8700863Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:36.8701645Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:36.8702569Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.8703439Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.8704264Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:36.8705121Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:36.8705969Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:36.8706819Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:36.8707732Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:36.8708630Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:36.8709461Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:36.8710340Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:36.8711137Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:36.8711971Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:36.8712824Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:36.8713664Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:36.8714491Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:36.8715312Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:36.8716180Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:36.8717009Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:36.8717878Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:36.8718704Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:36.8719523Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:36.8720443Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:36.8721301Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:36.8722195Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:36.8723110Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:36.8724070Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:36.8724991Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:36.8725899Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:36.8726739Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:36.8727611Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:36.8728444Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:36.8729399Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:36.8730290Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:36.8731139Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:36.8731978Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:36.8732804Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:36.8733625Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:36.8734478Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:36.8735316Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:36.8736180Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:36.8737065Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:36.8737896Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:36.8738760Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:36.8739693Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:36.8740511Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:36.8741314Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:36.8742226Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:36.8743124Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:36.8743967Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:36.8744852Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:36.8745737Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:36.8746575Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:36.8747441Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:36.8748326Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:36.8749110Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:36.8749925Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:36.8750767Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:36.8751602Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:36.8752436Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:36.8753243Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:36.8754081Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:36.8755029Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:36.8755953Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.8756799Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:36.8757693Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:36.8758557Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:36.8759427Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:36.8760236Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:36.8761242Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:36.8762085Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:36.8762915Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:36.8763716Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:36.8764544Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:36.8765456Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:36.8766346Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:36.8767332Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:36.8768165Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:36.8769000Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:36.8769845Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:36.8770724Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:36.8771645Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:36.8772559Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:36.8773413Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:36.8774253Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:36.8775164Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:36.8776035Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:36.8776951Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:36.8777806Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:36.8778657Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:36.8779504Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:36.8780299Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:36.8781245Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:36.8782114Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:36.8782916Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:36.8783752Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:36.8784669Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:36.8785597Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:36.8786506Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:36.8787401Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:36.8788363Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:36.8789222Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:36.8790096Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:36.8790974Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:36.8791915Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:36.8792783Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:36.8793662Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:36.8794484Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:36.8795312Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:36.8796357Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:36.8797214Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:36.8798110Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:36.8798994Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:36.8799817Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:36.8800622Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:36.8801440Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:36.8802252Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:36.8803066Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:36.8803899Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:36.8804726Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:36.8805625Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:36.8806569Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:36.8807467Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:36.8808276Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:36.8809045Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:36.8809961Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:36.8810790Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:36.8811665Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:36.8812515Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:36.8813370Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:36.8814241Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:36.8815062Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:36.8815940Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:36.8816804Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:36.8817743Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:36.8818631Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:36.8819469Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:36.8820194Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:36.8821158Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:36.8821984Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:36.8822744Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:36.8823543Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:36.8824325Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:36.8825130Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:36.8826019Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:36.8826788Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:36.8827551Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:36.8828378Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:36.8829231Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:36.8830080Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:36.8830962Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:36.8831854Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:36.8832683Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:36.8833482Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:36.8834294Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:36.8835053Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:36.8835813Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:36.8836612Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:36.8837402Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:36.8838227Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:36.8839078Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:36.8839902Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:36.8840695Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:36.8841487Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:36.8842296Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:36.8844252Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:36.8846695Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:36.8847457Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:36.8848420Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:36.8849246Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:36.8850099Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:36.8851065Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:36.8851906Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:36.8852852Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:36.8853760Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:36.8854659Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:36.8855560Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:36.8856485Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:36.8857524Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:36.8858457Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:36.8859290Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:36.8860215Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:36.8861152Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:36.8862032Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:36.8862919Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:36.8863890Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:36.8864678Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:36.8865430Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:36.8866171Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:36.8866899Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:36.8867615Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:36.8868355Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:36.8869258Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:36.8870114Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:36.8870959Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:36.8871681Z V1204 11:22:46.919000 100983 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:36.8872354Z V1204 11:22:46.920000 100983 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.8873026Z I1204 11:22:46.920000 100983 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.8873574Z V1204 11:22:46.920000 100983 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:36.8874060Z V1204 11:22:46.920000 100983 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:36.8874579Z I1204 11:22:46.921000 100983 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 2 2025-12-04T11:24:36.8875348Z V1204 11:22:46.934000 100983 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:36.8875798Z V1204 11:22:46.938000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.8876312Z V1204 11:22:46.938000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:36.8877269Z V1204 11:22:46.938000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.8878025Z V1204 11:22:46.938000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.8879090Z V1204 11:22:46.938000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.8880107Z V1204 11:22:46.938000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.8881090Z V1204 11:22:46.938000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.8881932Z V1204 11:22:46.938000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:36.8882760Z V1204 11:22:46.938000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:36.8883828Z V1204 11:22:46.938000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:36.8884672Z V1204 11:22:46.938000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:36.8885603Z V1204 11:22:46.938000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.8886523Z V1204 11:22:46.938000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.8887109Z V1204 11:22:46.938000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.8887579Z V1204 11:22:46.938000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.8887972Z V1204 11:22:46.938000 100983 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.8888542Z V1204 11:22:46.941000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:36.8889116Z V1204 11:22:46.941000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:36.8890122Z V1204 11:22:46.942000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.8890636Z V1204 11:22:46.942000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.8891640Z V1204 11:22:46.943000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.8892152Z V1204 11:22:46.944000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.8893185Z V1204 11:22:46.945000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:36.8893669Z V1204 11:22:46.945000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.8894622Z V1204 11:22:46.946000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:36.8895078Z V1204 11:22:46.946000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.8896157Z V1204 11:22:46.948000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:36.8896596Z V1204 11:22:46.948000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.8897582Z V1204 11:22:46.950000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:36.8898021Z V1204 11:22:46.950000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.8898911Z V1204 11:22:46.952000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.8899374Z V1204 11:22:46.952000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.8900265Z V1204 11:22:46.953000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.8900728Z V1204 11:22:46.954000 100983 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.8901242Z V1204 11:22:46.954000 100983 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.8901879Z V1204 11:22:46.955000 100983 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 2 2025-12-04T11:24:36.8902619Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:36.8902726Z warnings.warn( 2025-12-04T11:24:36.8903222Z V1204 11:22:46.955000 100983 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:36.8904073Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-041aad81ad32b1c4.xml - 2025-12-04T11:24:36.8904265Z =========================== short test summary info ============================ 2025-12-04T11:24:36.8905164Z FAILED [0.3294s] inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False - AssertionError: Scalars are not equal! 2025-12-04T11:24:36.8905175Z 2025-12-04T11:24:36.8905286Z Expected 1 but got 0. 2025-12-04T11:24:36.8905409Z Absolute difference: 1 2025-12-04T11:24:36.8905520Z Relative difference: 1.0 2025-12-04T11:24:36.8905527Z 2025-12-04T11:24:36.8905746Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:36.8906653Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:36.8906698Z 2025-12-04T11:24:36.8906976Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:36.8907169Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:24:36.8907370Z ================== 1 failed, 36 deselected, 2 rerun in 6.66s =================== 2025-12-04T11:24:36.8907478Z --- Logging error --- 2025-12-04T11:24:36.8907615Z Traceback (most recent call last): 2025-12-04T11:24:36.8907927Z File "/opt/conda/envs/py_3.10/lib/python3.10/logging/__init__.py", line 1103, in emit 2025-12-04T11:24:36.8908075Z stream.write(msg + self.terminator) 2025-12-04T11:24:36.8908211Z ValueError: I/O operation on closed file. 2025-12-04T11:24:36.8908311Z Call stack: 2025-12-04T11:24:36.8908849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/remote_cache.py", line 432, in dump_cache_stats 2025-12-04T11:24:36.8909002Z log.info("Cache Metrics:%s", out.getvalue()) 2025-12-04T11:24:36.8909116Z Message: 'Cache Metrics:%s' 2025-12-04T11:24:36.8909274Z Arguments: (' None\n',) 2025-12-04T11:24:36.8909373Z Got exit code 1 2025-12-04T11:24:36.8909482Z Retrying single test... 2025-12-04T11:24:36.8910159Z Test results will be stored in test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-043eb8ddb5d84fd4.xml 2025-12-04T11:24:36.8910326Z ============================= test session starts ============================== 2025-12-04T11:24:36.8910690Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:24:36.8910797Z cachedir: .pytest_cache 2025-12-04T11:24:36.8911321Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:24:36.8911459Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:24:36.8911568Z configfile: pytest.ini 2025-12-04T11:24:36.8912174Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:24:36.8913375Z collecting ... /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:52: PytestCollectionWarning: cannot collect test class 'TestDecomposeAddMM' because it has a __init__ constructor (from: test/inductor/test_decompose_mem_bound_mm.py) 2025-12-04T11:24:36.8913565Z class TestDecomposeAddMM(torch.nn.Module): 2025-12-04T11:24:36.8913728Z collected 37 items / 36 deselected / 1 selected 2025-12-04T11:24:36.8914662Z stepcurrent: skipping 29 already run items. Running only test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:36.8914793Z Running 1 items in this shard 2025-12-04T11:24:36.8914800Z 2025-12-04T11:24:36.8915605Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False ('RERUN', {'yellow': True}) [5.8744s] [100%] 2025-12-04T11:24:36.8916409Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False ('RERUN', {'yellow': True}) [0.3346s] [100%] 2025-12-04T11:24:36.8917143Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False FAILED [0.3213s] [100%] 2025-12-04T11:24:36.8917149Z 2025-12-04T11:24:36.8917291Z ==================================== RERUNS ==================================== 2025-12-04T11:24:36.8917750Z _ TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False _ 2025-12-04T11:24:36.8917905Z Traceback (most recent call last): 2025-12-04T11:24:36.8918475Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 384, in test_decompose_mm_mixed_precision 2025-12-04T11:24:36.8918624Z self.assertEqual( 2025-12-04T11:24:36.8919089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:36.8919265Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:36.8919802Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:36.8920010Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:36.8920156Z AssertionError: Scalars are not equal! 2025-12-04T11:24:36.8920161Z 2025-12-04T11:24:36.8920266Z Expected 1 but got 0. 2025-12-04T11:24:36.8920373Z Absolute difference: 1 2025-12-04T11:24:36.8920493Z Relative difference: 1.0 2025-12-04T11:24:36.8920501Z 2025-12-04T11:24:36.8920718Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:36.8921574Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:36.8921628Z 2025-12-04T11:24:36.8921902Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:36.8922124Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:36.8922250Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:36.8922370Z stats [('calls_captured', 1)] 2025-12-04T11:24:36.8923141Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:36.8923365Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:36.8923466Z graph_break [] 2025-12-04T11:24:36.8923699Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:36.8924202Z I1204 11:23:02.873000 101180 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:36.8924811Z I1204 11:23:03.122000 101180 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:36.8925251Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.8925776Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:36.8926732Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.8927469Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.8927975Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:36.8928463Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:36.8928883Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.8929920Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.8930697Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:36.8931217Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:36.8931635Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.8932032Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.8932590Z I1204 11:23:05.325000 101180 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:36.8933300Z V1204 11:23:05.803000 101180 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:36.8934026Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:36.8934719Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:36.8935012Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.8935288Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.8935566Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.8936044Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:36.8936933Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.8937771Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.8938393Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:36.8939204Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.8939954Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.8940403Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:36.8940713Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.8941287Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:36.8943631Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.8945935Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.8946654Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:36.8947567Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:36.8948301Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:36.8949118Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:36.8949930Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:36.8950663Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:36.8951376Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:36.8952113Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:36.8952829Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:36.8953580Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:36.8954258Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:36.8954975Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:36.8955861Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:36.8956650Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:36.8957833Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:36.8958609Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:36.8987669Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:36.8988759Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:36.8989536Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:36.8990329Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:36.8991164Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:36.8992047Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:36.8992768Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:36.8993554Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:36.8994336Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:36.8995114Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:36.8996157Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:36.8996991Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:36.8997943Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:36.8998852Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:36.8999714Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:36.9000618Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:36.9001476Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:36.9002305Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:36.9003351Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:36.9004131Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:36.9004970Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:36.9005726Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:36.9006569Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:36.9007332Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:36.9008101Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:36.9008904Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:36.9009668Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:36.9010423Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:36.9011269Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:36.9012031Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:36.9012811Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:36.9013606Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:36.9014401Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:36.9015228Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:36.9016006Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:36.9016781Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:36.9017673Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:36.9018490Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:36.9019273Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:36.9020042Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:36.9020866Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:36.9021635Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:36.9022424Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:36.9023236Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:36.9024066Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:36.9024902Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:36.9025688Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:36.9026570Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:36.9027351Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:36.9028126Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:36.9028872Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:36.9029707Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:36.9030674Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:36.9031468Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:36.9032275Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:36.9033104Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:36.9033868Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:36.9035083Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:36.9035898Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:36.9036763Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:36.9037654Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:36.9038487Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:36.9039305Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:36.9040194Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:36.9041074Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:36.9041884Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:36.9042825Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.9043652Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:36.9044619Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.9045425Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:36.9046397Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.9047243Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:36.9048127Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:36.9048884Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:36.9049634Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:36.9050511Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:36.9051359Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:36.9052127Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:36.9052944Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:36.9053740Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:36.9054584Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:36.9055489Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:36.9056417Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:36.9057341Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:36.9058130Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:36.9058962Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:36.9059851Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:36.9060647Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:36.9061435Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:36.9062375Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:36.9063266Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:36.9064144Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:36.9065012Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:36.9065812Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:36.9066639Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:36.9067374Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:36.9068163Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:36.9068963Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:36.9069762Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:36.9070598Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:36.9071464Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:36.9072284Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:36.9073102Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:36.9073838Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:36.9074630Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:36.9075555Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:36.9076475Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:36.9077415Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:36.9078256Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:36.9079120Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:36.9079954Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:36.9080802Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:36.9081704Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:36.9082585Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:36.9083379Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:36.9084189Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:36.9085026Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:36.9085854Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:36.9086644Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:36.9087468Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:36.9088300Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:36.9089093Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:36.9089871Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:36.9090692Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:36.9091515Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:36.9092354Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:36.9093214Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:36.9094018Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:36.9094941Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:36.9095732Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:36.9096774Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:36.9097852Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:36.9098718Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:36.9099480Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:36.9100257Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:36.9101047Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:36.9101874Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:36.9102736Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:36.9103579Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:36.9104459Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:36.9105313Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:36.9106074Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:36.9106987Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:36.9107838Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:36.9108665Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:36.9109459Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:36.9110346Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:36.9111157Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:36.9111926Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:36.9112707Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:36.9113466Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:36.9114224Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:36.9115035Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:36.9115894Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:36.9116743Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:36.9117556Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:36.9118357Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:36.9119116Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:36.9119990Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:36.9120790Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:36.9121645Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:36.9122524Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:36.9123348Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:36.9124180Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:36.9124974Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:36.9125838Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:36.9126703Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:36.9127510Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:36.9128291Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:36.9129171Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:36.9129970Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:36.9130823Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:36.9131708Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:36.9132519Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:36.9133291Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:36.9134113Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:36.9134924Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:36.9135741Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:36.9136561Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:36.9137584Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:36.9138596Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:36.9139586Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:36.9140556Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:36.9141544Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:36.9142573Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:36.9143611Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:36.9144594Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:36.9145591Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:36.9146440Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:36.9147297Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:36.9148126Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:36.9149035Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:36.9149880Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:36.9150665Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:36.9151465Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:36.9152330Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:36.9153130Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:36.9153932Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:36.9154727Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:36.9155535Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:36.9156420Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:36.9157238Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:36.9158096Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:36.9158859Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:36.9159718Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:36.9160496Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:36.9161278Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:36.9162115Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:36.9162988Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:36.9163796Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:36.9164617Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:36.9165367Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:36.9166162Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:36.9166973Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:36.9167805Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:36.9168786Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:36.9169602Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:36.9170345Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:36.9171200Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:36.9172014Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:36.9172880Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:36.9173686Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:36.9174531Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:36.9175505Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:36.9176381Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:36.9177315Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:36.9178103Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:36.9179067Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:36.9179866Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:36.9180656Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:36.9181478Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:36.9182308Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:36.9183186Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:36.9184035Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:36.9184793Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:36.9185593Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:36.9186403Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:36.9187159Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:36.9187948Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:36.9188766Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:36.9189600Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:36.9190429Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:36.9191326Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.9192204Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.9192973Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:36.9193826Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:36.9194710Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:36.9195562Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:36.9196576Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:36.9197475Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:36.9198330Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:36.9199203Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:36.9200055Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:36.9201252Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:36.9202079Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:36.9202912Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:36.9203742Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:36.9204574Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:36.9205414Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:36.9206243Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:36.9207669Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:36.9208554Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:36.9209372Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:36.9210264Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:36.9211132Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:36.9212049Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:36.9212978Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:36.9213877Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:36.9214791Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:36.9215706Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:36.9216546Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:36.9217516Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:36.9218346Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:36.9219283Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:36.9220165Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:36.9221012Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:36.9221848Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:36.9222675Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:36.9223532Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:36.9224390Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:36.9225217Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:36.9226038Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:36.9226877Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:36.9227698Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:36.9228588Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:36.9229478Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:36.9230296Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:36.9231109Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:36.9232008Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:36.9232922Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:36.9233778Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:36.9234637Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:36.9235517Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:36.9236352Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:36.9237218Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:36.9238107Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:36.9238904Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:36.9239740Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:36.9240628Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:36.9241434Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:36.9242226Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:36.9243040Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:36.9243882Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:36.9244870Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:36.9245760Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:36.9246615Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:36.9247494Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:36.9248367Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:36.9249241Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:36.9250071Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:36.9251058Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:36.9251893Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:36.9252739Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:36.9253531Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:36.9254362Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:36.9255270Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:36.9256201Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:36.9257247Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:36.9258054Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:36.9258905Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:36.9259744Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:36.9260655Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:36.9261588Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:36.9262474Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:36.9263326Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:36.9264171Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:36.9265098Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:36.9265983Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:36.9266862Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:36.9267666Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:36.9268522Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:36.9269370Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:36.9270167Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:36.9271123Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:36.9272018Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:36.9272867Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:36.9273662Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:36.9274540Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:36.9275473Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:36.9276385Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:36.9277304Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:36.9278246Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:36.9279118Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:36.9279980Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:36.9280881Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:36.9281806Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:36.9282688Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:36.9283547Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:36.9284369Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:36.9285210Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:36.9286088Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:36.9286951Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:36.9287835Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:36.9288761Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:36.9289545Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:36.9290310Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:36.9291135Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:36.9291939Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:36.9292790Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:36.9293602Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:36.9294393Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:36.9295280Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:36.9296381Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:36.9297335Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:36.9298208Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:36.9298992Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:36.9299854Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:36.9300687Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:36.9301565Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:36.9302429Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:36.9303288Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:36.9304162Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:36.9305064Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:36.9305924Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:36.9306774Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:36.9307645Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:36.9308536Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:36.9309427Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:36.9310153Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:36.9311100Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:36.9311921Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:36.9312703Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:36.9313479Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:36.9314299Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:36.9315107Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:36.9315994Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:36.9317701Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:36.9319498Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:36.9321491Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:36.9323498Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:36.9325518Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:36.9327687Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:36.9329764Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:36.9331652Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:36.9333512Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:36.9335470Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:36.9337331Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:36.9339143Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:36.9340976Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:36.9342845Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:36.9344800Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:36.9346837Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:36.9348822Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:36.9350711Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:36.9352502Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:36.9354410Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:36.9359108Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:36.9364928Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:36.9366841Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:36.9369042Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:36.9370930Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:36.9372925Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:36.9375256Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:36.9377295Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:36.9379397Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:36.9381507Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:36.9383561Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:36.9385652Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:36.9387816Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:36.9390119Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:36.9392191Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:36.9394112Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:36.9396421Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:36.9398558Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:36.9400611Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:36.9402736Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:36.9405058Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:36.9406787Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:36.9408398Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:36.9410119Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:36.9411762Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:36.9413504Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:36.9415181Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:36.9417253Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:36.9419202Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:36.9421163Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:36.9422851Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:36.9424493Z V1204 11:23:05.814000 101180 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.9426089Z I1204 11:23:05.814000 101180 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:36.9427223Z V1204 11:23:05.814000 101180 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:36.9428347Z V1204 11:23:05.814000 101180 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:36.9429538Z I1204 11:23:05.815000 101180 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:36.9431344Z V1204 11:23:05.854000 101180 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:36.9432397Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.9433554Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:36.9435795Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.9437603Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.9440066Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.9442344Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.9444667Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.9446706Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:36.9448621Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:36.9451087Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:36.9453021Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:36.9455241Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.9457469Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.9458839Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.9459843Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.9460749Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:36.9462105Z V1204 11:23:05.862000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:36.9463438Z V1204 11:23:05.863000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:36.9465831Z V1204 11:23:05.863000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.9466992Z V1204 11:23:05.864000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.9469426Z V1204 11:23:05.896000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:36.9470686Z V1204 11:23:05.896000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.9472918Z V1204 11:23:05.898000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:36.9473986Z V1204 11:23:05.898000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.9476199Z V1204 11:23:05.899000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:36.9477280Z V1204 11:23:05.900000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.9479480Z V1204 11:23:05.901000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:36.9480485Z V1204 11:23:05.901000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.9482511Z V1204 11:23:05.904000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:36.9483523Z V1204 11:23:05.904000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.9485654Z V1204 11:23:05.907000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.9486709Z V1204 11:23:05.907000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.9488876Z V1204 11:23:05.908000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:36.9489917Z V1204 11:23:05.908000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:36.9491116Z V1204 11:23:05.909000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:36.9492494Z V1204 11:23:05.910000 101180 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:36.9494217Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:36.9494471Z warnings.warn( 2025-12-04T11:24:36.9495581Z V1204 11:23:05.910000 101180 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:36.9496766Z _ TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False _ 2025-12-04T11:24:36.9497102Z Traceback (most recent call last): 2025-12-04T11:24:36.9498404Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 384, in test_decompose_mm_mixed_precision 2025-12-04T11:24:36.9498632Z self.assertEqual( 2025-12-04T11:24:36.9499698Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:36.9500140Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:36.9501381Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:36.9501899Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:36.9502193Z AssertionError: Scalars are not equal! 2025-12-04T11:24:36.9502211Z 2025-12-04T11:24:36.9502459Z Expected 1 but got 0. 2025-12-04T11:24:36.9502691Z Absolute difference: 1 2025-12-04T11:24:36.9502931Z Relative difference: 1.0 2025-12-04T11:24:36.9502963Z 2025-12-04T11:24:36.9503437Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:36.9505386Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:36.9505400Z 2025-12-04T11:24:36.9506022Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:36.9506510Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:36.9506754Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:36.9507201Z stats [('calls_captured', 1)] 2025-12-04T11:24:36.9508940Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:36.9509446Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:36.9509660Z graph_break [] 2025-12-04T11:24:36.9510149Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:36.9511292Z I1204 11:23:02.873000 101180 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:36.9512648Z I1204 11:23:03.122000 101180 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:36.9513670Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:36.9514797Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:36.9517029Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:36.9518706Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:36.9519819Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:36.9520960Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:36.9521915Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.9524292Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:36.9525998Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:36.9527074Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:36.9528086Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.9528950Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:36.9530292Z I1204 11:23:05.325000 101180 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:36.9531929Z V1204 11:23:05.803000 101180 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:36.9533560Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:36.9535098Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:36.9535723Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.9536360Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.9537103Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.9538168Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:36.9540066Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:36.9541944Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:36.9543383Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:36.9545146Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:36.9546881Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:36.9547945Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:36.9548605Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:36.9549930Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:36.9555343Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.9560786Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:36.9562315Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:36.9564497Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:36.9566167Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:36.9568064Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:36.9569878Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:36.9571587Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:36.9573246Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:36.9574910Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:36.9576560Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:36.9578311Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:36.9579910Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:36.9581493Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:36.9583589Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:36.9585415Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:36.9587993Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:36.9589844Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:36.9655577Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:36.9658195Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:36.9659977Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:36.9661754Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:36.9663725Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:36.9665694Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:36.9667363Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:36.9670271Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:36.9672157Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:36.9674010Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:36.9676048Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:36.9678061Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:36.9680130Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:36.9682178Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:36.9684268Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:36.9686478Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:36.9688513Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:36.9690464Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:36.9692869Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:36.9694711Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:36.9696832Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:36.9698594Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:36.9700546Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:36.9702309Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:36.9703967Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:36.9705809Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:36.9707667Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:36.9709456Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:36.9711451Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:36.9713282Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:36.9715128Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:36.9716973Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:36.9718784Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:36.9720714Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:36.9722660Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:36.9724455Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:36.9726363Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:36.9728211Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:36.9730049Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:36.9731933Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:36.9733858Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:36.9735700Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:36.9737560Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:36.9739511Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:36.9741428Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:36.9743455Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:36.9745322Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:36.9747369Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:36.9749223Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:36.9751025Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:36.9752815Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:36.9754734Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:36.9756947Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:36.9758870Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:36.9760812Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:36.9762811Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:36.9764590Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:36.9767415Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:36.9769368Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:36.9771446Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:36.9773555Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:36.9775506Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:36.9777459Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:36.9779495Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:36.9781661Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:36.9783560Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:36.9785807Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.9787777Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:36.9790084Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.9791984Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:36.9794208Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:36.9796429Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:36.9798534Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:36.9800312Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:36.9802055Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:36.9804004Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:36.9805963Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:36.9807778Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:36.9809688Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:36.9811493Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:36.9813453Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:36.9815473Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:36.9817681Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:36.9819751Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:36.9821526Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:36.9823451Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:36.9825490Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:36.9827360Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:36.9829155Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:36.9831237Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:36.9833296Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:36.9835400Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:36.9837469Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:36.9839363Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:36.9841271Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:36.9842989Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:36.9844928Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:36.9846841Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:36.9848727Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:36.9850697Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:36.9852667Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:36.9854632Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:36.9856603Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:36.9858402Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:36.9860291Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:36.9862482Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:36.9864686Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:36.9866895Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:36.9868824Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:36.9870867Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:36.9872924Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:36.9874906Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:36.9877086Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:36.9879113Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:36.9880976Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:36.9882981Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:36.9884952Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:36.9886950Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:36.9888784Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:36.9890685Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:36.9892675Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:36.9894588Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:36.9896547Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:36.9898485Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:36.9900462Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:36.9902455Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:36.9904545Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:36.9906371Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:36.9908521Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:36.9910376Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:36.9912502Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:36.9914451Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:36.9916308Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:36.9918126Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:36.9920054Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:36.9921930Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:36.9923900Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:36.9925901Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:36.9927894Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:36.9929857Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:36.9931975Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:36.9933769Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:36.9935921Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:36.9937944Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:36.9939870Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:36.9941767Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:36.9943740Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:36.9945569Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:36.9947399Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:36.9949291Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:36.9951084Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:36.9952837Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:36.9954705Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:36.9956785Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:36.9958817Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:36.9960761Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:36.9962656Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:36.9964469Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:36.9966400Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:36.9968330Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:36.9970376Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:36.9972469Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:36.9974395Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:36.9976398Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:36.9978334Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:36.9990491Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:36.9992552Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:36.9994598Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:36.9996707Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:36.9998733Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:37.0000642Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:37.0002572Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:37.0004747Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:37.0006708Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:37.0008493Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:37.0010448Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:37.0012338Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:37.0014209Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:37.0016156Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:37.0018611Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:37.0021032Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:37.0023346Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:37.0025666Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:37.0027995Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:37.0030407Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:37.0032854Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:37.0035273Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:37.0037649Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:37.0039636Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:37.0041626Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:37.0043621Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:37.0045818Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:37.0047818Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:37.0049688Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:37.0051543Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:37.0053343Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:37.0055264Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:37.0057248Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:37.0059150Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:37.0061023Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:37.0063151Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:37.0065104Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:37.0067032Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:37.0068840Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:37.0070822Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:37.0072730Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:37.0074558Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:37.0076565Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:37.0078539Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:37.0080419Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:37.0082435Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:37.0084163Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:37.0086038Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:37.0087918Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:37.0089902Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:37.0092161Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:37.0094111Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:37.0095863Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:37.0097990Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:37.0099890Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:37.0101871Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:37.0103758Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:37.0105626Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:37.0107914Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:37.0109937Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:37.0112032Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:37.0113926Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:37.0116084Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:37.0117990Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:37.0119915Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:37.0121888Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:37.0123849Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:37.0125893Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:37.0127879Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:37.0129647Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:37.0131659Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:37.0133547Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:37.0135360Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:37.0137287Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:37.0139131Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:37.0141134Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:37.0143001Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:37.0145075Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.0147208Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.0149211Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:37.0151240Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:37.0153236Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:37.0155275Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:37.0157227Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:37.0159376Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:37.0161368Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:37.0163444Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:37.0165334Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:37.0167292Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:37.0169311Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:37.0171255Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:37.0173234Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:37.0175168Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:37.0177258Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:37.0179201Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:37.0181237Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:37.0183172Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:37.0185090Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:37.0187275Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:37.0189318Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:37.0191422Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:37.0193595Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:37.0195794Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:37.0198145Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:37.0200298Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:37.0202316Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:37.0204355Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:37.0206348Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:37.0208625Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:37.0210727Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:37.0212741Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:37.0214692Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:37.0216674Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:37.0218647Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:37.0220651Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:37.0222537Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:37.0224528Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:37.0226437Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:37.0228358Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:37.0230349Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:37.0232471Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:37.0234380Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:37.0236236Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:37.0238326Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:37.0240413Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:37.0242355Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:37.0244357Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:37.0246414Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:37.0248381Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:37.0250360Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:37.0252452Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:37.0254265Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:37.0256140Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:37.0258146Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:37.0260044Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:37.0261950Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:37.0263816Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:37.0265781Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:37.0267977Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:37.0270092Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.0272061Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:37.0274105Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:37.0276125Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:37.0278114Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:37.0279990Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:37.0282310Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:37.0284268Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:37.0286195Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:37.0288026Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:37.0289935Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:37.0292029Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:37.0294106Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:37.0296471Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:37.0298521Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:37.0300460Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:37.0302413Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:37.0304465Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:37.0306581Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:37.0308713Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:37.0310663Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:37.0312630Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:37.0314742Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:37.0316766Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:37.0318764Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:37.0320716Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:37.0322699Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:37.0324645Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:37.0326506Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:37.0328674Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:37.0330680Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:37.0332531Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:37.0334445Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:37.0336532Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:37.0338732Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:37.0340877Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:37.0342906Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:37.0345139Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:37.0347128Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:37.0349165Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:37.0351201Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:37.0353363Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:37.0355355Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:37.0357357Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:37.0359271Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:37.0361186Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:37.0363235Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:37.0365209Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:37.0367283Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:37.0369336Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:37.0371101Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:37.0372927Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:37.0374799Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:37.0376683Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:37.0378610Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:37.0380519Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:37.0382448Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:37.0384539Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:37.0386803Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:37.0388903Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:37.0390461Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:37.0391242Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:37.0392161Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:37.0392993Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:37.0393851Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:37.0394723Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:37.0395571Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:37.0396821Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:37.0397658Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:37.0398575Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:37.0399461Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:37.0400359Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:37.0401238Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:37.0402069Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:37.0402811Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:37.0403793Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:37.0404630Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:37.0405396Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:37.0406187Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:37.0406980Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:37.0407781Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:37.0408696Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:37.0409469Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:37.0410233Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:37.0411060Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:37.0411911Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:37.0412763Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:37.0413636Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:37.0414548Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:37.0415368Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:37.0416168Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:37.0417039Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:37.0417792Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:37.0418555Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:37.0419348Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:37.0420147Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:37.0420964Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:37.0421829Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:37.0422653Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:37.0423441Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:37.0424230Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:37.0425022Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:37.0426998Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:37.0429437Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:37.0430201Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:37.0431153Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:37.0431986Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:37.0432819Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:37.0433791Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:37.0434632Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:37.0435576Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:37.0436486Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:37.0437381Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:37.0438283Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:37.0439187Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:37.0440194Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:37.0441101Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:37.0441947Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:37.0442870Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:37.0443802Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:37.0444687Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:37.0445569Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:37.0446540Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:37.0447306Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:37.0448050Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:37.0448792Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:37.0449512Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:37.0450231Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:37.0450954Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:37.0451875Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:37.0452718Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:37.0453579Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:37.0454300Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:37.0454965Z V1204 11:23:05.814000 101180 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.0455635Z I1204 11:23:05.814000 101180 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.0456152Z V1204 11:23:05.814000 101180 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:37.0456648Z V1204 11:23:05.814000 101180 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:37.0457212Z I1204 11:23:05.815000 101180 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:37.0457996Z V1204 11:23:05.854000 101180 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:37.0458441Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.0458959Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:37.0459919Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.0460665Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.0461744Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.0462743Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.0463743Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.0464581Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:37.0465410Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:37.0466478Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:37.0467320Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:37.0468250Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.0469152Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.0469740Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.0470210Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.0470611Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.0471183Z V1204 11:23:05.862000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:37.0471763Z V1204 11:23:05.863000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:37.0472774Z V1204 11:23:05.863000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.0473290Z V1204 11:23:05.864000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.0474296Z V1204 11:23:05.896000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.0474796Z V1204 11:23:05.896000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.0475784Z V1204 11:23:05.898000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:37.0476268Z V1204 11:23:05.898000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.0477220Z V1204 11:23:05.899000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:37.0477675Z V1204 11:23:05.900000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.0478599Z V1204 11:23:05.901000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:37.0479029Z V1204 11:23:05.901000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.0479909Z V1204 11:23:05.904000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:37.0480356Z V1204 11:23:05.904000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.0481248Z V1204 11:23:05.907000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.0481707Z V1204 11:23:05.907000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.0482594Z V1204 11:23:05.908000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.0483042Z V1204 11:23:05.908000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.0483565Z V1204 11:23:05.909000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.0484171Z V1204 11:23:05.910000 101180 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:37.0484925Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:37.0485030Z warnings.warn( 2025-12-04T11:24:37.0485508Z V1204 11:23:05.910000 101180 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:37.0485743Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:37.0485860Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:37.0485988Z stats [('calls_captured', 1)] 2025-12-04T11:24:37.0486214Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:37.0486979Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:37.0487094Z graph_break [] 2025-12-04T11:24:37.0487315Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:37.0487768Z V1204 11:23:06.129000 101180 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpvrodqez3 2025-12-04T11:24:37.0488240Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.0488731Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:37.0489718Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.0490447Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.0490948Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:37.0491436Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:37.0491859Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.0492928Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.0493680Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:37.0494169Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:37.0494589Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.0494988Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.0495708Z V1204 11:23:06.403000 101180 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:37.0496599Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:37.0497409Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:37.0497693Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.0497988Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.0498264Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.0498736Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:37.0499572Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.0500389Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.0501021Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:37.0501784Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.0502590Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.0503077Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:37.0503372Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.0503959Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:37.0506291Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.0508664Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.0509313Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:37.0510240Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:37.0510980Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:37.0511839Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:37.0512626Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:37.0513352Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:37.0514072Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:37.0514796Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:37.0515526Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:37.0516258Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:37.0516951Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:37.0517666Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:37.0518537Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:37.0519372Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:37.0520562Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:37.0521359Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:37.0550385Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:37.0551416Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:37.0552192Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:37.0552963Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:37.0553818Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:37.0554667Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:37.0555503Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:37.0556304Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:37.0557102Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:37.0557882Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:37.0558736Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:37.0559616Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:37.0560483Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:37.0561350Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:37.0562200Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:37.0563111Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:37.0563971Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:37.0564811Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:37.0565828Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:37.0566588Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:37.0567435Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:37.0568181Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:37.0569036Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:37.0569795Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:37.0570498Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:37.0571335Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:37.0572110Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:37.0572873Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:37.0573702Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:37.0574479Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:37.0575254Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:37.0576056Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:37.0576890Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:37.0577682Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:37.0578468Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:37.0579233Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:37.0580050Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:37.0580863Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:37.0581636Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:37.0582419Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:37.0583235Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:37.0584018Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:37.0584778Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:37.0585604Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:37.0586404Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:37.0587263Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:37.0588092Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:37.0588960Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:37.0589748Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:37.0590512Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:37.0591301Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:37.0592106Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:37.0593049Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:37.0593839Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:37.0594631Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:37.0595484Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:37.0596388Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:37.0597660Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:37.0598460Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:37.0599341Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:37.0600229Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:37.0601050Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:37.0601875Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:37.0602765Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:37.0603660Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:37.0604517Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:37.0605470Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.0606292Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:37.0607246Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.0608098Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:37.0609034Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.0609866Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:37.0610738Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:37.0611515Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:37.0612269Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:37.0613148Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:37.0613997Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:37.0614761Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:37.0615592Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:37.0616378Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:37.0617284Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:37.0618147Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:37.0619118Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:37.0620009Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:37.0620780Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:37.0621623Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:37.0622503Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:37.0623317Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:37.0624121Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:37.0625032Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:37.0625889Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:37.0626755Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:37.0627634Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:37.0628428Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:37.0629268Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:37.0629988Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:37.0630791Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:37.0631593Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:37.0632382Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:37.0633228Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:37.0634054Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:37.0634916Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:37.0635754Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:37.0636502Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:37.0637295Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:37.0638203Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:37.0639134Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:37.0640091Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:37.0640912Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:37.0641732Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:37.0642581Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:37.0643412Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:37.0644328Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:37.0645207Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:37.0645989Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:37.0646814Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:37.0647644Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:37.0648484Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:37.0649254Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:37.0650049Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:37.0650917Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:37.0651731Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:37.0652520Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:37.0653329Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:37.0654161Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:37.0654991Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:37.0655898Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:37.0656672Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:37.0657619Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:37.0658418Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:37.0659283Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:37.0660117Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:37.0660936Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:37.0661710Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:37.0662488Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:37.0663272Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:37.0664116Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:37.0664965Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:37.0665813Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:37.0666684Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:37.0667555Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:37.0668348Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:37.0669246Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:37.0670103Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:37.0670920Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:37.0671762Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:37.0672594Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:37.0673376Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:37.0674141Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:37.0674910Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:37.0675688Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:37.0676456Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:37.0677247Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:37.0678092Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:37.0678946Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:37.0679769Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:37.0680570Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:37.0681330Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:37.0682140Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:37.0682980Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:37.0683865Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:37.0684758Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:37.0685572Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:37.0686422Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:37.0687249Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:37.0688068Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:37.0688902Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:37.0689691Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:37.0690495Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:37.0691364Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:37.0692177Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:37.0693021Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:37.0693880Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:37.0694705Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:37.0695465Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:37.0696494Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:37.0697355Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:37.0698166Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:37.0699063Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:37.0700059Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:37.0701076Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:37.0702050Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:37.0703030Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:37.0704044Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:37.0705056Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:37.0706061Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:37.0707049Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:37.0708044Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:37.0708885Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:37.0709770Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:37.0710586Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:37.0711515Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:37.0712356Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:37.0713152Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:37.0713927Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:37.0714679Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:37.0715526Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:37.0716342Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:37.0717149Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:37.0717939Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:37.0718835Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:37.0719656Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:37.0720501Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:37.0721271Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:37.0722080Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:37.0722861Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:37.0723632Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:37.0724481Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:37.0725354Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:37.0726146Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:37.0726980Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:37.0727707Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:37.0728503Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:37.0729294Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:37.0730134Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:37.0731102Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:37.0731940Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:37.0732669Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:37.0733505Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:37.0734328Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:37.0735182Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:37.0736028Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:37.0736836Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:37.0737863Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:37.0738732Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:37.0739579Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:37.0740382Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:37.0741319Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:37.0742128Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:37.0742909Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:37.0743740Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:37.0744565Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:37.0745423Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:37.0746257Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:37.0747025Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:37.0747878Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:37.0748675Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:37.0749442Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:37.0750230Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:37.0751008Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:37.0751884Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:37.0752674Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:37.0753550Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.0754419Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.0755210Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:37.0756067Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:37.0756938Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:37.0757806Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:37.0758658Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:37.0759563Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:37.0760396Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:37.0761270Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:37.0762062Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:37.0762922Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:37.0763775Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:37.0764590Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:37.0765432Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:37.0766246Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:37.0767101Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:37.0767947Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:37.0768816Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:37.0769603Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:37.0770414Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:37.0771319Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:37.0772179Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:37.0773086Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:37.0773997Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:37.0774912Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:37.0775831Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:37.0776731Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:37.0777639Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:37.0778496Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:37.0779376Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:37.0780327Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:37.0781215Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:37.0782057Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:37.0782897Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:37.0783769Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:37.0784585Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:37.0785453Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:37.0786234Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:37.0787068Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:37.0787898Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:37.0788725Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:37.0789616Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:37.0790489Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:37.0791320Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:37.0792120Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:37.0793034Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:37.0793932Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:37.0794766Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:37.0795654Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:37.0796726Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:37.0797584Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:37.0798439Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:37.0799334Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:37.0800169Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:37.0800986Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:37.0801820Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:37.0802606Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:37.0803415Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:37.0804221Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:37.0805193Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:37.0806133Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:37.0807019Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.0807860Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:37.0808752Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:37.0809619Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:37.0810470Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:37.0811324Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:37.0812308Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:37.0813200Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:37.0814033Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:37.0814834Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:37.0815659Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:37.0816594Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:37.0817583Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:37.0818518Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:37.0819339Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:37.0820181Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:37.0821035Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:37.0821948Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:37.0822877Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:37.0823759Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:37.0824601Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:37.0825454Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:37.0826365Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:37.0827242Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:37.0828135Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:37.0828981Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:37.0829834Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:37.0830666Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:37.0831470Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:37.0832435Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:37.0833308Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:37.0834111Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:37.0834910Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:37.0835787Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:37.0836713Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:37.0837661Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:37.0838542Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:37.0839511Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:37.0840368Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:37.0841247Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:37.0842130Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:37.0843068Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:37.0843952Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:37.0844838Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:37.0845678Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:37.0846508Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:37.0847400Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:37.0848256Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:37.0849186Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:37.0850077Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:37.0850817Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:37.0851592Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:37.0852409Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:37.0853229Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:37.0854072Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:37.0854903Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:37.0855692Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:37.0856572Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:37.0857588Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:37.0858479Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:37.0859304Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:37.0860110Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:37.0861019Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:37.0861853Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:37.0862726Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:37.0863581Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:37.0864426Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:37.0865346Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:37.0866179Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:37.0867012Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:37.0867847Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:37.0868745Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:37.0869620Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:37.0870484Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:37.0871225Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:37.0872160Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:37.0873000Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:37.0873765Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:37.0874561Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:37.0875350Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:37.0876177Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:37.0877089Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:37.0877852Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:37.0878614Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:37.0879442Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:37.0880299Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:37.0881181Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:37.0882057Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:37.0882924Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:37.0883710Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:37.0884521Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:37.0885337Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:37.0886121Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:37.0886886Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:37.0887647Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:37.0888457Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:37.0889277Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:37.0890143Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:37.0890962Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:37.0891776Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:37.0892535Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:37.0893367Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:37.0895345Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:37.0898073Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:37.0898849Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:37.0899773Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:37.0900568Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:37.0901398Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:37.0902421Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:37.0903263Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:37.0904182Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:37.0905095Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:37.0905997Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:37.0906901Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:37.0907802Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:37.0908856Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:37.0909779Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:37.0910626Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:37.0911549Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:37.0912475Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:37.0913395Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:37.0914284Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:37.0915249Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:37.0915985Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:37.0916698Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:37.0917443Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:37.0918169Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:37.0918923Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:37.0919656Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:37.0920525Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:37.0921369Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:37.0922224Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:37.0922950Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:37.0923611Z V1204 11:23:06.413000 101180 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.0924311Z I1204 11:23:06.414000 101180 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.0924840Z V1204 11:23:06.414000 101180 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:37.0925322Z V1204 11:23:06.414000 101180 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:37.0925829Z I1204 11:23:06.414000 101180 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:37.0926606Z V1204 11:23:06.427000 101180 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:37.0927048Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.0927559Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:37.0928541Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.0929289Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.0930326Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.0931285Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.0932289Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.0933159Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:37.0933988Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:37.0935032Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:37.0935868Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:37.0936795Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.0937758Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.0938382Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.0938840Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.0939244Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.0939820Z V1204 11:23:06.434000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:37.0940402Z V1204 11:23:06.435000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:37.0941405Z V1204 11:23:06.435000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.0941952Z V1204 11:23:06.435000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.0942960Z V1204 11:23:06.437000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.0943464Z V1204 11:23:06.437000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.0944416Z V1204 11:23:06.438000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:37.0944869Z V1204 11:23:06.438000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.0945813Z V1204 11:23:06.439000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:37.0946266Z V1204 11:23:06.440000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.0947213Z V1204 11:23:06.441000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:37.0947639Z V1204 11:23:06.441000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.0948487Z V1204 11:23:06.443000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:37.0948935Z V1204 11:23:06.444000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.0949829Z V1204 11:23:06.446000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.0950295Z V1204 11:23:06.446000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.0951181Z V1204 11:23:06.447000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.0951675Z V1204 11:23:06.447000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.0952184Z V1204 11:23:06.448000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.0952792Z V1204 11:23:06.448000 101180 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:37.0953549Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:37.0953653Z warnings.warn( 2025-12-04T11:24:37.0954146Z V1204 11:23:06.449000 101180 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:37.0954295Z =================================== FAILURES =================================== 2025-12-04T11:24:37.0954746Z _ TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False _ 2025-12-04T11:24:37.0954882Z Traceback (most recent call last): 2025-12-04T11:24:37.0955476Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 384, in test_decompose_mm_mixed_precision 2025-12-04T11:24:37.0955588Z self.assertEqual( 2025-12-04T11:24:37.0956057Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:37.0956221Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:37.0956775Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:37.0956983Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:37.0957116Z AssertionError: Scalars are not equal! 2025-12-04T11:24:37.0957127Z 2025-12-04T11:24:37.0957249Z Expected 1 but got 0. 2025-12-04T11:24:37.0957361Z Absolute difference: 1 2025-12-04T11:24:37.0957473Z Relative difference: 1.0 2025-12-04T11:24:37.0957492Z 2025-12-04T11:24:37.0957706Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:37.0958554Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:37.0958560Z 2025-12-04T11:24:37.0958849Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:37.0959115Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:37.0959228Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:37.0959358Z stats [('calls_captured', 1)] 2025-12-04T11:24:37.0960117Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:37.0960354Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:37.0960453Z graph_break [] 2025-12-04T11:24:37.0960673Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:37.0961191Z I1204 11:23:02.873000 101180 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:37.0961786Z I1204 11:23:03.122000 101180 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:37.0962240Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.0962727Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:37.0963669Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.0964443Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.0964966Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:37.0965466Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:37.0965883Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.0966925Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.0967699Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:37.0968176Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:37.0968608Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.0968993Z V1204 11:23:03.124000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.0969561Z I1204 11:23:05.325000 101180 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:37.0970275Z V1204 11:23:05.803000 101180 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:37.0970989Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:37.0971665Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:37.0971947Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.0972271Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.0972547Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.0973023Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:37.0973859Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.0974677Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.0975314Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:37.0976081Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.0976901Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.0977395Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:37.0977723Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.0978312Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:37.0980650Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.0982995Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.0983650Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:37.0984577Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:37.0985305Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:37.0986138Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:37.0986918Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:37.0987673Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:37.0988390Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:37.0989109Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:37.0989839Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:37.0990571Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:37.0991268Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:37.0991955Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:37.0992841Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:37.0993657Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:37.0994858Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:37.0995670Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:37.1024923Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:37.1025958Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:37.1026728Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:37.1027504Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:37.1028359Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:37.1029203Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:37.1029938Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:37.1030751Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:37.1031585Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:37.1032364Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:37.1033222Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:37.1034070Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:37.1034973Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:37.1035831Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:37.1036677Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:37.1037587Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:37.1038444Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:37.1039277Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:37.1040312Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:37.1041075Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:37.1041923Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:37.1042669Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:37.1043521Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:37.1044279Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:37.1044997Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:37.1045794Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:37.1046573Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:37.1047370Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:37.1048204Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:37.1048974Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:37.1049752Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:37.1050538Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:37.1051332Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:37.1052116Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:37.1052901Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:37.1053661Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:37.1054472Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:37.1055255Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:37.1056062Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:37.1056828Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:37.1057680Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:37.1058465Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:37.1059233Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:37.1060062Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:37.1060862Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:37.1061697Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:37.1062514Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:37.1063411Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:37.1064193Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:37.1064947Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:37.1065712Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:37.1066608Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:37.1067549Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:37.1068348Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:37.1069136Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:37.1069981Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:37.1070735Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:37.1071963Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:37.1072761Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:37.1073644Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:37.1074533Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:37.1075352Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:37.1076187Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:37.1077038Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:37.1077929Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:37.1078761Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:37.1079737Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.1080567Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:37.1081536Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.1082340Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:37.1083298Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.1084122Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:37.1084997Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:37.1085768Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:37.1086515Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:37.1087370Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:37.1088246Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:37.1089000Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:37.1089828Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:37.1090611Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:37.1091464Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:37.1092328Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:37.1093267Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:37.1094151Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:37.1094962Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:37.1095795Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:37.1096889Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:37.1097716Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:37.1098491Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:37.1099451Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:37.1100317Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:37.1101185Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:37.1102065Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:37.1102858Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:37.1103665Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:37.1104424Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:37.1105223Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:37.1106027Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:37.1106830Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:37.1107660Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:37.1108492Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:37.1109327Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:37.1110172Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:37.1110925Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:37.1111758Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:37.1112682Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:37.1113602Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:37.1114528Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:37.1115377Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:37.1116202Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:37.1117051Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:37.1117876Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:37.1118790Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:37.1119644Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:37.1120454Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:37.1121280Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:37.1122107Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:37.1122950Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:37.1123726Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:37.1124526Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:37.1125353Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:37.1126171Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:37.1126989Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:37.1127794Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:37.1128628Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:37.1129456Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:37.1130336Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:37.1131141Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:37.1132026Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:37.1132822Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:37.1133685Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:37.1134511Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:37.1135304Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:37.1136102Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:37.1136929Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:37.1137714Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:37.1138554Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:37.1139399Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:37.1140250Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:37.1141067Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:37.1141935Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:37.1142731Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:37.1143664Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:37.1144506Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:37.1145335Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:37.1146148Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:37.1147008Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:37.1147794Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:37.1148561Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:37.1149344Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:37.1150112Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:37.1150861Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:37.1151655Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:37.1152537Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:37.1153394Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:37.1154215Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:37.1155026Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:37.1155789Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:37.1156601Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:37.1157413Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:37.1158297Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:37.1159212Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:37.1160034Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:37.1160884Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:37.1161681Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:37.1162528Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:37.1163369Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:37.1164162Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:37.1164963Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:37.1165834Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:37.1166647Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:37.1167467Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:37.1168359Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:37.1169192Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:37.1169952Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:37.1170787Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:37.1171582Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:37.1172394Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:37.1173214Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:37.1174196Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:37.1175242Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:37.1176220Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:37.1177256Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:37.1178232Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:37.1179275Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:37.1180273Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:37.1181262Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:37.1182264Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:37.1183106Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:37.1183947Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:37.1184788Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:37.1185712Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:37.1186554Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:37.1187349Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:37.1188131Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:37.1188896Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:37.1189694Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:37.1190510Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:37.1191340Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:37.1192131Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:37.1193027Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:37.1193844Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:37.1194671Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:37.1195461Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:37.1196424Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:37.1197213Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:37.1197983Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:37.1198834Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:37.1199679Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:37.1200538Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:37.1201355Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:37.1202086Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:37.1202885Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:37.1203686Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:37.1204529Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:37.1205476Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:37.1206324Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:37.1207058Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:37.1207928Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:37.1208752Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:37.1209600Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:37.1210421Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:37.1211263Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:37.1212217Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:37.1213088Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:37.1213931Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:37.1214733Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:37.1215643Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:37.1216495Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:37.1217341Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:37.1218178Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:37.1219009Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:37.1219869Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:37.1220711Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:37.1221453Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:37.1222312Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:37.1223112Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:37.1223906Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:37.1224700Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:37.1225476Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:37.1226329Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:37.1227150Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:37.1228032Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.1228910Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.1229700Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:37.1230551Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:37.1231405Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:37.1232287Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:37.1233144Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:37.1234050Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:37.1234888Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:37.1235770Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:37.1236571Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:37.1237406Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:37.1238268Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:37.1239086Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:37.1239963Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:37.1240778Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:37.1241632Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:37.1242451Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:37.1243350Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:37.1244143Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:37.1244948Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:37.1245850Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:37.1246711Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:37.1247591Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:37.1248542Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:37.1249459Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:37.1250380Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:37.1251297Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:37.1252135Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:37.1252991Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:37.1253832Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:37.1254790Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:37.1255709Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:37.1256552Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:37.1257451Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:37.1258277Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:37.1259091Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:37.1259987Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:37.1260775Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:37.1261605Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:37.1262424Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:37.1263257Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:37.1264116Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:37.1265026Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:37.1265859Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:37.1266663Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:37.1267576Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:37.1268474Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:37.1269311Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:37.1270170Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:37.1271092Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:37.1271964Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:37.1272819Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:37.1273717Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:37.1274501Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:37.1275373Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:37.1276217Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:37.1277023Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:37.1277818Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:37.1278623Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:37.1279482Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:37.1280435Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:37.1281368Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.1282205Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:37.1283101Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:37.1283967Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:37.1284823Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:37.1285633Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:37.1286609Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:37.1287491Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:37.1288352Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:37.1289154Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:37.1289988Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:37.1290898Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:37.1291844Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:37.1292784Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:37.1293606Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:37.1294435Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:37.1295284Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:37.1296322Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:37.1297404Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:37.1298291Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:37.1299138Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:37.1299992Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:37.1300899Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:37.1301774Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:37.1302635Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:37.1303502Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:37.1304393Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:37.1305231Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:37.1306042Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:37.1306973Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:37.1307893Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:37.1308710Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:37.1309519Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:37.1310399Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:37.1311339Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:37.1312254Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:37.1313164Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:37.1314116Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:37.1314972Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:37.1315857Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:37.1316739Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:37.1317680Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:37.1318541Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:37.1319413Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:37.1320354Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:37.1321188Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:37.1322084Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:37.1322949Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:37.1323847Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:37.1324762Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:37.1325509Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:37.1326286Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:37.1327101Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:37.1327927Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:37.1328749Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:37.1329607Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:37.1330401Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:37.1331299Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:37.1332253Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:37.1333144Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:37.1333965Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:37.1334729Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:37.1335641Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:37.1336511Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:37.1337450Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:37.1338304Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:37.1339146Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:37.1340034Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:37.1340890Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:37.1341723Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:37.1342553Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:37.1343435Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:37.1344318Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:37.1345145Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:37.1345912Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:37.1346848Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:37.1347675Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:37.1348440Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:37.1349226Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:37.1350008Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:37.1350796Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:37.1351699Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:37.1353121Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:37.1353883Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:37.1354716Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:37.1355578Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:37.1356424Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:37.1357342Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:37.1358194Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:37.1358978Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:37.1359780Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:37.1360591Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:37.1361347Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:37.1362138Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:37.1362909Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:37.1363697Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:37.1364518Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:37.1365385Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:37.1366213Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:37.1366995Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:37.1367754Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:37.1368594Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:37.1370588Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:37.1373042Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:37.1373809Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:37.1374732Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:37.1375525Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:37.1376359Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:37.1377397Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:37.1378279Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:37.1379208Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:37.1380114Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:37.1381019Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:37.1381921Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:37.1382838Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:37.1383828Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:37.1384738Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:37.1385618Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:37.1386546Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:37.1387477Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:37.1388355Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:37.1389272Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:37.1390240Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:37.1390979Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:37.1391693Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:37.1392438Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:37.1393160Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:37.1393876Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:37.1394635Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:37.1395511Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:37.1396493Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:37.1397354Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:37.1398077Z V1204 11:23:05.813000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:37.1398745Z V1204 11:23:05.814000 101180 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.1399417Z I1204 11:23:05.814000 101180 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.1399995Z V1204 11:23:05.814000 101180 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:37.1400480Z V1204 11:23:05.814000 101180 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:37.1401021Z I1204 11:23:05.815000 101180 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:37.1401806Z V1204 11:23:05.854000 101180 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:37.1402249Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.1402762Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:37.1403707Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.1404492Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.1405537Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.1406495Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.1407490Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.1408336Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:37.1409207Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:37.1410249Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:37.1411083Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:37.1412013Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.1412919Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.1413508Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.1413965Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.1414366Z V1204 11:23:05.859000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.1414969Z V1204 11:23:05.862000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:37.1415549Z V1204 11:23:05.863000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:37.1416555Z V1204 11:23:05.863000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.1417153Z V1204 11:23:05.864000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.1418210Z V1204 11:23:05.896000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.1418713Z V1204 11:23:05.896000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.1419671Z V1204 11:23:05.898000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:37.1420125Z V1204 11:23:05.898000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.1421073Z V1204 11:23:05.899000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:37.1421526Z V1204 11:23:05.900000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.1422453Z V1204 11:23:05.901000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:37.1422910Z V1204 11:23:05.901000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.1423764Z V1204 11:23:05.904000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:37.1424211Z V1204 11:23:05.904000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.1425104Z V1204 11:23:05.907000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.1425563Z V1204 11:23:05.907000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.1426455Z V1204 11:23:05.908000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.1426914Z V1204 11:23:05.908000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.1427429Z V1204 11:23:05.909000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.1428039Z V1204 11:23:05.910000 101180 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:37.1428820Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:37.1428925Z warnings.warn( 2025-12-04T11:24:37.1429422Z V1204 11:23:05.910000 101180 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:37.1429647Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:37.1429761Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:37.1429890Z stats [('calls_captured', 1)] 2025-12-04T11:24:37.1430119Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:37.1430880Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:37.1430991Z graph_break [] 2025-12-04T11:24:37.1431251Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:37.1431708Z V1204 11:23:06.129000 101180 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpvrodqez3 2025-12-04T11:24:37.1432146Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.1432642Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:37.1433596Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.1434324Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.1434831Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:37.1435315Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:37.1435776Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.1436806Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.1437555Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:37.1438054Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:37.1438484Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.1438883Z V1204 11:23:06.151000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.1439598Z V1204 11:23:06.403000 101180 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:37.1440313Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:37.1441027Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:37.1441341Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.1441632Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.1441907Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.1442380Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:37.1443212Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.1444027Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.1444702Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:37.1445471Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.1446238Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.1446691Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:37.1446988Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.1447581Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:37.1449967Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.1452291Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.1452944Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:37.1453874Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:37.1454599Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:37.1455427Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:37.1456239Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:37.1457072Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:37.1457793Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:37.1458519Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:37.1459246Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:37.1459977Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:37.1460710Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:37.1461401Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:37.1462289Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:37.1463082Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:37.1464237Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:37.1465042Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:37.1493894Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:37.1494981Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:37.1495749Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:37.1497091Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:37.1498119Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:37.1498966Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:37.1499700Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:37.1500481Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:37.1501279Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:37.1502056Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:37.1502923Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:37.1503809Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:37.1504670Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:37.1505531Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:37.1506387Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:37.1507296Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:37.1508150Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:37.1508978Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:37.1510026Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:37.1510825Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:37.1511676Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:37.1512492Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:37.1513343Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:37.1514136Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:37.1514858Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:37.1515651Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:37.1516402Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:37.1517170Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:37.1518008Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:37.1518782Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:37.1519593Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:37.1520381Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:37.1521146Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:37.1521929Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:37.1522715Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:37.1523483Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:37.1524295Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:37.1525110Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:37.1525924Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:37.1526689Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:37.1527499Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:37.1528275Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:37.1529038Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:37.1529887Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:37.1530685Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:37.1531522Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:37.1532308Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:37.1533172Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:37.1533960Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:37.1534741Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:37.1535506Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:37.1536310Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:37.1537306Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:37.1538107Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:37.1538909Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:37.1539761Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:37.1540510Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:37.1541828Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:37.1542655Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:37.1543542Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:37.1544431Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:37.1545261Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:37.1546101Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:37.1546952Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:37.1547846Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:37.1548651Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:37.1549602Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.1550433Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:37.1551453Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.1552255Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:37.1553192Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.1554021Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:37.1554888Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:37.1555660Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:37.1556409Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:37.1557286Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:37.1558169Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:37.1558933Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:37.1559763Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:37.1560546Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:37.1561403Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:37.1562293Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:37.1563236Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:37.1564091Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:37.1564868Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:37.1565708Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:37.1566582Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:37.1567422Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:37.1568195Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:37.1569098Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:37.1569962Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:37.1570848Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:37.1571710Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:37.1572502Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:37.1573340Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:37.1574089Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:37.1574882Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:37.1575686Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:37.1576491Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:37.1577393Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:37.1578265Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:37.1579100Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:37.1579903Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:37.1580647Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:37.1581438Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:37.1582373Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:37.1583329Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:37.1584255Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:37.1585075Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:37.1585899Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:37.1586746Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:37.1587576Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:37.1588486Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:37.1589368Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:37.1590184Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:37.1591014Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:37.1591844Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:37.1592682Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:37.1593453Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:37.1594282Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:37.1595113Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:37.1595916Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:37.1596915Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:37.1597728Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:37.1598567Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:37.1599466Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:37.1600349Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:37.1601142Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:37.1602071Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:37.1613207Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:37.1614228Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:37.1615064Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:37.1615993Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:37.1616805Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:37.1617655Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:37.1618432Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:37.1619254Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:37.1620104Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:37.1621002Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:37.1621827Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:37.1622674Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:37.1623425Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:37.1624316Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:37.1625163Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:37.1626016Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:37.1626807Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:37.1627637Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:37.1628401Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:37.1629169Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:37.1629937Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:37.1630698Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:37.1631441Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:37.1632247Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:37.1633122Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:37.1633962Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:37.1634774Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:37.1635558Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:37.1636343Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:37.1637161Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:37.1637957Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:37.1638797Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:37.1639667Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:37.1640479Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:37.1641307Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:37.1642118Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:37.1642936Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:37.1643755Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:37.1644545Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:37.1645328Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:37.1646198Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:37.1646989Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:37.1648184Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:37.1649085Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:37.1649892Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:37.1650644Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:37.1651453Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:37.1652279Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:37.1653068Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:37.1653882Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:37.1654843Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:37.1655840Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:37.1656820Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:37.1657865Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:37.1658836Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:37.1659824Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:37.1660823Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:37.1661793Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:37.1662779Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:37.1663616Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:37.1664466Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:37.1665316Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:37.1666216Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:37.1667056Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:37.1667828Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:37.1668640Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:37.1669401Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:37.1670197Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:37.1670979Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:37.1671762Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:37.1672554Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:37.1673434Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:37.1674287Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:37.1675100Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:37.1675850Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:37.1676668Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:37.1677437Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:37.1678209Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:37.1679037Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:37.1679901Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:37.1680724Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:37.1681540Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:37.1682268Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:37.1683044Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:37.1683833Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:37.1684683Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:37.1685633Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:37.1686425Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:37.1687150Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:37.1687991Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:37.1688802Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:37.1689689Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:37.1690488Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:37.1691293Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:37.1692226Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:37.1693094Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:37.1693941Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:37.1694719Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:37.1695655Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:37.1696659Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:37.1697515Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:37.1698322Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:37.1699142Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:37.1700011Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:37.1700886Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:37.1701627Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:37.1702431Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:37.1703229Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:37.1703979Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:37.1704760Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:37.1705584Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:37.1706407Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:37.1707192Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:37.1708052Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.1708928Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.1709696Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:37.1710549Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:37.1711418Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:37.1712266Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:37.1713164Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:37.1714048Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:37.1714883Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:37.1715747Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:37.1716568Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:37.1717382Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:37.1718194Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:37.1719003Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:37.1719825Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:37.1720648Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:37.1721508Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:37.1722324Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:37.1723168Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:37.1723952Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:37.1724759Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:37.1725640Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:37.1726501Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:37.1727394Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:37.1728336Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:37.1729232Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:37.1730154Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:37.1731047Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:37.1731881Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:37.1732775Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:37.1733605Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:37.1734530Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:37.1735406Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:37.1736250Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:37.1737138Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:37.1737989Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:37.1738802Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:37.1739651Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:37.1740442Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:37.1741255Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:37.1742080Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:37.1742885Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:37.1743788Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:37.1744693Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:37.1745500Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:37.1746297Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:37.1747181Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:37.1748108Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:37.1748932Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:37.1749786Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:37.1750649Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:37.1751507Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:37.1752360Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:37.1753288Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:37.1754125Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:37.1755021Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:37.1755873Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:37.1756838Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:37.1757834Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:37.1758673Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:37.1759610Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:37.1760640Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:37.1761741Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.1762652Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:37.1763619Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:37.1764522Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:37.1765451Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:37.1766318Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:37.1767374Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:37.1768316Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:37.1769187Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:37.1770059Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:37.1770967Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:37.1771969Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:37.1772919Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:37.1773904Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:37.1774791Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:37.1775664Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:37.1776573Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:37.1777698Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:37.1778755Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:37.1779670Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:37.1780560Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:37.1781490Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:37.1782447Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:37.1783451Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:37.1784348Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:37.1785247Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:37.1786144Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:37.1787051Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:37.1787917Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:37.1788940Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:37.1789896Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:37.1790741Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:37.1791628Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:37.1792530Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:37.1793587Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:37.1794553Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:37.1795626Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:37.1796821Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:37.1797724Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:37.1798709Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:37.1799659Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:37.1800740Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:37.1801642Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:37.1802568Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:37.1803399Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:37.1804398Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:37.1805316Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:37.1806243Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:37.1807210Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:37.1808149Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:37.1808997Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:37.1809817Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:37.1810717Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:37.1811553Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:37.1812464Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:37.1813387Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:37.1814247Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:37.1815302Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:37.1816294Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:37.1817361Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:37.1818251Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:37.1819123Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:37.1820037Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:37.1820897Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:37.1821864Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:37.1822763Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:37.1823696Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:37.1824639Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:37.1825581Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:37.1826445Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:37.1827364Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:37.1828281Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:37.1829243Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:37.1830242Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:37.1831075Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:37.1832142Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:37.1832996Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:37.1833917Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:37.1834752Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:37.1835635Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:37.1836527Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:37.1837453Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:37.1838295Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:37.1839124Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:37.1840045Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:37.1840960Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:37.1841851Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:37.1842842Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:37.1843770Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:37.1844651Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:37.1845483Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:37.1846387Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:37.1847186Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:37.1848146Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:37.1848945Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:37.1849773Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:37.1850674Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:37.1851647Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:37.1852609Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:37.1853443Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:37.1854289Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:37.1855119Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:37.1857270Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:37.1859740Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:37.1860658Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:37.1861618Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:37.1862510Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:37.1863377Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:37.1864372Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:37.1865364Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:37.1866355Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:37.1867363Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:37.1868279Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:37.1869274Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:37.1870222Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:37.1871357Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:37.1872292Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:37.1873221Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:37.1874231Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:37.1875189Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:37.1876213Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:37.1877156Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:37.1878215Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:37.1878994Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:37.1879781Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:37.1880536Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:37.1881330Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:37.1882195Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:37.1882986Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:37.1883944Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:37.1884820Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:37.1885773Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:37.1886547Z V1204 11:23:06.412000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:37.1887268Z V1204 11:23:06.413000 101180 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.1888025Z I1204 11:23:06.414000 101180 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.1888543Z V1204 11:23:06.414000 101180 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:37.1889101Z V1204 11:23:06.414000 101180 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:37.1889681Z I1204 11:23:06.414000 101180 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:37.1890618Z V1204 11:23:06.427000 101180 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:37.1891099Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.1891672Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:37.1892724Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.1893483Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.1894671Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.1895675Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.1896974Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.1897927Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:37.1898850Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:37.1899983Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:37.1900911Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:37.1901893Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.1902876Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.1903533Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.1904037Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.1904532Z V1204 11:23:06.432000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.1905157Z V1204 11:23:06.434000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:37.1905814Z V1204 11:23:06.435000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:37.1906907Z V1204 11:23:06.435000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.1907496Z V1204 11:23:06.435000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.1908654Z V1204 11:23:06.437000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.1909208Z V1204 11:23:06.437000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.1910243Z V1204 11:23:06.438000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:37.1910735Z V1204 11:23:06.438000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.1911711Z V1204 11:23:06.439000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:37.1912187Z V1204 11:23:06.440000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.1913138Z V1204 11:23:06.441000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:37.1913614Z V1204 11:23:06.441000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.1914478Z V1204 11:23:06.443000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:37.1914911Z V1204 11:23:06.444000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.1915818Z V1204 11:23:06.446000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.1916270Z V1204 11:23:06.446000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.1917210Z V1204 11:23:06.447000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.1917659Z V1204 11:23:06.447000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.1918173Z V1204 11:23:06.448000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.1918769Z V1204 11:23:06.448000 101180 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:37.1919511Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:37.1919635Z warnings.warn( 2025-12-04T11:24:37.1920124Z V1204 11:23:06.449000 101180 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:37.1920353Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:37.1920484Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:37.1920602Z stats [('calls_captured', 1)] 2025-12-04T11:24:37.1920858Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:37.1921632Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:37.1921734Z graph_break [] 2025-12-04T11:24:37.1921970Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:37.1922418Z V1204 11:23:06.465000 101180 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpye5dhuvu 2025-12-04T11:24:37.1922858Z V1204 11:23:06.487000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.1923372Z V1204 11:23:06.487000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:37.1924317Z V1204 11:23:06.487000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.1925062Z V1204 11:23:06.487000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.1925583Z V1204 11:23:06.487000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:37.1926080Z V1204 11:23:06.487000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:37.1926533Z V1204 11:23:06.487000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.1927564Z V1204 11:23:06.487000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.1928319Z V1204 11:23:06.487000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:37.1928794Z V1204 11:23:06.487000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:37.1929232Z V1204 11:23:06.487000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.1929647Z V1204 11:23:06.487000 101180 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.1930367Z V1204 11:23:06.726000 101180 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:37.1931094Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:37.1931761Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:37.1932055Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.1932337Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.1932614Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.1933103Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:37.1933966Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.1934794Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.1935416Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:37.1936199Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.1937012Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.1937468Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:37.1937776Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.1938347Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:37.1940698Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.1943090Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.1943749Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:37.1944695Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:37.1945435Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:37.1946249Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:37.1947043Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:37.1947759Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:37.1948476Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:37.1949212Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:37.1949956Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:37.1950707Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:37.1951386Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:37.1952083Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:37.1952954Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:37.1953741Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:37.1954940Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:37.1955757Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:37.1984748Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:37.1985816Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:37.1986609Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:37.1987392Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:37.1988235Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:37.1989097Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:37.1989820Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:37.1990606Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:37.1991386Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:37.1992160Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:37.1993061Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:37.1993925Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:37.1994799Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:37.1995648Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:37.1996675Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:37.1997637Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:37.1998508Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:37.1999324Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:37.2000334Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:37.2001117Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:37.2001963Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:37.2002775Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:37.2003609Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:37.2004382Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:37.2005091Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:37.2005885Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:37.2006646Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:37.2007396Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:37.2008247Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:37.2009050Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:37.2009890Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:37.2010665Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:37.2011434Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:37.2012227Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:37.2013005Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:37.2013826Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:37.2014629Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:37.2015431Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:37.2016201Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:37.2017040Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:37.2017872Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:37.2018673Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:37.2019455Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:37.2020271Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:37.2021096Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:37.2021922Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:37.2022711Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:37.2023591Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:37.2024398Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:37.2025175Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:37.2025950Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:37.2026772Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:37.2027693Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:37.2028487Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:37.2029320Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:37.2030149Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:37.2030910Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:37.2032095Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:37.2032900Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:37.2033769Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:37.2034684Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:37.2035511Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:37.2036326Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:37.2037187Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:37.2038069Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:37.2038886Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:37.2039821Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.2040683Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:37.2041676Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.2042475Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:37.2043413Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.2044223Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:37.2045133Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:37.2045896Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:37.2046667Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:37.2047498Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:37.2048343Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:37.2049121Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:37.2049938Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:37.2050761Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:37.2051598Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:37.2052470Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:37.2053403Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:37.2054266Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:37.2055051Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:37.2055870Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:37.2056789Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:37.2057684Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:37.2058477Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:37.2059364Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:37.2060222Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:37.2061140Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:37.2062008Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:37.2062820Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:37.2063616Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:37.2064352Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:37.2065137Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:37.2065949Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:37.2066770Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:37.2067604Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:37.2068449Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:37.2069277Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:37.2070097Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:37.2070832Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:37.2071640Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:37.2072588Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:37.2073558Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:37.2074501Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:37.2075308Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:37.2076137Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:37.2077008Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:37.2077849Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:37.2078751Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:37.2079602Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:37.2080398Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:37.2081213Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:37.2082054Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:37.2082915Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:37.2083699Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:37.2084489Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:37.2085321Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:37.2086114Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:37.2086893Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:37.2087711Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:37.2088559Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:37.2089432Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:37.2090293Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:37.2091071Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:37.2091963Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:37.2092746Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:37.2093655Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:37.2094474Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:37.2095275Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:37.2096187Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:37.2097073Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:37.2097877Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:37.2098773Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:37.2099631Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:37.2100466Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:37.2101314Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:37.2102178Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:37.2102952Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:37.2103850Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:37.2104736Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:37.2105607Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:37.2106403Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:37.2107249Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:37.2108015Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:37.2108797Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:37.2109613Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:37.2110384Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:37.2111146Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:37.2111930Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:37.2112794Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:37.2113643Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:37.2114505Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:37.2115299Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:37.2116062Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:37.2116888Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:37.2117693Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:37.2118551Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:37.2119424Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:37.2120246Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:37.2121115Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:37.2122004Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:37.2122837Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:37.2123655Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:37.2124463Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:37.2125278Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:37.2126158Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:37.2126954Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:37.2127768Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:37.2128643Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:37.2129460Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:37.2130222Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:37.2131068Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:37.2131872Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:37.2132667Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:37.2133494Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:37.2134465Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:37.2135467Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:37.2136459Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:37.2137546Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:37.2138576Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:37.2139571Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:37.2140579Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:37.2141579Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:37.2142580Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:37.2143433Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:37.2144261Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:37.2145090Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:37.2146003Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:37.2146880Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:37.2147660Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:37.2148450Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:37.2149204Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:37.2150007Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:37.2150808Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:37.2151593Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:37.2152393Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:37.2153307Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:37.2154169Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:37.2154984Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:37.2155743Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:37.2156568Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:37.2157341Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:37.2158164Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:37.2159002Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:37.2159857Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:37.2160645Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:37.2161469Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:37.2162212Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:37.2163024Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:37.2163833Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:37.2164670Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:37.2165626Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:37.2166423Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:37.2167156Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:37.2168005Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:37.2168850Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:37.2169741Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:37.2170545Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:37.2171363Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:37.2172298Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:37.2173178Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:37.2174066Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:37.2174863Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:37.2175785Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:37.2176572Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:37.2177421Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:37.2178241Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:37.2179095Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:37.2179970Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:37.2180795Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:37.2181552Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:37.2182351Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:37.2183164Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:37.2183916Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:37.2184732Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:37.2185545Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:37.2186376Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:37.2187183Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:37.2188044Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.2188926Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.2189725Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:37.2190590Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:37.2191423Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:37.2192274Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:37.2193148Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:37.2194044Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:37.2194934Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:37.2195798Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:37.2196764Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:37.2197590Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:37.2198410Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:37.2199238Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:37.2200057Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:37.2200943Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:37.2201832Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:37.2202660Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:37.2203507Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:37.2204294Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:37.2205114Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:37.2206042Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:37.2206920Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:37.2207784Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:37.2208711Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:37.2209615Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:37.2210536Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:37.2211497Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:37.2212340Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:37.2213208Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:37.2214042Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:37.2214980Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:37.2215859Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:37.2216720Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:37.2217640Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:37.2218501Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:37.2219320Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:37.2220167Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:37.2220965Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:37.2221813Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:37.2222657Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:37.2223480Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:37.2224340Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:37.2225236Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:37.2226056Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:37.2226864Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:37.2227799Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:37.2228706Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:37.2229531Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:37.2230391Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:37.2231271Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:37.2232114Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:37.2232977Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:37.2233890Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:37.2234720Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:37.2235523Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:37.2236370Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:37.2237159Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:37.2237982Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:37.2238801Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:37.2239649Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:37.2240615Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:37.2241497Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.2242353Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:37.2243259Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:37.2244122Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:37.2244993Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:37.2245799Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:37.2246796Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:37.2247638Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:37.2248481Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:37.2249302Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:37.2250158Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:37.2251077Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:37.2251969Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:37.2252926Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:37.2253727Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:37.2254616Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:37.2255450Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:37.2256346Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:37.2257324Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:37.2258206Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:37.2259063Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:37.2259943Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:37.2260871Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:37.2261736Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:37.2262611Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:37.2263420Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:37.2264278Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:37.2265125Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:37.2265951Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:37.2266935Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:37.2267795Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:37.2268612Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:37.2269400Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:37.2270306Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:37.2271243Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:37.2272158Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:37.2273046Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:37.2274045Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:37.2274958Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:37.2275865Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:37.2276765Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:37.2277690Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:37.2278565Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:37.2279422Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:37.2280244Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:37.2281086Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:37.2281994Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:37.2282895Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:37.2283778Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:37.2284668Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:37.2285421Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:37.2286183Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:37.2287040Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:37.2287844Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:37.2288671Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:37.2289487Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:37.2290280Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:37.2291168Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:37.2292145Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:37.2293050Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:37.2293858Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:37.2294634Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:37.2295492Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:37.2296504Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:37.2297440Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:37.2298360Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:37.2299261Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:37.2300136Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:37.2300969Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:37.2301795Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:37.2302682Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:37.2303557Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:37.2304444Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:37.2305291Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:37.2306024Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:37.2306979Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:37.2307798Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:37.2308620Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:37.2309401Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:37.2310182Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:37.2310992Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:37.2311855Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:37.2312641Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:37.2314348Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:37.2316104Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:37.2317988Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:37.2319840Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:37.2321720Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:37.2323605Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:37.2325445Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:37.2327221Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:37.2328979Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:37.2330676Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:37.2332329Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:37.2333990Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:37.2335692Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:37.2337539Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:37.2339364Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:37.2341184Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:37.2342921Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:37.2344601Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:37.2346306Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:37.2349216Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:37.2353833Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:37.2357137Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:37.2358943Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:37.2360816Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:37.2362594Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:37.2364531Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:37.2366473Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:37.2368374Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:37.2370335Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:37.2372301Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:37.2374237Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:37.2376184Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:37.2378272Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:37.2380283Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:37.2382132Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:37.2384040Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:37.2386075Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:37.2388059Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:37.2389959Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:37.2391951Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:37.2393811Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:37.2395425Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:37.2397191Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:37.2398796Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:37.2400363Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:37.2401953Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:37.2403696Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:37.2405559Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:37.2407456Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:37.2409154Z V1204 11:23:06.735000 101180 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:37.2410675Z V1204 11:23:06.736000 101180 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.2412146Z I1204 11:23:06.736000 101180 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.2413440Z V1204 11:23:06.737000 101180 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:37.2414542Z V1204 11:23:06.737000 101180 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:37.2415675Z I1204 11:23:06.737000 101180 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 2 2025-12-04T11:24:37.2417135Z V1204 11:23:06.750000 101180 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:37.2418551Z V1204 11:23:06.754000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.2419685Z V1204 11:23:06.754000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:37.2421269Z V1204 11:23:06.754000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.2423117Z V1204 11:23:06.754000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.2425042Z V1204 11:23:06.754000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.2427226Z V1204 11:23:06.754000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.2429321Z V1204 11:23:06.754000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.2431282Z V1204 11:23:06.754000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:37.2433077Z V1204 11:23:06.754000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:37.2435066Z V1204 11:23:06.754000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:37.2437112Z V1204 11:23:06.754000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:37.2439005Z V1204 11:23:06.754000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.2440967Z V1204 11:23:06.754000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.2442589Z V1204 11:23:06.754000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.2443742Z V1204 11:23:06.754000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.2444709Z V1204 11:23:06.754000 101180 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.2445820Z V1204 11:23:06.757000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:37.2447096Z V1204 11:23:06.757000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:37.2448934Z V1204 11:23:06.758000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.2450627Z V1204 11:23:06.758000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.2452292Z V1204 11:23:06.759000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.2453948Z V1204 11:23:06.760000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.2455538Z V1204 11:23:06.761000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:37.2457166Z V1204 11:23:06.761000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.2458707Z V1204 11:23:06.762000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:37.2460247Z V1204 11:23:06.762000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.2461756Z V1204 11:23:06.764000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:37.2463242Z V1204 11:23:06.764000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.2464655Z V1204 11:23:06.766000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:37.2466094Z V1204 11:23:06.766000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.2467611Z V1204 11:23:06.768000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.2469108Z V1204 11:23:06.768000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.2470585Z V1204 11:23:06.769000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.2472072Z V1204 11:23:06.769000 101180 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.2473183Z V1204 11:23:06.770000 101180 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.2474417Z V1204 11:23:06.771000 101180 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 2 2025-12-04T11:24:37.2475873Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:37.2476840Z warnings.warn( 2025-12-04T11:24:37.2477492Z V1204 11:23:06.771000 101180 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:37.2478992Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-043eb8ddb5d84fd4.xml - 2025-12-04T11:24:37.2480195Z =========================== short test summary info ============================ 2025-12-04T11:24:37.2481410Z FAILED [0.3213s] inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False - AssertionError: Scalars are not equal! 2025-12-04T11:24:37.2482435Z 2025-12-04T11:24:37.2482548Z Expected 1 but got 0. 2025-12-04T11:24:37.2482842Z Absolute difference: 1 2025-12-04T11:24:37.2483133Z Relative difference: 1.0 2025-12-04T11:24:37.2483337Z 2025-12-04T11:24:37.2483556Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:37.2484757Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:37.2485735Z 2025-12-04T11:24:37.2486018Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:37.2486651Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:24:37.2487165Z ================== 1 failed, 36 deselected, 2 rerun in 6.57s =================== 2025-12-04T11:24:37.2487620Z --- Logging error --- 2025-12-04T11:24:37.2487931Z Traceback (most recent call last): 2025-12-04T11:24:37.2488478Z File "/opt/conda/envs/py_3.10/lib/python3.10/logging/__init__.py", line 1103, in emit 2025-12-04T11:24:37.2489066Z stream.write(msg + self.terminator) 2025-12-04T11:24:37.2489467Z ValueError: I/O operation on closed file. 2025-12-04T11:24:37.2489836Z Call stack: 2025-12-04T11:24:37.2490492Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/remote_cache.py", line 432, in dump_cache_stats 2025-12-04T11:24:37.2491304Z log.info("Cache Metrics:%s", out.getvalue()) 2025-12-04T11:24:37.2491711Z Message: 'Cache Metrics:%s' 2025-12-04T11:24:37.2492020Z Arguments: (' None\n',) 2025-12-04T11:24:37.2492309Z Got exit code 1 2025-12-04T11:24:37.2493246Z FAILED CONSISTENTLY: test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False 2025-12-04T11:24:37.2494572Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:24:37.2495819Z Test results will be stored in test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-555ea3dc15f2a8d0.xml 2025-12-04T11:24:37.2497007Z ============================= test session starts ============================== 2025-12-04T11:24:37.2497681Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:24:37.2498301Z cachedir: .pytest_cache 2025-12-04T11:24:37.2499002Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:24:37.2499799Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:24:37.2500157Z configfile: pytest.ini 2025-12-04T11:24:37.2500927Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:24:37.2502859Z collecting ... /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:52: PytestCollectionWarning: cannot collect test class 'TestDecomposeAddMM' because it has a __init__ constructor (from: test/inductor/test_decompose_mem_bound_mm.py) 2025-12-04T11:24:37.2504346Z class TestDecomposeAddMM(torch.nn.Module): 2025-12-04T11:24:37.2504778Z collected 37 items / 30 deselected / 7 selected 2025-12-04T11:24:37.2505196Z stepcurrent: skipping 30 already run items. 2025-12-04T11:24:37.2505652Z Running 7 items in this shard 2025-12-04T11:24:37.2505862Z 2025-12-04T11:24:37.2506690Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True ('RERUN', {'yellow': True}) [5.8886s] [ 14%] 2025-12-04T11:24:37.2508470Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True ('RERUN', {'yellow': True}) [0.3292s] [ 14%] 2025-12-04T11:24:37.2510106Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True FAILED [0.3139s] [ 14%] 2025-12-04T11:24:37.2510966Z 2025-12-04T11:24:37.2511108Z ==================================== RERUNS ==================================== 2025-12-04T11:24:37.2511833Z _ TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True _ 2025-12-04T11:24:37.2512544Z Traceback (most recent call last): 2025-12-04T11:24:37.2513346Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 384, in test_decompose_mm_mixed_precision 2025-12-04T11:24:37.2514160Z self.assertEqual( 2025-12-04T11:24:37.2514856Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:37.2515631Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:37.2516462Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:37.2517349Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:37.2517829Z AssertionError: Scalars are not equal! 2025-12-04T11:24:37.2518080Z 2025-12-04T11:24:37.2518187Z Expected 1 but got 0. 2025-12-04T11:24:37.2518479Z Absolute difference: 1 2025-12-04T11:24:37.2518776Z Relative difference: 1.0 2025-12-04T11:24:37.2518966Z 2025-12-04T11:24:37.2519195Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:37.2520383Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:37.2521373Z 2025-12-04T11:24:37.2521641Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:37.2522280Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:37.2522760Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:37.2523137Z stats [('calls_captured', 1)] 2025-12-04T11:24:37.2524118Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:37.2525237Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:37.2525691Z graph_break [] 2025-12-04T11:24:37.2526072Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:37.2526941Z I1204 11:23:22.427000 101377 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:37.2528190Z I1204 11:23:22.672000 101377 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:37.2529357Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.2530440Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:37.2532021Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.2533864Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.2535246Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:37.2536362Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:37.2537485Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.2539082Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.2541001Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:37.2542388Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:37.2543432Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.2544380Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.2545465Z I1204 11:23:24.889000 101377 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:37.2546865Z V1204 11:23:25.368000 101377 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:37.2548436Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:37.2549953Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:37.2551038Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.2551740Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.2552471Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.2553361Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:37.2554794Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.2556563Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.2558134Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:37.2559675Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.2561334Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.2562674Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:37.2563589Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.2564619Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:37.2567721Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.2572533Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.2575627Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:37.2577379Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:37.2579154Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:37.2580814Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:37.2582562Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:37.2584231Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:37.2585809Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:37.2587400Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:37.2588972Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:37.2590570Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:37.2592124Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:37.2593644Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:37.2595347Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:37.2597356Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:37.2599517Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:37.2601253Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:37.2631263Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:37.2661628Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:37.2663535Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:37.2665210Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:37.2666973Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:37.2668804Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:37.2670511Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:37.2672147Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:37.2673897Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:37.2675637Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:37.2677411Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:37.2679252Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:37.2681085Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:37.2682972Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:37.2684820Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:37.2686712Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:37.2688609Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:37.2690424Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:37.2692376Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:37.2694311Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:37.2696307Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:37.2698117Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:37.2699847Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:37.2701590Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:37.2703202Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:37.2704843Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:37.2706532Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:37.2708220Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:37.2709990Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:37.2711741Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:37.2713426Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:37.2715119Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:37.2716790Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:37.2718542Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:37.2720245Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:37.2721925Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:37.2723628Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:37.2725346Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:37.2727042Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:37.2728764Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:37.2730481Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:37.2732201Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:37.2733860Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:37.2735571Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:37.2737414Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:37.2739195Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:37.2740993Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:37.2742773Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:37.2744578Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:37.2746245Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:37.2747897Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:37.2749601Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:37.2751497Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:37.2753353Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:37.2755079Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:37.2756842Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:37.2758554Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:37.2760643Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:37.2762802Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:37.2764612Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:37.2766509Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:37.2768358Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:37.2770117Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:37.2771920Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:37.2773785Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:37.2775650Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:37.2777645Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.2779542Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:37.2781476Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.2783384Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:37.2785314Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.2787214Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:37.2789034Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:37.2790812Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:37.2794878Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:37.2796808Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:37.2800497Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:37.2802273Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:37.2803994Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:37.2805771Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:37.2807549Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:37.2809418Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:37.2811352Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:37.2813271Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:37.2815120Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:37.2816982Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:37.2818851Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:37.2820696Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:37.2822408Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:37.2824234Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:37.2826144Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:37.2828022Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:37.2829904Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:37.2831789Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:37.2833536Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:37.2835241Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:37.2836904Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:37.2838648Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:37.2840377Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:37.2842150Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:37.2843953Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:37.2845757Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:37.2847522Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:37.2849254Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:37.2850956Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:37.2852811Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:37.2854791Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:37.2856787Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:37.2858726Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:37.2860501Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:37.2862297Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:37.2864105Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:37.2866033Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:37.2867924Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:37.2869739Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:37.2871479Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:37.2873258Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:37.2875066Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:37.2876802Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:37.2878506Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:37.2880267Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:37.2882023Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:37.2883757Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:37.2885505Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:37.2887273Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:37.2889072Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:37.2890915Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:37.2892700Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:37.2894483Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:37.2896463Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:37.2898321Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:37.2900155Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:37.2901974Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:37.2903658Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:37.2905393Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:37.2907107Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:37.2908860Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:37.2910668Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:37.2912484Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:37.2914295Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:37.2916122Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:37.2917923Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:37.2919753Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:37.2921632Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:37.2923436Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:37.2925189Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:37.2926969Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:37.2928725Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:37.2930403Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:37.2932084Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:37.2933825Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:37.2935600Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:37.2937350Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:37.2939155Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:37.2940992Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:37.2942807Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:37.2944571Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:37.2946271Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:37.2947978Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:37.2949740Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:37.2951534Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:37.2953441Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:37.2963119Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:37.2964939Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:37.2966702Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:37.2968458Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:37.2970237Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:37.2971987Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:37.2973698Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:37.2975480Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:37.2977499Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:37.2979256Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:37.2981127Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:37.2982944Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:37.2984642Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:37.2986361Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:37.2988114Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:37.2989825Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:37.2991558Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:37.2993459Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:37.2995593Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:37.2997939Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:37.3000010Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:37.3002083Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:37.3004194Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:37.3006314Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:37.3008425Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:37.3010510Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:37.3012608Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:37.3014410Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:37.3016226Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:37.3018161Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:37.3020046Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:37.3021794Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:37.3023487Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:37.3025161Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:37.3026848Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:37.3028572Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:37.3030343Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:37.3032096Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:37.3033907Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:37.3035807Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:37.3037592Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:37.3039295Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:37.3040998Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:37.3042711Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:37.3044393Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:37.3046183Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:37.3048006Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:37.3049768Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:37.3051528Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:37.3053205Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:37.3054846Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:37.3056563Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:37.3058396Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:37.3060285Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:37.3062154Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:37.3063861Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:37.3065619Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:37.3067400Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:37.3069182Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:37.3070983Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:37.3072732Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:37.3074614Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:37.3076565Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:37.3078400Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:37.3080161Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:37.3082019Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:37.3083862Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:37.3085601Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:37.3087328Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:37.3089110Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:37.3090922Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:37.3092759Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:37.3094457Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:37.3096276Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:37.3098162Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:37.3099901Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:37.3101590Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:37.3103308Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:37.3105042Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:37.3106817Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:37.3108609Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.3110487Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.3112269Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:37.3114033Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:37.3115937Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:37.3117771Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:37.3119684Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:37.3121585Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:37.3123452Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:37.3125304Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:37.3127115Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:37.3128874Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:37.3130665Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:37.3132472Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:37.3134297Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:37.3136088Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:37.3137937Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:37.3139743Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:37.3141545Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:37.3143342Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:37.3145080Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:37.3146915Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:37.3148808Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:37.3150713Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:37.3152640Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:37.3154627Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:37.3156592Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:37.3158561Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:37.3160445Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:37.3162283Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:37.3164111Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:37.3166013Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:37.3168009Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:37.3169896Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:37.3171711Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:37.3173499Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:37.3175276Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:37.3177148Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:37.3178919Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:37.3180666Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:37.3182462Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:37.3184300Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:37.3186123Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:37.3188023Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:37.3189859Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:37.3191612Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:37.3193465Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:37.3195397Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:37.3197405Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:37.3199236Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:37.3201114Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:37.3203039Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:37.3204920Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:37.3205804Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:37.3206589Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:37.3207414Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:37.3208250Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:37.3209055Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:37.3209845Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:37.3210657Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:37.3211548Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:37.3212493Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:37.3213418Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.3214257Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:37.3215148Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:37.3216015Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:37.3216944Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:37.3217746Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:37.3218743Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:37.3219627Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:37.3220505Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:37.3221304Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:37.3222131Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:37.3223047Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:37.3223944Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:37.3224893Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:37.3225695Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:37.3226529Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:37.3227420Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:37.3228308Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:37.3229274Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:37.3230149Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:37.3230995Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:37.3231840Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:37.3232775Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:37.3233634Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:37.3234490Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:37.3235310Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:37.3236210Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:37.3237087Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:37.3237890Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:37.3238837Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:37.3239701Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:37.3240511Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:37.3241316Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:37.3242200Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:37.3243132Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:37.3244082Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:37.3244973Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:37.3245953Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:37.3246816Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:37.3247693Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:37.3248577Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:37.3249519Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:37.3250384Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:37.3251240Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:37.3252105Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:37.3252974Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:37.3253851Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:37.3254700Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:37.3255598Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:37.3256490Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:37.3257324Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:37.3258089Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:37.3258921Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:37.3259729Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:37.3260587Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:37.3261422Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:37.3262246Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:37.3263137Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:37.3264091Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:37.3264991Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:37.3265798Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:37.3266561Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:37.3267437Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:37.3268298Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:37.3269220Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:37.3270081Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:37.3270939Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:37.3271813Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:37.3272641Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:37.3273477Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:37.3274309Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:37.3275193Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:37.3276110Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:37.3276959Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:37.3277714Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:37.3278658Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:37.3279476Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:37.3280243Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:37.3281037Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:37.3281820Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:37.3282623Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:37.3283495Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:37.3284331Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:37.3285116Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:37.3285946Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:37.3286806Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:37.3287654Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:37.3288552Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:37.3289409Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:37.3290217Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:37.3291016Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:37.3291892Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:37.3292660Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:37.3293424Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:37.3294244Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:37.3295037Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:37.3295864Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:37.3296951Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:37.3297787Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:37.3298582Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:37.3299351Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:37.3300239Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:37.3302203Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:37.3304689Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:37.3305456Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:37.3306378Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:37.3307170Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:37.3308017Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:37.3309020Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:37.3309878Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:37.3310824Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:37.3311727Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:37.3312632Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:37.3313535Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:37.3314446Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:37.3315436Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:37.3316335Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:37.3317200Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:37.3318151Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:37.3319083Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:37.3319972Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:37.3320863Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:37.3321832Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:37.3322584Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:37.3323287Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:37.3324040Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:37.3324789Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:37.3325510Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:37.3326274Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:37.3327146Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:37.3328000Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:37.3328851Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:37.3329586Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:37.3330240Z V1204 11:23:25.379000 101377 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.3330911Z I1204 11:23:25.379000 101377 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.3331408Z V1204 11:23:25.379000 101377 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:37.3331921Z V1204 11:23:25.379000 101377 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:37.3332439Z I1204 11:23:25.380000 101377 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:37.3333232Z V1204 11:23:25.420000 101377 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:37.3333684Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.3334198Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:37.3335149Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.3335906Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.3337040Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.3338018Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.3339000Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.3339896Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:37.3340755Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:37.3341797Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:37.3342633Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:37.3343562Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.3344473Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.3345060Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.3345500Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.3345926Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.3346502Z V1204 11:23:25.427000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:37.3347118Z V1204 11:23:25.428000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:37.3348129Z V1204 11:23:25.429000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.3348648Z V1204 11:23:25.429000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.3349664Z V1204 11:23:25.461000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.3350184Z V1204 11:23:25.461000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.3351127Z V1204 11:23:25.463000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:37.3351593Z V1204 11:23:25.463000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.3352532Z V1204 11:23:25.465000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:37.3353016Z V1204 11:23:25.465000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.3353947Z V1204 11:23:25.466000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:37.3354406Z V1204 11:23:25.467000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.3355268Z V1204 11:23:25.469000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:37.3355705Z V1204 11:23:25.470000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.3356601Z V1204 11:23:25.472000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.3357064Z V1204 11:23:25.473000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.3357955Z V1204 11:23:25.473000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.3358417Z V1204 11:23:25.474000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.3358929Z V1204 11:23:25.474000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.3359522Z V1204 11:23:25.475000 101377 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:37.3360296Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:37.3360511Z warnings.warn( 2025-12-04T11:24:37.3361009Z V1204 11:23:25.476000 101377 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:37.3361453Z _ TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True _ 2025-12-04T11:24:37.3361593Z Traceback (most recent call last): 2025-12-04T11:24:37.3362164Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 384, in test_decompose_mm_mixed_precision 2025-12-04T11:24:37.3362271Z self.assertEqual( 2025-12-04T11:24:37.3362745Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:37.3362912Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:37.3363450Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:37.3363672Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:37.3363803Z AssertionError: Scalars are not equal! 2025-12-04T11:24:37.3363813Z 2025-12-04T11:24:37.3363935Z Expected 1 but got 0. 2025-12-04T11:24:37.3364040Z Absolute difference: 1 2025-12-04T11:24:37.3364154Z Relative difference: 1.0 2025-12-04T11:24:37.3364161Z 2025-12-04T11:24:37.3364391Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:37.3365233Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:37.3365240Z 2025-12-04T11:24:37.3365524Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:37.3365782Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:37.3365898Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:37.3366027Z stats [('calls_captured', 1)] 2025-12-04T11:24:37.3366783Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:37.3367036Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:37.3367153Z graph_break [] 2025-12-04T11:24:37.3367373Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:37.3367883Z I1204 11:23:22.427000 101377 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:37.3368477Z I1204 11:23:22.672000 101377 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:37.3368922Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.3369426Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:37.3370365Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.3371106Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.3371594Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:37.3372134Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:37.3372582Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.3373617Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.3374374Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:37.3374848Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:37.3375280Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.3375663Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.3376223Z I1204 11:23:24.889000 101377 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:37.3377013Z V1204 11:23:25.368000 101377 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:37.3377732Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:37.3378406Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:37.3378690Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.3379014Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.3379291Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.3379761Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:37.3380625Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.3381438Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.3382082Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:37.3382849Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.3383601Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.3384070Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:37.3384365Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.3384949Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:37.3387322Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.3390197Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.3390850Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:37.3391782Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:37.3392510Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:37.3393334Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:37.3394116Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:37.3394879Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:37.3395617Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:37.3396555Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:37.3397284Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:37.3398013Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:37.3398719Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:37.3399406Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:37.3400286Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:37.3401092Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:37.3402272Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:37.3403131Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:37.3432132Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:37.3433154Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:37.3433966Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:37.3434743Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:37.3435603Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:37.3436452Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:37.3437192Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:37.3437965Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:37.3438746Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:37.3439571Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:37.3440431Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:37.3441315Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:37.3442173Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:37.3443043Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:37.3443894Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:37.3444804Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:37.3445666Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:37.3446480Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:37.3447505Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:37.3448297Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:37.3449150Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:37.3449925Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:37.3450786Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:37.3451557Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:37.3452268Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:37.3453082Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:37.3453834Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:37.3454599Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:37.3455463Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:37.3456234Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:37.3457095Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:37.3457875Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:37.3458659Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:37.3459444Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:37.3460235Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:37.3460996Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:37.3461806Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:37.3462583Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:37.3463380Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:37.3464154Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:37.3464984Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:37.3465756Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:37.3466520Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:37.3467343Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:37.3468150Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:37.3468974Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:37.3469773Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:37.3470637Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:37.3471456Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:37.3472259Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:37.3473018Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:37.3473826Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:37.3474775Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:37.3475606Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:37.3476405Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:37.3477249Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:37.3477994Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:37.3479219Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:37.3480015Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:37.3480917Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:37.3481802Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:37.3482614Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:37.3483434Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:37.3484278Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:37.3485169Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:37.3485967Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:37.3486911Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.3487768Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:37.3488750Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.3489561Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:37.3490492Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.3491319Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:37.3492190Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:37.3492961Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:37.3493709Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:37.3494559Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:37.3495436Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:37.3496342Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:37.3497282Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:37.3498067Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:37.3498922Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:37.3499788Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:37.3500732Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:37.3501593Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:37.3502358Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:37.3503199Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:37.3504120Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:37.3504968Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:37.3505748Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:37.3506643Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:37.3507503Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:37.3508370Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:37.3509255Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:37.3510049Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:37.3510861Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:37.3511623Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:37.3512421Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:37.3513251Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:37.3514040Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:37.3514884Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:37.3515711Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:37.3516545Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:37.3517348Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:37.3518092Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:37.3518883Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:37.3519827Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:37.3520798Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:37.3521731Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:37.3522552Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:37.3523377Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:37.3524226Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:37.3525063Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:37.3525976Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:37.3526830Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:37.3527648Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:37.3528476Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:37.3529330Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:37.3530174Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:37.3530946Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:37.3531748Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:37.3532579Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:37.3533361Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:37.3534143Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:37.3534945Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:37.3535820Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:37.3536675Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:37.3537607Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:37.3538381Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:37.3539266Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:37.3540069Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:37.3540936Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:37.3541763Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:37.3542548Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:37.3543352Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:37.3544131Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:37.3544939Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:37.3545775Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:37.3546615Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:37.3547464Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:37.3548287Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:37.3549155Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:37.3549911Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:37.3550807Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:37.3551701Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:37.3552546Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:37.3553362Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:37.3554200Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:37.3554987Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:37.3555753Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:37.3556525Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:37.3557303Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:37.3558052Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:37.3558847Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:37.3559741Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:37.3560597Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:37.3561447Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:37.3562251Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:37.3563012Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:37.3563826Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:37.3564640Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:37.3565485Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:37.3566371Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:37.3567215Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:37.3568095Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:37.3568890Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:37.3569708Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:37.3570546Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:37.3571342Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:37.3572141Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:37.3573011Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:37.3573824Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:37.3574645Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:37.3575547Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:37.3576378Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:37.3577218Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:37.3578052Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:37.3578845Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:37.3579654Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:37.3580474Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:37.3581434Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:37.3582453Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:37.3583465Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:37.3584470Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:37.3585443Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:37.3586453Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:37.3587455Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:37.3588445Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:37.3589440Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:37.3590281Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:37.3591125Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:37.3591975Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:37.3592911Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:37.3593791Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:37.3594584Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:37.3595359Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:37.3596399Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:37.3597236Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:37.3598026Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:37.3598843Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:37.3599705Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:37.3600620Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:37.3601480Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:37.3602300Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:37.3603062Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:37.3603874Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:37.3604669Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:37.3605443Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:37.3606300Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:37.3607146Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:37.3607985Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:37.3608826Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:37.3609599Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:37.3610399Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:37.3611200Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:37.3612058Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:37.3613019Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:37.3613830Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:37.3614566Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:37.3615399Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:37.3616268Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:37.3617211Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:37.3618039Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:37.3618845Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:37.3619797Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:37.3620673Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:37.3621521Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:37.3622324Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:37.3623228Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:37.3624068Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:37.3624846Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:37.3625696Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:37.3626559Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:37.3627467Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:37.3628321Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:37.3629064Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:37.3629884Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:37.3630678Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:37.3631441Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:37.3632316Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:37.3633118Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:37.3633966Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:37.3634752Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:37.3635629Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.3636503Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.3637287Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:37.3638144Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:37.3638977Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:37.3639871Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:37.3640722Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:37.3641660Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:37.3642496Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:37.3643371Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:37.3644171Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:37.3645007Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:37.3645836Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:37.3646652Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:37.3647489Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:37.3648347Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:37.3649233Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:37.3650055Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:37.3650916Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:37.3651705Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:37.3652517Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:37.3653418Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:37.3654272Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:37.3655154Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:37.3656117Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:37.3657100Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:37.3658056Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:37.3658957Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:37.3659813Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:37.3660676Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:37.3661523Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:37.3662451Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:37.3663337Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:37.3664216Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:37.3665080Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:37.3665909Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:37.3666713Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:37.3667573Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:37.3668359Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:37.3669192Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:37.3670018Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:37.3670841Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:37.3671695Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:37.3672604Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:37.3673427Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:37.3674257Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:37.3675164Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:37.3676065Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:37.3676913Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:37.3677769Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:37.3678640Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:37.3679493Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:37.3680376Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:37.3681306Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:37.3682089Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:37.3682908Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:37.3683741Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:37.3684538Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:37.3685349Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:37.3686148Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:37.3687002Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:37.3687949Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:37.3688871Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.3689712Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:37.3690626Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:37.3691491Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:37.3692357Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:37.3693172Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:37.3694156Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:37.3694997Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:37.3695825Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:37.3696815Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:37.3697765Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:37.3698666Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:37.3699571Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:37.3700520Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:37.3701329Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:37.3702169Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:37.3703020Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:37.3703901Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:37.3704876Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:37.3705762Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:37.3706636Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:37.3707494Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:37.3708400Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:37.3709277Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:37.3710136Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:37.3710954Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:37.3711804Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:37.3712688Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:37.3713542Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:37.3714474Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:37.3715344Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:37.3716150Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:37.3716958Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:37.3717839Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:37.3718758Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:37.3719679Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:37.3720589Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:37.3721543Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:37.3722430Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:37.3723304Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:37.3724181Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:37.3725121Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:37.3725979Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:37.3726821Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:37.3727654Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:37.3728479Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:37.3729397Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:37.3730278Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:37.3731174Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:37.3732063Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:37.3732811Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:37.3733590Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:37.3734408Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:37.3735222Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:37.3736039Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:37.3736961Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:37.3737769Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:37.3738680Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:37.3739646Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:37.3740532Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:37.3741359Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:37.3742130Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:37.3743005Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:37.3743834Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:37.3744711Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:37.3745607Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:37.3746481Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:37.3747368Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:37.3748189Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:37.3749028Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:37.3749865Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:37.3750752Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:37.3751637Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:37.3752466Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:37.3753245Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:37.3754184Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:37.3755041Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:37.3755806Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:37.3756592Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:37.3757388Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:37.3758185Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:37.3759063Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:37.3759837Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:37.3760604Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:37.3761469Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:37.3762352Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:37.3763201Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:37.3764071Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:37.3764934Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:37.3765722Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:37.3766528Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:37.3767340Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:37.3768097Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:37.3768894Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:37.3769656Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:37.3770483Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:37.3771299Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:37.3772161Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:37.3772991Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:37.3773773Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:37.3774535Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:37.3775325Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:37.3777367Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:37.3779869Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:37.3780635Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:37.3781569Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:37.3782364Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:37.3783192Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:37.3784162Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:37.3785036Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:37.3785957Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:37.3786893Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:37.3787791Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:37.3788692Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:37.3789599Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:37.3790596Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:37.3791477Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:37.3792318Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:37.3793286Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:37.3794246Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:37.3795127Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:37.3796171Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:37.3797141Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:37.3797879Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:37.3798595Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:37.3799337Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:37.3800062Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:37.3800776Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:37.3801567Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:37.3802445Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:37.3803323Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:37.3804188Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:37.3804906Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:37.3805574Z V1204 11:23:25.379000 101377 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.3806250Z I1204 11:23:25.379000 101377 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.3806753Z V1204 11:23:25.379000 101377 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:37.3807238Z V1204 11:23:25.379000 101377 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:37.3807743Z I1204 11:23:25.380000 101377 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:37.3808527Z V1204 11:23:25.420000 101377 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:37.3809012Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.3809561Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:37.3810509Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.3811251Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.3812294Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.3813263Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.3814256Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.3815093Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:37.3815965Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:37.3817084Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:37.3817969Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:37.3818903Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.3819814Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.3820405Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.3820833Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.3821238Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.3821807Z V1204 11:23:25.427000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:37.3822422Z V1204 11:23:25.428000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:37.3823435Z V1204 11:23:25.429000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.3823986Z V1204 11:23:25.429000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.3825000Z V1204 11:23:25.461000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.3825502Z V1204 11:23:25.461000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.3826462Z V1204 11:23:25.463000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:37.3826917Z V1204 11:23:25.463000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.3827867Z V1204 11:23:25.465000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:37.3828319Z V1204 11:23:25.465000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.3829246Z V1204 11:23:25.466000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:37.3829707Z V1204 11:23:25.467000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.3830558Z V1204 11:23:25.469000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:37.3831047Z V1204 11:23:25.470000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.3831944Z V1204 11:23:25.472000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.3832403Z V1204 11:23:25.473000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.3833298Z V1204 11:23:25.473000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.3833760Z V1204 11:23:25.474000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.3834272Z V1204 11:23:25.474000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.3834851Z V1204 11:23:25.475000 101377 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:37.3835606Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:37.3835711Z warnings.warn( 2025-12-04T11:24:37.3836238Z V1204 11:23:25.476000 101377 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:37.3836465Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:37.3836608Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:37.3836737Z stats [('calls_captured', 1)] 2025-12-04T11:24:37.3836961Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:37.3837718Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:37.3837831Z graph_break [] 2025-12-04T11:24:37.3838051Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:37.3838501Z V1204 11:23:25.702000 101377 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmp24jtc_b_ 2025-12-04T11:24:37.3838947Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.3839442Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:37.3840392Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.3841123Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.3841620Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:37.3842104Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:37.3842555Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.3843600Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.3844371Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:37.3844860Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:37.3845279Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.3845681Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.3846397Z V1204 11:23:25.971000 101377 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:37.3847112Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:37.3847786Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:37.3848069Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.3848359Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.3848661Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.3849130Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:37.3849987Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.3850803Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.3851439Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:37.3852207Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.3852965Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.3853416Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:37.3853711Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.3854296Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:37.3856663Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.3859095Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.3859741Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:37.3860677Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:37.3861400Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:37.3862222Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:37.3863014Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:37.3863741Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:37.3864494Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:37.3865246Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:37.3865975Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:37.3866705Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:37.3867407Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:37.3868097Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:37.3868973Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:37.3869778Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:37.3870937Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:37.3871736Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:37.3901284Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:37.3902361Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:37.3903167Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:37.3903944Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:37.3904804Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:37.3905651Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:37.3906393Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:37.3907167Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:37.3907963Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:37.3908748Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:37.3909605Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:37.3910506Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:37.3911378Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:37.3912278Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:37.3913125Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:37.3914040Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:37.3914904Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:37.3915728Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:37.3916734Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:37.3917495Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:37.3918388Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:37.3919168Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:37.3920017Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:37.3920782Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:37.3921508Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:37.3922308Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:37.3923066Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:37.3923838Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:37.3924673Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:37.3925446Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:37.3926260Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:37.3927053Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:37.3927842Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:37.3928626Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:37.3929411Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:37.3930178Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:37.3930979Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:37.3931753Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:37.3932528Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:37.3933293Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:37.3934132Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:37.3934954Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:37.3935720Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:37.3936537Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:37.3937411Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:37.3938255Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:37.3939041Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:37.3939911Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:37.3940691Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:37.3941442Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:37.3942239Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:37.3943047Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:37.3944008Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:37.3944800Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:37.3945594Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:37.3946434Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:37.3947186Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:37.3948393Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:37.3949193Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:37.3950106Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:37.3951019Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:37.3951837Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:37.3952656Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:37.3953498Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:37.3954390Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:37.3955195Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:37.3956146Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.3956968Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:37.3957964Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.3958773Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:37.3959732Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.3960563Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:37.3961428Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:37.3962203Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:37.3962960Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:37.3963808Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:37.3964650Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:37.3965405Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:37.3966266Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:37.3967096Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:37.3967948Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:37.3968806Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:37.3969747Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:37.3970601Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:37.3971370Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:37.3972207Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:37.3973083Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:37.3973927Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:37.3974705Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:37.3975632Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:37.3976487Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:37.3977407Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:37.3978293Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:37.3979088Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:37.3979901Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:37.3980617Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:37.3981420Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:37.3982257Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:37.3983074Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:37.3983919Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:37.3984741Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:37.3985577Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:37.3986379Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:37.3987130Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:37.3987925Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:37.3988848Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:37.3989796Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:37.3990734Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:37.3991577Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:37.3992399Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:37.3993244Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:37.3994075Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:37.3994993Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:37.3995849Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:37.3996788Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:37.3997676Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:37.3998500Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:37.3999388Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:37.4000158Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:37.4000963Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:37.4001799Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:37.4002579Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:37.4003368Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:37.4004173Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:37.4004998Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:37.4005867Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:37.4006740Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:37.4007551Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:37.4008423Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:37.4009220Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:37.4010092Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:37.4010917Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:37.4011702Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:37.4012469Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:37.4013246Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:37.4014060Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:37.4014928Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:37.4015776Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:37.4016621Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:37.4017498Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:37.4018368Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:37.4019125Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:37.4020033Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:37.4020883Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:37.4021733Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:37.4022545Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:37.4023413Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:37.4024198Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:37.4024962Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:37.4025751Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:37.4026512Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:37.4027258Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:37.4028049Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:37.4028894Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:37.4029798Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:37.4030645Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:37.4031450Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:37.4032212Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:37.4033021Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:37.4033833Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:37.4034674Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:37.4035560Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:37.4036368Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:37.4037216Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:37.4038044Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:37.4038868Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:37.4039728Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:37.4040518Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:37.4041316Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:37.4042188Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:37.4043000Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:37.4043808Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:37.4044671Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:37.4045536Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:37.4046325Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:37.4047162Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:37.4047955Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:37.4048763Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:37.4049580Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:37.4050540Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:37.4051561Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:37.4052539Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:37.4053549Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:37.4054532Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:37.4055573Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:37.4056577Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:37.4057624Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:37.4058628Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:37.4059475Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:37.4060317Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:37.4061127Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:37.4062104Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:37.4062976Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:37.4063773Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:37.4064554Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:37.4065308Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:37.4066126Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:37.4066911Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:37.4067722Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:37.4068508Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:37.4069414Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:37.4070264Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:37.4071102Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:37.4071883Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:37.4072690Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:37.4073481Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:37.4074258Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:37.4075111Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:37.4075951Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:37.4076757Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:37.4077584Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:37.4078341Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:37.4079169Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:37.4079961Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:37.4080800Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:37.4081745Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:37.4082549Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:37.4083282Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:37.4084118Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:37.4084942Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:37.4085854Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:37.4086676Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:37.4087530Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:37.4088478Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:37.4089346Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:37.4090189Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:37.4090987Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:37.4091892Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:37.4092701Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:37.4093480Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:37.4094338Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:37.4095190Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:37.4096209Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:37.4097096Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:37.4097843Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:37.4098663Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:37.4099461Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:37.4100229Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:37.4101011Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:37.4101844Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:37.4102694Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:37.4103516Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:37.4104394Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.4105260Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.4106041Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:37.4106893Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:37.4107742Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:37.4108594Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:37.4109447Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:37.4110393Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:37.4111275Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:37.4112153Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:37.4112949Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:37.4113784Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:37.4114606Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:37.4115424Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:37.4116260Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:37.4117076Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:37.4117964Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:37.4118784Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:37.4119679Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:37.4120468Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:37.4121274Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:37.4122178Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:37.4123034Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:37.4123919Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:37.4124825Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:37.4125741Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:37.4126687Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:37.4127621Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:37.4128469Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:37.4129322Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:37.4130165Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:37.4131089Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:37.4131982Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:37.4132825Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:37.4133650Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:37.4134502Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:37.4135314Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:37.4136204Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:37.4137037Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:37.4137876Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:37.4138703Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:37.4139535Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:37.4140398Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:37.4141269Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:37.4142141Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:37.4142973Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:37.4143881Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:37.4144776Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:37.4145618Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:37.4146479Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:37.4147352Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:37.4148203Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:37.4149048Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:37.4150052Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:37.4150835Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:37.4151649Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:37.4152521Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:37.4153322Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:37.4154121Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:37.4154917Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:37.4155780Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:37.4156728Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:37.4157617Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.4158484Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:37.4159395Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:37.4160261Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:37.4161114Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:37.4161922Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:37.4162904Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:37.4163753Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:37.4164582Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:37.4165380Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:37.4166239Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:37.4167143Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:37.4168073Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:37.4169016Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:37.4169826Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:37.4170667Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:37.4171517Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:37.4172400Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:37.4173324Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:37.4174233Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:37.4175070Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:37.4175951Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:37.4176904Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:37.4177777Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:37.4178640Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:37.4179455Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:37.4180310Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:37.4181144Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:37.4181949Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:37.4182919Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:37.4183794Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:37.4184631Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:37.4185435Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:37.4186326Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:37.4187243Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:37.4188169Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:37.4189045Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:37.4189991Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:37.4190880Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:37.4191804Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:37.4192683Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:37.4193620Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:37.4194477Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:37.4195316Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:37.4196327Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:37.4197150Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:37.4198043Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:37.4198959Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:37.4199865Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:37.4200787Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:37.4201530Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:37.4202304Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:37.4203119Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:37.4203940Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:37.4204749Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:37.4205579Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:37.4206368Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:37.4207287Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:37.4208302Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:37.4209192Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:37.4210021Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:37.4210793Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:37.4211675Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:37.4212507Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:37.4213383Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:37.4214238Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:37.4215113Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:37.4216006Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:37.4216914Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:37.4217751Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:37.4218583Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:37.4219472Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:37.4220359Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:37.4221195Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:37.4221932Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:37.4222903Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:37.4223769Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:37.4224536Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:37.4225321Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:37.4226102Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:37.4226896Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:37.4227778Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:37.4228552Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:37.4229315Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:37.4230141Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:37.4231024Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:37.4231874Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:37.4232773Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:37.4233639Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:37.4234429Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:37.4235237Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:37.4236052Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:37.4236810Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:37.4237574Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:37.4238327Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:37.4239163Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:37.4240008Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:37.4240868Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:37.4241689Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:37.4242470Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:37.4243228Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:37.4244037Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:37.4246003Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:37.4248493Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:37.4249258Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:37.4250178Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:37.4250975Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:37.4251807Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:37.4252777Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:37.4253615Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:37.4254536Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:37.4255474Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:37.4256405Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:37.4257370Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:37.4258273Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:37.4259278Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:37.4260157Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:37.4261005Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:37.4261929Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:37.4262859Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:37.4263774Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:37.4264662Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:37.4265660Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:37.4266399Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:37.4267114Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:37.4267859Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:37.4268582Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:37.4269300Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:37.4270033Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:37.4270955Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:37.4271798Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:37.4272687Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:37.4273400Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:37.4274062Z V1204 11:23:25.981000 101377 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.4274737Z I1204 11:23:25.981000 101377 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.4275235Z V1204 11:23:25.982000 101377 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:37.4275716Z V1204 11:23:25.982000 101377 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:37.4276223Z I1204 11:23:25.982000 101377 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:37.4277002Z V1204 11:23:25.995000 101377 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:37.4277447Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.4277984Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:37.4278941Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.4279720Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.4280763Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.4281726Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.4282721Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.4283562Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:37.4284394Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:37.4285472Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:37.4286340Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:37.4287272Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.4288176Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.4288763Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.4289195Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.4289593Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.4290164Z V1204 11:23:26.002000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:37.4290741Z V1204 11:23:26.002000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:37.4291749Z V1204 11:23:26.003000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.4292294Z V1204 11:23:26.003000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.4293303Z V1204 11:23:26.004000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.4293836Z V1204 11:23:26.005000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.4294784Z V1204 11:23:26.006000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:37.4295239Z V1204 11:23:26.006000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.4296335Z V1204 11:23:26.007000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:37.4296791Z V1204 11:23:26.007000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.4297772Z V1204 11:23:26.009000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:37.4298204Z V1204 11:23:26.009000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.4299055Z V1204 11:23:26.011000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:37.4299566Z V1204 11:23:26.011000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.4300509Z V1204 11:23:26.013000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.4300977Z V1204 11:23:26.014000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.4301873Z V1204 11:23:26.014000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.4302333Z V1204 11:23:26.015000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.4302844Z V1204 11:23:26.015000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.4303423Z V1204 11:23:26.016000 101377 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:37.4304175Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:37.4304279Z warnings.warn( 2025-12-04T11:24:37.4304770Z V1204 11:23:26.016000 101377 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:37.4304919Z =================================== FAILURES =================================== 2025-12-04T11:24:37.4305363Z _ TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True _ 2025-12-04T11:24:37.4305540Z Traceback (most recent call last): 2025-12-04T11:24:37.4306109Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 384, in test_decompose_mm_mixed_precision 2025-12-04T11:24:37.4306218Z self.assertEqual( 2025-12-04T11:24:37.4306688Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:37.4306890Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:37.4307441Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:37.4307649Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:37.4307782Z AssertionError: Scalars are not equal! 2025-12-04T11:24:37.4307790Z 2025-12-04T11:24:37.4307909Z Expected 1 but got 0. 2025-12-04T11:24:37.4308018Z Absolute difference: 1 2025-12-04T11:24:37.4308129Z Relative difference: 1.0 2025-12-04T11:24:37.4308148Z 2025-12-04T11:24:37.4308367Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:37.4309213Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:37.4309219Z 2025-12-04T11:24:37.4309500Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:37.4309727Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:37.4309839Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:37.4309970Z stats [('calls_captured', 1)] 2025-12-04T11:24:37.4310732Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:37.4310998Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:37.4311099Z graph_break [] 2025-12-04T11:24:37.4311321Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:37.4311864Z I1204 11:23:22.427000 101377 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:37.4312460Z I1204 11:23:22.672000 101377 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:37.4312913Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.4313403Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:37.4314340Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.4315086Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.4322610Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:37.4323224Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:37.4323645Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.4324675Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.4325540Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:37.4326015Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:37.4326501Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.4326878Z V1204 11:23:22.674000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.4327439Z I1204 11:23:24.889000 101377 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:37.4328148Z V1204 11:23:25.368000 101377 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:37.4328859Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:37.4329527Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:37.4329807Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.4330089Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.4330359Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.4330824Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:37.4331688Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.4332533Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.4333163Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:37.4333921Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.4334672Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.4335121Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:37.4335409Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.4335990Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:37.4338416Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.4340789Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.4341464Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:37.4342379Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:37.4343106Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:37.4343921Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:37.4344703Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:37.4345415Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:37.4346137Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:37.4346886Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:37.4347606Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:37.4348366Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:37.4349050Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:37.4349729Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:37.4350596Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:37.4351395Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:37.4352558Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:37.4353344Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:37.4382348Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:37.4383369Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:37.4384168Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:37.4384963Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:37.4385803Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:37.4386644Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:37.4387357Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:37.4388131Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:37.4388911Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:37.4389684Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:37.4390531Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:37.4391368Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:37.4392251Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:37.4393105Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:37.4393978Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:37.4394878Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:37.4395736Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:37.4396681Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:37.4397695Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:37.4398454Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:37.4399291Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:37.4400100Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:37.4400979Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:37.4401736Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:37.4402437Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:37.4403228Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:37.4403979Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:37.4404745Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:37.4405576Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:37.4406336Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:37.4407112Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:37.4407931Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:37.4408707Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:37.4409517Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:37.4410299Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:37.4411053Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:37.4411857Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:37.4412633Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:37.4413398Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:37.4414165Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:37.4414966Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:37.4415766Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:37.4416616Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:37.4417494Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:37.4418296Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:37.4419117Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:37.4419914Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:37.4420780Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:37.4421559Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:37.4422319Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:37.4423077Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:37.4423918Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:37.4424837Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:37.4425662Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:37.4426449Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:37.4427295Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:37.4428041Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:37.4429236Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:37.4430029Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:37.4430896Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:37.4431810Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:37.4432648Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:37.4433470Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:37.4434309Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:37.4435198Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:37.4435998Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:37.4436937Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.4437761Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:37.4438712Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.4439549Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:37.4440480Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.4441332Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:37.4442196Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:37.4442953Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:37.4443706Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:37.4444532Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:37.4445385Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:37.4446145Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:37.4446965Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:37.4447780Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:37.4448651Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:37.4449504Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:37.4450432Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:37.4451288Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:37.4452052Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:37.4452889Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:37.4453759Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:37.4454561Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:37.4455357Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:37.4456252Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:37.4457201Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:37.4458069Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:37.4458944Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:37.4459737Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:37.4460545Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:37.4461259Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:37.4462053Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:37.4462847Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:37.4463666Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:37.4464541Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:37.4465369Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:37.4466193Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:37.4467001Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:37.4467741Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:37.4468533Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:37.4469442Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:37.4470366Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:37.4471321Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:37.4472136Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:37.4472974Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:37.4473818Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:37.4474637Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:37.4475551Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:37.4476403Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:37.4477178Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:37.4477994Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:37.4478813Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:37.4479679Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:37.4480481Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:37.4481274Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:37.4482106Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:37.4482891Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:37.4483666Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:37.4484467Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:37.4485301Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:37.4486134Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:37.4487045Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:37.4487824Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:37.4488732Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:37.4489531Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:37.4490396Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:37.4491223Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:37.4492010Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:37.4492787Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:37.4493568Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:37.4494349Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:37.4495219Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:37.4496243Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:37.4497144Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:37.4497974Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:37.4498844Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:37.4499609Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:37.4500513Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:37.4501374Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:37.4502187Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:37.4503060Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:37.4503889Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:37.4504717Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:37.4505488Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:37.4506260Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:37.4507042Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:37.4507786Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:37.4508585Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:37.4509436Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:37.4510292Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:37.4511153Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:37.4511984Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:37.4512767Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:37.4513581Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:37.4514399Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:37.4515252Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:37.4516145Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:37.4516960Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:37.4517791Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:37.4518596Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:37.4519451Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:37.4520293Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:37.4521117Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:37.4521914Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:37.4522777Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:37.4523597Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:37.4524417Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:37.4525285Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:37.4526110Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:37.4526911Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:37.4527736Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:37.4528555Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:37.4529359Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:37.4530181Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:37.4531138Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:37.4532150Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:37.4533124Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:37.4534102Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:37.4535102Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:37.4536115Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:37.4537238Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:37.4538239Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:37.4539229Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:37.4540071Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:37.4540906Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:37.4541724Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:37.4542648Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:37.4543521Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:37.4544345Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:37.4545129Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:37.4545883Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:37.4546704Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:37.4547493Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:37.4548304Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:37.4549094Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:37.4549995Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:37.4550812Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:37.4551653Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:37.4552419Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:37.4553255Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:37.4554045Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:37.4554815Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:37.4555665Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:37.4556508Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:37.4557302Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:37.4558134Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:37.4558863Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:37.4559690Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:37.4560514Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:37.4561357Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:37.4562302Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:37.4563098Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:37.4563845Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:37.4564682Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:37.4565508Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:37.4566361Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:37.4567210Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:37.4568020Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:37.4569005Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:37.4569880Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:37.4570718Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:37.4571525Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:37.4572434Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:37.4573240Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:37.4574013Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:37.4574840Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:37.4575696Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:37.4576587Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:37.4577480Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:37.4578222Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:37.4579044Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:37.4579845Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:37.4580616Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:37.4581406Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:37.4582180Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:37.4583061Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:37.4583846Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:37.4584759Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.4585627Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.4586405Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:37.4587259Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:37.4588101Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:37.4588962Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:37.4589820Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:37.4590722Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:37.4591592Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:37.4592498Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:37.4593301Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:37.4594122Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:37.4594956Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:37.4595772Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:37.4596757Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:37.4597576Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:37.4598433Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:37.4599309Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:37.4600176Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:37.4601012Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:37.4601823Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:37.4602728Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:37.4603592Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:37.4604477Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:37.4605391Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:37.4606308Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:37.4607227Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:37.4608175Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:37.4609066Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:37.4609926Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:37.4610768Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:37.4611698Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:37.4612585Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:37.4613428Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:37.4614241Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:37.4615074Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:37.4615926Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:37.4616789Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:37.4617664Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:37.4618503Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:37.4619326Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:37.4620161Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:37.4621022Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:37.4621902Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:37.4622728Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:37.4623564Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:37.4624479Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:37.4625408Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:37.4626250Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:37.4627104Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:37.4627978Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:37.4628828Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:37.4629683Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:37.4630575Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:37.4631358Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:37.4632205Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:37.4633047Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:37.4633866Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:37.4634678Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:37.4635478Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:37.4636342Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:37.4637294Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:37.4638183Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.4639025Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:37.4639930Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:37.4640835Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:37.4641690Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:37.4642503Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:37.4643478Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:37.4644328Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:37.4645156Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:37.4645970Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:37.4646801Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:37.4647746Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:37.4648655Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:37.4649621Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:37.4650435Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:37.4651267Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:37.4652119Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:37.4652999Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:37.4653909Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:37.4654805Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:37.4655643Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:37.4656538Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:37.4657541Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:37.4658417Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:37.4659278Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:37.4660103Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:37.4660954Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:37.4661792Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:37.4662604Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:37.4663540Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:37.4664449Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:37.4665255Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:37.4666090Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:37.4666968Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:37.4667888Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:37.4668820Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:37.4669701Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:37.4670654Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:37.4671512Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:37.4672495Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:37.4673409Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:37.4674348Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:37.4675206Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:37.4676057Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:37.4676896Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:37.4677731Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:37.4678621Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:37.4679477Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:37.4680405Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:37.4681292Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:37.4682471Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:37.4683255Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:37.4684070Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:37.4684890Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:37.4685704Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:37.4686535Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:37.4687322Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:37.4688193Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:37.4689184Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:37.4690096Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:37.4690911Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:37.4691676Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:37.4692552Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:37.4693382Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:37.4694245Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:37.4695104Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:37.4695951Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:37.4697107Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:37.4697939Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:37.4698815Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:37.4699645Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:37.4700530Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:37.4701417Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:37.4702243Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:37.4702983Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:37.4703914Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:37.4704787Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:37.4705559Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:37.4706408Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:37.4707193Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:37.4707979Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:37.4708859Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:37.4709633Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:37.4710398Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:37.4711229Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:37.4712085Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:37.4712975Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:37.4713862Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:37.4714760Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:37.4715554Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:37.4716358Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:37.4717175Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:37.4717928Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:37.4718687Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:37.4719446Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:37.4720242Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:37.4721093Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:37.4721983Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:37.4722806Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:37.4723590Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:37.4724349Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:37.4725145Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:37.4727125Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:37.4729612Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:37.4730382Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:37.4731323Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:37.4732134Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:37.4732965Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:37.4733941Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:37.4734786Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:37.4735702Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:37.4736601Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:37.4737591Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:37.4738528Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:37.4739432Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:37.4740434Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:37.4741324Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:37.4742178Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:37.4743106Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:37.4744035Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:37.4744919Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:37.4745841Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:37.4746809Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:37.4747578Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:37.4748291Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:37.4749032Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:37.4749758Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:37.4750475Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:37.4751209Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:37.4752080Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:37.4752919Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:37.4753816Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:37.4754565Z V1204 11:23:25.378000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:37.4755239Z V1204 11:23:25.379000 101377 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.4755911Z I1204 11:23:25.379000 101377 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.4756395Z V1204 11:23:25.379000 101377 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:37.4756891Z V1204 11:23:25.379000 101377 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:37.4757400Z I1204 11:23:25.380000 101377 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:37.4758177Z V1204 11:23:25.420000 101377 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:37.4758619Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.4759123Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:37.4760107Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.4760851Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.4761924Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.4762881Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.4763883Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.4764724Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:37.4765551Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:37.4766592Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:37.4767473Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:37.4768427Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.4769333Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.4769916Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.4770342Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.4770748Z V1204 11:23:25.424000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.4771324Z V1204 11:23:25.427000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:37.4771904Z V1204 11:23:25.428000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:37.4772911Z V1204 11:23:25.429000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.4773431Z V1204 11:23:25.429000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.4774475Z V1204 11:23:25.461000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.4774982Z V1204 11:23:25.461000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.4775961Z V1204 11:23:25.463000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:37.4776417Z V1204 11:23:25.463000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.4777437Z V1204 11:23:25.465000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:37.4777898Z V1204 11:23:25.465000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.4778822Z V1204 11:23:25.466000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:37.4779254Z V1204 11:23:25.467000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.4780098Z V1204 11:23:25.469000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:37.4780546Z V1204 11:23:25.470000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.4781474Z V1204 11:23:25.472000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.4781971Z V1204 11:23:25.473000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.4782864Z V1204 11:23:25.473000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.4783325Z V1204 11:23:25.474000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.4783835Z V1204 11:23:25.474000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.4784418Z V1204 11:23:25.475000 101377 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:37.4785173Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:37.4785276Z warnings.warn( 2025-12-04T11:24:37.4785756Z V1204 11:23:25.476000 101377 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:37.4785999Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:37.4786114Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:37.4786242Z stats [('calls_captured', 1)] 2025-12-04T11:24:37.4786464Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:37.4787226Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:37.4787380Z graph_break [] 2025-12-04T11:24:37.4787606Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:37.4788056Z V1204 11:23:25.702000 101377 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmp24jtc_b_ 2025-12-04T11:24:37.4788527Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.4789023Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:37.4789979Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.4790716Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.4791222Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:37.4791712Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:37.4792136Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.4793175Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.4793955Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:37.4794446Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:37.4794901Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.4795303Z V1204 11:23:25.725000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.4796195Z V1204 11:23:25.971000 101377 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:37.4796913Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:37.4797598Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:37.4797880Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.4798173Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.4798449Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.4798919Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:37.4799752Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.4800568Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.4801274Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:37.4802042Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.4802846Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.4803301Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:37.4803597Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.4804192Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:37.4806528Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.4808867Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.4809591Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:37.4810525Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:37.4811249Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:37.4812069Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:37.4812861Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:37.4813586Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:37.4814301Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:37.4815021Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:37.4815747Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:37.4816520Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:37.4817289Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:37.4818024Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:37.4818898Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:37.4819703Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:37.4820877Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:37.4821675Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:37.4850594Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:37.4851684Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:37.4852451Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:37.4853221Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:37.4854098Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:37.4854945Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:37.4855701Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:37.4856472Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:37.4857332Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:37.4858108Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:37.4858968Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:37.4859809Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:37.4860668Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:37.4861561Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:37.4862435Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:37.4863339Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:37.4864194Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:37.4865009Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:37.4866033Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:37.4866796Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:37.4867647Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:37.4868390Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:37.4869233Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:37.4870027Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:37.4870737Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:37.4871574Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:37.4872328Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:37.4873096Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:37.4873935Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:37.4874708Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:37.4875483Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:37.4876253Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:37.4877072Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:37.4877857Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:37.4878672Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:37.4879438Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:37.4880243Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:37.4881025Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:37.4881790Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:37.4882572Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:37.4883382Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:37.4884158Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:37.4884964Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:37.4885792Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:37.4886631Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:37.4887462Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:37.4888265Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:37.4889131Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:37.4889916Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:37.4890677Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:37.4891438Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:37.4892243Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:37.4893210Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:37.4894032Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:37.4894820Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:37.4895665Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:37.4896582Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:37.4897843Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:37.4898645Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:37.4899531Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:37.4900416Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:37.4901319Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:37.4902154Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:37.4903048Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:37.4903945Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:37.4904742Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:37.4905691Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.4906520Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:37.4907473Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.4908287Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:37.4909219Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.4910088Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:37.4910996Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:37.4911765Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:37.4912512Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:37.4913367Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:37.4914218Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:37.4914979Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:37.4915809Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:37.4916596Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:37.4917473Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:37.4918336Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:37.4919302Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:37.4920155Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:37.4920925Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:37.4921763Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:37.4922641Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:37.4923460Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:37.4924233Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:37.4925135Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:37.4926029Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:37.4926922Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:37.4927794Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:37.4928588Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:37.4929399Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:37.4930116Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:37.4930917Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:37.4931717Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:37.4932506Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:37.4933379Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:37.4934207Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:37.4935141Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:37.4935944Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:37.4936700Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:37.4937559Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:37.4938476Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:37.4939415Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:37.4940350Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:37.4941212Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:37.4942032Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:37.4942916Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:37.4943747Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:37.4944666Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:37.4945520Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:37.4946300Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:37.4947130Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:37.4947956Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:37.4948805Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:37.4949611Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:37.4950418Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:37.4951274Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:37.4952054Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:37.4952841Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:37.4953649Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:37.4954483Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:37.4955317Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:37.4956188Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:37.4956962Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:37.4957873Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:37.4958724Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:37.4959594Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:37.4960419Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:37.4961203Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:37.4961968Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:37.4962765Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:37.4963543Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:37.4964384Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:37.4965260Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:37.4966108Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:37.4966966Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:37.4967840Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:37.4968599Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:37.4969510Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:37.4970373Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:37.4971183Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:37.4971996Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:37.4972829Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:37.4973642Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:37.4974435Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:37.4975223Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:37.4975986Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:37.4976733Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:37.4977597Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:37.4978456Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:37.4979318Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:37.4980136Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:37.4980945Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:37.4981748Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:37.4982566Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:37.4983409Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:37.4984255Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:37.4985148Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:37.4985963Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:37.4986815Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:37.4987607Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:37.4988425Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:37.4989299Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:37.4990154Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:37.4990952Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:37.4991817Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:37.4992629Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:37.4993446Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:37.4994309Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:37.4995139Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:37.4995894Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:37.4996881Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:37.4997740Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:37.4998544Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:37.4999401Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:37.5000359Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:37.5001381Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:37.5002360Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:37.5003343Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:37.5004321Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:37.5005328Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:37.5006366Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:37.5007398Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:37.5008394Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:37.5009226Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:37.5010066Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:37.5010879Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:37.5011808Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:37.5012643Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:37.5013434Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:37.5014238Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:37.5014996Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:37.5015832Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:37.5016623Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:37.5017484Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:37.5018279Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:37.5019178Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:37.5019996Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:37.5020812Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:37.5021627Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:37.5022443Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:37.5023261Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:37.5024029Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:37.5024876Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:37.5025719Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:37.5026522Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:37.5027344Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:37.5028075Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:37.5028871Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:37.5029687Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:37.5030524Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:37.5031499Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:37.5032307Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:37.5033038Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:37.5033878Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:37.5034706Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:37.5035562Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:37.5036383Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:37.5037188Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:37.5038179Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:37.5039079Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:37.5039928Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:37.5040731Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:37.5041643Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:37.5042452Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:37.5043232Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:37.5044057Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:37.5044877Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:37.5045776Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:37.5046612Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:37.5047385Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:37.5048208Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:37.5048999Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:37.5049768Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:37.5050553Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:37.5051326Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:37.5052166Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:37.5052952Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:37.5053864Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.5054769Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.5055555Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:37.5056410Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:37.5057302Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:37.5058170Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:37.5059025Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:37.5059934Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:37.5060768Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:37.5061680Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:37.5062477Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:37.5063340Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:37.5064165Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:37.5064982Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:37.5065827Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:37.5066637Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:37.5067492Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:37.5068308Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:37.5069166Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:37.5069990Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:37.5070826Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:37.5071725Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:37.5072580Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:37.5073465Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:37.5074373Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:37.5075286Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:37.5076204Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:37.5077110Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:37.5077991Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:37.5078846Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:37.5079732Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:37.5080656Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:37.5081545Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:37.5082390Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:37.5083222Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:37.5084048Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:37.5084852Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:37.5085750Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:37.5086560Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:37.5087393Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:37.5088221Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:37.5089048Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:37.5089918Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:37.5090794Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:37.5091623Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:37.5092428Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:37.5093346Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:37.5094274Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:37.5095116Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:37.5096169Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:37.5097095Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:37.5097954Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:37.5098803Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:37.5099702Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:37.5100483Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:37.5101298Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:37.5102188Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:37.5103021Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:37.5103825Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:37.5104627Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:37.5105484Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:37.5106436Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:37.5107337Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.5108177Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:37.5109072Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:37.5109976Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:37.5110831Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:37.5111679Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:37.5112659Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:37.5113507Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:37.5114342Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:37.5115139Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:37.5115969Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:37.5116875Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:37.5117777Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:37.5118771Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:37.5119621Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:37.5120458Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:37.5121309Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:37.5122193Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:37.5123123Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:37.5124002Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:37.5124845Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:37.5125702Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:37.5126647Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:37.5127520Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:37.5128414Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:37.5129235Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:37.5130095Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:37.5130928Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:37.5131747Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:37.5132687Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:37.5133565Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:37.5134404Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:37.5135236Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:37.5136123Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:37.5137118Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:37.5138046Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:37.5138930Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:37.5139884Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:37.5140740Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:37.5141625Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:37.5142569Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:37.5143512Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:37.5144406Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:37.5145242Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:37.5146077Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:37.5146912Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:37.5147808Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:37.5148661Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:37.5149558Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:37.5150445Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:37.5151220Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:37.5152026Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:37.5152844Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:37.5153661Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:37.5154474Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:37.5155298Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:37.5156086Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:37.5156960Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:37.5157922Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:37.5158839Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:37.5159663Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:37.5160460Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:37.5161337Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:37.5162162Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:37.5163045Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:37.5163899Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:37.5164743Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:37.5165621Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:37.5166478Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:37.5167306Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:37.5168166Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:37.5169050Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:37.5169931Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:37.5170767Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:37.5171508Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:37.5172444Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:37.5173275Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:37.5174040Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:37.5174862Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:37.5175647Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:37.5176469Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:37.5177413Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:37.5178186Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:37.5178954Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:37.5179785Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:37.5180646Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:37.5181490Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:37.5182371Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:37.5183271Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:37.5184089Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:37.5184897Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:37.5185710Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:37.5186471Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:37.5187232Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:37.5187992Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:37.5188792Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:37.5189602Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:37.5190566Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:37.5191388Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:37.5192210Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:37.5192968Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:37.5193764Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:37.5195751Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:37.5198512Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:37.5199346Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:37.5200334Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:37.5201134Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:37.5201971Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:37.5202947Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:37.5203800Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:37.5204734Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:37.5205643Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:37.5206538Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:37.5207476Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:37.5208386Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:37.5209421Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:37.5210303Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:37.5211149Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:37.5212076Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:37.5213010Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:37.5213888Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:37.5214775Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:37.5215779Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:37.5216547Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:37.5217325Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:37.5218071Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:37.5218796Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:37.5219516Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:37.5220250Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:37.5221125Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:37.5221966Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:37.5222832Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:37.5223589Z V1204 11:23:25.980000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:37.5224259Z V1204 11:23:25.981000 101377 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.5224954Z I1204 11:23:25.981000 101377 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.5225454Z V1204 11:23:25.982000 101377 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:37.5225934Z V1204 11:23:25.982000 101377 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:37.5226437Z I1204 11:23:25.982000 101377 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:37.5227224Z V1204 11:23:25.995000 101377 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:37.5227670Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.5228177Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:37.5229127Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.5229872Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.5230953Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.5231945Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.5232942Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.5233780Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:37.5234605Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:37.5235646Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:37.5236480Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:37.5237437Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.5238343Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.5238959Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.5239386Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.5239791Z V1204 11:23:25.999000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.5240369Z V1204 11:23:26.002000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:37.5240944Z V1204 11:23:26.002000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:37.5241953Z V1204 11:23:26.003000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.5242471Z V1204 11:23:26.003000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.5243478Z V1204 11:23:26.004000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.5244012Z V1204 11:23:26.005000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.5244993Z V1204 11:23:26.006000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:37.5245450Z V1204 11:23:26.006000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.5246403Z V1204 11:23:26.007000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:37.5246855Z V1204 11:23:26.007000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.5247783Z V1204 11:23:26.009000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:37.5248212Z V1204 11:23:26.009000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.5249067Z V1204 11:23:26.011000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:37.5249512Z V1204 11:23:26.011000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.5250406Z V1204 11:23:26.013000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.5250896Z V1204 11:23:26.014000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.5251792Z V1204 11:23:26.014000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.5252285Z V1204 11:23:26.015000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.5252799Z V1204 11:23:26.015000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.5253373Z V1204 11:23:26.016000 101377 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:37.5254125Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:37.5254233Z warnings.warn( 2025-12-04T11:24:37.5254727Z V1204 11:23:26.016000 101377 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:37.5254959Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:37.5255071Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:37.5255199Z stats [('calls_captured', 1)] 2025-12-04T11:24:37.5255427Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:37.5256183Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:37.5256294Z graph_break [] 2025-12-04T11:24:37.5256512Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:37.5257070Z V1204 11:23:26.033000 101377 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpffvs4woc 2025-12-04T11:24:37.5257528Z V1204 11:23:26.054000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.5258057Z V1204 11:23:26.054000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:37.5259015Z V1204 11:23:26.054000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.5259742Z V1204 11:23:26.054000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.5260247Z V1204 11:23:26.054000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:37.5260739Z V1204 11:23:26.054000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:37.5261161Z V1204 11:23:26.054000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.5262204Z V1204 11:23:26.054000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.5262950Z V1204 11:23:26.054000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:37.5263441Z V1204 11:23:26.054000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:37.5263887Z V1204 11:23:26.054000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.5264293Z V1204 11:23:26.054000 101377 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.5265010Z V1204 11:23:26.287000 101377 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:37.5265752Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:37.5266432Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:37.5266712Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.5267006Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.5267285Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.5267758Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:37.5268590Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.5269407Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.5270043Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:37.5270843Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.5271640Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.5272093Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:37.5272389Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.5272985Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:37.5275319Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.5277644Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.5278315Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:37.5279239Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:37.5279995Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:37.5280817Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:37.5281594Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:37.5282324Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:37.5283040Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:37.5283767Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:37.5284494Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:37.5285229Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:37.5285952Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:37.5286638Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:37.5287554Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:37.5288343Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:37.5289502Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:37.5290306Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:37.5319497Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:37.5320537Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:37.5321310Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:37.5322078Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:37.5322930Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:37.5323819Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:37.5324590Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:37.5325364Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:37.5326156Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:37.5326930Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:37.5327788Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:37.5328641Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:37.5329510Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:37.5330373Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:37.5331251Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:37.5332163Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:37.5333063Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:37.5333894Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:37.5334897Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:37.5335663Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:37.5336516Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:37.5337321Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:37.5338171Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:37.5338932Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:37.5339690Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:37.5340511Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:37.5341259Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:37.5342022Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:37.5342856Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:37.5343630Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:37.5344409Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:37.5345195Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:37.5345960Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:37.5346743Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:37.5347561Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:37.5348324Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:37.5349159Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:37.5349940Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:37.5350715Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:37.5351486Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:37.5352295Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:37.5353080Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:37.5353842Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:37.5354671Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:37.5355507Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:37.5356373Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:37.5357162Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:37.5358026Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:37.5358811Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:37.5359570Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:37.5360333Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:37.5361141Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:37.5362069Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:37.5362886Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:37.5363674Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:37.5364521Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:37.5365297Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:37.5366504Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:37.5367303Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:37.5368180Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:37.5369064Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:37.5369878Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:37.5370700Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:37.5371575Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:37.5372491Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:37.5373293Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:37.5374243Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.5375066Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:37.5376039Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.5376840Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:37.5377829Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.5378657Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:37.5379567Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:37.5380343Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:37.5381123Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:37.5381975Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:37.5382823Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:37.5383585Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:37.5384424Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:37.5385211Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:37.5386057Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:37.5386915Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:37.5387881Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:37.5388766Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:37.5389535Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:37.5390371Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:37.5391247Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:37.5392058Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:37.5392831Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:37.5393733Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:37.5394587Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:37.5395497Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:37.5396537Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:37.5397388Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:37.5398193Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:37.5398913Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:37.5399718Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:37.5400520Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:37.5401327Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:37.5402151Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:37.5403020Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:37.5403851Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:37.5404698Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:37.5405449Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:37.5406243Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:37.5407175Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:37.5408100Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:37.5409035Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:37.5409854Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:37.5410678Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:37.5411571Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:37.5412401Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:37.5413343Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:37.5414201Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:37.5414981Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:37.5415804Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:37.5416629Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:37.5417528Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:37.5418302Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:37.5419139Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:37.5419970Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:37.5420787Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:37.5421573Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:37.5422376Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:37.5423209Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:37.5424044Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:37.5424918Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:37.5425690Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:37.5426566Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:37.5427393Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:37.5428266Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:37.5429115Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:37.5429903Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:37.5430674Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:37.5431460Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:37.5432238Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:37.5433081Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:37.5433922Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:37.5434766Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:37.5435626Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:37.5436519Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:37.5437280Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:37.5438194Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:37.5439041Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:37.5439853Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:37.5440662Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:37.5441492Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:37.5442276Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:37.5443073Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:37.5443866Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:37.5444666Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:37.5445409Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:37.5446199Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:37.5447053Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:37.5447911Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:37.5448731Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:37.5449534Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:37.5450296Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:37.5451139Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:37.5452069Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:37.5452916Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:37.5453802Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:37.5454609Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:37.5455457Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:37.5456254Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:37.5457132Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:37.5457973Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:37.5458759Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:37.5459600Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:37.5460477Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:37.5461315Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:37.5462128Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:37.5462992Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:37.5463816Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:37.5464576Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:37.5465409Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:37.5466200Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:37.5467029Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:37.5467878Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:37.5468838Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:37.5469851Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:37.5470817Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:37.5471793Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:37.5472772Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:37.5473779Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:37.5474776Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:37.5475792Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:37.5476791Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:37.5477650Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:37.5478485Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:37.5479302Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:37.5480228Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:37.5481065Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:37.5481855Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:37.5482633Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:37.5483434Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:37.5484270Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:37.5485060Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:37.5485871Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:37.5486654Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:37.5487552Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:37.5488378Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:37.5489206Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:37.5489966Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:37.5490775Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:37.5491597Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:37.5492374Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:37.5493251Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:37.5494090Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:37.5494892Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:37.5495717Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:37.5496587Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:37.5497439Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:37.5498230Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:37.5499072Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:37.5500081Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:37.5500938Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:37.5501673Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:37.5502509Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:37.5503343Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:37.5504197Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:37.5505023Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:37.5505829Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:37.5506785Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:37.5507696Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:37.5508544Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:37.5509376Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:37.5510283Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:37.5511091Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:37.5511874Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:37.5512704Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:37.5513533Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:37.5514388Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:37.5515272Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:37.5516015Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:37.5516862Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:37.5517663Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:37.5518424Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:37.5519216Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:37.5519989Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:37.5520836Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:37.5521621Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:37.5522499Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.5523399Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.5524184Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:37.5525067Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:37.5525912Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:37.5526758Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:37.5527619Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:37.5528524Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:37.5529360Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:37.5530236Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:37.5531033Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:37.5531905Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:37.5532757Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:37.5533576Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:37.5534418Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:37.5535233Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:37.5536087Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:37.5536973Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:37.5537843Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:37.5538632Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:37.5539480Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:37.5540382Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:37.5541270Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:37.5542149Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:37.5543055Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:37.5543974Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:37.5544893Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:37.5545806Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:37.5546642Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:37.5547529Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:37.5548401Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:37.5549325Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:37.5550208Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:37.5551045Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:37.5551876Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:37.5552701Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:37.5553508Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:37.5554374Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:37.5555154Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:37.5556013Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:37.5556845Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:37.5557708Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:37.5558575Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:37.5559458Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:37.5560289Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:37.5561088Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:37.5562004Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:37.5562895Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:37.5563770Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:37.5564660Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:37.5565528Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:37.5566378Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:37.5567232Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:37.5568126Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:37.5568905Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:37.5569723Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:37.5570561Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:37.5571392Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:37.5572192Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:37.5573030Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:37.5573889Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:37.5574835Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:37.5575724Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.5576569Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:37.5577527Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:37.5578391Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:37.5579247Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:37.5580105Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:37.5581117Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:37.5581980Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:37.5582808Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:37.5583607Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:37.5584438Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:37.5585344Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:37.5586243Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:37.5587175Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:37.5588028Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:37.5588867Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:37.5589746Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:37.5590629Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:37.5591553Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:37.5592438Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:37.5593285Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:37.5594137Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:37.5595044Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:37.5595947Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:37.5596980Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:37.5597802Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:37.5598654Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:37.5599492Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:37.5600307Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:37.5601248Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:37.5602127Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:37.5602933Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:37.5603744Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:37.5604674Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:37.5605638Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:37.5606568Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:37.5607444Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:37.5608404Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:37.5609264Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:37.5610139Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:37.5611022Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:37.5611958Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:37.5612855Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:37.5613737Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:37.5614574Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:37.5615392Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:37.5616292Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:37.5617223Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:37.5618124Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:37.5619007Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:37.5619747Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:37.5620562Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:37.5621372Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:37.5622218Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:37.5623030Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:37.5623853Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:37.5624643Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:37.5625520Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:37.5626481Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:37.5627366Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:37.5628218Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:37.5628983Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:37.5629893Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:37.5630721Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:37.5631589Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:37.5632446Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:37.5633286Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:37.5634178Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:37.5635001Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:37.5635831Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:37.5636702Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:37.5637588Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:37.5638496Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:37.5639331Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:37.5640068Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:37.5641005Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:37.5641836Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:37.5642601Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:37.5643385Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:37.5644202Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:37.5645025Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:37.5645899Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:37.5646668Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:37.5647425Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:37.5648257Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:37.5649117Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:37.5649966Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:37.5650847Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:37.5651721Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:37.5652536Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:37.5653344Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:37.5654185Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:37.5654945Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:37.5655709Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:37.5656483Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:37.5657339Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:37.5658157Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:37.5659024Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:37.5659847Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:37.5660675Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:37.5661469Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:37.5662279Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:37.5664241Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:37.5666689Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:37.5667455Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:37.5668406Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:37.5669196Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:37.5670053Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:37.5671021Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:37.5671860Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:37.5672781Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:37.5673690Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:37.5674588Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:37.5675485Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:37.5676423Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:37.5677420Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:37.5678333Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:37.5679169Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:37.5680097Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:37.5681030Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:37.5681912Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:37.5682796Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:37.5683759Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:37.5684528Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:37.5685238Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:37.5685981Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:37.5686738Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:37.5687452Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:37.5688186Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:37.5689054Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:37.5689895Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:37.5690752Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:37.5691469Z V1204 11:23:26.296000 101377 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:37.5692167Z V1204 11:23:26.297000 101377 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.5692839Z I1204 11:23:26.297000 101377 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.5693374Z V1204 11:23:26.297000 101377 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:37.5693859Z V1204 11:23:26.297000 101377 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:37.5694362Z I1204 11:23:26.298000 101377 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 2 2025-12-04T11:24:37.5695142Z V1204 11:23:26.310000 101377 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:37.5695590Z V1204 11:23:26.315000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.5696246Z V1204 11:23:26.315000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:37.5697265Z V1204 11:23:26.315000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.5698009Z V1204 11:23:26.315000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.5699053Z V1204 11:23:26.315000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.5700068Z V1204 11:23:26.315000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.5701104Z V1204 11:23:26.315000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.5701940Z V1204 11:23:26.315000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:37.5702766Z V1204 11:23:26.315000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:37.5703811Z V1204 11:23:26.315000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:37.5704650Z V1204 11:23:26.315000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:37.5705577Z V1204 11:23:26.315000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.5706583Z V1204 11:23:26.315000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.5707213Z V1204 11:23:26.315000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.5707636Z V1204 11:23:26.315000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.5708041Z V1204 11:23:26.315000 101377 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.5708613Z V1204 11:23:26.317000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:37.5709191Z V1204 11:23:26.318000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:37.5710199Z V1204 11:23:26.318000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.5710720Z V1204 11:23:26.319000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.5711735Z V1204 11:23:26.320000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.5712237Z V1204 11:23:26.320000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.5713225Z V1204 11:23:26.321000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:37.5713685Z V1204 11:23:26.321000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.5714634Z V1204 11:23:26.323000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:37.5715114Z V1204 11:23:26.323000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.5716041Z V1204 11:23:26.324000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:37.5716470Z V1204 11:23:26.325000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.5717319Z V1204 11:23:26.327000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:37.5717767Z V1204 11:23:26.327000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.5718662Z V1204 11:23:26.329000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.5719122Z V1204 11:23:26.329000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.5720015Z V1204 11:23:26.330000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.5720504Z V1204 11:23:26.330000 101377 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.5728293Z V1204 11:23:26.331000 101377 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.5729007Z V1204 11:23:26.331000 101377 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 2 2025-12-04T11:24:37.5729756Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:37.5729854Z warnings.warn( 2025-12-04T11:24:37.5730337Z V1204 11:23:26.332000 101377 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:37.5731192Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-555ea3dc15f2a8d0.xml - 2025-12-04T11:24:37.5731368Z =========================== short test summary info ============================ 2025-12-04T11:24:37.5732265Z FAILED [0.3139s] inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True - AssertionError: Scalars are not equal! 2025-12-04T11:24:37.5732276Z 2025-12-04T11:24:37.5732378Z Expected 1 but got 0. 2025-12-04T11:24:37.5732483Z Absolute difference: 1 2025-12-04T11:24:37.5732596Z Relative difference: 1.0 2025-12-04T11:24:37.5732602Z 2025-12-04T11:24:37.5732819Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:37.5733672Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:37.5733680Z 2025-12-04T11:24:37.5734026Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:37.5734209Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:24:37.5734411Z ================== 1 failed, 30 deselected, 2 rerun in 6.57s =================== 2025-12-04T11:24:37.5734515Z --- Logging error --- 2025-12-04T11:24:37.5734641Z Traceback (most recent call last): 2025-12-04T11:24:37.5734982Z File "/opt/conda/envs/py_3.10/lib/python3.10/logging/__init__.py", line 1103, in emit 2025-12-04T11:24:37.5735106Z stream.write(msg + self.terminator) 2025-12-04T11:24:37.5735241Z ValueError: I/O operation on closed file. 2025-12-04T11:24:37.5735331Z Call stack: 2025-12-04T11:24:37.5735846Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/remote_cache.py", line 432, in dump_cache_stats 2025-12-04T11:24:37.5736000Z log.info("Cache Metrics:%s", out.getvalue()) 2025-12-04T11:24:37.5736107Z Message: 'Cache Metrics:%s' 2025-12-04T11:24:37.5736214Z Arguments: (' None\n',) 2025-12-04T11:24:37.5736308Z Got exit code 1 2025-12-04T11:24:37.5736410Z Retrying single test... 2025-12-04T11:24:37.5737224Z Test results will be stored in test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-01d96e43057cf9b8.xml 2025-12-04T11:24:37.5737388Z ============================= test session starts ============================== 2025-12-04T11:24:37.5737738Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:24:37.5737850Z cachedir: .pytest_cache 2025-12-04T11:24:37.5738366Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:24:37.5738494Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:24:37.5738597Z configfile: pytest.ini 2025-12-04T11:24:37.5739234Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:24:37.5740436Z collecting ... /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:52: PytestCollectionWarning: cannot collect test class 'TestDecomposeAddMM' because it has a __init__ constructor (from: test/inductor/test_decompose_mem_bound_mm.py) 2025-12-04T11:24:37.5740617Z class TestDecomposeAddMM(torch.nn.Module): 2025-12-04T11:24:37.5740769Z collected 37 items / 36 deselected / 1 selected 2025-12-04T11:24:37.5741701Z stepcurrent: skipping 30 already run items. Running only test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:37.5741815Z Running 1 items in this shard 2025-12-04T11:24:37.5741821Z 2025-12-04T11:24:37.5742625Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True ('RERUN', {'yellow': True}) [5.9469s] [100%] 2025-12-04T11:24:37.5743426Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True ('RERUN', {'yellow': True}) [0.3269s] [100%] 2025-12-04T11:24:37.5744141Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True FAILED [0.3176s] [100%] 2025-12-04T11:24:37.5744147Z 2025-12-04T11:24:37.5744282Z ==================================== RERUNS ==================================== 2025-12-04T11:24:37.5744717Z _ TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True _ 2025-12-04T11:24:37.5744840Z Traceback (most recent call last): 2025-12-04T11:24:37.5745399Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 384, in test_decompose_mm_mixed_precision 2025-12-04T11:24:37.5745509Z self.assertEqual( 2025-12-04T11:24:37.5746003Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:37.5746171Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:37.5746709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:37.5746912Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:37.5747065Z AssertionError: Scalars are not equal! 2025-12-04T11:24:37.5747080Z 2025-12-04T11:24:37.5747184Z Expected 1 but got 0. 2025-12-04T11:24:37.5747286Z Absolute difference: 1 2025-12-04T11:24:37.5747399Z Relative difference: 1.0 2025-12-04T11:24:37.5747404Z 2025-12-04T11:24:37.5747620Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:37.5748458Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:37.5748468Z 2025-12-04T11:24:37.5748736Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:37.5748951Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:37.5749064Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:37.5749171Z stats [('calls_captured', 1)] 2025-12-04T11:24:37.5749925Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:37.5750149Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:37.5750240Z graph_break [] 2025-12-04T11:24:37.5750451Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:37.5750986Z I1204 11:23:42.011000 101574 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:37.5751578Z I1204 11:23:42.259000 101574 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:37.5752047Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.5752532Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:37.5753482Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.5754204Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.5754693Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:37.5755183Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:37.5755595Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.5756621Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.5757358Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:37.5757858Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:37.5758282Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.5758659Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.5759249Z I1204 11:23:44.464000 101574 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:37.5759955Z V1204 11:23:44.949000 101574 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:37.5760667Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:37.5761325Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:37.5761602Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.5761876Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.5762144Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.5762614Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:37.5763431Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.5764238Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.5764891Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:37.5765682Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.5766428Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.5766872Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:37.5767158Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.5767740Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:37.5770083Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.5772409Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.5773062Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:37.5773995Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:37.5774716Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:37.5775528Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:37.5776314Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:37.5777093Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:37.5777807Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:37.5778537Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:37.5779245Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:37.5780029Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:37.5780744Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:37.5781425Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:37.5782296Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:37.5783075Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:37.5784253Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:37.5785040Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:37.5814162Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:37.5815189Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:37.5815953Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:37.5816776Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:37.5817660Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:37.5818544Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:37.5819269Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:37.5820033Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:37.5820832Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:37.5821595Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:37.5822451Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:37.5823273Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:37.5824126Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:37.5825016Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:37.5825859Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:37.5826794Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:37.5827644Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:37.5828464Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:37.5829465Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:37.5830227Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:37.5831059Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:37.5831796Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:37.5832672Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:37.5833428Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:37.5834163Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:37.5834950Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:37.5835690Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:37.5836439Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:37.5837265Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:37.5838027Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:37.5838793Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:37.5839559Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:37.5840347Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:37.5841121Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:37.5841893Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:37.5842670Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:37.5843466Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:37.5844239Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:37.5845007Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:37.5845770Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:37.5846577Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:37.5847343Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:37.5848127Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:37.5848948Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:37.5849779Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:37.5850609Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:37.5851385Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:37.5852253Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:37.5853018Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:37.5853770Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:37.5854518Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:37.5855318Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:37.5856275Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:37.5857129Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:37.5857978Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:37.5858808Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:37.5859556Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:37.5860764Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:37.5861555Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:37.5862443Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:37.5863330Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:37.5864191Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:37.5865030Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:37.5865878Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:37.5866772Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:37.5867572Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:37.5868527Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.5869352Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:37.5870326Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.5871126Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:37.5872106Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.5872923Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:37.5873828Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:37.5874603Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:37.5875351Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:37.5876211Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:37.5877061Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:37.5877835Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:37.5878654Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:37.5879442Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:37.5880328Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:37.5881221Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:37.5882163Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:37.5883020Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:37.5883802Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:37.5884631Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:37.5885516Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:37.5886333Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:37.5887109Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:37.5888042Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:37.5888903Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:37.5889817Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:37.5890683Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:37.5891480Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:37.5892295Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:37.5893018Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:37.5893822Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:37.5894624Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:37.5895430Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:37.5896422Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:37.5897365Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:37.5898204Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:37.5899006Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:37.5899758Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:37.5900561Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:37.5901491Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:37.5902415Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:37.5903359Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:37.5904212Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:37.5905037Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:37.5905925Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:37.5906755Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:37.5907672Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:37.5908532Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:37.5909327Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:37.5910139Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:37.5910962Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:37.5911806Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:37.5912623Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:37.5913457Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:37.5914293Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:37.5915080Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:37.5915851Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:37.5916655Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:37.5917487Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:37.5918320Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:37.5919198Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:37.5919972Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:37.5920905Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:37.5921695Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:37.5922589Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:37.5923411Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:37.5924202Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:37.5924970Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:37.5925748Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:37.5926540Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:37.5927371Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:37.5928248Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:37.5929123Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:37.5929951Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:37.5930824Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:37.5931576Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:37.5932484Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:37.5933326Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:37.5934151Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:37.5934942Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:37.5935769Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:37.5936582Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:37.5937405Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:37.5938225Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:37.5938992Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:37.5939751Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:37.5940531Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:37.5941385Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:37.5942240Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:37.5943053Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:37.5943861Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:37.5944654Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:37.5945511Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:37.5946315Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:37.5947160Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:37.5948046Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:37.5948860Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:37.5949708Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:37.5950498Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:37.5951330Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:37.5952182Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:37.5952981Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:37.5953806Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:37.5954671Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:37.5955482Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:37.5956300Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:37.5957175Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:37.5957983Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:37.5958737Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:37.5959572Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:37.5960393Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:37.5961596Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:37.5962423Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:37.5963396Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:37.5964403Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:37.5965379Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:37.5966356Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:37.5967334Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:37.5968344Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:37.5969377Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:37.5970465Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:37.5971460Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:37.5972313Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:37.5973145Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:37.5973965Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:37.5974889Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:37.5975723Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:37.5976523Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:37.5977399Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:37.5978220Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:37.5979024Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:37.5979809Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:37.5980613Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:37.5981398Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:37.5982298Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:37.5983117Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:37.5983938Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:37.5984726Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:37.5985538Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:37.5986355Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:37.5987129Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:37.5987977Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:37.5988821Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:37.5989628Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:37.5990447Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:37.5991189Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:37.5991976Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:37.5992800Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:37.5993667Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:37.5994613Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:37.5995421Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:37.5996370Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:37.5997233Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:37.5998058Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:37.5998915Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:37.5999732Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:37.6000536Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:37.6001560Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:37.6002473Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:37.6003326Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:37.6004109Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:37.6005024Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:37.6005832Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:37.6006612Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:37.6007438Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:37.6008260Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:37.6009179Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:37.6010036Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:37.6010777Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:37.6011595Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:37.6012392Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:37.6013157Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:37.6013945Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:37.6014731Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:37.6015557Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:37.6016341Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:37.6017314Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.6018209Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.6019029Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:37.6019881Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:37.6020728Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:37.6021576Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:37.6022437Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:37.6023345Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:37.6024186Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:37.6025096Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:37.6025924Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:37.6026762Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:37.6027585Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:37.6028419Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:37.6029244Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:37.6030056Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:37.6030910Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:37.6031726Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:37.6032591Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:37.6033407Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:37.6034237Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:37.6035160Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:37.6036023Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:37.6036910Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:37.6037825Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:37.6038740Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:37.6039659Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:37.6040581Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:37.6041471Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:37.6042374Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:37.6043215Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:37.6044140Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:37.6045032Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:37.6045876Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:37.6046709Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:37.6047523Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:37.6048341Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:37.6049224Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:37.6050011Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:37.6050878Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:37.6051704Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:37.6052528Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:37.6053395Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:37.6054291Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:37.6055104Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:37.6055900Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:37.6056805Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:37.6057785Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:37.6058663Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:37.6059521Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:37.6060397Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:37.6061240Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:37.6062091Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:37.6062983Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:37.6063763Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:37.6064580Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:37.6065447Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:37.6066254Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:37.6067076Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:37.6067877Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:37.6068734Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:37.6069681Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:37.6070574Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.6071415Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:37.6072301Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:37.6073197Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:37.6074093Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:37.6074900Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:37.6075879Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:37.6076722Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:37.6077554Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:37.6078352Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:37.6079184Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:37.6080102Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:37.6080991Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:37.6081961Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:37.6082806Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:37.6083643Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:37.6084493Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:37.6085376Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:37.6086293Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:37.6087178Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:37.6088029Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:37.6088862Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:37.6089811Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:37.6090713Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:37.6091575Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:37.6092388Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:37.6093245Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:37.6094094Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:37.6094890Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:37.6095824Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:37.6096828Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:37.6097760Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:37.6098568Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:37.6099508Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:37.6100450Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:37.6101359Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:37.6102247Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:37.6103196Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:37.6104048Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:37.6104923Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:37.6105851Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:37.6106824Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:37.6107688Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:37.6108548Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:37.6109369Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:37.6110204Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:37.6111103Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:37.6111968Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:37.6112867Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:37.6113793Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:37.6114546Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:37.6115333Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:37.6116144Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:37.6116952Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:37.6117769Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:37.6118591Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:37.6119381Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:37.6120274Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:37.6121226Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:37.6122143Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:37.6122989Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:37.6123762Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:37.6124634Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:37.6125467Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:37.6126340Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:37.6127194Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:37.6128043Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:37.6128917Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:37.6129773Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:37.6130611Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:37.6131468Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:37.6132353Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:37.6133235Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:37.6134080Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:37.6134803Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:37.6135734Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:37.6136559Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:37.6137378Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:37.6138202Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:37.6139007Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:37.6139817Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:37.6140683Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:37.6141454Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:37.6142217Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:37.6143047Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:37.6143907Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:37.6144752Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:37.6145673Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:37.6146533Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:37.6147351Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:37.6148156Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:37.6148967Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:37.6149735Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:37.6150495Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:37.6151266Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:37.6152052Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:37.6152864Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:37.6153763Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:37.6154612Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:37.6155403Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:37.6156167Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:37.6156975Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:37.6158948Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:37.6161379Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:37.6162189Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:37.6163112Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:37.6163938Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:37.6164785Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:37.6165739Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:37.6166576Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:37.6167498Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:37.6168401Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:37.6169300Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:37.6170250Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:37.6171191Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:37.6172181Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:37.6173066Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:37.6173911Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:37.6174837Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:37.6175774Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:37.6176657Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:37.6177593Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:37.6178606Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:37.6179365Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:37.6180094Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:37.6180839Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:37.6181565Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:37.6182290Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:37.6183027Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:37.6183903Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:37.6184761Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:37.6185609Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:37.6186360Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:37.6187054Z V1204 11:23:44.960000 101574 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.6187727Z I1204 11:23:44.960000 101574 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.6188225Z V1204 11:23:44.961000 101574 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:37.6188707Z V1204 11:23:44.961000 101574 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:37.6189211Z I1204 11:23:44.961000 101574 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:37.6189995Z V1204 11:23:45.002000 101574 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:37.6190441Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.6190952Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:37.6191907Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.6192667Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.6193742Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.6194752Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.6195739Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.6196740Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:37.6197569Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:37.6198613Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:37.6199445Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:37.6200438Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.6201385Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.6201975Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.6202418Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.6202808Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.6203383Z V1204 11:23:45.010000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:37.6203956Z V1204 11:23:45.011000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:37.6204964Z V1204 11:23:45.011000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.6205481Z V1204 11:23:45.012000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.6206489Z V1204 11:23:45.044000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.6207046Z V1204 11:23:45.045000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.6207986Z V1204 11:23:45.046000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:37.6208476Z V1204 11:23:45.046000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.6209430Z V1204 11:23:45.048000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:37.6209880Z V1204 11:23:45.048000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.6210804Z V1204 11:23:45.050000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:37.6211236Z V1204 11:23:45.050000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.6212100Z V1204 11:23:45.053000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:37.6212534Z V1204 11:23:45.053000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.6213419Z V1204 11:23:45.055000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.6213912Z V1204 11:23:45.056000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.6214847Z V1204 11:23:45.057000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.6215312Z V1204 11:23:45.057000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.6215825Z V1204 11:23:45.058000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.6216398Z V1204 11:23:45.058000 101574 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:37.6217225Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:37.6217334Z warnings.warn( 2025-12-04T11:24:37.6217830Z V1204 11:23:45.059000 101574 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:37.6218274Z _ TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True _ 2025-12-04T11:24:37.6218393Z Traceback (most recent call last): 2025-12-04T11:24:37.6218977Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 384, in test_decompose_mm_mixed_precision 2025-12-04T11:24:37.6219084Z self.assertEqual( 2025-12-04T11:24:37.6219554Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:37.6219722Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:37.6220262Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:37.6220593Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:37.6220728Z AssertionError: Scalars are not equal! 2025-12-04T11:24:37.6220735Z 2025-12-04T11:24:37.6220855Z Expected 1 but got 0. 2025-12-04T11:24:37.6220961Z Absolute difference: 1 2025-12-04T11:24:37.6221069Z Relative difference: 1.0 2025-12-04T11:24:37.6221075Z 2025-12-04T11:24:37.6221304Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:37.6222180Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:37.6222186Z 2025-12-04T11:24:37.6222455Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:37.6222692Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:37.6222804Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:37.6222933Z stats [('calls_captured', 1)] 2025-12-04T11:24:37.6223691Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:37.6223915Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:37.6224026Z graph_break [] 2025-12-04T11:24:37.6224247Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:37.6224759Z I1204 11:23:42.011000 101574 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:37.6225354Z I1204 11:23:42.259000 101574 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:37.6225794Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.6226332Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:37.6227295Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.6228038Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.6228526Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:37.6229007Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:37.6229440Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.6230469Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.6231226Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:37.6231698Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:37.6232124Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.6232508Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.6233095Z I1204 11:23:44.464000 101574 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:37.6233828Z V1204 11:23:44.949000 101574 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:37.6234562Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:37.6235242Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:37.6235521Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.6235801Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.6236090Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.6236561Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:37.6237391Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.6238204Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.6238824Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:37.6239638Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.6240411Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.6240871Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:37.6241168Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.6241748Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:37.6244083Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.6246406Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.6247052Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:37.6248002Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:37.6248727Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:37.6249570Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:37.6250365Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:37.6251078Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:37.6251804Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:37.6252528Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:37.6253247Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:37.6253974Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:37.6254650Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:37.6255373Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:37.6256272Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:37.6257155Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:37.6258311Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:37.6259102Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:37.6288139Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:37.6289159Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:37.6289947Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:37.6290712Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:37.6291557Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:37.6292451Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:37.6293214Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:37.6293987Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:37.6294771Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:37.6295557Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:37.6296586Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:37.6297496Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:37.6298356Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:37.6299217Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:37.6300063Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:37.6301021Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:37.6301890Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:37.6302740Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:37.6303764Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:37.6304526Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:37.6305379Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:37.6306126Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:37.6306961Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:37.6307729Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:37.6308485Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:37.6309330Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:37.6310078Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:37.6310845Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:37.6311677Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:37.6312440Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:37.6313229Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:37.6314005Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:37.6314784Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:37.6315564Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:37.6316386Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:37.6317156Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:37.6317986Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:37.6318779Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:37.6319546Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:37.6320334Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:37.6321145Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:37.6321922Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:37.6322685Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:37.6323490Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:37.6324332Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:37.6325178Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:37.6325987Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:37.6326851Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:37.6327636Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:37.6328397Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:37.6329150Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:37.6329965Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:37.6330890Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:37.6331699Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:37.6332526Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:37.6333372Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:37.6334154Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:37.6335336Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:37.6336145Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:37.6337084Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:37.6337988Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:37.6338805Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:37.6339628Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:37.6340521Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:37.6341444Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:37.6342244Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:37.6343186Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.6344031Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:37.6344993Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.6345813Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:37.6346752Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.6347588Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:37.6348500Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:37.6349262Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:37.6350064Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:37.6350901Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:37.6351755Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:37.6352520Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:37.6353351Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:37.6354138Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:37.6354978Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:37.6355854Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:37.6356806Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:37.6357713Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:37.6358488Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:37.6359325Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:37.6360205Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:37.6361018Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:37.6361794Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:37.6362681Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:37.6363563Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:37.6364457Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:37.6365334Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:37.6366161Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:37.6366968Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:37.6367688Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:37.6368479Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:37.6369298Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:37.6370092Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:37.6370932Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:37.6371758Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:37.6372622Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:37.6373459Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:37.6374195Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:37.6375005Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:37.6375913Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:37.6376898Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:37.6377843Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:37.6378659Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:37.6379478Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:37.6380349Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:37.6381194Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:37.6382121Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:37.6382989Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:37.6383771Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:37.6384599Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:37.6385426Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:37.6386272Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:37.6387042Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:37.6387830Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:37.6388700Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:37.6389510Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:37.6390294Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:37.6391095Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:37.6391930Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:37.6392759Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:37.6393621Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:37.6394407Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:37.6395280Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:37.6396267Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:37.6397143Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:37.6398006Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:37.6398797Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:37.6399555Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:37.6400359Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:37.6401146Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:37.6401991Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:37.6402835Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:37.6403685Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:37.6404551Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:37.6405443Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:37.6406215Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:37.6407110Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:37.6407962Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:37.6408778Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:37.6409586Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:37.6410423Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:37.6411193Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:37.6412022Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:37.6412797Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:37.6413580Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:37.6414351Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:37.6415146Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:37.6415994Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:37.6416832Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:37.6417722Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:37.6418518Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:37.6419286Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:37.6420139Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:37.6420983Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:37.6421824Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:37.6422693Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:37.6423514Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:37.6424354Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:37.6425165Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:37.6425983Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:37.6426819Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:37.6427604Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:37.6428417Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:37.6429301Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:37.6430126Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:37.6430956Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:37.6431807Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:37.6432631Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:37.6433382Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:37.6434214Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:37.6435004Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:37.6435830Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:37.6436660Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:37.6437645Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:37.6438659Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:37.6439630Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:37.6440604Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:37.6441576Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:37.6442570Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:37.6443573Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:37.6444592Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:37.6445598Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:37.6446463Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:37.6447297Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:37.6448108Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:37.6449032Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:37.6449868Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:37.6450650Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:37.6451434Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:37.6452224Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:37.6453070Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:37.6453860Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:37.6454665Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:37.6455450Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:37.6456345Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:37.6457244Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:37.6458065Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:37.6458828Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:37.6459643Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:37.6460476Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:37.6461247Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:37.6462111Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:37.6462969Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:37.6463763Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:37.6464604Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:37.6465332Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:37.6466130Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:37.6466926Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:37.6467747Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:37.6468737Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:37.6469579Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:37.6470324Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:37.6471161Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:37.6471992Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:37.6472849Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:37.6473656Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:37.6474475Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:37.6475415Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:37.6476332Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:37.6477176Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:37.6478005Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:37.6478909Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:37.6479716Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:37.6480496Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:37.6481314Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:37.6482152Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:37.6483014Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:37.6483847Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:37.6484622Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:37.6485553Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:37.6486361Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:37.6487114Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:37.6487914Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:37.6488690Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:37.6489540Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:37.6490330Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:37.6491209Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.6492110Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.6492885Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:37.6493787Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:37.6494625Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:37.6495491Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:37.6496529Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:37.6497492Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:37.6498336Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:37.6499203Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:37.6500014Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:37.6500901Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:37.6501779Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:37.6502598Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:37.6503440Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:37.6504250Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:37.6505098Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:37.6505937Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:37.6506788Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:37.6507598Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:37.6508446Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:37.6509348Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:37.6510247Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:37.6511129Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:37.6512041Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:37.6512947Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:37.6513886Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:37.6514793Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:37.6515650Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:37.6516538Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:37.6517380Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:37.6518332Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:37.6519208Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:37.6520066Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:37.6520884Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:37.6521720Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:37.6522531Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:37.6523397Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:37.6524178Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:37.6525025Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:37.6525859Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:37.6526711Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:37.6527586Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:37.6528465Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:37.6529295Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:37.6530101Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:37.6531015Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:37.6531913Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:37.6532791Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:37.6533691Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:37.6534563Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:37.6535412Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:37.6536259Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:37.6537209Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:37.6537999Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:37.6538803Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:37.6539655Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:37.6540448Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:37.6541295Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:37.6542101Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:37.6542981Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:37.6543929Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:37.6544813Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.6545666Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:37.6546540Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:37.6547420Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:37.6548277Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:37.6549123Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:37.6550130Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:37.6550977Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:37.6551803Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:37.6552593Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:37.6553438Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:37.6554348Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:37.6555250Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:37.6556189Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:37.6557033Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:37.6557868Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:37.6558733Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:37.6559624Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:37.6560538Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:37.6561434Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:37.6562274Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:37.6563127Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:37.6564035Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:37.6564929Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:37.6565835Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:37.6566647Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:37.6567512Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:37.6568348Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:37.6569152Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:37.6570097Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:37.6570972Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:37.6571772Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:37.6572560Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:37.6573485Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:37.6574408Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:37.6575354Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:37.6576228Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:37.6577247Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:37.6578106Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:37.6578968Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:37.6579864Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:37.6580780Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:37.6581693Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:37.6582564Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:37.6583400Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:37.6584225Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:37.6585117Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:37.6585970Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:37.6586853Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:37.6587745Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:37.6588486Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:37.6589288Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:37.6590103Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:37.6590959Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:37.6591772Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:37.6592583Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:37.6593389Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:37.6594267Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:37.6595229Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:37.6596274Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:37.6597095Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:37.6597923Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:37.6598826Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:37.6599666Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:37.6600526Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:37.6601395Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:37.6602239Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:37.6603126Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:37.6603944Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:37.6604762Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:37.6605649Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:37.6606526Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:37.6607468Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:37.6608296Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:37.6609033Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:37.6609966Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:37.6610792Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:37.6611560Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:37.6612336Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:37.6613168Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:37.6613965Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:37.6614874Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:37.6615643Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:37.6616402Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:37.6617297Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:37.6618141Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:37.6619004Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:37.6619880Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:37.6620745Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:37.6621569Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:37.6622380Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:37.6623221Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:37.6623966Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:37.6624748Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:37.6625511Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:37.6626309Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:37.6627125Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:37.6627987Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:37.6628819Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:37.6629626Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:37.6630433Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:37.6631239Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:37.6633227Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:37.6635639Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:37.6636419Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:37.6637360Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:37.6638170Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:37.6639028Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:37.6639999Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:37.6640839Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:37.6641764Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:37.6642676Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:37.6643565Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:37.6644484Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:37.6645381Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:37.6646416Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:37.6647329Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:37.6648170Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:37.6649093Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:37.6650019Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:37.6650916Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:37.6651788Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:37.6652770Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:37.6653511Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:37.6654273Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:37.6655018Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:37.6655762Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:37.6656490Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:37.6657270Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:37.6658165Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:37.6659007Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:37.6659878Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:37.6660598Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:37.6661250Z V1204 11:23:44.960000 101574 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.6661977Z I1204 11:23:44.960000 101574 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.6662489Z V1204 11:23:44.961000 101574 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:37.6662986Z V1204 11:23:44.961000 101574 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:37.6663490Z I1204 11:23:44.961000 101574 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:37.6664271Z V1204 11:23:45.002000 101574 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:37.6664722Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.6665221Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:37.6666191Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.6666933Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.6667976Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.6668967Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.6669994Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.6670833Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:37.6671659Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:37.6672701Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:37.6673526Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:37.6674463Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.6675349Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.6676013Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.6676438Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.6676842Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.6677413Z V1204 11:23:45.010000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:37.6677978Z V1204 11:23:45.011000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:37.6679006Z V1204 11:23:45.011000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.6679514Z V1204 11:23:45.012000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.6680537Z V1204 11:23:45.044000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.6681036Z V1204 11:23:45.045000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.6681988Z V1204 11:23:45.046000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:37.6682478Z V1204 11:23:45.046000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.6683418Z V1204 11:23:45.048000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:37.6683911Z V1204 11:23:45.048000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.6684815Z V1204 11:23:45.050000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:37.6685262Z V1204 11:23:45.050000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.6686119Z V1204 11:23:45.053000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:37.6686570Z V1204 11:23:45.053000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.6687470Z V1204 11:23:45.055000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.6687919Z V1204 11:23:45.056000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.6688817Z V1204 11:23:45.057000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.6689299Z V1204 11:23:45.057000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.6689855Z V1204 11:23:45.058000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.6690435Z V1204 11:23:45.058000 101574 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:37.6691188Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:37.6691290Z warnings.warn( 2025-12-04T11:24:37.6691767Z V1204 11:23:45.059000 101574 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:37.6692005Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:37.6692120Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:37.6692234Z stats [('calls_captured', 1)] 2025-12-04T11:24:37.6692472Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:37.6693225Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:37.6693337Z graph_break [] 2025-12-04T11:24:37.6693559Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:37.6693998Z V1204 11:23:45.296000 101574 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpgg44ryix 2025-12-04T11:24:37.6694448Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.6694944Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:37.6695933Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.6696934Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.6697433Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:37.6697935Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:37.6698357Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.6699405Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.6700148Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:37.6700641Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:37.6701066Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.6701453Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.6702237Z V1204 11:23:45.562000 101574 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:37.6703003Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:37.6703682Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:37.6703962Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.6704238Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.6704527Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.6704994Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:37.6705826Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.6706644Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.6707280Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:37.6708044Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.6708796Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.6709299Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:37.6709600Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.6710180Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:37.6712545Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.6714878Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.6715520Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:37.6716441Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:37.6717194Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:37.6718036Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:37.6718832Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:37.6719544Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:37.6720272Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:37.6720996Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:37.6721720Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:37.6722453Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:37.6723128Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:37.6723819Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:37.6724726Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:37.6725527Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:37.6726706Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:37.6727497Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:37.6756417Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:37.6757578Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:37.6758359Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:37.6759131Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:37.6759981Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:37.6760818Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:37.6761550Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:37.6762348Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:37.6763135Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:37.6763953Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:37.6764808Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:37.6765654Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:37.6766515Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:37.6767376Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:37.6768229Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:37.6769122Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:37.6770023Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:37.6770865Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:37.6771889Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:37.6772645Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:37.6773496Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:37.6774245Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:37.6775081Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:37.6775858Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:37.6776564Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:37.6777433Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:37.6778218Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:37.6778980Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:37.6779848Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:37.6780609Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:37.6781397Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:37.6782172Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:37.6782952Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:37.6783733Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:37.6784516Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:37.6785277Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:37.6786119Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:37.6786945Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:37.6787712Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:37.6788495Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:37.6789304Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:37.6790082Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:37.6790845Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:37.6791659Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:37.6792478Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:37.6793341Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:37.6794139Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:37.6795028Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:37.6795821Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:37.6796837Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:37.6797605Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:37.6798424Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:37.6799351Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:37.6800161Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:37.6800956Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:37.6801871Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:37.6802666Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:37.6803873Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:37.6804671Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:37.6805537Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:37.6806443Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:37.6807264Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:37.6808088Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:37.6808932Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:37.6809869Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:37.6810672Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:37.6811645Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.6812489Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:37.6813441Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.6814258Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:37.6815197Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.6816031Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:37.6816955Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:37.6817755Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:37.6818519Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:37.6819389Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:37.6820251Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:37.6821006Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:37.6821843Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:37.6822634Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:37.6823492Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:37.6824356Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:37.6825279Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:37.6826185Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:37.6826960Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:37.6827836Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:37.6828714Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:37.6829529Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:37.6830305Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:37.6831196Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:37.6832064Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:37.6832929Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:37.6833842Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:37.6834665Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:37.6835468Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:37.6836183Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:37.6836974Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:37.6837791Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:37.6838583Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:37.6839421Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:37.6840250Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:37.6841086Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:37.6841919Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:37.6842658Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:37.6843505Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:37.6844419Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:37.6845350Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:37.6846279Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:37.6847097Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:37.6847919Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:37.6848751Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:37.6849622Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:37.6850524Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:37.6851424Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:37.6852207Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:37.6853034Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:37.6853870Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:37.6854787Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:37.6855557Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:37.6856344Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:37.6857239Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:37.6858059Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:37.6858848Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:37.6859681Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:37.6860510Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:37.6861335Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:37.6862209Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:37.6862998Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:37.6863873Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:37.6864672Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:37.6865538Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:37.6866396Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:37.6867207Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:37.6867969Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:37.6868765Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:37.6869543Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:37.6870381Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:37.6871228Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:37.6872072Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:37.6872888Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:37.6873768Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:37.6874542Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:37.6875464Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:37.6876322Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:37.6877138Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:37.6877945Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:37.6878774Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:37.6879544Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:37.6880321Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:37.6881089Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:37.6881900Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:37.6882675Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:37.6883471Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:37.6884320Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:37.6885160Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:37.6885989Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:37.6886777Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:37.6887550Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:37.6888362Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:37.6889177Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:37.6890049Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:37.6890924Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:37.6891776Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:37.6892614Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:37.6893424Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:37.6894245Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:37.6895084Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:37.6895869Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:37.6896820Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:37.6897804Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:37.6898646Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:37.6899477Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:37.6900334Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:37.6901163Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:37.6901921Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:37.6902753Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:37.6903546Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:37.6904333Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:37.6905159Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:37.6906175Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:37.6907194Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:37.6908203Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:37.6909180Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:37.6910157Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:37.6911157Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:37.6912168Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:37.6913137Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:37.6914177Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:37.6915047Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:37.6915884Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:37.6916700Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:37.6917619Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:37.6918459Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:37.6919237Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:37.6920028Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:37.6920784Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:37.6921603Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:37.6922419Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:37.6923231Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:37.6924047Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:37.6924937Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:37.6925767Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:37.6926583Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:37.6927356Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:37.6928167Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:37.6928955Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:37.6929762Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:37.6930596Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:37.6931482Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:37.6932271Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:37.6933105Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:37.6933837Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:37.6934634Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:37.6935432Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:37.6936261Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:37.6937268Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:37.6938111Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:37.6938859Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:37.6939720Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:37.6940550Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:37.6941406Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:37.6942224Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:37.6943044Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:37.6943999Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:37.6944883Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:37.6945728Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:37.6946562Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:37.6947500Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:37.6948301Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:37.6949077Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:37.6949892Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:37.6950730Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:37.6951593Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:37.6952428Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:37.6953166Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:37.6954019Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:37.6954814Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:37.6955597Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:37.6956398Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:37.6957175Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:37.6958020Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:37.6958804Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:37.6959683Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.6960554Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.6961323Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:37.6962221Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:37.6963113Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:37.6963983Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:37.6964843Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:37.6965754Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:37.6966587Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:37.6967459Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:37.6968268Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:37.6969094Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:37.6969967Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:37.6970783Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:37.6971652Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:37.6972468Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:37.6973319Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:37.6974144Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:37.6974994Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:37.6975799Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:37.6976602Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:37.6977585Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:37.6978484Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:37.6979401Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:37.6980315Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:37.6981215Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:37.6982152Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:37.6983055Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:37.6983910Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:37.6984767Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:37.6985605Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:37.6986557Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:37.6987436Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:37.6988319Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:37.6989142Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:37.6989984Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:37.6990791Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:37.6991661Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:37.6992442Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:37.6993274Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:37.6994134Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:37.6994976Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:37.6995850Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:37.6996884Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:37.6997708Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:37.6998511Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:37.6999429Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:37.7000323Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:37.7001151Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:37.7002080Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:37.7002953Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:37.7003913Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:37.7004770Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:37.7005661Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:37.7006447Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:37.7007250Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:37.7008101Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:37.7008892Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:37.7009697Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:37.7010541Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:37.7011436Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:37.7012385Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:37.7013266Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.7014115Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:37.7014993Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:37.7015868Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:37.7016724Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:37.7017590Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:37.7018605Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:37.7019462Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:37.7020320Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:37.7021112Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:37.7021950Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:37.7022854Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:37.7023763Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:37.7024701Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:37.7025513Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:37.7026397Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:37.7027226Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:37.7028150Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:37.7029057Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:37.7029942Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:37.7030781Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:37.7031638Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:37.7032552Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:37.7033434Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:37.7034291Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:37.7035123Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:37.7035990Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:37.7036849Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:37.7037658Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:37.7038596Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:37.7039468Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:37.7040271Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:37.7041057Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:37.7041946Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:37.7042893Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:37.7043845Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:37.7044723Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:37.7045666Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:37.7046518Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:37.7047379Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:37.7048270Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:37.7049188Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:37.7050054Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:37.7050917Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:37.7051757Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:37.7052607Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:37.7053497Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:37.7054348Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:37.7055231Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:37.7056128Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:37.7056917Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:37.7057691Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:37.7058506Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:37.7059350Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:37.7060201Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:37.7061014Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:37.7061813Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:37.7062692Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:37.7063648Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:37.7064534Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:37.7065351Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:37.7066114Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:37.7067008Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:37.7067854Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:37.7068745Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:37.7069612Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:37.7070454Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:37.7071348Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:37.7072172Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:37.7073005Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:37.7073836Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:37.7074745Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:37.7075670Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:37.7076499Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:37.7077241Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:37.7078172Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:37.7079001Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:37.7079765Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:37.7080544Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:37.7081332Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:37.7082122Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:37.7083032Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:37.7083803Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:37.7084610Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:37.7085437Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:37.7086275Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:37.7092791Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:37.7093728Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:37.7094613Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:37.7095406Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:37.7096486Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:37.7097386Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:37.7098243Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:37.7099020Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:37.7099776Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:37.7100583Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:37.7101396Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:37.7102251Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:37.7103082Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:37.7103853Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:37.7104677Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:37.7105471Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:37.7107483Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:37.7109907Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:37.7110688Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:37.7111593Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:37.7112398Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:37.7113267Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:37.7114265Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:37.7115109Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:37.7116009Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:37.7116934Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:37.7117810Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:37.7118725Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:37.7119620Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:37.7120615Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:37.7121529Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:37.7122369Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:37.7123327Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:37.7124244Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:37.7125140Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:37.7126010Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:37.7126983Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:37.7127712Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:37.7128418Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:37.7129188Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:37.7129922Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:37.7130643Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:37.7131356Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:37.7132233Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:37.7133071Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:37.7133924Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:37.7134639Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:37.7135284Z V1204 11:23:45.572000 101574 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.7135963Z I1204 11:23:45.572000 101574 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.7136473Z V1204 11:23:45.572000 101574 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:37.7137028Z V1204 11:23:45.573000 101574 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:37.7137528Z I1204 11:23:45.573000 101574 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:37.7138330Z V1204 11:23:45.585000 101574 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:37.7138785Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.7139278Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:37.7140246Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.7140982Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.7142018Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.7142970Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.7143995Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.7144854Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:37.7145667Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:37.7146718Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:37.7147540Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:37.7148482Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.7149369Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.7149968Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.7150438Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.7150836Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.7151409Z V1204 11:23:45.592000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:37.7151994Z V1204 11:23:45.593000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:37.7153014Z V1204 11:23:45.594000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.7153518Z V1204 11:23:45.594000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.7154536Z V1204 11:23:45.595000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.7155034Z V1204 11:23:45.595000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.7155983Z V1204 11:23:45.596000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:37.7156437Z V1204 11:23:45.597000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.7157402Z V1204 11:23:45.598000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:37.7157898Z V1204 11:23:45.598000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.7158806Z V1204 11:23:45.600000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:37.7159246Z V1204 11:23:45.600000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.7160087Z V1204 11:23:45.602000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:37.7160531Z V1204 11:23:45.602000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.7161414Z V1204 11:23:45.604000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.7161864Z V1204 11:23:45.604000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.7162762Z V1204 11:23:45.605000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.7163203Z V1204 11:23:45.605000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.7163723Z V1204 11:23:45.606000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.7164329Z V1204 11:23:45.606000 101574 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:37.7165076Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:37.7165174Z warnings.warn( 2025-12-04T11:24:37.7165678Z V1204 11:23:45.607000 101574 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:37.7165839Z =================================== FAILURES =================================== 2025-12-04T11:24:37.7166277Z _ TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True _ 2025-12-04T11:24:37.7166399Z Traceback (most recent call last): 2025-12-04T11:24:37.7166975Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 384, in test_decompose_mm_mixed_precision 2025-12-04T11:24:37.7167079Z self.assertEqual( 2025-12-04T11:24:37.7167547Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:37.7167707Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:37.7168246Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:37.7168456Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:37.7168584Z AssertionError: Scalars are not equal! 2025-12-04T11:24:37.7168592Z 2025-12-04T11:24:37.7168697Z Expected 1 but got 0. 2025-12-04T11:24:37.7168810Z Absolute difference: 1 2025-12-04T11:24:37.7168918Z Relative difference: 1.0 2025-12-04T11:24:37.7168923Z 2025-12-04T11:24:37.7169175Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:37.7170020Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:37.7170056Z 2025-12-04T11:24:37.7170320Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:37.7170552Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:37.7170664Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:37.7170785Z stats [('calls_captured', 1)] 2025-12-04T11:24:37.7171544Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:37.7171762Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:37.7171871Z graph_break [] 2025-12-04T11:24:37.7172087Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:37.7172586Z I1204 11:23:42.011000 101574 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:37.7173192Z I1204 11:23:42.259000 101574 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:37.7173624Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.7174125Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:37.7175055Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.7175826Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.7176315Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:37.7176796Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:37.7177322Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.7178344Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.7179104Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:37.7179576Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:37.7180006Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.7180385Z V1204 11:23:42.262000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.7180937Z I1204 11:23:44.464000 101574 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:37.7181651Z V1204 11:23:44.949000 101574 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:37.7182391Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:37.7183061Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:37.7183371Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.7183643Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.7183923Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.7184388Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:37.7185209Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.7186022Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.7186644Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:37.7187416Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.7188158Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.7188613Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:37.7188939Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.7189516Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:37.7191883Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.7194208Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.7194857Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:37.7195782Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:37.7196697Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:37.7197582Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:37.7198431Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:37.7199139Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:37.7199864Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:37.7200581Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:37.7201313Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:37.7202045Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:37.7202721Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:37.7203414Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:37.7204282Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:37.7205125Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:37.7206315Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:37.7207163Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:37.7236083Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:37.7237195Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:37.7237967Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:37.7238740Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:37.7239593Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:37.7240437Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:37.7241150Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:37.7241934Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:37.7242750Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:37.7243540Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:37.7244837Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:37.7245684Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:37.7246544Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:37.7247413Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:37.7248263Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:37.7249161Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:37.7250029Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:37.7250880Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:37.7251936Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:37.7252698Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:37.7253545Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:37.7254288Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:37.7255124Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:37.7258705Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:37.7259422Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:37.7260216Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:37.7260982Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:37.7261801Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:37.7262657Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:37.7263492Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:37.7264296Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:37.7265076Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:37.7265847Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:37.7266652Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:37.7267432Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:37.7268205Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:37.7269004Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:37.7269789Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:37.7270613Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:37.7271378Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:37.7272207Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:37.7272977Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:37.7273761Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:37.7274719Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:37.7275529Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:37.7276366Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:37.7277155Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:37.7278061Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:37.7278835Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:37.7279652Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:37.7280401Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:37.7281225Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:37.7282143Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:37.7282943Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:37.7283744Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:37.7284572Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:37.7285342Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:37.7286548Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:37.7287362Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:37.7288226Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:37.7289103Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:37.7289929Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:37.7290787Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:37.7291643Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:37.7292522Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:37.7293355Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:37.7294299Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.7295154Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:37.7296311Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.7297192Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:37.7298141Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.7298955Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:37.7299835Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:37.7300595Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:37.7301354Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:37.7302189Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:37.7303105Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:37.7303884Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:37.7304697Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:37.7305495Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:37.7306333Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:37.7307259Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:37.7308181Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:37.7309039Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:37.7309868Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:37.7311643Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:37.7313525Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:37.7315358Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:37.7317094Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:37.7318912Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:37.7320818Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:37.7322678Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:37.7324598Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:37.7326407Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:37.7328141Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:37.7329840Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:37.7331479Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:37.7333207Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:37.7334944Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:37.7336693Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:37.7338619Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:37.7340391Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:37.7342171Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:37.7343905Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:37.7345589Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:37.7347461Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:37.7349436Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:37.7351421Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:37.7353303Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:37.7355077Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:37.7356900Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:37.7358713Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:37.7360602Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:37.7362507Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:37.7364328Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:37.7366073Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:37.7367860Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:37.7369681Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:37.7371434Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:37.7373172Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:37.7374949Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:37.7376716Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:37.7378541Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:37.7380279Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:37.7382065Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:37.7383870Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:37.7385718Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:37.7387511Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:37.7389324Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:37.7391134Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:37.7392945Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:37.7394811Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:37.7396708Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:37.7398489Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:37.7400161Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:37.7401868Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:37.7403619Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:37.7405432Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:37.7407318Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:37.7409105Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:37.7410923Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:37.7412728Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:37.7414532Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:37.7416468Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:37.7418330Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:37.7420087Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:37.7421891Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:37.7423740Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:37.7425429Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:37.7427111Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:37.7428795Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:37.7430464Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:37.7432199Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:37.7433985Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:37.7435818Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:37.7437628Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:37.7439383Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:37.7441087Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:37.7442852Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:37.7444619Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:37.7446399Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:37.7448299Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:37.7450129Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:37.7451947Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:37.7453724Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:37.7455481Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:37.7457330Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:37.7459109Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:37.7460817Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:37.7462624Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:37.7464447Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:37.7466242Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:37.7468066Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:37.7469883Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:37.7471590Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:37.7473307Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:37.7475064Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:37.7476845Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:37.7478584Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:37.7480501Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:37.7482626Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:37.7484740Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:37.7486849Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:37.7488939Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:37.7491034Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:37.7493169Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:37.7495293Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:37.7497631Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:37.7499612Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:37.7501482Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:37.7503277Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:37.7505158Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:37.7507057Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:37.7508823Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:37.7510518Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:37.7512274Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:37.7513986Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:37.7515721Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:37.7517489Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:37.7519202Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:37.7521069Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:37.7522927Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:37.7524706Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:37.7526429Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:37.7528141Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:37.7529882Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:37.7531578Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:37.7533335Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:37.7535156Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:37.7537028Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:37.7538798Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:37.7540496Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:37.7542165Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:37.7543901Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:37.7545701Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:37.7547623Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:37.7549503Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:37.7551216Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:37.7552953Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:37.7554768Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:37.7556586Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:37.7558405Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:37.7560170Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:37.7562068Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:37.7564014Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:37.7565873Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:37.7567636Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:37.7569470Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:37.7571347Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:37.7573043Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:37.7574783Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:37.7576565Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:37.7578472Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:37.7580357Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:37.7582050Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:37.7583744Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:37.7585527Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:37.7587230Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:37.7588942Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:37.7590651Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:37.7592405Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:37.7594165Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:37.7595963Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.7597983Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.7599769Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:37.7601536Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:37.7603382Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:37.7605279Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:37.7607125Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:37.7609032Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:37.7610914Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:37.7612756Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:37.7614626Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:37.7616375Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:37.7618214Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:37.7620047Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:37.7621843Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:37.7623669Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:37.7625460Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:37.7627262Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:37.7629069Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:37.7630856Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:37.7632602Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:37.7634422Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:37.7636310Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:37.7638178Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:37.7640161Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:37.7642112Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:37.7644059Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:37.7646020Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:37.7647904Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:37.7649777Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:37.7651604Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:37.7653491Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:37.7655447Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:37.7657387Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:37.7659236Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:37.7661030Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:37.7662804Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:37.7664599Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:37.7666386Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:37.7668140Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:37.7669930Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:37.7671720Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:37.7673534Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:37.7675454Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:37.7677282Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:37.7679044Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:37.7680892Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:37.7682820Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:37.7684729Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:37.7686555Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:37.7688446Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:37.7690327Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:37.7692149Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:37.7694061Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:37.7695883Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:37.7697828Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:37.7699621Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:37.7701396Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:37.7703140Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:37.7704889Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:37.7706690Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:37.7708695Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:37.7710667Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.7712534Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:37.7714401Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:37.7716299Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:37.7718211Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:37.7720011Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:37.7721949Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:37.7723951Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:37.7725776Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:37.7727587Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:37.7729332Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:37.7731201Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:37.7733140Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:37.7735103Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:37.7737025Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:37.7738793Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:37.7740600Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:37.7742452Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:37.7744421Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:37.7746343Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:37.7748190Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:37.7750016Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:37.7751908Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:37.7753862Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:37.7755728Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:37.7757526Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:37.7759351Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:37.7761186Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:37.7763004Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:37.7764879Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:37.7766803Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:37.7768611Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:37.7770355Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:37.7772169Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:37.7774110Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:37.7776063Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:37.7778163Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:37.7780131Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:37.7782073Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:37.7783946Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:37.7785828Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:37.7787842Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:37.7789770Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:37.7791611Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:37.7793452Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:37.7795254Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:37.7797329Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:37.7799212Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:37.7801103Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:37.7803019Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:37.7804795Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:37.7806435Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:37.7808152Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:37.7809904Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:37.7811667Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:37.7813466Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:37.7815215Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:37.7817087Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:37.7819052Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:37.7821027Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:37.7822921Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:37.7824626Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:37.7826403Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:37.7828281Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:37.7830115Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:37.7832004Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:37.7833837Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:37.7835690Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:37.7837540Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:37.7839326Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:37.7841123Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:37.7842961Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:37.7844863Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:37.7846724Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:37.7848463Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:37.7850263Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:37.7852139Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:37.7853859Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:37.7855546Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:37.7857334Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:37.7859061Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:37.7860850Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:37.7862670Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:37.7864338Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:37.7866092Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:37.7867895Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:37.7869737Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:37.7871608Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:37.7873491Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:37.7875288Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:37.7877004Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:37.7878762Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:37.7880470Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:37.7882164Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:37.7883830Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:37.7885521Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:37.7887267Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:37.7889085Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:37.7890954Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:37.7892699Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:37.7894362Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:37.7896281Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:37.7899288Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:37.7903832Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:37.7907159Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:37.7908970Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:37.7910820Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:37.7912574Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:37.7914507Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:37.7916509Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:37.7918394Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:37.7920342Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:37.7922278Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:37.7924192Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:37.7926181Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:37.7928221Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:37.7930229Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:37.7932118Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:37.7934010Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:37.7936026Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:37.7938031Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:37.7939936Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:37.7941917Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:37.7943759Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:37.7945347Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:37.7946932Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:37.7948530Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:37.7950109Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:37.7951710Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:37.7953451Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:37.7955305Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:37.7957136Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:37.7958840Z V1204 11:23:44.959000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:37.7960374Z V1204 11:23:44.960000 101574 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.7961839Z I1204 11:23:44.960000 101574 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.7963124Z V1204 11:23:44.961000 101574 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:37.7964231Z V1204 11:23:44.961000 101574 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:37.7965379Z I1204 11:23:44.961000 101574 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:37.7966795Z V1204 11:23:45.002000 101574 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:37.7968171Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.7969262Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:37.7970837Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.7972673Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.7974587Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.7976720Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.7978849Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.7980816Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:37.7982644Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:37.7984613Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:37.7986612Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:37.7988508Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.7990505Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.7992117Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.7993257Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.7994218Z V1204 11:23:45.007000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.7995355Z V1204 11:23:45.010000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:37.7996777Z V1204 11:23:45.011000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:37.7998558Z V1204 11:23:45.011000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.8000208Z V1204 11:23:45.012000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.8001861Z V1204 11:23:45.044000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.8003514Z V1204 11:23:45.045000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.8005112Z V1204 11:23:45.046000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:37.8006652Z V1204 11:23:45.046000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.8008184Z V1204 11:23:45.048000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:37.8009729Z V1204 11:23:45.048000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.8011244Z V1204 11:23:45.050000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:37.8012777Z V1204 11:23:45.050000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.8014199Z V1204 11:23:45.053000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:37.8015631Z V1204 11:23:45.053000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.8017177Z V1204 11:23:45.055000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.8018698Z V1204 11:23:45.056000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.8020263Z V1204 11:23:45.057000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.8021741Z V1204 11:23:45.057000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.8022855Z V1204 11:23:45.058000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.8024096Z V1204 11:23:45.058000 101574 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:37.8025606Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:37.8026594Z warnings.warn( 2025-12-04T11:24:37.8027238Z V1204 11:23:45.059000 101574 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:37.8028086Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:37.8028608Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:37.8028938Z stats [('calls_captured', 1)] 2025-12-04T11:24:37.8029384Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:37.8030515Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:37.8031521Z graph_break [] 2025-12-04T11:24:37.8031897Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:37.8032703Z V1204 11:23:45.296000 101574 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpgg44ryix 2025-12-04T11:24:37.8033735Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.8034806Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:37.8036381Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.8038187Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.8039549Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:37.8040706Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:37.8041747Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.8043338Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.8045241Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:37.8046601Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:37.8047638Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.8048612Z V1204 11:23:45.319000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.8049853Z V1204 11:23:45.562000 101574 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:37.8051416Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:37.8052931Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:37.8054029Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.8054736Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.8055440Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.8056330Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:37.8057848Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.8059627Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.8061198Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:37.8062719Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.8064384Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.8065713Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:37.8066597Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.8067614Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:37.8070668Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.8075457Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.8078533Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:37.8080255Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:37.8082037Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:37.8083714Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:37.8085449Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:37.8087123Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:37.8088688Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:37.8090299Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:37.8091882Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:37.8093471Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:37.8095031Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:37.8096714Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:37.8098469Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:37.8100279Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:37.8102412Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:37.8104157Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:37.8134077Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:37.8164489Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:37.8166389Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:37.8168079Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:37.8169886Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:37.8171705Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:37.8173399Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:37.8175030Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:37.8176728Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:37.8178483Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:37.8180297Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:37.8182135Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:37.8183975Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:37.8185830Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:37.8187663Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:37.8189584Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:37.8191472Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:37.8193291Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:37.8195293Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:37.8197350Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:37.8199135Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:37.8200846Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:37.8202573Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:37.8204314Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:37.8205925Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:37.8207558Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:37.8209248Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:37.8210894Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:37.8212623Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:37.8214412Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:37.8216082Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:37.8217852Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:37.8219540Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:37.8221237Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:37.8222938Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:37.8224666Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:37.8226370Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:37.8228095Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:37.8229823Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:37.8231503Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:37.8233249Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:37.8234979Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:37.8236657Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:37.8238386Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:37.8240142Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:37.8241906Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:37.8243652Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:37.8245451Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:37.8247239Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:37.8248957Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:37.8250593Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:37.8251424Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:37.8252338Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:37.8253141Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:37.8253967Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:37.8254802Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:37.8255563Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:37.8256779Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:37.8257646Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:37.8258546Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:37.8259446Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:37.8260257Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:37.8261081Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:37.8261934Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:37.8262812Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:37.8263624Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:37.8264560Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.8265396Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:37.8266390Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.8267201Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:37.8268133Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.8268951Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:37.8269834Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:37.8270627Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:37.8271393Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:37.8272237Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:37.8273131Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:37.8273894Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:37.8274754Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:37.8275552Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:37.8276391Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:37.8277265Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:37.8278196Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:37.8279065Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:37.8279833Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:37.8280675Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:37.8281554Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:37.8282395Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:37.8283175Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:37.8284059Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:37.8284929Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:37.8285801Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:37.8286716Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:37.8287513Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:37.8288308Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:37.8289068Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:37.8289852Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:37.8290759Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:37.8291555Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:37.8292401Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:37.8293227Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:37.8294053Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:37.8294872Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:37.8295612Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:37.8296640Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:37.8297621Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:37.8298639Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:37.8299566Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:37.8300377Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:37.8301217Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:37.8302055Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:37.8302970Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:37.8303871Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:37.8304734Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:37.8305549Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:37.8306378Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:37.8307244Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:37.8308074Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:37.8308857Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:37.8309647Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:37.8310494Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:37.8311272Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:37.8312058Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:37.8312864Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:37.8313679Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:37.8314555Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:37.8315415Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:37.8316201Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:37.8317080Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:37.8317876Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:37.8318779Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:37.8319591Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:37.8320388Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:37.8321170Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:37.8321962Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:37.8322769Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:37.8323612Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:37.8324451Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:37.8325283Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:37.8326116Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:37.8326978Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:37.8327743Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:37.8328635Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:37.8329489Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:37.8330334Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:37.8331126Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:37.8331966Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:37.8332736Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:37.8333514Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:37.8334324Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:37.8335101Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:37.8335841Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:37.8336623Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:37.8337600Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:37.8338447Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:37.8339305Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:37.8340096Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:37.8340870Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:37.8341679Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:37.8342488Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:37.8343342Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:37.8344214Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:37.8345042Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:37.8345905Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:37.8346718Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:37.8347539Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:37.8348376Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:37.8349166Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:37.8349950Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:37.8350863Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:37.8351663Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:37.8352485Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:37.8353378Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:37.8354211Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:37.8354986Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:37.8355807Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:37.8356607Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:37.8357400Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:37.8358238Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:37.8359198Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:37.8360220Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:37.8361202Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:37.8362200Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:37.8363188Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:37.8364182Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:37.8365191Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:37.8366170Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:37.8367227Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:37.8368063Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:37.8368897Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:37.8369737Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:37.8370653Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:37.8371532Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:37.8372311Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:37.8373106Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:37.8373868Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:37.8374689Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:37.8375480Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:37.8376270Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:37.8377122Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:37.8378012Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:37.8378895Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:37.8379713Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:37.8380479Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:37.8381286Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:37.8382062Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:37.8382877Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:37.8383710Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:37.8384564Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:37.8385379Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:37.8386223Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:37.8386985Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:37.8387779Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:37.8388587Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:37.8389416Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:37.8390378Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:37.8391177Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:37.8391922Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:37.8392768Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:37.8393579Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:37.8394475Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:37.8395281Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:37.8396320Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:37.8397300Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:37.8398194Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:37.8399124Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:37.8399926Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:37.8400829Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:37.8401662Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:37.8402465Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:37.8403329Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:37.8404171Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:37.8405030Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:37.8405870Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:37.8406617Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:37.8407428Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:37.8408235Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:37.8408989Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:37.8409788Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:37.8410604Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:37.8411446Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:37.8412225Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:37.8413090Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.8413974Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.8414776Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:37.8415645Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:37.8416479Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:37.8417435Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:37.8418293Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:37.8419223Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:37.8420071Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:37.8420933Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:37.8421745Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:37.8422566Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:37.8423400Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:37.8424213Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:37.8425037Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:37.8425867Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:37.8426737Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:37.8427569Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:37.8428421Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:37.8429226Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:37.8430031Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:37.8430966Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:37.8431819Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:37.8432687Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:37.8433639Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:37.8434539Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:37.8435505Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:37.8436407Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:37.8437255Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:37.8438111Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:37.8438940Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:37.8439873Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:37.8440745Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:37.8441603Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:37.8442455Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:37.8443291Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:37.8444097Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:37.8444959Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:37.8445741Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:37.8446561Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:37.8447431Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:37.8448246Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:37.8449123Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:37.8450026Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:37.8450864Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:37.8451696Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:37.8452596Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:37.8453502Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:37.8454330Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:37.8455209Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:37.8456072Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:37.8456978Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:37.8457830Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:37.8458756Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:37.8459562Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:37.8460364Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:37.8461215Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:37.8462004Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:37.8462845Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:37.8463650Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:37.8464487Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:37.8465474Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:37.8466364Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.8467248Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:37.8468127Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:37.8469000Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:37.8469861Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:37.8470678Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:37.8471661Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:37.8472495Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:37.8473331Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:37.8474122Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:37.8474991Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:37.8475896Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:37.8476792Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:37.8477728Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:37.8478524Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:37.8479406Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:37.8480239Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:37.8481131Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:37.8482068Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:37.8482959Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:37.8483858Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:37.8484709Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:37.8485625Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:37.8486487Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:37.8487366Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:37.8488168Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:37.8489031Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:37.8489866Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:37.8490707Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:37.8491644Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:37.8492499Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:37.8493313Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:37.8494104Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:37.8494998Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:37.8495955Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:37.8497086Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:37.8498022Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:37.8498960Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:37.8499876Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:37.8500746Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:37.8501642Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:37.8502565Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:37.8503443Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:37.8504289Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:37.8505121Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:37.8505951Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:37.8506832Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:37.8507738Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:37.8508617Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:37.8509508Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:37.8517658Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:37.8518549Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:37.8519513Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:37.8520314Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:37.8521128Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:37.8521974Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:37.8522776Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:37.8523704Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:37.8524654Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:37.8525549Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:37.8526354Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:37.8527130Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:37.8527991Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:37.8528828Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:37.8529684Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:37.8530535Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:37.8531432Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:37.8532303Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:37.8533130Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:37.8533947Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:37.8534790Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:37.8535690Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:37.8536584Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:37.8537510Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:37.8538272Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:37.8539221Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:37.8540059Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:37.8540832Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:37.8541599Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:37.8542391Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:37.8543179Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:37.8544043Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:37.8544812Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:37.8545559Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:37.8546391Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:37.8547264Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:37.8548118Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:37.8548985Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:37.8549835Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:37.8550632Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:37.8551458Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:37.8552276Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:37.8553017Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:37.8553878Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:37.8554636Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:37.8555453Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:37.8556270Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:37.8557120Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:37.8557950Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:37.8558721Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:37.8559488Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:37.8560278Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:37.8562250Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:37.8564715Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:37.8565489Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:37.8566398Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:37.8567243Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:37.8568073Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:37.8569024Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:37.8569901Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:37.8570802Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:37.8571739Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:37.8572617Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:37.8573519Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:37.8574415Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:37.8575395Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:37.8576289Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:37.8577179Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:37.8578116Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:37.8579066Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:37.8579959Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:37.8580825Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:37.8581795Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:37.8582529Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:37.8583255Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:37.8584006Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:37.8584708Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:37.8585433Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:37.8586188Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:37.8587068Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:37.8587934Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:37.8588780Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:37.8589507Z V1204 11:23:45.571000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:37.8590152Z V1204 11:23:45.572000 101574 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.8590830Z I1204 11:23:45.572000 101574 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.8591308Z V1204 11:23:45.572000 101574 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:37.8591786Z V1204 11:23:45.573000 101574 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:37.8592293Z I1204 11:23:45.573000 101574 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:37.8593057Z V1204 11:23:45.585000 101574 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:37.8593509Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.8594032Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:37.8594991Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.8595727Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.8596934Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.8597911Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.8598956Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.8599806Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:37.8600661Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:37.8601755Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:37.8602579Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:37.8603512Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.8604400Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.8604994Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.8605418Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.8605805Z V1204 11:23:45.590000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.8606387Z V1204 11:23:45.592000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:37.8606951Z V1204 11:23:45.593000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:37.8608003Z V1204 11:23:45.594000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.8608507Z V1204 11:23:45.594000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.8609530Z V1204 11:23:45.595000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.8610024Z V1204 11:23:45.595000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.8610959Z V1204 11:23:45.596000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:37.8611455Z V1204 11:23:45.597000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.8612386Z V1204 11:23:45.598000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:37.8612851Z V1204 11:23:45.598000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.8613752Z V1204 11:23:45.600000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:37.8614215Z V1204 11:23:45.600000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.8615064Z V1204 11:23:45.602000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:37.8615522Z V1204 11:23:45.602000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.8616420Z V1204 11:23:45.604000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.8616920Z V1204 11:23:45.604000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.8617825Z V1204 11:23:45.605000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.8618270Z V1204 11:23:45.605000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.8618782Z V1204 11:23:45.606000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.8619368Z V1204 11:23:45.606000 101574 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:37.8620099Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:37.8620210Z warnings.warn( 2025-12-04T11:24:37.8620687Z V1204 11:23:45.607000 101574 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:37.8620911Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:37.8621070Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:37.8621187Z stats [('calls_captured', 1)] 2025-12-04T11:24:37.8621416Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:37.8622172Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:37.8622273Z graph_break [] 2025-12-04T11:24:37.8622498Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:37.8622945Z V1204 11:23:45.624000 101574 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpia0fwtf1 2025-12-04T11:24:37.8623392Z V1204 11:23:45.647000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.8623889Z V1204 11:23:45.647000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:37.8624858Z V1204 11:23:45.647000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.8625589Z V1204 11:23:45.647000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.8626077Z V1204 11:23:45.647000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:37.8626563Z V1204 11:23:45.647000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:37.8627011Z V1204 11:23:45.647000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.8628035Z V1204 11:23:45.647000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.8628807Z V1204 11:23:45.647000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:37.8629279Z V1204 11:23:45.647000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:37.8629698Z V1204 11:23:45.647000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.8630081Z V1204 11:23:45.647000 101574 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.8630800Z V1204 11:23:45.881000 101574 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:37.8631520Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:37.8632174Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:37.8632455Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.8632727Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.8633001Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.8633468Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:37.8634311Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.8635135Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.8635756Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:37.8636525Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.8637275Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.8637777Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:37.8638066Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.8638637Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:37.8641017Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.8643354Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.8644002Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:37.8644916Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:37.8645650Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:37.8646470Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:37.8647267Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:37.8647979Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:37.8648708Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:37.8649470Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:37.8650183Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:37.8650925Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:37.8651605Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:37.8652304Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:37.8653177Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:37.8654002Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:37.8655195Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:37.8655975Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:37.8684953Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:37.8685990Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:37.8686806Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:37.8687576Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:37.8688418Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:37.8689275Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:37.8689994Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:37.8690812Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:37.8691594Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:37.8692385Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:37.8693240Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:37.8694113Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:37.8694993Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:37.8695869Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:37.8696928Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:37.8697834Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:37.8698708Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:37.8699524Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:37.8700527Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:37.8701298Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:37.8702136Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:37.8702954Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:37.8703792Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:37.8704578Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:37.8705284Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:37.8706077Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:37.8706884Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:37.8707635Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:37.8708488Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:37.8709246Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:37.8710084Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:37.8710862Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:37.8711664Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:37.8712468Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:37.8713247Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:37.8714022Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:37.8714828Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:37.8715623Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:37.8716391Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:37.8717163Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:37.8717980Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:37.8718797Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:37.8719575Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:37.8720380Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:37.8721194Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:37.8722019Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:37.8722834Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:37.8723708Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:37.8724478Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:37.8725277Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:37.8726024Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:37.8726839Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:37.8727786Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:37.8728592Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:37.8729386Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:37.8730217Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:37.8730985Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:37.8732173Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:37.8732982Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:37.8733848Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:37.8734781Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:37.8735599Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:37.8736405Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:37.8737332Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:37.8738217Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:37.8739071Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:37.8740006Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.8740845Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:37.8741831Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.8742645Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:37.8743606Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.8744421Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:37.8745306Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:37.8746063Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:37.8746827Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:37.8747663Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:37.8748518Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:37.8749275Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:37.8750085Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:37.8750918Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:37.8751751Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:37.8752621Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:37.8753544Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:37.8754409Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:37.8755215Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:37.8756033Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:37.8756919Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:37.8757749Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:37.8758545Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:37.8759457Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:37.8760325Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:37.8761190Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:37.8762052Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:37.8762871Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:37.8763665Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:37.8764401Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:37.8765190Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:37.8766010Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:37.8766836Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:37.8767663Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:37.8768500Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:37.8769322Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:37.8770140Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:37.8770909Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:37.8771715Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:37.8772621Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:37.8773582Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:37.8774516Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:37.8775369Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:37.8776200Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:37.8777100Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:37.8777951Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:37.8778861Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:37.8779729Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:37.8780512Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:37.8781326Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:37.8782208Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:37.8783042Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:37.8783827Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:37.8784613Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:37.8785456Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:37.8786232Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:37.8787036Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:37.8787851Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:37.8788665Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:37.8789535Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:37.8790398Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:37.8791214Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:37.8792090Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:37.8792871Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:37.8793748Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:37.8794564Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:37.8795364Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:37.8796271Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:37.8797064Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:37.8797843Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:37.8798723Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:37.8799584Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:37.8800415Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:37.8801253Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:37.8802109Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:37.8802924Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:37.8803823Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:37.8804674Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:37.8805525Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:37.8806324Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:37.8807211Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:37.8807985Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:37.8808764Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:37.8809535Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:37.8810314Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:37.8811062Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:37.8811839Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:37.8812708Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:37.8813556Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:37.8814412Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:37.8815204Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:37.8815982Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:37.8816797Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:37.8817684Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:37.8818650Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:37.8819526Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:37.8820348Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:37.8821222Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:37.8822030Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:37.8822850Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:37.8823710Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:37.8824515Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:37.8825299Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:37.8826176Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:37.8826972Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:37.8827794Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:37.8828652Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:37.8829470Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:37.8830272Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:37.8831103Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:37.8831907Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:37.8832701Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:37.8833537Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:37.8834500Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:37.8835557Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:37.8836549Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:37.8837558Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:37.8838545Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:37.8839573Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:37.8840586Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:37.8841563Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:37.8842568Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:37.8843414Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:37.8844246Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:37.8845074Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:37.8845981Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:37.8846866Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:37.8847652Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:37.8848448Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:37.8849209Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:37.8850014Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:37.8850818Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:37.8851647Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:37.8852445Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:37.8853333Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:37.8854194Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:37.8855018Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:37.8855806Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:37.8856628Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:37.8857463Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:37.8858249Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:37.8859092Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:37.8859947Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:37.8860735Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:37.8861572Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:37.8862302Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:37.8863127Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:37.8863935Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:37.8864758Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:37.8865712Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:37.8866511Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:37.8867296Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:37.8868133Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:37.8868939Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:37.8869829Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:37.8870640Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:37.8871480Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:37.8872416Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:37.8873298Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:37.8874140Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:37.8874927Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:37.8875850Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:37.8876644Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:37.8877434Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:37.8878248Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:37.8879125Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:37.8879980Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:37.8880805Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:37.8881565Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:37.8882372Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:37.8883212Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:37.8883963Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:37.8884762Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:37.8885565Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:37.8886400Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:37.8887236Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:37.8888100Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.8888984Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.8889755Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:37.8890625Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:37.8891462Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:37.8892314Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:37.8893180Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:37.8894071Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:37.8894949Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:37.8895815Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:37.8896767Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:37.8897645Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:37.8898484Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:37.8899385Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:37.8900213Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:37.8901046Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:37.8901934Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:37.8902771Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:37.8903662Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:37.8904467Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:37.8905272Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:37.8906163Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:37.8907037Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:37.8907902Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:37.8908832Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:37.8909730Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:37.8910658Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:37.8911611Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:37.8912446Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:37.8913312Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:37.8914141Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:37.8915073Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:37.8915981Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:37.8916835Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:37.8917650Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:37.8918507Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:37.8919319Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:37.8920199Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:37.8921007Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:37.8921828Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:37.8922663Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:37.8923485Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:37.8924359Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:37.8925235Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:37.8926050Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:37.8926891Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:37.8927795Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:37.8928708Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:37.8929534Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:37.8930406Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:37.8931304Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:37.8932142Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:37.8933005Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:37.8933911Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:37.8934710Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:37.8935537Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:37.8936389Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:37.8937235Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:37.8938029Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:37.8938842Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:37.8939689Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:37.8940654Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:37.8941535Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.8942389Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:37.8943314Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:37.8944191Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:37.8945044Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:37.8945846Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:37.8946836Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:37.8947698Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:37.8948539Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:37.8949323Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:37.8950189Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:37.8951088Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:37.8952010Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:37.8952958Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:37.8953755Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:37.8954609Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:37.8955440Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:37.8956332Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:37.8957241Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:37.8958137Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:37.8959022Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:37.8959862Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:37.8960781Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:37.8961635Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:37.8962513Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:37.8963358Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:37.8964225Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:37.8965060Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:37.8965853Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:37.8966842Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:37.8967726Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:37.8968546Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:37.8969338Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:37.8970231Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:37.8971153Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:37.8972066Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:37.8972959Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:37.8973888Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:37.8974759Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:37.8975652Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:37.8976543Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:37.8977528Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:37.8978403Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:37.8979245Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:37.8980105Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:37.8980947Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:37.8981826Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:37.8982727Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:37.8983623Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:37.8984558Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:37.8985303Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:37.8986064Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:37.8986888Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:37.8987695Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:37.8988521Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:37.8989331Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:37.8990134Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:37.8991015Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:37.8992000Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:37.8992897Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:37.8993704Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:37.8994486Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:37.8995354Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:37.8996404Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:37.8997263Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:37.8998129Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:37.8999036Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:37.8999923Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:37.9000793Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:37.9001613Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:37.9002462Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:37.9003337Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:37.9004240Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:37.9005071Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:37.9005798Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:37.9006747Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:37.9007621Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:37.9008399Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:37.9009174Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:37.9009961Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:37.9010754Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:37.9011620Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:37.9012446Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:37.9013194Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:37.9014035Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:37.9014907Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:37.9015769Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:37.9016691Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:37.9017609Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:37.9018405Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:37.9019205Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:37.9020031Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:37.9020782Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:37.9021556Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:37.9022309Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:37.9023092Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:37.9024396Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:37.9025248Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:37.9026083Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:37.9026859Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:37.9027633Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:37.9028469Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:37.9030436Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:37.9032913Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:37.9033691Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:37.9034605Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:37.9035420Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:37.9036255Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:37.9037216Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:37.9038069Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:37.9038975Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:37.9039895Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:37.9040818Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:37.9041733Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:37.9042631Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:37.9043612Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:37.9044510Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:37.9045376Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:37.9046315Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:37.9047259Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:37.9048156Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:37.9049065Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:37.9050045Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:37.9050779Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:37.9051484Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:37.9052236Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:37.9052956Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:37.9053680Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:37.9054403Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:37.9055288Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:37.9056161Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:37.9057078Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:37.9057812Z V1204 11:23:45.890000 101574 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:37.9058464Z V1204 11:23:45.891000 101574 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.9059146Z I1204 11:23:45.891000 101574 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.9059632Z V1204 11:23:45.892000 101574 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:37.9060170Z V1204 11:23:45.892000 101574 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:37.9060678Z I1204 11:23:45.892000 101574 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 2 2025-12-04T11:24:37.9061447Z V1204 11:23:45.905000 101574 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:37.9061906Z V1204 11:23:45.909000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.9062432Z V1204 11:23:45.909000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:37.9063397Z V1204 11:23:45.909000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.9064172Z V1204 11:23:45.909000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.9065205Z V1204 11:23:45.909000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.9066176Z V1204 11:23:45.909000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.9067165Z V1204 11:23:45.909000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.9068025Z V1204 11:23:45.909000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:37.9068834Z V1204 11:23:45.909000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:37.9069886Z V1204 11:23:45.909000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:37.9070817Z V1204 11:23:45.909000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:37.9071760Z V1204 11:23:45.909000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.9072657Z V1204 11:23:45.909000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.9073263Z V1204 11:23:45.909000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.9073691Z V1204 11:23:45.909000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.9074112Z V1204 11:23:45.909000 101574 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.9074696Z V1204 11:23:45.912000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:37.9075261Z V1204 11:23:45.912000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:37.9076327Z V1204 11:23:45.913000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.9076832Z V1204 11:23:45.913000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.9077882Z V1204 11:23:45.914000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.9078382Z V1204 11:23:45.915000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.9079328Z V1204 11:23:45.916000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:37.9079795Z V1204 11:23:45.916000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.9080724Z V1204 11:23:45.917000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:37.9081195Z V1204 11:23:45.917000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.9082099Z V1204 11:23:45.919000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:37.9082545Z V1204 11:23:45.919000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.9083398Z V1204 11:23:45.921000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:37.9083865Z V1204 11:23:45.921000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.9084772Z V1204 11:23:45.923000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.9085220Z V1204 11:23:45.923000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.9086121Z V1204 11:23:45.924000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.9086571Z V1204 11:23:45.925000 101574 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.9087094Z V1204 11:23:45.925000 101574 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.9087717Z V1204 11:23:45.926000 101574 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 2 2025-12-04T11:24:37.9088456Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:37.9088574Z warnings.warn( 2025-12-04T11:24:37.9089056Z V1204 11:23:45.926000 101574 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:37.9089911Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-01d96e43057cf9b8.xml - 2025-12-04T11:24:37.9090113Z =========================== short test summary info ============================ 2025-12-04T11:24:37.9091012Z FAILED [0.3176s] inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True - AssertionError: Scalars are not equal! 2025-12-04T11:24:37.9091025Z 2025-12-04T11:24:37.9091144Z Expected 1 but got 0. 2025-12-04T11:24:37.9091252Z Absolute difference: 1 2025-12-04T11:24:37.9091392Z Relative difference: 1.0 2025-12-04T11:24:37.9091411Z 2025-12-04T11:24:37.9091629Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:37.9092478Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:37.9092484Z 2025-12-04T11:24:37.9092762Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:37.9092944Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:24:37.9093155Z ================== 1 failed, 36 deselected, 2 rerun in 6.63s =================== 2025-12-04T11:24:37.9093263Z --- Logging error --- 2025-12-04T11:24:37.9093386Z Traceback (most recent call last): 2025-12-04T11:24:37.9093710Z File "/opt/conda/envs/py_3.10/lib/python3.10/logging/__init__.py", line 1103, in emit 2025-12-04T11:24:37.9093838Z stream.write(msg + self.terminator) 2025-12-04T11:24:37.9093973Z ValueError: I/O operation on closed file. 2025-12-04T11:24:37.9094080Z Call stack: 2025-12-04T11:24:37.9094597Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/remote_cache.py", line 432, in dump_cache_stats 2025-12-04T11:24:37.9094747Z log.info("Cache Metrics:%s", out.getvalue()) 2025-12-04T11:24:37.9094872Z Message: 'Cache Metrics:%s' 2025-12-04T11:24:37.9094977Z Arguments: (' None\n',) 2025-12-04T11:24:37.9095086Z Got exit code 1 2025-12-04T11:24:37.9095189Z Retrying single test... 2025-12-04T11:24:37.9095857Z Test results will be stored in test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-563749739dc7a57a.xml 2025-12-04T11:24:37.9096215Z ============================= test session starts ============================== 2025-12-04T11:24:37.9096568Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:24:37.9096679Z cachedir: .pytest_cache 2025-12-04T11:24:37.9097277Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:24:37.9097402Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:24:37.9097524Z configfile: pytest.ini 2025-12-04T11:24:37.9098115Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:24:37.9099313Z collecting ... /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:52: PytestCollectionWarning: cannot collect test class 'TestDecomposeAddMM' because it has a __init__ constructor (from: test/inductor/test_decompose_mem_bound_mm.py) 2025-12-04T11:24:37.9099547Z class TestDecomposeAddMM(torch.nn.Module): 2025-12-04T11:24:37.9099699Z collected 37 items / 36 deselected / 1 selected 2025-12-04T11:24:37.9100640Z stepcurrent: skipping 30 already run items. Running only test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:37.9100754Z Running 1 items in this shard 2025-12-04T11:24:37.9100760Z 2025-12-04T11:24:37.9101562Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True ('RERUN', {'yellow': True}) [5.9179s] [100%] 2025-12-04T11:24:37.9102405Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True ('RERUN', {'yellow': True}) [0.3430s] [100%] 2025-12-04T11:24:37.9103116Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True FAILED [0.3278s] [100%] 2025-12-04T11:24:37.9103123Z 2025-12-04T11:24:37.9103277Z ==================================== RERUNS ==================================== 2025-12-04T11:24:37.9103758Z _ TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True _ 2025-12-04T11:24:37.9103883Z Traceback (most recent call last): 2025-12-04T11:24:37.9104465Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 384, in test_decompose_mm_mixed_precision 2025-12-04T11:24:37.9104570Z self.assertEqual( 2025-12-04T11:24:37.9105043Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:37.9105212Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:37.9105748Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:37.9105971Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:37.9106102Z AssertionError: Scalars are not equal! 2025-12-04T11:24:37.9106108Z 2025-12-04T11:24:37.9106224Z Expected 1 but got 0. 2025-12-04T11:24:37.9106343Z Absolute difference: 1 2025-12-04T11:24:37.9106455Z Relative difference: 1.0 2025-12-04T11:24:37.9106460Z 2025-12-04T11:24:37.9106690Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:37.9107542Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:37.9107548Z 2025-12-04T11:24:37.9107817Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:37.9108055Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:37.9108210Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:37.9108341Z stats [('calls_captured', 1)] 2025-12-04T11:24:37.9109102Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:37.9109325Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:37.9109438Z graph_break [] 2025-12-04T11:24:37.9109658Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:37.9110172Z I1204 11:24:01.606000 101771 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:37.9110769Z I1204 11:24:01.855000 101771 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:37.9111211Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.9111719Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:37.9112698Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.9113443Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.9113933Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:37.9114453Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:37.9114890Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.9115942Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.9116708Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:37.9117185Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:37.9117623Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.9118007Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.9118576Z I1204 11:24:04.067000 101771 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:37.9119306Z V1204 11:24:04.547000 101771 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:37.9120018Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:37.9120696Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:37.9120980Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.9121257Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.9121580Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.9122051Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:37.9122888Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.9123705Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.9124334Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:37.9125117Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.9125917Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.9126380Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:37.9126676Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.9127264Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:37.9129670Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.9132014Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.9132658Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:37.9133582Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:37.9134309Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:37.9135115Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:37.9135910Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:37.9136618Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:37.9137443Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:37.9138166Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:37.9138880Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:37.9139610Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:37.9140290Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:37.9141021Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:37.9141893Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:37.9142688Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:37.9143847Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:37.9144676Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:37.9173657Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:37.9174717Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:37.9175498Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:37.9176266Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:37.9177183Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:37.9178032Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:37.9178803Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:37.9179575Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:37.9180357Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:37.9181175Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:37.9182034Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:37.9182917Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:37.9183778Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:37.9184635Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:37.9185487Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:37.9186382Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:37.9187247Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:37.9188054Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:37.9189068Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:37.9189858Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:37.9190712Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:37.9191454Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:37.9192287Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:37.9193060Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:37.9193767Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:37.9194603Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:37.9195352Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:37.9196285Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:37.9197183Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:37.9197950Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:37.9198784Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:37.9199559Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:37.9200344Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:37.9201130Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:37.9201925Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:37.9202692Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:37.9203486Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:37.9204279Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:37.9205043Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:37.9205880Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:37.9206691Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:37.9207466Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:37.9208225Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:37.9209039Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:37.9209853Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:37.9210719Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:37.9211516Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:37.9212378Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:37.9213193Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:37.9213955Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:37.9214729Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:37.9215544Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:37.9216462Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:37.9217331Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:37.9218129Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:37.9218973Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:37.9219716Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:37.9220900Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:37.9221744Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:37.9222609Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:37.9223514Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:37.9224323Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:37.9225144Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:37.9225990Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:37.9226910Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:37.9227711Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:37.9228674Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.9229508Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:37.9230505Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.9231320Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:37.9232253Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.9233079Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:37.9233949Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:37.9234710Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:37.9235469Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:37.9236304Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:37.9237159Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:37.9237948Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:37.9238773Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:37.9239555Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:37.9240388Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:37.9241254Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:37.9242214Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:37.9243082Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:37.9243854Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:37.9244725Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:37.9245598Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:37.9246450Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:37.9247224Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:37.9248108Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:37.9248986Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:37.9249858Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:37.9250742Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:37.9251537Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:37.9252342Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:37.9253064Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:37.9253891Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:37.9254714Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:37.9255504Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:37.9256344Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:37.9257230Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:37.9258103Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:37.9258908Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:37.9259648Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:37.9260454Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:37.9261404Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:37.9262385Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:37.9263316Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:37.9264136Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:37.9264962Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:37.9265805Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:37.9266647Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:37.9267539Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:37.9268408Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:37.9269192Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:37.9270053Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:37.9270881Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:37.9271711Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:37.9272499Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:37.9273291Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:37.9274170Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:37.9274957Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:37.9275748Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:37.9276586Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:37.9277418Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:37.9278249Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:37.9279137Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:37.9279929Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:37.9280802Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:37.9281601Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:37.9282468Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:37.9283295Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:37.9284080Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:37.9284840Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:37.9285668Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:37.9286450Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:37.9287291Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:37.9288135Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:37.9288983Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:37.9289806Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:37.9290688Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:37.9291455Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:37.9292351Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:37.9293232Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:37.9294051Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:37.9294890Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:37.9295719Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:37.9296638Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:37.9297477Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:37.9298256Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:37.9299037Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:37.9299780Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:37.9300577Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:37.9301422Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:37.9302326Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:37.9303152Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:37.9303945Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:37.9304727Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:37.9305544Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:37.9306408Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:37.9307250Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:37.9308124Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:37.9308990Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:37.9309832Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:37.9310678Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:37.9311497Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:37.9312326Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:37.9313116Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:37.9313902Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:37.9314781Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:37.9315577Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:37.9316405Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:37.9317268Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:37.9318123Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:37.9318873Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:37.9319699Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:37.9320492Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:37.9321274Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:37.9322137Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:37.9323085Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:37.9324095Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:37.9325110Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:37.9326090Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:37.9327161Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:37.9328158Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:37.9329163Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:37.9330139Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:37.9331148Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:37.9331985Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:37.9332820Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:37.9333635Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:37.9334589Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:37.9335422Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:37.9336202Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:37.9337044Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:37.9337801Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:37.9338656Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:37.9339439Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:37.9340239Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:37.9341053Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:37.9341948Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:37.9342805Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:37.9343620Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:37.9344385Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:37.9345200Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:37.9345990Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:37.9346763Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:37.9347599Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:37.9348449Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:37.9349238Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:37.9350111Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:37.9350845Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:37.9351646Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:37.9352434Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:37.9353264Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:37.9354268Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:37.9355064Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:37.9355808Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:37.9356643Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:37.9357495Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:37.9358352Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:37.9359189Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:37.9360010Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:37.9360950Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:37.9361836Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:37.9362680Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:37.9363475Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:37.9364384Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:37.9365188Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:37.9365997Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:37.9366817Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:37.9367653Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:37.9368511Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:37.9369343Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:37.9370115Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:37.9370930Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:37.9371722Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:37.9372479Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:37.9373307Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:37.9374084Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:37.9374951Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:37.9375738Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:37.9376614Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.9377534Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.9378316Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:37.9379177Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:37.9380012Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:37.9380875Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:37.9381768Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:37.9382675Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:37.9383509Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:37.9384374Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:37.9385186Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:37.9386048Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:37.9386875Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:37.9387690Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:37.9388524Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:37.9389366Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:37.9390214Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:37.9391070Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:37.9391927Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:37.9392734Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:37.9393539Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:37.9394444Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:37.9395303Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:37.9396318Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:37.9397230Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:37.9398189Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:37.9399119Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:37.9400022Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:37.9400877Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:37.9401728Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:37.9402625Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:37.9403543Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:37.9404419Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:37.9405316Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:37.9406135Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:37.9407015Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:37.9407825Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:37.9408691Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:37.9409475Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:37.9410292Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:37.9411135Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:37.9411959Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:37.9412833Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:37.9413708Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:37.9414568Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:37.9415368Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:37.9416282Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:37.9417256Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:37.9418084Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:37.9418993Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:37.9419860Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:37.9420714Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:37.9421600Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:37.9422497Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:37.9423325Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:37.9424126Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:37.9424976Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:37.9425773Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:37.9426587Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:37.9427389Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:37.9428242Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:37.9429188Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:37.9430064Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.9430953Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:37.9431830Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:37.9432715Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:37.9433577Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:37.9434394Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:37.9435412Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:37.9436259Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:37.9437092Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:37.9437910Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:37.9438757Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:37.9439692Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:37.9440596Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:37.9441538Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:37.9442359Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:37.9443198Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:37.9444028Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:37.9444921Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:37.9445831Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:37.9446777Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:37.9447623Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:37.9448482Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:37.9449398Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:37.9450252Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:37.9451166Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:37.9451966Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:37.9452835Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:37.9453697Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:37.9454513Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:37.9455479Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:37.9456351Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:37.9457208Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:37.9458002Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:37.9458901Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:37.9459824Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:37.9460748Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:37.9461628Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:37.9462613Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:37.9463470Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:37.9464337Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:37.9465231Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:37.9466153Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:37.9467057Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:37.9467901Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:37.9468728Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:37.9469578Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:37.9470467Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:37.9471351Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:37.9472233Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:37.9473127Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:37.9473870Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:37.9474643Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:37.9475459Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:37.9476268Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:37.9477076Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:37.9477885Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:37.9478713Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:37.9479598Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:37.9480558Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:37.9481441Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:37.9482265Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:37.9483069Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:37.9483934Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:37.9484773Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:37.9485661Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:37.9486527Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:37.9487403Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:37.9488288Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:37.9489111Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:37.9489932Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:37.9490781Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:37.9491665Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:37.9492554Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:37.9493383Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:37.9494115Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:37.9495076Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:37.9495903Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:37.9496804Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:37.9497645Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:37.9498441Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:37.9499300Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:37.9500183Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:37.9500953Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:37.9501759Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:37.9502592Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:37.9503491Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:37.9504355Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:37.9505230Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:37.9506099Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:37.9506889Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:37.9507696Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:37.9508509Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:37.9509254Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:37.9510030Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:37.9510833Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:37.9511642Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:37.9512452Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:37.9513315Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:37.9514140Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:37.9514913Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:37.9515716Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:37.9516509Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:37.9518512Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:37.9520958Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:37.9521735Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:37.9522646Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:37.9523456Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:37.9524289Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:37.9525256Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:37.9526102Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:37.9527050Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:37.9527960Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:37.9528838Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:37.9529749Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:37.9530651Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:37.9531683Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:37.9532564Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:37.9533403Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:37.9534353Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:37.9535275Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:37.9536192Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:37.9537120Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:37.9538101Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:37.9538836Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:37.9539550Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:37.9540292Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:37.9541008Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:37.9541737Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:37.9542459Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:37.9543381Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:37.9544223Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:37.9545080Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:37.9545793Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:37.9546448Z V1204 11:24:04.558000 101771 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.9547161Z I1204 11:24:04.558000 101771 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:37.9547651Z V1204 11:24:04.558000 101771 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:37.9548145Z V1204 11:24:04.558000 101771 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:37.9548649Z I1204 11:24:04.559000 101771 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:37.9549460Z V1204 11:24:04.598000 101771 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:37.9549908Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.9550406Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:37.9551400Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.9552150Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.9553197Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.9554162Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.9555165Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.9556003Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:37.9556826Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:37.9557901Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:37.9558722Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:37.9559659Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.9560551Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.9561184Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.9561610Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.9562012Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:37.9562584Z V1204 11:24:04.606000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:37.9563192Z V1204 11:24:04.607000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:37.9564225Z V1204 11:24:04.607000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.9564763Z V1204 11:24:04.608000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.9565784Z V1204 11:24:04.640000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:37.9566287Z V1204 11:24:04.640000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.9567242Z V1204 11:24:04.642000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:37.9567700Z V1204 11:24:04.642000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.9568647Z V1204 11:24:04.643000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:37.9569115Z V1204 11:24:04.644000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.9570029Z V1204 11:24:04.645000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:37.9570477Z V1204 11:24:04.645000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.9571358Z V1204 11:24:04.648000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:37.9571806Z V1204 11:24:04.648000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.9572695Z V1204 11:24:04.651000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.9573142Z V1204 11:24:04.651000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.9574050Z V1204 11:24:04.652000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:37.9574528Z V1204 11:24:04.652000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:37.9575059Z V1204 11:24:04.653000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:37.9575638Z V1204 11:24:04.654000 101771 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:37.9576395Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:37.9576499Z warnings.warn( 2025-12-04T11:24:37.9577130Z V1204 11:24:04.654000 101771 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:37.9577595Z _ TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True _ 2025-12-04T11:24:37.9577721Z Traceback (most recent call last): 2025-12-04T11:24:37.9578291Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 384, in test_decompose_mm_mixed_precision 2025-12-04T11:24:37.9578443Z self.assertEqual( 2025-12-04T11:24:37.9578903Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:37.9579084Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:37.9579622Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:37.9579831Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:37.9579980Z AssertionError: Scalars are not equal! 2025-12-04T11:24:37.9579988Z 2025-12-04T11:24:37.9580096Z Expected 1 but got 0. 2025-12-04T11:24:37.9580222Z Absolute difference: 1 2025-12-04T11:24:37.9580333Z Relative difference: 1.0 2025-12-04T11:24:37.9580342Z 2025-12-04T11:24:37.9580558Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:37.9581421Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:37.9581426Z 2025-12-04T11:24:37.9581698Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:37.9581934Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:37.9582050Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:37.9582167Z stats [('calls_captured', 1)] 2025-12-04T11:24:37.9582946Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:37.9583202Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:37.9583305Z graph_break [] 2025-12-04T11:24:37.9583540Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:37.9584043Z I1204 11:24:01.606000 101771 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:37.9584651Z I1204 11:24:01.855000 101771 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:37.9585086Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:37.9585577Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:37.9586541Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:37.9587305Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:37.9587808Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:37.9588297Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:37.9588716Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.9589780Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:37.9590529Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:37.9591040Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:37.9591461Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.9591863Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:37.9592420Z I1204 11:24:04.067000 101771 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:37.9593138Z V1204 11:24:04.547000 101771 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:37.9593866Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:37.9594529Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:37.9594819Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.9595094Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.9595366Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.9595846Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:37.9596869Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:37.9597701Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:37.9598323Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:37.9599103Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:37.9599850Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:37.9600303Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:37.9600681Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:37.9601258Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:37.9603656Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.9606017Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:37.9606678Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:37.9607586Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:37.9608325Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:37.9609144Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:37.9609938Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:37.9610648Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:37.9611369Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:37.9612147Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:37.9612860Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:37.9613609Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:37.9614287Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:37.9614971Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:37.9615861Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:37.9616699Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:37.9617948Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:37.9618728Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:37.9647608Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:37.9648637Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:37.9649439Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:37.9650226Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:37.9651063Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:37.9651913Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:37.9652630Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:37.9653417Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:37.9654235Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:37.9655008Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:37.9655868Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:37.9656735Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:37.9657679Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:37.9658568Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:37.9659433Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:37.9660327Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:37.9661184Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:37.9662011Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:37.9663018Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:37.9663787Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:37.9664628Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:37.9665419Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:37.9666255Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:37.9667032Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:37.9667734Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:37.9668532Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:37.9669296Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:37.9670081Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:37.9670925Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:37.9671683Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:37.9672499Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:37.9673273Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:37.9674074Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:37.9674870Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:37.9675639Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:37.9676410Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:37.9677206Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:37.9677995Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:37.9678757Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:37.9679522Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:37.9680341Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:37.9681139Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:37.9681915Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:37.9682723Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:37.9683535Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:37.9684360Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:37.9685178Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:37.9686050Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:37.9686818Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:37.9687589Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:37.9688365Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:37.9689187Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:37.9690131Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:37.9690926Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:37.9691731Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:37.9692555Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:37.9693314Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:37.9694497Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:37.9695301Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:37.9696313Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:37.9697325Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:37.9698158Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:37.9698965Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:37.9699828Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:37.9700705Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:37.9701575Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:37.9702506Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.9703330Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:37.9704347Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.9705154Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:37.9706139Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:37.9706950Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:37.9707833Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:37.9708591Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:37.9709356Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:37.9710194Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:37.9711040Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:37.9711810Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:37.9712630Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:37.9713474Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:37.9714311Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:37.9715176Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:37.9716098Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:37.9716956Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:37.9717777Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:37.9718602Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:37.9719491Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:37.9720327Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:37.9721117Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:37.9722037Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:37.9722898Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:37.9723777Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:37.9724641Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:37.9725459Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:37.9726255Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:37.9726984Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:37.9727772Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:37.9728582Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:37.9729418Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:37.9730249Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:37.9731093Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:37.9731906Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:37.9732723Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:37.9733495Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:37.9734304Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:37.9735212Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:37.9736164Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:37.9737185Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:37.9738037Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:37.9738874Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:37.9739711Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:37.9740559Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:37.9741465Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:37.9742326Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:37.9743129Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:37.9743939Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:37.9744784Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:37.9745649Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:37.9746432Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:37.9747222Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:37.9748055Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:37.9748859Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:37.9749665Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:37.9750484Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:37.9751305Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:37.9752177Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:37.9753046Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:37.9753864Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:37.9754758Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:37.9755540Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:37.9756424Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:37.9757243Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:37.9758050Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:37.9758804Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:37.9759590Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:37.9760385Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:37.9761245Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:37.9762105Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:37.9762941Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:37.9763780Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:37.9764635Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:37.9765437Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:37.9766339Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:37.9767182Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:37.9768035Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:37.9768837Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:37.9769709Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:37.9770478Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:37.9771253Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:37.9772025Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:37.9772792Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:37.9773548Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:37.9774326Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:37.9775187Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:37.9776033Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:37.9776954Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:37.9777750Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:37.9778514Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:37.9779339Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:37.9780139Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:37.9780998Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:37.9781911Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:37.9782735Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:37.9783568Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:37.9784391Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:37.9785223Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:37.9786076Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:37.9786880Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:37.9787670Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:37.9788553Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:37.9789352Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:37.9790165Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:37.9791038Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:37.9791849Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:37.9792645Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:37.9793468Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:37.9794278Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:37.9795066Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:37.9795888Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:37.9796992Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:37.9798051Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:37.9799036Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:37.9800037Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:37.9801023Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:37.9802062Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:37.9803072Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:37.9804042Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:37.9805035Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:37.9805890Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:37.9806718Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:37.9807551Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:37.9808465Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:37.9809357Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:37.9810142Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:37.9810916Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:37.9811686Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:37.9812485Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:37.9813288Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:37.9814118Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:37.9814921Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:37.9815805Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:37.9816682Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:37.9817571Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:37.9818365Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:37.9819187Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:37.9819960Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:37.9820743Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:37.9821585Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:37.9822438Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:37.9823228Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:37.9824051Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:37.9824798Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:37.9825612Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:37.9826417Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:37.9827243Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:37.9828198Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:37.9828997Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:37.9829767Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:37.9830616Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:37.9831426Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:37.9832321Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:37.9833131Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:37.9833980Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:37.9834923Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:37.9835790Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:37.9836646Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:37.9837435Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:37.9838352Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:37.9839155Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:37.9839940Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:37.9840758Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:37.9841615Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:37.9842490Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:37.9843311Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:37.9844064Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:37.9844867Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:37.9845779Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:37.9846535Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:37.9847324Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:37.9848139Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:37.9848970Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:37.9849803Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:37.9850666Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.9851550Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.9852326Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:37.9853181Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:37.9854038Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:37.9854887Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:37.9855755Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:37.9856651Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:37.9857588Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:37.9858464Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:37.9859278Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:37.9860104Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:37.9860927Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:37.9861797Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:37.9862624Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:37.9863456Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:37.9864359Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:37.9865195Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:37.9866087Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:37.9866876Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:37.9867693Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:37.9868581Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:37.9869443Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:37.9870313Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:37.9871232Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:37.9877588Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:37.9878593Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:37.9879594Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:37.9880432Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:37.9881302Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:37.9882131Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:37.9883062Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:37.9884021Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:37.9884859Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:37.9885684Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:37.9886536Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:37.9887354Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:37.9888233Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:37.9889015Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:37.9889838Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:37.9890661Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:37.9891488Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:37.9892346Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:37.9893225Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:37.9894036Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:37.9894872Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:37.9895769Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:37.9896898Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:37.9897733Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:37.9898589Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:37.9899527Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:37.9900363Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:37.9901221Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:37.9902144Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:37.9902933Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:37.9903733Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:37.9904600Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:37.9905401Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:37.9906195Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:37.9907000Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:37.9907844Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:37.9908798Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:37.9909674Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:37.9910519Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:37.9911448Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:37.9912310Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:37.9913169Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:37.9913959Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:37.9914948Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:37.9915815Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:37.9916654Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:37.9917432Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:37.9918281Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:37.9919190Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:37.9920106Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:37.9921044Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:37.9921835Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:37.9922688Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:37.9923517Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:37.9924392Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:37.9925298Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:37.9926176Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:37.9927022Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:37.9927892Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:37.9928802Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:37.9929654Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:37.9930514Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:37.9931324Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:37.9932205Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:37.9933047Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:37.9933835Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:37.9934810Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:37.9935668Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:37.9936507Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:37.9937347Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:37.9938221Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:37.9939138Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:37.9940052Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:37.9940934Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:37.9941864Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:37.9942729Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:37.9943639Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:37.9944515Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:37.9945438Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:37.9946291Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:37.9947145Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:37.9947995Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:37.9948826Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:37.9949699Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:37.9950578Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:37.9951462Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:37.9952376Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:37.9953124Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:37.9953880Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:37.9954699Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:37.9955503Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:37.9956328Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:37.9957136Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:37.9957916Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:37.9958804Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:37.9959780Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:37.9960675Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:37.9961479Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:37.9962252Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:37.9963111Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:37.9963964Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:37.9964829Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:37.9965677Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:37.9966559Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:37.9967433Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:37.9968297Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:37.9969114Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:37.9969943Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:37.9970824Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:37.9971705Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:37.9972540Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:37.9973261Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:37.9974199Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:37.9975040Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:37.9975822Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:37.9976595Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:37.9977438Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:37.9978234Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:37.9979097Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:37.9979919Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:37.9980664Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:37.9981501Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:37.9982368Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:37.9983221Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:37.9984138Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:37.9984991Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:37.9985783Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:37.9986572Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:37.9987399Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:37.9988143Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:37.9988906Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:37.9989669Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:37.9990456Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:37.9991313Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:37.9992166Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:37.9992997Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:37.9993764Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:37.9994515Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:37.9995344Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:37.9997849Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:38.0000485Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:38.0001252Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:38.0002178Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:38.0002970Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:38.0003817Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:38.0004779Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:38.0005628Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:38.0006533Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:38.0007438Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:38.0008378Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:38.0009275Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:38.0010178Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:38.0011165Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:38.0012060Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:38.0012936Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:38.0013868Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:38.0014780Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:38.0015678Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:38.0016564Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:38.0017624Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:38.0018371Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:38.0019066Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:38.0019823Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:38.0020532Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:38.0021246Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:38.0021979Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:38.0022843Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:38.0023694Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:38.0024571Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:38.0025301Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:38.0025951Z V1204 11:24:04.558000 101771 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:38.0026619Z I1204 11:24:04.558000 101771 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:38.0027117Z V1204 11:24:04.558000 101771 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:38.0027633Z V1204 11:24:04.558000 101771 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:38.0028151Z I1204 11:24:04.559000 101771 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:38.0028922Z V1204 11:24:04.598000 101771 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:38.0029378Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:38.0029924Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:38.0030876Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:38.0031666Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:38.0032699Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:38.0033672Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:38.0034654Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:38.0035517Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:38.0036331Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:38.0037393Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:38.0038253Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:38.0039187Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:38.0040094Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:38.0040686Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:38.0041128Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:38.0041549Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:38.0042138Z V1204 11:24:04.606000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:38.0042705Z V1204 11:24:04.607000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:38.0043713Z V1204 11:24:04.607000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:38.0044264Z V1204 11:24:04.608000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0045281Z V1204 11:24:04.640000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:38.0045827Z V1204 11:24:04.640000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0046767Z V1204 11:24:04.642000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:38.0047237Z V1204 11:24:04.642000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0048180Z V1204 11:24:04.643000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:38.0048636Z V1204 11:24:04.644000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0049565Z V1204 11:24:04.645000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:38.0050001Z V1204 11:24:04.645000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0050869Z V1204 11:24:04.648000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:38.0051307Z V1204 11:24:04.648000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0052251Z V1204 11:24:04.651000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:38.0052704Z V1204 11:24:04.651000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0053590Z V1204 11:24:04.652000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:38.0054052Z V1204 11:24:04.652000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0054569Z V1204 11:24:04.653000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:38.0055166Z V1204 11:24:04.654000 101771 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:38.0055940Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:38.0056058Z warnings.warn( 2025-12-04T11:24:38.0056543Z V1204 11:24:04.654000 101771 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:38.0056769Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:38.0056946Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:38.0057067Z stats [('calls_captured', 1)] 2025-12-04T11:24:38.0057330Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:38.0058109Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:38.0058215Z graph_break [] 2025-12-04T11:24:38.0058450Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:38.0058928Z V1204 11:24:04.883000 101771 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpsdwdsb1j 2025-12-04T11:24:38.0059369Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:38.0059872Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:38.0060806Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:38.0061541Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:38.0062033Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:38.0062515Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:38.0062942Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:38.0063971Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:38.0064729Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:38.0065239Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:38.0065666Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:38.0066046Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:38.0066757Z V1204 11:24:05.166000 101771 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:38.0067483Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:38.0068147Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:38.0068469Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:38.0068746Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:38.0069018Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:38.0069499Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:38.0070314Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:38.0071168Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:38.0071796Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:38.0072586Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:38.0073343Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:38.0073795Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:38.0074108Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:38.0074684Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:38.0077035Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:38.0079334Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:38.0080096Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:38.0081010Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:38.0081751Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:38.0082560Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:38.0083391Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:38.0084117Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:38.0084828Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:38.0085564Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:38.0086301Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:38.0087045Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:38.0087750Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:38.0088432Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:38.0089324Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:38.0090118Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:38.0091332Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:38.0092122Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:38.0121249Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:38.0122399Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:38.0123206Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:38.0123992Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:38.0124876Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:38.0125731Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:38.0126449Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:38.0127234Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:38.0128015Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:38.0128794Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:38.0129660Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:38.0130493Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:38.0131362Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:38.0132237Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:38.0133101Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:38.0133994Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:38.0134850Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:38.0135681Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:38.0136725Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:38.0137549Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:38.0138390Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:38.0139179Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:38.0140014Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:38.0140804Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:38.0141524Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:38.0142313Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:38.0143076Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:38.0143827Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:38.0144678Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:38.0145432Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:38.0146212Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:38.0146993Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:38.0147758Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:38.0148585Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:38.0149365Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:38.0150135Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:38.0150932Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:38.0151710Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:38.0152523Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:38.0153287Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:38.0154102Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:38.0154896Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:38.0155671Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:38.0156512Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:38.0157312Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:38.0158146Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:38.0158933Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:38.0159802Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:38.0160574Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:38.0161343Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:38.0162095Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:38.0162895Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:38.0163878Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:38.0164673Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:38.0165475Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:38.0166303Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:38.0167060Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:38.0168276Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:38.0169079Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:38.0169939Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:38.0170857Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:38.0171689Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:38.0172537Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:38.0173400Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:38.0174278Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:38.0175092Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:38.0176031Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:38.0176931Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:38.0177902Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:38.0178707Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:38.0179658Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:38.0180510Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:38.0181392Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:38.0182150Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:38.0182898Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:38.0183748Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:38.0184626Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:38.0185399Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:38.0186209Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:38.0187028Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:38.0187870Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:38.0188768Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:38.0189699Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:38.0190553Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:38.0191335Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:38.0192165Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:38.0193053Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:38.0193860Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:38.0194666Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:38.0195552Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:38.0196637Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:38.0197527Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:38.0198391Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:38.0199200Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:38.0199995Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:38.0200796Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:38.0201578Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:38.0202379Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:38.0203222Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:38.0204056Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:38.0204943Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:38.0205759Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:38.0206579Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:38.0207316Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:38.0208117Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:38.0209049Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:38.0209975Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:38.0210918Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:38.0211766Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:38.0212602Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:38.0213439Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:38.0214266Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:38.0215182Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:38.0216035Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:38.0216913Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:38.0217727Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:38.0218569Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:38.0219436Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:38.0220228Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:38.0221066Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:38.0221900Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:38.0222698Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:38.0223475Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:38.0224302Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:38.0225124Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:38.0225973Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:38.0226834Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:38.0227604Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:38.0228524Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:38.0229314Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:38.0230190Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:38.0231003Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:38.0231802Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:38.0232587Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:38.0233365Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:38.0234155Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:38.0235003Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:38.0235857Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:38.0236716Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:38.0237548Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:38.0238395Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:38.0239154Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:38.0240062Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:38.0240903Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:38.0241726Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:38.0242522Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:38.0243366Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:38.0244165Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:38.0244929Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:38.0245711Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:38.0246471Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:38.0247227Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:38.0248034Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:38.0248898Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:38.0249736Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:38.0250577Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:38.0251381Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:38.0252149Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:38.0253008Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:38.0253805Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:38.0254660Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:38.0255538Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:38.0256350Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:38.0257262Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:38.0258055Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:38.0258890Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:38.0259754Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:38.0260557Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:38.0261337Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:38.0262217Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:38.0263011Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:38.0263826Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:38.0264731Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:38.0265541Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:38.0266307Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:38.0267149Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:38.0267959Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:38.0268772Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:38.0269592Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:38.0270561Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:38.0271571Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:38.0272560Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:38.0273521Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:38.0274501Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:38.0275502Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:38.0276537Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:38.0277527Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:38.0278519Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:38.0279370Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:38.0280230Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:38.0281055Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:38.0281969Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:38.0282859Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:38.0283636Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:38.0284417Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:38.0285212Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:38.0286008Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:38.0286803Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:38.0287590Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:38.0288389Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:38.0289280Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:38.0290101Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:38.0290926Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:38.0291710Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:38.0292533Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:38.0293301Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:38.0294083Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:38.0294923Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:38.0295760Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:38.0296801Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:38.0297675Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:38.0298421Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:38.0299271Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:38.0300079Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:38.0300948Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:38.0301895Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:38.0302702Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:38.0303436Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:38.0304294Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:38.0305102Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:38.0305970Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:38.0306778Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:38.0307582Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:38.0308578Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:38.0309445Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:38.0310297Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:38.0311079Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:38.0311996Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:38.0312831Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:38.0313619Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:38.0314436Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:38.0315288Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:38.0316164Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:38.0317014Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:38.0317765Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:38.0318574Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:38.0319392Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:38.0320148Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:38.0320936Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:38.0321726Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:38.0322551Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:38.0323353Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:38.0324251Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:38.0325131Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:38.0325902Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:38.0326751Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:38.0327603Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:38.0328488Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:38.0329353Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:38.0330242Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:38.0331116Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:38.0331982Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:38.0332810Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:38.0333642Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:38.0334465Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:38.0335292Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:38.0336113Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:38.0336995Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:38.0337838Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:38.0338652Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:38.0339518Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:38.0340353Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:38.0341176Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:38.0342071Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:38.0342944Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:38.0343810Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:38.0344785Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:38.0345683Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:38.0346599Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:38.0347545Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:38.0348385Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:38.0349356Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:38.0350185Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:38.0351126Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:38.0352001Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:38.0352851Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:38.0353690Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:38.0354513Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:38.0355339Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:38.0356222Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:38.0357028Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:38.0357847Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:38.0358676Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:38.0359511Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:38.0360370Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:38.0361297Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:38.0362108Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:38.0362929Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:38.0363854Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:38.0364774Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:38.0365626Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:38.0366487Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:38.0367368Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:38.0368205Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:38.0369071Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:38.0369951Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:38.0370745Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:38.0371548Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:38.0372414Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:38.0373222Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:38.0374014Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:38.0374834Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:38.0375688Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:38.0376686Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:38.0377618Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:38.0378460Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:38.0379385Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:38.0380253Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:38.0381153Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:38.0381954Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:38.0382949Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:38.0383783Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:38.0384622Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:38.0385407Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:38.0386231Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:38.0387139Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:38.0388030Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:38.0389017Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:38.0389814Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:38.0390653Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:38.0391485Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:38.0392359Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:38.0393312Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:38.0394192Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:38.0395043Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:38.0395907Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:38.0397010Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:38.0397933Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:38.0398808Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:38.0399612Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:38.0400462Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:38.0401313Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:38.0402109Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:38.0403062Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:38.0403920Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:38.0404793Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:38.0405587Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:38.0406468Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:38.0407401Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:38.0408314Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:38.0409245Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:38.0410177Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:38.0411043Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:38.0411944Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:38.0412826Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:38.0413792Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:38.0414650Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:38.0415502Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:38.0416323Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:38.0417239Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:38.0418121Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:38.0418987Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:38.0419871Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:38.0420761Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:38.0421564Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:38.0422326Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:38.0423153Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:38.0423956Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:38.0424784Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:38.0425634Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:38.0426422Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:38.0427313Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:38.0428288Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:38.0429186Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:38.0430023Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:38.0430800Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:38.0431658Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:38.0432484Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:38.0433360Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:38.0434210Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:38.0435063Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:38.0435938Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:38.0436806Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:38.0437625Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:38.0438454Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:38.0439338Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:38.0440217Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:38.0441064Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:38.0441822Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:38.0442761Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:38.0443570Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:38.0444377Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:38.0445158Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:38.0445969Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:38.0446773Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:38.0447635Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:38.0448424Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:38.0449176Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:38.0450015Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:38.0450858Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:38.0451708Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:38.0452593Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:38.0453482Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:38.0454285Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:38.0455076Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:38.0455908Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:38.0456659Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:38.0457514Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:38.0458285Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:38.0459071Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:38.0459935Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:38.0460791Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:38.0461677Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:38.0462448Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:38.0463206Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:38.0464012Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:38.0465977Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:38.0468397Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:38.0469202Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:38.0470127Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:38.0470924Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:38.0471764Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:38.0472723Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:38.0473613Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:38.0474515Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:38.0475430Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:38.0476343Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:38.0477243Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:38.0478196Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:38.0479190Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:38.0480083Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:38.0480912Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:38.0481852Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:38.0482768Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:38.0483648Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:38.0484534Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:38.0485537Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:38.0486291Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:38.0486991Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:38.0487747Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:38.0488457Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:38.0489172Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:38.0489950Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:38.0490825Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:38.0491683Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:38.0492565Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:38.0493301Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:38.0493982Z V1204 11:24:05.176000 101771 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:38.0494651Z I1204 11:24:05.176000 101771 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:38.0495146Z V1204 11:24:05.176000 101771 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:38.0495632Z V1204 11:24:05.177000 101771 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:38.0496336Z I1204 11:24:05.177000 101771 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:38.0497170Z V1204 11:24:05.189000 101771 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:38.0497643Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:38.0498142Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:38.0499098Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:38.0499854Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:38.0500971Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:38.0501944Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:38.0502928Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:38.0503784Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:38.0504653Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:38.0505705Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:38.0506572Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:38.0507521Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:38.0508473Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:38.0509064Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:38.0509502Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:38.0509896Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:38.0510486Z V1204 11:24:05.196000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:38.0511056Z V1204 11:24:05.197000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:38.0512074Z V1204 11:24:05.197000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:38.0512596Z V1204 11:24:05.198000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0513609Z V1204 11:24:05.199000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:38.0514159Z V1204 11:24:05.199000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0515105Z V1204 11:24:05.200000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:38.0515577Z V1204 11:24:05.201000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0516517Z V1204 11:24:05.202000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:38.0516985Z V1204 11:24:05.202000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0517902Z V1204 11:24:05.204000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:38.0518370Z V1204 11:24:05.204000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0519232Z V1204 11:24:05.206000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:38.0519666Z V1204 11:24:05.206000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0520621Z V1204 11:24:05.208000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:38.0521078Z V1204 11:24:05.208000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0521996Z V1204 11:24:05.209000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:38.0522461Z V1204 11:24:05.209000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0522978Z V1204 11:24:05.210000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:38.0523573Z V1204 11:24:05.210000 101771 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:38.0524315Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:38.0524439Z warnings.warn( 2025-12-04T11:24:38.0524919Z V1204 11:24:05.211000 101771 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:38.0525072Z =================================== FAILURES =================================== 2025-12-04T11:24:38.0525535Z _ TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True _ 2025-12-04T11:24:38.0525662Z Traceback (most recent call last): 2025-12-04T11:24:38.0526233Z File "/var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py", line 384, in test_decompose_mm_mixed_precision 2025-12-04T11:24:38.0526355Z self.assertEqual( 2025-12-04T11:24:38.0526816Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T11:24:38.0527029Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T11:24:38.0527563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T11:24:38.0527771Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T11:24:38.0527915Z AssertionError: Scalars are not equal! 2025-12-04T11:24:38.0527925Z 2025-12-04T11:24:38.0528029Z Expected 1 but got 0. 2025-12-04T11:24:38.0528136Z Absolute difference: 1 2025-12-04T11:24:38.0528258Z Relative difference: 1.0 2025-12-04T11:24:38.0528264Z 2025-12-04T11:24:38.0528477Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:38.0529339Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:38.0529345Z 2025-12-04T11:24:38.0529617Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:38.0529842Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:38.0529999Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:38.0530113Z stats [('calls_captured', 1)] 2025-12-04T11:24:38.0530884Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:38.0531106Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:38.0531204Z graph_break [] 2025-12-04T11:24:38.0531434Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:38.0531932Z I1204 11:24:01.606000 101771 site-packages/torch/_inductor/async_compile.py:258] [0/0] Creating 'subprocess' pool with 16 workers 2025-12-04T11:24:38.0532567Z I1204 11:24:01.855000 101771 site-packages/torch/_inductor/compile_worker/subproc_pool.py:170] [0/0] Suppressing compile worker output due to config 2025-12-04T11:24:38.0533006Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:38.0533500Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:38.0534487Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:38.0535216Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:38.0535722Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:38.0536205Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:38.0536627Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:38.0537736Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:38.0538484Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:38.0538972Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:38.0539391Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:38.0539827Z V1204 11:24:01.857000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:38.0540383Z I1204 11:24:04.067000 101771 site-packages/torch/_inductor/analysis/device_info.py:204] [0/0] Device Tesla T4 not in datasheet, returning None 2025-12-04T11:24:38.0541099Z V1204 11:24:04.547000 101771 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:38.0541825Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:38.0542488Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:38.0542784Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:38.0543092Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:38.0543367Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:38.0543848Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:38.0544664Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:38.0545494Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:38.0546153Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:38.0546938Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:38.0547721Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:38.0548170Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:38.0548477Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:38.0549054Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:38.0551409Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:38.0553731Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:38.0554422Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:38.0555334Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:38.0556071Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:38.0556886Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:38.0557665Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:38.0558395Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:38.0559142Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:38.0559878Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:38.0560594Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:38.0561369Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:38.0562050Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:38.0562765Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:38.0563648Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:38.0564439Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:38.0565633Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:38.0566414Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:38.0595390Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:38.0596697Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:38.0597463Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:38.0598248Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:38.0599154Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:38.0600013Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:38.0600772Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:38.0601558Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:38.0602341Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:38.0603118Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:38.0603995Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:38.0604827Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:38.0605701Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:38.0606550Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:38.0607454Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:38.0608354Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:38.0609209Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:38.0610032Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:38.0611034Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:38.0611845Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:38.0612688Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:38.0613445Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:38.0614277Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:38.0615129Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:38.0615852Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:38.0616669Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:38.0617506Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:38.0618260Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:38.0619108Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:38.0619877Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:38.0620654Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:38.0621435Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:38.0622194Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:38.0622988Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:38.0623806Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:38.0624579Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:38.0625374Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:38.0626147Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:38.0626926Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:38.0627732Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:38.0628556Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:38.0629325Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:38.0630129Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:38.0630944Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:38.0631746Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:38.0632616Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:38.0633399Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:38.0634278Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:38.0635052Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:38.0635827Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:38.0636576Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:38.0637378Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:38.0638315Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:38.0639136Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:38.0639978Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:38.0640810Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:38.0641565Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:38.0642747Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:38.0643589Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:38.0644458Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:38.0645342Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:38.0646198Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:38.0647014Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:38.0647906Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:38.0648782Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:38.0649591Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:38.0650533Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:38.0651357Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:38.0652335Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:38.0653136Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:38.0654091Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:38.0654909Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:38.0655849Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:38.0656611Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:38.0657434Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:38.0658269Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:38.0659118Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:38.0659928Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:38.0660743Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:38.0661539Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:38.0662403Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:38.0663278Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:38.0664230Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:38.0665091Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:38.0665880Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:38.0666706Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:38.0667600Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:38.0668403Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:38.0669195Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:38.0670090Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:38.0670947Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:38.0671861Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:38.0672728Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:38.0673540Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:38.0674343Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:38.0675083Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:38.0675904Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:38.0676703Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:38.0677512Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:38.0678364Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:38.0679217Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:38.0680070Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:38.0680888Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:38.0681623Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:38.0682421Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:38.0683354Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:38.0684278Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:38.0685220Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:38.0686031Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:38.0686872Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:38.0687742Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:38.0688585Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:38.0689491Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:38.0690350Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:38.0691146Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:38.0691990Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:38.0692825Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:38.0693656Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:38.0694470Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:38.0695258Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:38.0696328Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:38.0697189Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:38.0697969Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:38.0698790Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:38.0699606Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:38.0700443Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:38.0701297Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:38.0702068Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:38.0702952Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:38.0703778Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:38.0704661Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:38.0705474Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:38.0706269Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:38.0707027Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:38.0707848Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:38.0708642Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:38.0709470Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:38.0710367Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:38.0711199Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:38.0712077Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:38.0712935Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:38.0713694Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:38.0714607Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:38.0715449Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:38.0716278Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:38.0717069Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:38.0717907Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:38.0718675Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:38.0719474Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:38.0720261Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:38.0721027Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:38.0721787Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:38.0722565Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:38.0723431Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:38.0724308Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:38.0725137Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:38.0725926Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:38.0726711Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:38.0727545Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:38.0728374Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:38.0729237Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:38.0730107Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:38.0730934Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:38.0731775Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:38.0732568Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:38.0733396Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:38.0734225Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:38.0735052Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:38.0735838Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:38.0736714Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:38.0737567Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:38.0738379Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:38.0739255Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:38.0740103Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:38.0740867Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:38.0741683Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:38.0742515Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:38.0743309Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:38.0744161Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:38.0745131Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:38.0746131Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:38.0747111Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:38.0748080Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:38.0749063Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:38.0750063Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:38.0751073Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:38.0752076Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:38.0753063Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:38.0753908Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:38.0754735Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:38.0755563Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:38.0756505Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:38.0757355Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:38.0758134Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:38.0758941Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:38.0759718Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:38.0760548Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:38.0761344Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:38.0762139Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:38.0762941Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:38.0763830Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:38.0764645Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:38.0765470Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:38.0766224Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:38.0767047Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:38.0767876Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:38.0768655Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:38.0769491Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:38.0770331Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:38.0771127Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:38.0771978Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:38.0772717Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:38.0773503Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:38.0774350Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:38.0775184Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:38.0776165Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:38.0777026Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:38.0777762Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:38.0778617Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:38.0779433Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:38.0780301Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:38.0781112Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:38.0781928Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:38.0782872Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:38.0783782Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:38.0784633Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:38.0785417Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:38.0786332Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:38.0787128Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:38.0787946Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:38.0788765Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:38.0789585Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:38.0790487Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:38.0791312Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:38.0792089Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:38.0792900Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:38.0793707Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:38.0794459Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:38.0795244Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:38.0796200Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:38.0797116Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:38.0797919Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:38.0798788Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:38.0799733Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:38.0800508Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:38.0801363Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:38.0802214Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:38.0803078Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:38.0803986Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:38.0804878Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:38.0805730Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:38.0806638Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:38.0807450Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:38.0808314Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:38.0809136Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:38.0809964Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:38.0810789Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:38.0811620Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:38.0812466Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:38.0813298Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:38.0814149Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:38.0814946Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:38.0815799Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:38.0816691Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:38.0817616Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:38.0818485Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:38.0819409Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:38.0820350Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:38.0821269Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:38.0822188Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:38.0823057Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:38.0823934Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:38.0824801Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:38.0825740Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:38.0826626Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:38.0827480Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:38.0828300Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:38.0829124Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:38.0829939Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:38.0830788Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:38.0831612Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:38.0832430Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:38.0833264Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:38.0834083Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:38.0834948Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:38.0835840Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:38.0836703Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:38.0837512Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:38.0838405Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:38.0839340Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:38.0840170Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:38.0841048Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:38.0841931Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:38.0842768Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:38.0843630Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:38.0844516Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:38.0845313Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:38.0846112Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:38.0846947Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:38.0847777Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:38.0848581Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:38.0849397Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:38.0850241Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:38.0851201Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:38.0852112Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:38.0852976Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:38.0853855Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:38.0854752Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:38.0855625Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:38.0856456Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:38.0857504Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:38.0858345Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:38.0859191Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:38.0859973Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:38.0860806Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:38.0861722Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:38.0862611Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:38.0863563Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:38.0864399Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:38.0865243Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:38.0866072Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:38.0866947Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:38.0867869Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:38.0868846Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:38.0869702Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:38.0870547Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:38.0871493Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:38.0872353Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:38.0873249Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:38.0874052Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:38.0874902Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:38.0875751Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:38.0876554Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:38.0877502Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:38.0878356Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:38.0879172Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:38.0879987Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:38.0880868Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:38.0881801Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:38.0882708Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:38.0883600Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:38.0884568Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:38.0885445Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:38.0886314Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:38.0887228Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:38.0888150Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:38.0889039Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:38.0889894Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:38.0890709Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:38.0891544Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:38.0892423Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:38.0893291Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:38.0894170Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:38.0895054Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:38.0895810Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:38.0897013Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:38.0897851Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:38.0898651Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:38.0899483Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:38.0900297Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:38.0901148Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:38.0902034Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:38.0902985Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:38.0903928Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:38.0904736Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:38.0905551Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:38.0906424Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:38.0907254Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:38.0908124Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:38.0908982Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:38.0909841Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:38.0910708Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:38.0911540Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:38.0912353Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:38.0913242Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:38.0914117Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:38.0914997Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:38.0915838Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:38.0916566Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:38.0917546Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:38.0918357Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:38.0919137Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:38.0919955Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:38.0920736Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:38.0921572Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:38.0922443Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:38.0923229Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:38.0923980Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:38.0924825Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:38.0925668Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:38.0926519Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:38.0927413Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:38.0928261Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:38.0929096Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:38.0929894Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:38.0930717Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:38.0931467Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:38.0932232Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:38.0933044Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:38.0933834Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:38.0934662Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:38.0935547Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:38.0936387Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:38.0937257Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:38.0938022Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:38.0938833Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:38.0940792Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:38.0943234Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:38.0944008Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:38.0944950Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:38.0945746Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:38.0946583Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:38.0947536Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:38.0948395Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:38.0949329Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:38.0950249Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:38.0951129Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:38.0952058Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:38.0952970Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:38.0953997Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:38.0954888Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:38.0955714Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:38.0956656Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:38.0957578Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:38.0958473Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:38.0959349Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:38.0960320Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:38.0961104Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:38.0961812Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:38.0962565Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:38.0963275Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:38.0964007Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:38.0964734Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:38.0965643Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:38.0966495Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:38.0967349Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:38.0968113Z V1204 11:24:04.557000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:38.0968773Z V1204 11:24:04.558000 101771 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:38.0969486Z I1204 11:24:04.558000 101771 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:38.0969974Z V1204 11:24:04.558000 101771 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:38.0970458Z V1204 11:24:04.558000 101771 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:38.0970976Z I1204 11:24:04.559000 101771 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 0 2025-12-04T11:24:38.0971747Z V1204 11:24:04.598000 101771 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:38.0972212Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:38.0972712Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:38.0973666Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:38.0974421Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:38.0975457Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:38.0976469Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:38.0977506Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:38.0978362Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:38.0979172Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:38.0980260Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:38.0981088Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:38.0982055Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:38.0982950Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:38.0983571Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:38.0984013Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:38.0984405Z V1204 11:24:04.603000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:38.0984998Z V1204 11:24:04.606000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:38.0985561Z V1204 11:24:04.607000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:38.0986593Z V1204 11:24:04.607000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:38.0987097Z V1204 11:24:04.608000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0988105Z V1204 11:24:04.640000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:38.0988621Z V1204 11:24:04.640000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0989595Z V1204 11:24:04.642000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:38.0990073Z V1204 11:24:04.642000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0991006Z V1204 11:24:04.643000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:38.0991467Z V1204 11:24:04.644000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0992374Z V1204 11:24:04.645000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:38.0992806Z V1204 11:24:04.645000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0993695Z V1204 11:24:04.648000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:38.0994131Z V1204 11:24:04.648000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0995031Z V1204 11:24:04.651000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:38.0995505Z V1204 11:24:04.651000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0996571Z V1204 11:24:04.652000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:38.0997077Z V1204 11:24:04.652000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.0997592Z V1204 11:24:04.653000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:38.0998187Z V1204 11:24:04.654000 101771 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 0 2025-12-04T11:24:38.0998926Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:38.0999045Z warnings.warn( 2025-12-04T11:24:38.0999524Z V1204 11:24:04.654000 101771 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:38.0999757Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:38.0999883Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:38.0999999Z stats [('calls_captured', 1)] 2025-12-04T11:24:38.1000225Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:38.1000994Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:38.1001093Z graph_break [] 2025-12-04T11:24:38.1001323Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:38.1001762Z V1204 11:24:04.883000 101771 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpsdwdsb1j 2025-12-04T11:24:38.1002206Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:38.1002751Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:38.1003692Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:38.1004429Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:38.1004920Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:38.1005421Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:38.1005842Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:38.1006909Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:38.1007663Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:38.1008137Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:38.1008619Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:38.1009007Z V1204 11:24:04.906000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:38.1009723Z V1204 11:24:05.166000 101771 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:38.1010476Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:38.1011138Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:38.1011430Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:38.1011706Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:38.1011982Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:38.1012464Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:38.1013281Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:38.1014109Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:38.1014731Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:38.1015509Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:38.1016287Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:38.1016741Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:38.1017127Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:38.1017702Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:38.1020054Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:38.1022402Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:38.1023091Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:38.1024003Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:38.1024779Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:38.1025588Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:38.1026382Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:38.1027091Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:38.1027808Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:38.1028547Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:38.1029263Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:38.1030004Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:38.1030682Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:38.1031365Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:38.1032282Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:38.1033070Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:38.1034263Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:38.1035045Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:38.1064070Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:38.1065126Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:38.1065896Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:38.1066676Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:38.1067516Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:38.1068369Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:38.1069116Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:38.1069907Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:38.1070687Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:38.1071464Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:38.1072331Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:38.1073172Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:38.1074075Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:38.1074920Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:38.1075782Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:38.1076708Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:38.1077595Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:38.1078423Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:38.1079423Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:38.1080192Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:38.1081031Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:38.1081793Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:38.1082629Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:38.1083400Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:38.1084106Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:38.1084927Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:38.1085688Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:38.1086437Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:38.1087277Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:38.1088037Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:38.1088828Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:38.1089653Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:38.1090418Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:38.1091207Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:38.1092009Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:38.1092787Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:38.1093610Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:38.1094406Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:38.1095167Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:38.1095932Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:38.1096965Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:38.1097739Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:38.1098518Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:38.1099326Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:38.1100140Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:38.1101025Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:38.1101817Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:38.1102703Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:38.1103472Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:38.1104241Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:38.1105030Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:38.1105850Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:38.1106770Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:38.1107594Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:38.1108399Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:38.1109230Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:38.1110037Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:38.1111228Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:38.1112034Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:38.1112898Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:38.1113785Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:38.1114615Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:38.1115422Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:38.1116280Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:38.1117189Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:38.1118004Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:38.1118938Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:38.1119762Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:38.1120753Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:38.1121590Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:38.1122535Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:38.1123347Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:38.1124256Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:38.1125020Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:38.1125807Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:38.1126649Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:38.1127501Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:38.1128278Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:38.1129096Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:38.1129895Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:38.1130731Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:38.1131606Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:38.1132530Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:38.1133487Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:38.1134275Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:38.1135105Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:38.1135999Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:38.1136805Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:38.1137696Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:38.1138584Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:38.1139447Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:38.1140364Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:38.1141240Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:38.1142553Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:38.1143359Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:38.1144093Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:38.1144880Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:38.1145691Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:38.1146505Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:38.1147335Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:38.1148182Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:38.1148998Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:38.1149870Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:38.1150612Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:38.1151424Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:38.1152337Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:38.1153264Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:38.1154242Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:38.1155054Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:38.1155888Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:38.1156756Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:38.1157598Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:38.1158523Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:38.1159378Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:38.1160170Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:38.1160980Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:38.1161824Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:38.1162658Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:38.1163436Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:38.1164223Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:38.1165052Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:38.1165877Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:38.1166650Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:38.1167459Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:38.1168276Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:38.1169124Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:38.1170020Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:38.1170793Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:38.1171688Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:38.1172499Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:38.1173376Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:38.1174213Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:38.1175011Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:38.1175767Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:38.1176549Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:38.1177422Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:38.1178254Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:38.1179112Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:38.1179948Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:38.1180783Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:38.1181685Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:38.1182454Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:38.1183352Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:38.1184193Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:38.1185023Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:38.1185852Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:38.1186690Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:38.1187454Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:38.1188256Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:38.1189030Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:38.1189818Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:38.1190576Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:38.1191351Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:38.1192216Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:38.1193061Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:38.1193893Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:38.1194682Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:38.1195443Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:38.1196418Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:38.1197282Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:38.1198138Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:38.1199008Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:38.1199832Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:38.1200668Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:38.1201504Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:38.1202340Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:38.1203166Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:38.1203972Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:38.1204790Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:38.1205675Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:38.1206526Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:38.1207340Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:38.1208218Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:38.1209029Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:38.1209804Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:38.1210624Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:38.1211424Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:38.1212215Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:38.1213067Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:38.1214035Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:38.1215034Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:38.1216019Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:38.1217064Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:38.1218094Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:38.1219088Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:38.1220096Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:38.1221098Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:38.1222098Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:38.1222976Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:38.1223802Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:38.1224635Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:38.1225550Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:38.1226404Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:38.1227187Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:38.1227978Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:38.1228737Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:38.1229570Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:38.1230372Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:38.1231160Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:38.1231962Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:38.1232846Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:38.1233677Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:38.1234525Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:38.1235283Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:38.1236105Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:38.1236910Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:38.1237691Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:38.1238563Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:38.1239415Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:38.1240205Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:38.1241028Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:38.1241774Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:38.1242564Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:38.1243371Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:38.1244195Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:38.1245152Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:38.1245982Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:38.1246719Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:38.1247574Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:38.1248390Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:38.1249254Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:38.1250096Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:38.1250916Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:38.1251856Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:38.1252763Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:38.1253618Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:38.1254431Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:38.1255356Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:38.1256148Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:38.1256988Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:38.1257814Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:38.1258642Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:38.1259509Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:38.1260326Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:38.1267319Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:38.1268321Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:38.1269139Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:38.1269896Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:38.1270689Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:38.1271467Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:38.1272342Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:38.1273132Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:38.1273998Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:38.1274935Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:38.1275711Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:38.1276598Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:38.1277442Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:38.1278291Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:38.1279149Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:38.1280041Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:38.1280883Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:38.1281741Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:38.1282534Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:38.1283370Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:38.1284222Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:38.1285045Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:38.1285862Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:38.1286682Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:38.1287516Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:38.1288368Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:38.1289214Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:38.1290000Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:38.1290840Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:38.1291736Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:38.1292628Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:38.1293492Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:38.1294408Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:38.1295304Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:38.1296388Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:38.1297372Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:38.1298207Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:38.1299064Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:38.1299962Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:38.1300897Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:38.1301769Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:38.1302610Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:38.1303432Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:38.1304252Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:38.1305112Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:38.1305958Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:38.1306746Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:38.1307601Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:38.1308426Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:38.1309288Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:38.1310149Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:38.1311035Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:38.1311845Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:38.1312653Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:38.1313546Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:38.1314457Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:38.1315276Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:38.1316159Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:38.1317035Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:38.1317870Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:38.1318721Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:38.1319600Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:38.1320423Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:38.1321225Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:38.1322063Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:38.1322890Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:38.1323683Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:38.1324493Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:38.1325365Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:38.1326318Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:38.1327196Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:38.1328034Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:38.1328916Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:38.1329775Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:38.1330632Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:38.1331433Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:38.1332452Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:38.1333281Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:38.1334115Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:38.1334893Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:38.1335717Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:38.1336674Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:38.1337614Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:38.1338559Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:38.1339391Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:38.1340244Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:38.1341106Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:38.1341992Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:38.1342906Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:38.1343784Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:38.1344628Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:38.1345462Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:38.1346378Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:38.1347237Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:38.1348127Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:38.1348936Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:38.1349782Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:38.1350615Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:38.1351410Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:38.1352357Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:38.1353245Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:38.1354050Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:38.1354834Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:38.1355740Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:38.1356702Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:38.1357606Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:38.1358495Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:38.1359427Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:38.1360296Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:38.1361156Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:38.1362031Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:38.1362953Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:38.1363807Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:38.1364687Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:38.1365504Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:38.1366329Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:38.1367200Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:38.1368061Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:38.1368964Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:38.1369852Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:38.1370601Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:38.1371379Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:38.1372195Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:38.1373026Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:38.1373845Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:38.1374651Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:38.1375438Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:38.1376326Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:38.1377333Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:38.1378229Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:38.1379028Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:38.1379799Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:38.1380691Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:38.1381517Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:38.1382385Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:38.1383236Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:38.1384087Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:38.1384991Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:38.1385827Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:38.1386641Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:38.1387500Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:38.1388381Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:38.1389351Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:38.1390188Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:38.1390906Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:38.1391855Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:38.1392669Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:38.1393442Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:38.1394205Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:38.1394984Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:38.1395781Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:38.1396853Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:38.1397637Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:38.1398372Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:38.1399211Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:38.1400047Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:38.1400956Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:38.1401837Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:38.1402686Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:38.1403528Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:38.1404317Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:38.1405186Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:38.1405932Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:38.1406693Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:38.1407457Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:38.1408248Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:38.1409072Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:38.1409928Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:38.1410761Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:38.1411534Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:38.1412337Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:38.1413146Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:38.1415118Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:38.1417619Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:38.1418418Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:38.1419370Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:38.1420170Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:38.1421046Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:38.1422009Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:38.1422869Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:38.1423780Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:38.1424700Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:38.1425588Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:38.1426488Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:38.1427408Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:38.1428394Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:38.1429323Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:38.1430155Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:38.1431099Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:38.1432021Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:38.1432938Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:38.1433824Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:38.1434796Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:38.1435583Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:38.1436287Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:38.1437047Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:38.1437785Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:38.1438507Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:38.1439242Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:38.1440119Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:38.1440975Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:38.1441823Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:38.1442557Z V1204 11:24:05.175000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:38.1443213Z V1204 11:24:05.176000 101771 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:38.1443881Z I1204 11:24:05.176000 101771 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:38.1444418Z V1204 11:24:05.176000 101771 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:38.1444904Z V1204 11:24:05.177000 101771 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:38.1445419Z I1204 11:24:05.177000 101771 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 1 2025-12-04T11:24:38.1446183Z V1204 11:24:05.189000 101771 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:38.1446645Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:38.1447145Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:38.1448134Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:38.1448894Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:38.1449925Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:38.1450931Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:38.1451961Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:38.1452813Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:38.1453623Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:38.1454679Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:38.1455513Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:38.1456456Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:38.1457405Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:38.1458035Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:38.1458472Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:38.1458863Z V1204 11:24:05.194000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:38.1459446Z V1204 11:24:05.196000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:38.1460008Z V1204 11:24:05.197000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:38.1461014Z V1204 11:24:05.197000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:38.1461561Z V1204 11:24:05.198000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.1462572Z V1204 11:24:05.199000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:38.1463084Z V1204 11:24:05.199000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.1464055Z V1204 11:24:05.200000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:38.1464522Z V1204 11:24:05.201000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.1465465Z V1204 11:24:05.202000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:38.1465944Z V1204 11:24:05.202000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.1466864Z V1204 11:24:05.204000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:38.1467295Z V1204 11:24:05.204000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.1468166Z V1204 11:24:05.206000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:38.1468603Z V1204 11:24:05.206000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.1469503Z V1204 11:24:05.208000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:38.1469955Z V1204 11:24:05.208000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.1470846Z V1204 11:24:05.209000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:38.1471309Z V1204 11:24:05.209000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.1471858Z V1204 11:24:05.210000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:38.1472450Z V1204 11:24:05.210000 101771 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 1 2025-12-04T11:24:38.1473184Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:38.1473299Z warnings.warn( 2025-12-04T11:24:38.1473785Z V1204 11:24:05.211000 101771 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:38.1474008Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:24:38.1474137Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:24:38.1474257Z stats [('calls_captured', 1)] 2025-12-04T11:24:38.1474481Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:24:38.1475284Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 6), ('pad_mm_bench', 3), ('fxgraph_cache_miss', 1), ('decompose_mm', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1)] 2025-12-04T11:24:38.1475385Z graph_break [] 2025-12-04T11:24:38.1475616Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:24:38.1476055Z V1204 11:24:05.228000 101771 site-packages/torch/_inductor/utils.py:1327] Using inductor cache dir /tmp/tmpmbib7otv 2025-12-04T11:24:38.1476492Z V1204 11:24:05.249000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] TRACED GRAPH 2025-12-04T11:24:38.1477024Z V1204 11:24:05.249000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] ===== BEFORE PRE GRAD ===== 2025-12-04T11:24:38.1477964Z V1204 11:24:05.249000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:38.1478731Z V1204 11:24:05.249000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] def forward(self, L_input1_: "f32[20480, 5][5, 1]cuda:0", L_input2_: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:38.1479219Z V1204 11:24:05.249000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input1_ = L_input1_ 2025-12-04T11:24:38.1479703Z V1204 11:24:05.249000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] l_input2_ = L_input2_ 2025-12-04T11:24:38.1480139Z V1204 11:24:05.249000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:38.1481157Z V1204 11:24:05.249000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:38.1481915Z V1204 11:24:05.249000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] output: "bf16[20480, 2][2, 1]cuda:0" = torch.mm(l_input1_, l_input2_); l_input1_ = l_input2_ = None 2025-12-04T11:24:38.1482388Z V1204 11:24:05.249000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] return (output,) 2025-12-04T11:24:38.1482819Z V1204 11:24:05.249000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:38.1483201Z V1204 11:24:05.249000 101771 site-packages/torch/_inductor/compile_fx.py:2414] [0/0] [__pre_grad_graphs] 2025-12-04T11:24:38.1483920Z V1204 11:24:05.495000 101771 site-packages/torch/_inductor/compile_fx.py:895] [0/0] FX cache status: use_cache=True, local=True, remote=False, aot_mode=False, force_disable_caches=False 2025-12-04T11:24:38.1484673Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] FX graph cache hash details for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil: 2025-12-04T11:24:38.1485340Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [qwr4y4uncg2rdkw3pwgybtseexjtvdkkzepjntdi4z75ajv4ng3] gm: GraphModule() 2025-12-04T11:24:38.1485634Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:38.1485912Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:38.1486187Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:38.1486669Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] def forward(self, primals_1, primals_2): 2025-12-04T11:24:38.1487486Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:38.1488341Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] convert_element_type_1 = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:38.1488963Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] mm = torch.ops.aten.mm.default(convert_element_type_1, convert_element_type) 2025-12-04T11:24:38.1489730Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] permute = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:38.1490518Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] permute_1 = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:38.1490972Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] return (mm, permute, permute_1) 2025-12-04T11:24:38.1491286Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] 2025-12-04T11:24:38.1491888Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] # To see more debug info, please use `graph_module.print_readable()` 2025-12-04T11:24:38.1494239Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [3muebtmfttazr34eyvlphxyztuno2ozpf4ltpkase62ntmm45fd] example_inputs[0]: TensorMetadata(dtype=torch.float32, shape=torch.Size([20480, 5]), stride=(5, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:38.1496744Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ndtydxqtw6jljyvq7zs4uzqjkqrpxbz2ra5mscprj4ckj6j4aab] example_inputs[1]: TensorMetadata(dtype=torch.float32, shape=torch.Size([5, 2]), stride=(2, 1), device=device(type='cuda', index=0), layout=torch.strided, memory_format=torch.contiguous_format, storage_offset=0, storage_bytes=None, requires_grad=False, is_quantized=False, is_conj=False, is_neg=False, is_inference=False, is_sparse=False, is_coalesced=None, dense_dim=None, sparse_dim=None) 2025-12-04T11:24:38.1497466Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] cache_key_tag: 2025-12-04T11:24:38.1498384Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [lmglpn4zi7vob56n34r2j2rk7flv5xfgrcvmo7xcpirqsitygqx] fx_kwargs[boxed_forward_device_index]: BoxedDeviceIndex(value=None) 2025-12-04T11:24:38.1499124Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[cpp_wrapper]: False 2025-12-04T11:24:38.1500001Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [xq2hdkbfkbcuye6rgtypayrkhqf4cntij2dsd24rei3lsknakkf] fx_kwargs[cudagraphs]: BoxedBool(value=False) 2025-12-04T11:24:38.1500791Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[extern_node_serializer]: None 2025-12-04T11:24:38.1501520Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[fx_wrapper]: False 2025-12-04T11:24:38.1502235Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_backward]: False 2025-12-04T11:24:38.1502974Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] fx_kwargs[is_inference]: False 2025-12-04T11:24:38.1503689Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] fx_kwargs[layout_opt]: None 2025-12-04T11:24:38.1504478Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] fx_kwargs[static_input_idxs]: [] 2025-12-04T11:24:38.1505154Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inputs_to_check[0]: 0 2025-12-04T11:24:38.1505834Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inputs_to_check[1]: 1 2025-12-04T11:24:38.1506766Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [du4vyrfyozrfxcf6kk6ma7oqwatapifazeelfsawmsiu6gjdtxp] deterministic_algorithms_settings: (False, False, True) 2025-12-04T11:24:38.1507557Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [pyawus3dzq5k52f53obyevhjmttghvob2hr5d7g4uml5s7av6wb] cuda_matmul_settings: ('none', True, True) 2025-12-04T11:24:38.1508784Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [xa67e3kcqsohektjugueuqeaic4q6gpoqw3n2opqvsd3gou2ljr] torch_version: ��.*]cX�FC`5�{�3��ϛ�D���63�i� 2025-12-04T11:24:38.1509566Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [nciv4tx3omifkqskf67bgetzk22hsmyo34vd4e4cvtcgtm6ymct] system_info[device]: {'name': 'Tesla T4'} 2025-12-04T11:24:38.1538523Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [rpvxtyafu2a7z4z3r4ecmheibfouh6ibhsc7wxcnnwlmmbqtlof] system_info[version]: {'triton': '3.5.10355c3e3452fa4500122244b8fff8864f22bc05c417a3d6f5dada9f2e92f8040-c4d799c32eb6a3b92ad36ad7ea5645efc5b2fc6cd5c73dadc3a678d57dd2ccf4-c37149275a03d063fc1c339cd76a19da1c17e3d555410372e6cad29eedef6202-23d635e690d670bf61798e1259674b78c0ed5ba222ab6a455f329f27a758fc2d-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-03bffcc16c9d6be5ea8fde645caf3a6e3c0ac892eec49051a79d4cc99b66b9c7-e68505098eef3e7c0b050cb91da2587ab6c10d455ca0b941ff06fa8068e16305-318dbf7101b6ea9ebccfc57046fd8d963fe1d837c487005b37edf471a3207a9d-25cb0bee9547488335de2d495af738298ba6d4c20f1d37941dc17751c57a211e-e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855-3e3d33fbab70c7c05fc70e2e2fdd763dca0b847f62150592b99f8886a419ee64-83ef58f2371da7ad8e01822c2afc82f9a2d6516c2249ee0bbd8873bb20616be0-2e11b1d935d007c8ff5f1fb831cbdb6ece15fe48601bdbf3bc6c0f0b844702da-5c1281b67c0d949da34ccf1c3b68804a2caa2665953984d837d753e91af611fe-5d15c5bebef8d7aa51b21fd187e5faa95eba4a213254355bc69e0648013599f7-30106ed84518c6ca7aca08e2c0ee188755f512cc0cb2d7da8914cc48c1ad6dcc-adb54e71d0ffc3bdac437fbc97929769fbbe4ebd03e6361c6357f2a24f7c5954-27b2a5d1e8db008bacefe6019f63922bbd65926de90bb1b527ee597477d2f365-a610dc5c215589aab7a784e1c07acef3e16d53ef00f08de793899964956f4e2a-18572e33e474a820799036f2b2f8c3e54d8a526386356716cccf2bf32a832376-2cdca74c4297804dcf499a7e9d4315ab87edfe2d72f536a8fdc02f28a3e7dacd-b53abe93473eb37d88bc378692065c9a8b1bf54b6417cb1911a13d10918c6d20-f60c2bb2d8eebe1c191f4b8b819844414dd1bce243645635a094f9f92665a58e-08abee21ce6230a873ed0831f70f9570b7ce39969dbf9b2f28ae1a1992ee1cc7-8e4b8599f819f32bcabae6fd118dbbccfbec0ba9e1909224d39c5fe32fbb491f-3db4bee9427c7eb0e2105aff484bdacc819357d298e8f6e89c372ae9c3625bdf-59cf295f3aab4fa62b96a627aa9fec1302950133750de59e542c7b4c9e5b80b6-5305890c3b133def44e2f3d3405e0fb1fd6ce78d0a28b2127670a195bbe11c66', 'cuda': '12.4'} 2025-12-04T11:24:38.1539590Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [e56pno3aqdzuhra63t3yayh647zlzfisr7kd6dumnro23eazxas] system_info[hash]: 01b097f2b927d926ac9e10e9ea01881a3b463d32d7616a3658a01987cd0cc971 2025-12-04T11:24:38.1540359Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[TYPE_CHECKING]: False 2025-12-04T11:24:38.1541175Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_padding]: True 2025-12-04T11:24:38.1542019Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[can_inplace_pad_graph_input]: False 2025-12-04T11:24:38.1542871Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_auto_functionalized_v2]: True 2025-12-04T11:24:38.1543618Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug]: False 2025-12-04T11:24:38.1544411Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_progress]: True 2025-12-04T11:24:38.1545226Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[verbose_progress]: False 2025-12-04T11:24:38.1546000Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[worker_log_path]: None 2025-12-04T11:24:38.1546860Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [mxibia26nanvqq4lqvdfub66benrqh5fqtsyzzj2qnwy7srv2s3] inductor_config[precompilation_timeout_seconds]: 3600 2025-12-04T11:24:38.1547690Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[remote_gemm_autotune_cache]: False 2025-12-04T11:24:38.1548562Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bundle_triton_into_fx_graph_cache]: True 2025-12-04T11:24:38.1549413Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[non_blocking_remote_cache_write]: True 2025-12-04T11:24:38.1550267Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bundled_autotune_remote_cache]: None 2025-12-04T11:24:38.1551161Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_skip_cache_dynamic_shape_guards]: False 2025-12-04T11:24:38.1552016Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[unsafe_marked_cacheable_functions]: {} 2025-12-04T11:24:38.1552871Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[sleep_sec_TESTING_ONLY]: None 2025-12-04T11:24:38.1553879Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [pikr7bbcoixfzftsazp5ggufhdklj24babfry77bl4nuvyrrcp4] inductor_config[triton_kernel_default_layout_constraint]: needs_fixed_stride_order 2025-12-04T11:24:38.1554648Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper]: False 2025-12-04T11:24:38.1555489Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp_wrapper_build_separate]: False 2025-12-04T11:24:38.1556244Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fx_wrapper]: False 2025-12-04T11:24:38.1557112Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp_cache_precompile_headers]: True 2025-12-04T11:24:38.1557874Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[online_softmax]: True 2025-12-04T11:24:38.1558591Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[dce]: False 2025-12-04T11:24:38.1559411Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[static_weight_shapes]: True 2025-12-04T11:24:38.1560179Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[size_asserts]: True 2025-12-04T11:24:38.1560933Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[nan_asserts]: False 2025-12-04T11:24:38.1561810Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_triton_nan_asserts]: False 2025-12-04T11:24:38.1562570Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[scalar_asserts]: True 2025-12-04T11:24:38.1563347Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[alignment_asserts]: True 2025-12-04T11:24:38.1564132Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pick_loop_orders]: True 2025-12-04T11:24:38.1564905Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[inplace_buffers]: True 2025-12-04T11:24:38.1565701Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[allow_buffer_reuse]: True 2025-12-04T11:24:38.1566475Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[memory_planning]: False 2025-12-04T11:24:38.1567255Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[use_fast_math]: False 2025-12-04T11:24:38.1568050Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [x75won4jmsgeb63pcvwr2y4eteyzzdhmf5rv6xhjppie4hx2yu5] inductor_config[memory_pool]: intermediates 2025-12-04T11:24:38.1568861Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_harness]: True 2025-12-04T11:24:38.1569640Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[epilogue_fusion]: True 2025-12-04T11:24:38.1570405Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[prologue_fusion]: True 2025-12-04T11:24:38.1571224Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[epilogue_fusion_first]: False 2025-12-04T11:24:38.1571993Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pattern_matcher]: True 2025-12-04T11:24:38.1572802Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[b2b_gemm_pass]: False 2025-12-04T11:24:38.1573612Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[custom_partitioner_fn]: None 2025-12-04T11:24:38.1574415Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[pre_grad_custom_pass]: None 2025-12-04T11:24:38.1575277Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[_post_fusion_custom_pass]: None 2025-12-04T11:24:38.1576067Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_cat_fx_passes]: True 2025-12-04T11:24:38.1577040Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[efficient_conv_bn_eval_fx_passes]: False 2025-12-04T11:24:38.1577812Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[is_predispatch]: False 2025-12-04T11:24:38.1578582Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[group_fusion]: False 2025-12-04T11:24:38.1579332Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[batch_fusion]: True 2025-12-04T11:24:38.1580139Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[pre_grad_fusion_options]: {} 2025-12-04T11:24:38.1581084Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tmb245gnmtdy2k7ei7jigz6u7ykbzvh5zmz6bhvkp5hl4w7evvo] inductor_config[post_grad_fusion_options]: {'decompose_mm_pass': {}} 2025-12-04T11:24:38.1581878Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_locality]: True 2025-12-04T11:24:38.1582692Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[dynamic_scale_rblock]: True 2025-12-04T11:24:38.1583521Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_fuse_int_mm_with_mul]: False 2025-12-04T11:24:38.1584314Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_mixed_mm]: True 2025-12-04T11:24:38.1585507Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwmmbkdkarexuhbigurz5lfnhx64tht7fznecjkrvznh6rzivbv] inductor_config[fx_passes_numeric_check]: {'pre_grad': False, 'precision': 0.0001, 'num_iterations': 1, 'requires_optimizer': True} 2025-12-04T11:24:38.1586324Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v2td5s4lnsvyxvaevy4chx6kc5h3mm2axazbgwimqule5zrzao7] inductor_config[mixed_mm_choice]: heuristic 2025-12-04T11:24:38.1587194Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_compute_comm_overlap]: False 2025-12-04T11:24:38.1588078Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[reorder_for_compute_comm_overlap_passes]: [] 2025-12-04T11:24:38.1588959Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[reorder_prefetch_limit]: None 2025-12-04T11:24:38.1589772Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[reorder_for_peak_memory]: True 2025-12-04T11:24:38.1590628Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[reorder_for_peak_memory_debug]: False 2025-12-04T11:24:38.1591539Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[size_threshold_for_succ_based_strategy]: 0 2025-12-04T11:24:38.1592358Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_gathers_fx]: none 2025-12-04T11:24:38.1593326Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_gathers_fx_bucket_size_determinator]: None 2025-12-04T11:24:38.1594158Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_reduce_scatters_fx]: none 2025-12-04T11:24:38.1595132Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_reduce_scatters_fx_bucket_size_determinator]: None 2025-12-04T11:24:38.1595933Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t3u4yj5mzijyfjvypyqngc4gf3wv6433necbugezv54jsexzrfp] inductor_config[bucket_all_reduces_fx]: none 2025-12-04T11:24:38.1597087Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[bucket_all_reduces_fx_bucket_size_determinator]: None 2025-12-04T11:24:38.1597903Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [lxxtoqhcoepwfokeiibd575gnxo3uzwiv4hmpomlwkpzqz3qzsh] inductor_config[estimate_op_runtime]: default 2025-12-04T11:24:38.1598796Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[runtime_estimations_mms_benchmark]: False 2025-12-04T11:24:38.1599556Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yezuzjtg4h3jjur4jwtwiehbyixa7eonq4tqsqmwqve2lvvmrem] inductor_config[intra_node_bw]: 300 2025-12-04T11:24:38.1600373Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[inter_node_bw]: 25 2025-12-04T11:24:38.1601230Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_experimental_benchmarker]: True 2025-12-04T11:24:38.1602073Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[distributed_max_autotune_gemm]: False 2025-12-04T11:24:38.1602846Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune]: False 2025-12-04T11:24:38.1603661Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_pointwise]: False 2025-12-04T11:24:38.1604466Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_gemm]: False 2025-12-04T11:24:38.1605348Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[autotune_num_choices_displayed]: 10 2025-12-04T11:24:38.1606225Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_report_choices_stats]: True 2025-12-04T11:24:38.1607150Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[max_autotune_prune_choices_based_on_shared_mem]: True 2025-12-04T11:24:38.1608043Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton_disable_device_detection]: False 2025-12-04T11:24:38.1608832Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[graph_partition]: True 2025-12-04T11:24:38.1609691Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[custom_should_partition_ops]: [] 2025-12-04T11:24:38.1610584Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[max_autotune_allow_flexible_layouts]: False 2025-12-04T11:24:38.1611389Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_same_precision]: False 2025-12-04T11:24:38.1612179Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[multi_kernel_hints]: [] 2025-12-04T11:24:38.1613074Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [2y7luesktjrque3nr7qtxnum2mkbeegzdrsvkm3rvdlhqboajhx] inductor_config[max_autotune_gemm_backends]: ATEN,TRITON,CPP 2025-12-04T11:24:38.1613935Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [uqlsbif4zxd75vt522p52txyuguieipi2lwz5g5awt56lccqk7s] inductor_config[max_autotune_conv_backends]: ATEN,TRITON 2025-12-04T11:24:38.1614813Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_gemm_search_space]: DEFAULT 2025-12-04T11:24:38.1615677Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[max_autotune_flex_search_space]: DEFAULT 2025-12-04T11:24:38.1616520Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_by_default]: False 2025-12-04T11:24:38.1617424Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[selective_decompose]: False 2025-12-04T11:24:38.1618163Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_dce]: True 2025-12-04T11:24:38.1618951Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_pre_grad_passes]: True 2025-12-04T11:24:38.1619748Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_joint_graph_passes]: True 2025-12-04T11:24:38.1620546Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_post_grad_passes]: True 2025-12-04T11:24:38.1621411Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cutedsl_enable_autotuning]: False 2025-12-04T11:24:38.1622245Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_fallback_to_aten]: False 2025-12-04T11:24:38.1623060Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [wft6ljqsfr3x4m7fa5zuyb7cwknky4irrxz4bjr6uzr2yiopxqj] inductor_config[unbacked_symint_fallback]: 8192 2025-12-04T11:24:38.1623906Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[search_autotune_cache]: False 2025-12-04T11:24:38.1624643Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[save_args]: False 2025-12-04T11:24:38.1625468Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_in_subproc]: False 2025-12-04T11:24:38.1626392Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [iglov24t7x5ruci344aer2tm6nqshi4veuw4wxlssxtu46cx76m] inductor_config[max_autotune_subproc_result_timeout_seconds]: 60.0 2025-12-04T11:24:38.1627309Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_graceful_timeout_seconds]: 0.0 2025-12-04T11:24:38.1628249Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zslw6pp37dzmhi5lhweftlhhdttfjade3t5j3y3vfk3ouze7nhw] inductor_config[max_autotune_subproc_terminate_timeout_seconds]: 0.0 2025-12-04T11:24:38.1629061Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[autotune_multi_device]: False 2025-12-04T11:24:38.1629893Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tuax46wac7rfv2trf5gcps6vleo3cq44lbnrdxtprvo3ljjaddj] inductor_config[collective_benchmark_nruns]: 50 2025-12-04T11:24:38.1630728Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [5uk25jozsokfxmtekuawzrutfda644f563ena2lt6lv3xraleqh] inductor_config[collective_benchmark_timeout]: 30.0 2025-12-04T11:24:38.1631562Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_tuning]: False 2025-12-04T11:24:38.1632508Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[coordinate_descent_check_all_directions]: False 2025-12-04T11:24:38.1633366Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[coordinate_descent_search_radius]: 1 2025-12-04T11:24:38.1634155Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[autoheuristic_collect]: 2025-12-04T11:24:38.1634966Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jwbrgxes7vjqumngs5hyj6gn5nytv2whnppnzngvaagfmawhkkd] inductor_config[autoheuristic_use]: mixed_mm 2025-12-04T11:24:38.1635810Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[run_jit_post_compile_hook]: False 2025-12-04T11:24:38.1636644Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jvchmi66fvqzlemhr5fcqorz5trfdtdalzfagtj2aolmimwqhdq] inductor_config[autoheuristic_log_path]: DEFAULT 2025-12-04T11:24:38.1637458Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4p2fdjlvxrcw7c7fvzm5huhtqxnro4kvkx56f7p5zyrxqkwooov] inductor_config[layout_opt_default]: 1 2025-12-04T11:24:38.1638251Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[layout_optimization]: True 2025-12-04T11:24:38.1639080Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_layout_optimization]: False 2025-12-04T11:24:38.1639906Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[keep_output_stride]: True 2025-12-04T11:24:38.1640683Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[warn_mix_layout]: False 2025-12-04T11:24:38.1641528Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [lkkae3meylaixfif4thncru4hjqeaislawjoghffrbwuscaagei] inductor_config[realize_reads_threshold]: 4 2025-12-04T11:24:38.1642344Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [rr5m5hsocoyodldz7vcvaizdwvm2rt34evmqdxvng7wz3tufvo6] inductor_config[realize_opcount_threshold]: 30 2025-12-04T11:24:38.1643190Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[realize_acc_reads_threshold]: 8 2025-12-04T11:24:38.1644050Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[realize_acc_reads_size_threshold]: None 2025-12-04T11:24:38.1644826Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_random]: False 2025-12-04T11:24:38.1645714Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[fallback_embedding_bag_byte_unpack]: False 2025-12-04T11:24:38.1646497Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[implicit_fallbacks]: True 2025-12-04T11:24:38.1647369Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_unaligned_fallback_output]: False 2025-12-04T11:24:38.1648175Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[inductor_choices_class]: None 2025-12-04T11:24:38.1649030Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aggressive_fusion]: False 2025-12-04T11:24:38.1649790Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_fusion]: False 2025-12-04T11:24:38.1650573Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_fusion]: False 2025-12-04T11:24:38.1651365Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[enabled_metric_tables]: 2025-12-04T11:24:38.1652193Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_ordering_after_fusion]: True 2025-12-04T11:24:38.1653083Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[loop_index_inversion_in_fusion]: True 2025-12-04T11:24:38.1653914Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[score_fusion_memory_threshold]: 10 2025-12-04T11:24:38.1654752Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[benchmark_epilogue_fusion]: True 2025-12-04T11:24:38.1655709Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[max_epilogue_benchmarked_choices]: 1 2025-12-04T11:24:38.1656471Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_size]: 64 2025-12-04T11:24:38.1657485Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jykiys6ynafs3zdylwa5ggq6j655mxeh42d6mtdi22gffkrmiac] inductor_config[max_fusion_buffer_group_pairwise_attempts]: 64 2025-12-04T11:24:38.1658332Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[max_fusion_unique_io_buffers]: None 2025-12-04T11:24:38.1659157Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[max_pointwise_cat_inputs]: 8 2025-12-04T11:24:38.1659957Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_pointwise_cat]: False 2025-12-04T11:24:38.1660810Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[unroll_reductions_threshold]: 8 2025-12-04T11:24:38.1661584Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[comment_origin]: False 2025-12-04T11:24:38.1662350Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[conv_1x1_as_mm]: False 2025-12-04T11:24:38.1663139Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[split_reductions]: True 2025-12-04T11:24:38.1663900Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[deterministic]: False 2025-12-04T11:24:38.1664696Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[min_num_split]: 0 2025-12-04T11:24:38.1665475Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_kernel]: False 2025-12-04T11:24:38.1666330Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[constant_and_index_propagation]: True 2025-12-04T11:24:38.1667172Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[always_keep_tensor_constants]: False 2025-12-04T11:24:38.1667992Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[assert_indirect_indexing]: True 2025-12-04T11:24:38.1668824Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[compute_all_bounds]: False 2025-12-04T11:24:38.1669590Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[combo_kernels]: False 2025-12-04T11:24:38.1670414Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[benchmark_combo_kernel]: False 2025-12-04T11:24:38.1671239Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernels_autotune]: 1 2025-12-04T11:24:38.1672097Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[combo_kernel_allow_mixed_sizes]: 1 2025-12-04T11:24:38.1672998Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[combo_kernel_foreach_dynamic_shapes]: True 2025-12-04T11:24:38.1673813Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [7c6i6h6bfell5u33q6rcv25lpgmk4jah3uhjjx6bjevvjnshoim] inductor_config[combo_kernel_max_num_args]: 250 2025-12-04T11:24:38.1674653Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[joint_graph_constant_folding]: True 2025-12-04T11:24:38.1675446Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_index_asserts]: False 2025-12-04T11:24:38.1676281Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_precision_casts]: False 2025-12-04T11:24:38.1677103Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[emulate_divison_rounding]: False 2025-12-04T11:24:38.1677899Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[is_nightly_or_source]: True 2025-12-04T11:24:38.1678683Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[developer_warnings]: True 2025-12-04T11:24:38.1679571Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[optimize_scatter_upon_const_tensor]: True 2025-12-04T11:24:38.1680399Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[add_pre_grad_passes]: None 2025-12-04T11:24:38.1681216Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[remove_pre_grad_passes]: None 2025-12-04T11:24:38.1682085Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [eyt4i73byifiidlcgmugo4juf3sqznr7bv6k2xujnk6hfzd7vcn] inductor_config[small_memory_access_threshold]: 16777216 2025-12-04T11:24:38.1682900Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[worker_suppress_logging]: True 2025-12-04T11:24:38.1683666Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[log_tlparse]: False 2025-12-04T11:24:38.1684516Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_fuse_ddp_communication]: False 2025-12-04T11:24:38.1685324Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [5fxczt3ciyxitdhizb7sfsgn7fhpczcqsngttnt5ot2wyctk7co] inductor_config[_fuse_ddp_bucket_size]: 25 2025-12-04T11:24:38.1686110Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_micro_pipeline_tp]: False 2025-12-04T11:24:38.1686932Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_collective.auto_select]: False 2025-12-04T11:24:38.1687922Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4vdewewvaarnygruqwzavmkvu4lqggolypo2tq5ohtx2kcelkky] inductor_config[_collective.one_shot_all_reduce_threshold_bytes]: 131072 2025-12-04T11:24:38.1688952Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aten_distributed_optimizations.enable_overlap_scheduling]: False 2025-12-04T11:24:38.1689938Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.collective_bucketing]: None 2025-12-04T11:24:38.1690905Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.insert_overlap_deps]: None 2025-12-04T11:24:38.1691891Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_compute_pre_fetch]: None 2025-12-04T11:24:38.1692890Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.custom_runtime_estimation]: None 2025-12-04T11:24:38.1693897Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [y5k6t5wy5da5t7gcf4bkajyleeovwikw6pyjiaf6ieqev62zxrj] inductor_config[aten_distributed_optimizations.collective_estimator]: analytical 2025-12-04T11:24:38.1694876Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_gb]: None 2025-12-04T11:24:38.1695869Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aten_distributed_optimizations.max_memory_increase_ratio]: None 2025-12-04T11:24:38.1696948Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[quiesce_async_compile_pool]: False 2025-12-04T11:24:38.1697778Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [smfbtbb3fuwobubxbj6g3jio7u6ufdkgz35qhppjgt3yxptkxha] inductor_config[quiesce_async_compile_time]: 60 2025-12-04T11:24:38.1698602Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[use_static_cuda_launcher]: True 2025-12-04T11:24:38.1699512Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[static_launch_user_defined_triton_kernels]: False 2025-12-04T11:24:38.1700362Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[strict_static_cuda_launcher]: False 2025-12-04T11:24:38.1701209Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[global_cache_dir]: None 2025-12-04T11:24:38.1701988Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[kernel_name_max_ops]: 10 2025-12-04T11:24:38.1702758Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[shape_padding]: True 2025-12-04T11:24:38.1703558Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[comprehensive_padding]: True 2025-12-04T11:24:38.1704393Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_channels_last]: False 2025-12-04T11:24:38.1705193Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_dynamic_shapes]: False 2025-12-04T11:24:38.1706040Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[disable_padding_cpu]: True 2025-12-04T11:24:38.1706925Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[expand_dimension_for_pointwise_nodes]: False 2025-12-04T11:24:38.1707748Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[padding_alignment_bytes]: 128 2025-12-04T11:24:38.1708574Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [dnnw5ks3yxrp7mwvihb2hh4tqx35ye637xt33x64kw4fvz2nyzg] inductor_config[padding_stride_threshold]: 1024 2025-12-04T11:24:38.1709338Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[pad_outputs]: False 2025-12-04T11:24:38.1710162Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[bw_outputs_user_visible]: True 2025-12-04T11:24:38.1710930Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[force_shape_pad]: False 2025-12-04T11:24:38.1711715Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[permute_fusion]: False 2025-12-04T11:24:38.1712596Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profiler_mark_wrapper_call]: False 2025-12-04T11:24:38.1713445Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[generate_intermediate_hooks]: False 2025-12-04T11:24:38.1714246Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[debug_ir_traceback]: False 2025-12-04T11:24:38.1715062Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_raise_error_for_testing]: False 2025-12-04T11:24:38.1715804Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[_profile_var]: 2025-12-04T11:24:38.1716587Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth]: False 2025-12-04T11:24:38.1717430Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[profile_bandwidth_regex]: 2025-12-04T11:24:38.1718255Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[profile_bandwidth_output]: None 2025-12-04T11:24:38.1719200Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[profile_bandwidth_with_do_bench_using_profiling]: False 2025-12-04T11:24:38.1720038Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[disable_cpp_codegen]: False 2025-12-04T11:24:38.1720778Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing]: False 2025-12-04T11:24:38.1721663Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[freezing_discard_parameters]: False 2025-12-04T11:24:38.1722473Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[decompose_mem_bound_mm]: False 2025-12-04T11:24:38.1723339Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[wrap_inductor_compiled_regions]: False 2025-12-04T11:24:38.1724146Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_aligned_inputs]: False 2025-12-04T11:24:38.1724954Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[assume_32bit_indexing]: False 2025-12-04T11:24:38.1725906Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[unsafe_ignore_unsupported_triton_autotune_args]: False 2025-12-04T11:24:38.1726780Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[check_stack_no_cycles_TESTING_ONLY]: False 2025-12-04T11:24:38.1727632Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_linear_binary_folding]: False 2025-12-04T11:24:38.1728448Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[annotate_training]: False 2025-12-04T11:24:38.1729371Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[enable_caching_generated_triton_templates]: True 2025-12-04T11:24:38.1730165Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[autotune_lookup_table]: {} 2025-12-04T11:24:38.1730953Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [c2gaopsp6oqinpzvhlaz4gsqnq3shl5e54b7j6dsgwjadh5uatx] inductor_config[file_lock_timeout]: 600 2025-12-04T11:24:38.1731772Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[enable_autograd_for_aot]: False 2025-12-04T11:24:38.1732590Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[_debug_cpu_to_tpu_pallas]: False 2025-12-04T11:24:38.1733494Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[pallas_take_first_jax_device_only]: True 2025-12-04T11:24:38.1734321Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[torchinductor_worker_logpath]: 2025-12-04T11:24:38.1735078Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [sz3im5ogc6asp7g4uqocnovype63tkdexzfrniv6hn2oank3biu] inductor_config[cpp.threads]: -1 2025-12-04T11:24:38.1735908Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.no_redundant_loops]: True 2025-12-04T11:24:38.1736718Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.dynamic_threads]: False 2025-12-04T11:24:38.1737578Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.simdlen]: None 2025-12-04T11:24:38.1738366Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [xgnfe6mw7nii5zpxhlblgsehzrcqmjqpqswcwvf5adwbhz7aj2h] inductor_config[cpp.min_chunk_size]: 512 2025-12-04T11:24:38.1739156Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [c7zj4qytmety6keurs3hsh5wn7foxp3dqx4kym2ucszzcb2ngrf] inductor_config[cpp.cxx]: (None, 'g++') 2025-12-04T11:24:38.1739990Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_kernel_profile]: False 2025-12-04T11:24:38.1740796Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.weight_prepack]: True 2025-12-04T11:24:38.1741668Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:38.1742553Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.inject_log1p_bug_TESTING_ONLY]: None 2025-12-04T11:24:38.1743323Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.vec_isa_ok]: None 2025-12-04T11:24:38.1744173Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[cpp.descriptive_names]: original_aten 2025-12-04T11:24:38.1745054Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[cpp.max_horizontal_fusion_size]: 16 2025-12-04T11:24:38.1745905Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.fallback_scatter_reduce_sum]: True 2025-12-04T11:24:38.1746773Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_unsafe_math_opt_flag]: False 2025-12-04T11:24:38.1747662Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ijs44lspkinjvhcs7uff7n3noc53jvsp4yfljjh22mafhb7khxe] inductor_config[cpp.enable_floating_point_contract_flag]: off 2025-12-04T11:24:38.1748517Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_tiling_heuristics]: True 2025-12-04T11:24:38.1749413Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_grouped_gemm_template]: False 2025-12-04T11:24:38.1750206Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cpp.gemm_max_k_slices]: 1 2025-12-04T11:24:38.1751041Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_cache_blocking]: None 2025-12-04T11:24:38.1751888Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cpp.gemm_thread_factors]: None 2025-12-04T11:24:38.1752718Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.enable_loop_tail_vec]: True 2025-12-04T11:24:38.1753580Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.enable_concat_linear]: False 2025-12-04T11:24:38.1754411Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_decompose_tanh]: False 2025-12-04T11:24:38.1755252Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.use_small_dequant_buffer]: False 2025-12-04T11:24:38.1756085Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cpp.force_inline_kernel]: False 2025-12-04T11:24:38.1756940Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cpp.use_constexpr_for_int_array]: True 2025-12-04T11:24:38.1757733Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraphs]: False 2025-12-04T11:24:38.1758556Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_trees]: True 2025-12-04T11:24:38.1759448Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_skip_dynamic_graphs]: False 2025-12-04T11:24:38.1760313Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.cudagraph_capture_sizes]: None 2025-12-04T11:24:38.1761219Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.slow_path_cudagraph_asserts]: True 2025-12-04T11:24:38.1762144Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_trees_history_recording]: False 2025-12-04T11:24:38.1763044Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.cudagraph_support_input_mutation]: True 2025-12-04T11:24:38.1763961Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljdqgtysl3vdf7j6attlz5gmjg2ncihnveojfyubosplmkrjgra] inductor_config[triton.cudagraph_unexpected_rerecord_limit]: 128 2025-12-04T11:24:38.1764868Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yttmfmxblgcbsvbokguzowcorrcxz5uunxtcvsbe6nijgcx45he] inductor_config[triton.cudagraph_dynamic_shape_warn_limit]: 8 2025-12-04T11:24:38.1765733Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraph_sync]: False 2025-12-04T11:24:38.1766600Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cudagraphs_warmup]: False 2025-12-04T11:24:38.1767433Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cudagraph_or_error]: False 2025-12-04T11:24:38.1768410Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.reorder_for_reducing_graph_partitions]: True 2025-12-04T11:24:38.1769293Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.fast_path_cudagraph_asserts]: False 2025-12-04T11:24:38.1770161Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_cudagraph_warmup]: False 2025-12-04T11:24:38.1770990Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_graph]: False 2025-12-04T11:24:38.1771818Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.debug_sync_kernel]: False 2025-12-04T11:24:38.1772637Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.dense_indexing]: False 2025-12-04T11:24:38.1773493Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.coalesce_tiling_analysis]: True 2025-12-04T11:24:38.1774289Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.max_tiles]: None 2025-12-04T11:24:38.1775105Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.prefer_nd_tiling]: False 2025-12-04T11:24:38.1775941Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_pointwise]: True 2025-12-04T11:24:38.1776784Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.autotune_cublasLt]: True 2025-12-04T11:24:38.1777713Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.autotune_at_compile_time]: None 2025-12-04T11:24:38.1778604Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.autotune_with_sample_inputs]: False 2025-12-04T11:24:38.1779413Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.tile_reductions]: False 2025-12-04T11:24:38.1780224Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.native_matmul]: False 2025-12-04T11:24:38.1781160Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_pointwise_fusion]: True 2025-12-04T11:24:38.1782077Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.tiling_prevents_reduction_fusion]: True 2025-12-04T11:24:38.1782903Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.unique_kernel_names]: True 2025-12-04T11:24:38.1783793Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.unique_user_kernel_names]: False 2025-12-04T11:24:38.1784680Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [yrty22bseefglnysuoec4ji7j2rnaggdj3g33zzj7avogwfmgdw] inductor_config[triton.descriptive_names]: original_aten 2025-12-04T11:24:38.1785550Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.persistent_reductions]: True 2025-12-04T11:24:38.1786414Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.cooperative_reductions]: False 2025-12-04T11:24:38.1787303Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.force_cooperative_reductions]: False 2025-12-04T11:24:38.1788104Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [vrl5ktomgtzox5xucd3np6vug3vyj6hwwzahqijuwpmamlv7ohi] inductor_config[triton.multi_kernel]: 0 2025-12-04T11:24:38.1788911Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.divisible_by_16]: True 2025-12-04T11:24:38.1789748Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [fv6slhtedtydps5s5u2etitscliblzcidyitqf7krsv4e23fzk6] inductor_config[triton.min_split_scan_rblock]: 256 2025-12-04T11:24:38.1790546Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.store_cubin]: False 2025-12-04T11:24:38.1791336Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[triton.spill_threshold]: 16 2025-12-04T11:24:38.1792149Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_block_ptr]: False 2025-12-04T11:24:38.1793019Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.use_tensor_descriptor]: False 2025-12-04T11:24:38.1793984Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.transpose_discontiguous_tensor_descriptor]: True 2025-12-04T11:24:38.1794866Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.inject_relu_bug_TESTING_ONLY]: None 2025-12-04T11:24:38.1795712Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.codegen_upcast_to_fp32]: True 2025-12-04T11:24:38.1796750Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_persistent_tma_matmul]: False 2025-12-04T11:24:38.1797680Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_template_tma_store]: False 2025-12-04T11:24:38.1798551Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.enable_epilogue_subtiling]: True 2025-12-04T11:24:38.1799354Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.skip_l1_cache]: False 2025-12-04T11:24:38.1800383Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.disallow_failing_autotune_kernels_TESTING_ONLY]: False 2025-12-04T11:24:38.1801225Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [j6c55jha5r2sdys2rwq7uqhtleea5dgjcye7nicfgft36v7xfvp] inductor_config[triton.num_decompose_k_splits]: 10 2025-12-04T11:24:38.1802104Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [jffvide67gguonizth6bla7qwy6egn73yfn66335sv5b7i2rx3p] inductor_config[triton.decompose_k_threshold]: 32 2025-12-04T11:24:38.1802896Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.enable_pdl]: False 2025-12-04T11:24:38.1803720Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[triton.mix_order_reduction]: True 2025-12-04T11:24:38.1804631Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[triton.mix_order_reduction_initial_xblock]: 1 2025-12-04T11:24:38.1805521Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[triton.mix_order_reduction_split_size]: None 2025-12-04T11:24:38.1806472Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[triton.mix_order_reduction_autotune_split_size]: False 2025-12-04T11:24:38.1807269Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.output_path]: 2025-12-04T11:24:38.1808116Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_compile]: False 2025-12-04T11:24:38.1808988Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.debug_symbols]: False 2025-12-04T11:24:38.1809868Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [6fxyf5ymh244xdypwkhtsbszab4nnfsgmul2kmyqmw422i5h54e] inductor_config[aot_inductor.compile_wrapper_opt_level]: O1 2025-12-04T11:24:38.1810789Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[aot_inductor.debug_intermediate_value_printer]: 0 2025-12-04T11:24:38.1811667Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.filtered_kernel_names]: None 2025-12-04T11:24:38.1812515Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_in_spec]: 2025-12-04T11:24:38.1813354Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[aot_inductor.serialized_out_spec]: 2025-12-04T11:24:38.1814307Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_runtime_constant_folding]: False 2025-12-04T11:24:38.1815167Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.force_mmap_weights]: False 2025-12-04T11:24:38.1816070Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.use_consts_asm_build]: True 2025-12-04T11:24:38.1816951Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.package]: False 2025-12-04T11:24:38.1817853Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_cpp_only]: None 2025-12-04T11:24:38.1818706Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.dynamic_linkage]: True 2025-12-04T11:24:38.1819506Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.metadata]: {} 2025-12-04T11:24:38.1820457Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.raise_error_on_ignored_optimization]: True 2025-12-04T11:24:38.1821321Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.dump_aoti_minifier]: False 2025-12-04T11:24:38.1822145Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [pr5nr4a7dthirgd2ljo3d2xakc63ywxugusu6mkmr6gmpeliyib] inductor_config[aot_inductor.repro_level]: 2 2025-12-04T11:24:38.1822931Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.presets]: {} 2025-12-04T11:24:38.1823808Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.allow_stack_allocation]: False 2025-12-04T11:24:38.1824742Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.use_minimal_arrayref_interface]: False 2025-12-04T11:24:38.1825700Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.weight_use_caching_allocator]: False 2025-12-04T11:24:38.1826593Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.package_constants_in_so]: True 2025-12-04T11:24:38.1827532Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.package_constants_on_disk_format]: None 2025-12-04T11:24:38.1828396Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.precompile_headers]: True 2025-12-04T11:24:38.1829262Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.embed_kernel_binary]: None 2025-12-04T11:24:38.1830191Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.emit_multi_arch_kernel]: None 2025-12-04T11:24:38.1831126Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.model_name_for_generated_files]: None 2025-12-04T11:24:38.1831983Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [4bryyl4ahh5whyg3zwqebpwmjnx6w77nqgqbdjlowju6lkqtn7w] inductor_config[aot_inductor.custom_ops_to_c_shims]: {} 2025-12-04T11:24:38.1832866Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.custom_op_libs]: None 2025-12-04T11:24:38.1833690Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor.enable_lto]: False 2025-12-04T11:24:38.1834563Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[aot_inductor.link_libtorch]: True 2025-12-04T11:24:38.1835433Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.cross_target_platform]: None 2025-12-04T11:24:38.1836298Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library]: None 2025-12-04T11:24:38.1837182Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[aot_inductor.aoti_shim_library_path]: None 2025-12-04T11:24:38.1838069Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[aot_inductor_mode.compile_standalone]: False 2025-12-04T11:24:38.1838822Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.arch]: None 2025-12-04T11:24:38.1839586Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.version]: None 2025-12-04T11:24:38.1840411Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tvyftmtdmezlejo2xllu7awzv4pzc4vm4fub4b3gpl5jptjkosi] inductor_config[cuda.compile_opt_level]: -O1 2025-12-04T11:24:38.1841213Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_cuda_lto]: False 2025-12-04T11:24:38.1842068Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_ptxas_info]: False 2025-12-04T11:24:38.1842882Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.enable_debug_info]: False 2025-12-04T11:24:38.1843667Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.use_fast_math]: False 2025-12-04T11:24:38.1844559Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_max_profiling_configs]: None 2025-12-04T11:24:38.1845503Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ty4d7ntvjwumcgotd4j6w7bwokf5njhzmtvqvxa32jjub6k2ty2] inductor_config[cuda.cutlass_max_profiling_swizzle_options]: [1, 2, 4, 8] 2025-12-04T11:24:38.1846431Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_epilogue_fusion_enabled]: False 2025-12-04T11:24:38.1847231Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_tma_only]: False 2025-12-04T11:24:38.1848006Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cuda_cxx]: None 2025-12-04T11:24:38.1848895Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [aghvyrrgwvxijco2pk5wzc3cgmmthrbmgxitiibxuuscxdwrjd3] inductor_config[cuda.cutlass_backend_min_gemm_size]: 1 2025-12-04T11:24:38.1849730Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.generate_test_runner]: False 2025-12-04T11:24:38.1850624Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_allowlist_regex]: None 2025-12-04T11:24:38.1851479Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[cuda.cutlass_op_denylist_regex]: None 2025-12-04T11:24:38.1852338Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ngkkx5e6z7erl6da23zb2cmsctz4yvaqyameyg5hbqln4wrhh7x] inductor_config[cuda.cutlass_instantiation_level]: 0 2025-12-04T11:24:38.1853213Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.cutlass_hash_with_compile_cmd]: False 2025-12-04T11:24:38.1854054Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.cutlass_prescreening]: True 2025-12-04T11:24:38.1854868Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ly46nlihymo3siersryfadlchkmxk6ohljz4l7vognsjg2qurpp] inductor_config[cuda.cutlass_enabled_ops]: all 2025-12-04T11:24:38.1855713Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.use_binary_remote_cache]: True 2025-12-04T11:24:38.1856586Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.upload_to_binary_remote_cache]: False 2025-12-04T11:24:38.1857563Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[cuda.binary_remote_cache_force_write]: False 2025-12-04T11:24:38.1858413Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[cuda.enable_caching_codegen]: True 2025-12-04T11:24:38.1859136Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[rocm.arch]: [] 2025-12-04T11:24:38.1860087Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [gzctoy3drvth5kwqmdxb4tjn2picfdjsdu33nbniulhx5hsi3lv] inductor_config[rocm.ck_supported_arch]: ['gfx90a', 'gfx942', 'gfx950'] 2025-12-04T11:24:38.1860898Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [klfqjprnpfhcdurgvuikvc4rpd5ynkpk77toousr5h3u5roty6p] inductor_config[rocm.compile_opt_level]: -O2 2025-12-04T11:24:38.1861718Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.is_debug]: False 2025-12-04T11:24:38.1862496Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.save_temps]: False 2025-12-04T11:24:38.1863273Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.use_fast_math]: True 2025-12-04T11:24:38.1864076Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[rocm.flush_denormals]: True 2025-12-04T11:24:38.1864972Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.print_kernel_resource_usage]: False 2025-12-04T11:24:38.1865762Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.rocm_home]: None 2025-12-04T11:24:38.1866542Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_dir]: None 2025-12-04T11:24:38.1867380Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.generate_test_runner]: False 2025-12-04T11:24:38.1868226Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.n_max_profiling_configs]: None 2025-12-04T11:24:38.1869075Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_max_profiling_configs]: None 2025-12-04T11:24:38.1869971Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.ck_tile_max_profiling_configs]: None 2025-12-04T11:24:38.1870823Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[rocm.use_preselected_instances]: False 2025-12-04T11:24:38.1871619Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[rocm.kBatch_sweep]: None 2025-12-04T11:24:38.1872416Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.split_k_threshold]: 16 2025-12-04T11:24:38.1873274Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ebt2ncs4f5y7dn7btzi76mnouepvzad474tmp5iju4wiuumjl4s] inductor_config[rocm.contiguous_threshold]: 16 2025-12-04T11:24:38.1874025Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [bsvfcwwoczx2rlkdz2eta6doujsymyihmi46hhwk6clrrvwcb6m] inductor_config[cpu_backend]: cpp 2025-12-04T11:24:38.1874787Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[cuda_backend]: triton 2025-12-04T11:24:38.1875554Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [caw4ly2z672k6kjfahoxwpajp5idhhtrpgf3ma2clylcp7c7aid] inductor_config[xpu_backend]: triton 2025-12-04T11:24:38.1876343Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [ljhgflgihidopsfsdcbqynv27nceykby3nutyd5jlcpq7n6e7l4] inductor_config[halide.cpu_target]: host 2025-12-04T11:24:38.1877167Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [wx7vmsmrdpk5ue2txlywp3lj3faqmdjphs5fgg2ehzsyno7uovg] inductor_config[halide.gpu_target]: host-cuda 2025-12-04T11:24:38.1878051Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [svgytlua5wcyeia7wq7e6zgh5tsueikrnzchmdmouvmkpfsc2zq] inductor_config[halide.scheduler_cuda]: Anderson2021 2025-12-04T11:24:38.1878888Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [k5ogk6345jvklsnu7g2njqstiz2g6pm5wmqpgg3kasrmuqwjvl6] inductor_config[halide.scheduler_cpu]: Adams2019 2025-12-04T11:24:38.1879659Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.asserts]: False 2025-12-04T11:24:38.1880446Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.debug]: False 2025-12-04T11:24:38.1881260Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[halide.scan_kernels]: False 2025-12-04T11:24:38.1883259Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [zwewsbwzgzypcnzixgl7ybbc4tk5kq36yeo267m422vyiuhdyiv] inductor_config[_save_config_ignore]: ['trace.upload_tar', 'joint_custom_pre_pass', 'joint_custom_post_pass', 'pre_grad_custom_pass', 'aot_inductor.repro_level', 'aot_inductor.dump_aoti_minifier', 'post_grad_custom_pre_pass', 'post_grad_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass'] 2025-12-04T11:24:38.1885699Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [6trwnwm4voevl4joplmkcssruwgd46kgqfejamut6kq662kstpd] inductor_config[_cache_config_ignore_prefix]: ['trace', 'cuda.cutlass_dir', 'worker_start_method', 'compile_threads', 'post_grad_custom_post_pass', 'post_grad_custom_pre_pass', 'joint_custom_pre_pass', 'joint_custom_post_pass', '_fuse_ddp_communication_passes', '_pre_fusion_custom_pass', 'always_complex_memory_overlap_TESTING_ONLY', 'fx_graph_cache', 'fx_graph_remote_cache', 'autotune_local_cache', 'autotune_remote_cache'] 2025-12-04T11:24:38.1886463Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [h25wqx6vliw4j5rtzzbv6latydxyei3deyg6v7wzvnzryfktuki] inductor_config[external_matmul]: [] 2025-12-04T11:24:38.1887386Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[write_are_deterministic_algorithms_enabled]: True 2025-12-04T11:24:38.1888192Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[lookup_table.table]: None 2025-12-04T11:24:38.1889046Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[lookup_table.check_src_hash]: True 2025-12-04T11:24:38.1890051Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_extern_kernel_in_multi_template]: False 2025-12-04T11:24:38.1890908Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.max_mm_configs]: None 2025-12-04T11:24:38.1891817Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_dtype_assert]: False 2025-12-04T11:24:38.1892738Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.runtime_triton_shape_assert]: False 2025-12-04T11:24:38.1893625Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.static_cpp_dtype_assert]: False 2025-12-04T11:24:38.1894558Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_name_regex]: None 2025-12-04T11:24:38.1895470Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.autotune_choice_desc_regex]: None 2025-12-04T11:24:38.1896621Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.graphsafe_rng_func_ignores_fallback_random]: False 2025-12-04T11:24:38.1897641Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] inductor_config[test_configs.track_memory_lifecycle]: None 2025-12-04T11:24:38.1898481Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.use_libtorch]: False 2025-12-04T11:24:38.1899462Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [cev5uo2jlwdhw2uyzcm7vr6cl23azjfw437f5r5lskm7spucos6] inductor_config[test_configs.assume_bucketing_reduces_latency]: True 2025-12-04T11:24:38.1900380Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.force_filter_reduction_configs]: False 2025-12-04T11:24:38.1901277Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [v3hzzlv4tjgvp3pyhmzagjd25orl6n7nynoa7svlhhwk73b7u3c] inductor_config[test_configs.distort_benchmarking_result]: 2025-12-04T11:24:38.1902156Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_pre_grad_graph]: False 2025-12-04T11:24:38.1903129Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] inductor_config[test_configs.bisect_keep_custom_backend_for_inductor]: False 2025-12-04T11:24:38.1903882Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_pre_pass: None 2025-12-04T11:24:38.1904578Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [esstihe2nyydk4mhzpvox3qkajyu5y5t23hk3fi2me7jn75xi3o] precompile_enabled: False 2025-12-04T11:24:38.1905342Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] post_grad_custom_post_pass: None 2025-12-04T11:24:38.1906092Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_pre_pass: None 2025-12-04T11:24:38.1906829Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] joint_custom_post_pass: None 2025-12-04T11:24:38.1907554Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _pre_fusion_custom_pass: None 2025-12-04T11:24:38.1908420Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [nk3qjerriqqc77fquy5nbegbf4gnlzzbxbtxwvyxvcdzt65xl2a] _fuse_ddp_communication_passes[0]: fuse_ddp_with_concat_op 2025-12-04T11:24:38.1909275Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [t46i2lzpuxqpmemjedva3sub75arja6fqed4duz4kp2bb7d3sgc] _fuse_ddp_communication_passes[1]: schedule_comm_wait 2025-12-04T11:24:38.1910119Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [74x2jtykapblkbwkh24fsfbwq4iejjkibyckoc2bmgj6llnf57s] custom_backend_passes: (None, None, None, None, None) 2025-12-04T11:24:38.1910964Z V1204 11:24:05.504000 101771 site-packages/torch/_inductor/codecache.py:983] [0/0] [tquy2we2efmowuj4wuqzcfcfdcrkzkzmwdae6hprj7fa64jpusq] _custom_partitioner_fn: None 2025-12-04T11:24:38.1911616Z V1204 11:24:05.505000 101771 site-packages/torch/_inductor/compile_fx.py:931] [0/0] FX cache key generated: fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:38.1912305Z I1204 11:24:05.505000 101771 site-packages/torch/_inductor/codecache.py:1613] [0/0] fx graph cache miss for key fith2ggvcm3f2gbtjs7mhrq4nmfp5cmqfm7ggiio7u2zzyn7dyil 2025-12-04T11:24:38.1912822Z V1204 11:24:05.505000 101771 site-packages/torch/_inductor/compile_fx.py:1000] [0/0] FX cache miss, compiling and saving to cache 2025-12-04T11:24:38.1913311Z V1204 11:24:05.506000 101771 site-packages/torch/_inductor/triton_bundler.py:140] [0/0] TritonBundler.begin_compile is called 2025-12-04T11:24:38.1913833Z I1204 11:24:05.506000 101771 site-packages/torch/_inductor/compile_fx.py:1233] [0/0] Step 3: torchinductor compiling FORWARDS graph 2 2025-12-04T11:24:38.1914623Z V1204 11:24:05.519000 101771 site-packages/torch/_inductor/fx_passes/decompose_mem_bound_mm.py:214] [0/0] Decompose aten.mm.default with input shape: torch.Size([20480, 5]), torch.Size([5, 2]) 2025-12-04T11:24:38.1915086Z V1204 11:24:05.523000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] TRACED GRAPH 2025-12-04T11:24:38.1915582Z V1204 11:24:05.523000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] ===== AFTER POST GRAD ===== 2025-12-04T11:24:38.1916540Z V1204 11:24:05.523000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/fx/_lazy_graph_module.py class GraphModule(torch.nn.Module): 2025-12-04T11:24:38.1917297Z V1204 11:24:05.523000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] def forward(self, primals_1: "f32[20480, 5][5, 1]cuda:0", primals_2: "f32[5, 2][2, 1]cuda:0"): 2025-12-04T11:24:38.1918328Z V1204 11:24:05.523000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] # File: /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:48 in forward, code: output = torch.mm(input1, input2) 2025-12-04T11:24:38.1919309Z V1204 11:24:05.523000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type: "bf16[5, 2][2, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_2, torch.bfloat16); primals_2 = None 2025-12-04T11:24:38.1920297Z V1204 11:24:05.523000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] convert_element_type_1: "bf16[20480, 5][5, 1]cuda:0" = torch.ops.prims.convert_element_type.default(primals_1, torch.bfloat16); primals_1 = None 2025-12-04T11:24:38.1921185Z V1204 11:24:05.523000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default: "bf16[20480, 5, 1][5, 1, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type_1, 2) 2025-12-04T11:24:38.1921999Z V1204 11:24:05.523000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] unsqueeze_default_1: "bf16[1, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.unsqueeze.default(convert_element_type, 0) 2025-12-04T11:24:38.1923061Z V1204 11:24:05.523000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] mul_tensor: "bf16[20480, 5, 2][10, 2, 1]cuda:0" = torch.ops.aten.mul.Tensor(unsqueeze_default, unsqueeze_default_1); unsqueeze_default = unsqueeze_default_1 = None 2025-12-04T11:24:38.1923884Z V1204 11:24:05.523000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] sum_dim_int_list: "bf16[20480, 2][2, 1]cuda:0" = torch.ops.aten.sum.dim_IntList(mul_tensor, [-2]); mul_tensor = None 2025-12-04T11:24:38.1924857Z V1204 11:24:05.523000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute: "bf16[5, 20480][1, 5]cuda:0" = torch.ops.aten.permute.default(convert_element_type_1, [1, 0]); convert_element_type_1 = None 2025-12-04T11:24:38.1925747Z V1204 11:24:05.523000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] permute_1: "bf16[2, 5][1, 2]cuda:0" = torch.ops.aten.permute.default(convert_element_type, [1, 0]); convert_element_type = None 2025-12-04T11:24:38.1926330Z V1204 11:24:05.523000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:38.1926794Z V1204 11:24:05.523000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:38.1927185Z V1204 11:24:05.523000 101771 site-packages/torch/_inductor/compile_fx.py:1313] [0/0] [__post_grad_graphs] 2025-12-04T11:24:38.1927800Z V1204 11:24:05.526000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_1 : [num_users=1] = placeholder[target=primals_1] 2025-12-04T11:24:38.1928370Z V1204 11:24:05.526000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %primals_2 : [num_users=1] = placeholder[target=primals_2] 2025-12-04T11:24:38.1929398Z V1204 11:24:05.527000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_2, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:38.1929901Z V1204 11:24:05.527000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.1930913Z V1204 11:24:05.528000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %convert_element_type_1 : [num_users=2] = call_function[target=torch.ops.prims.convert_element_type.default](args = (%primals_1, torch.bfloat16), kwargs = {}) 2025-12-04T11:24:38.1931427Z V1204 11:24:05.529000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.1932370Z V1204 11:24:05.530000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type_1, 2), kwargs = {}) 2025-12-04T11:24:38.1932834Z V1204 11:24:05.530000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.1933771Z V1204 11:24:05.531000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %unsqueeze_default_1 : [num_users=1] = call_function[target=torch.ops.aten.unsqueeze.default](args = (%convert_element_type, 0), kwargs = {}) 2025-12-04T11:24:38.1934267Z V1204 11:24:05.531000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.1935178Z V1204 11:24:05.533000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %mul_tensor : [num_users=1] = call_function[target=torch.ops.aten.mul.Tensor](args = (%unsqueeze_default, %unsqueeze_default_1), kwargs = {}) 2025-12-04T11:24:38.1935608Z V1204 11:24:05.533000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.1936470Z V1204 11:24:05.535000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %sum_dim_int_list : [num_users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%mul_tensor, [-2]), kwargs = {}) 2025-12-04T11:24:38.1936976Z V1204 11:24:05.535000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.1937884Z V1204 11:24:05.537000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type_1, [1, 0]), kwargs = {}) 2025-12-04T11:24:38.1938373Z V1204 11:24:05.538000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.1939279Z V1204 11:24:05.538000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering %permute_1 : [num_users=1] = call_function[target=torch.ops.aten.permute.default](args = (%convert_element_type, [1, 0]), kwargs = {}) 2025-12-04T11:24:38.1939725Z V1204 11:24:05.539000 101771 site-packages/torch/_inductor/graph.py:1319] [0/0] via 2025-12-04T11:24:38.1940269Z V1204 11:24:05.539000 101771 site-packages/torch/_inductor/graph.py:1631] [0/0] lowering return (sum_dim_int_list, permute, permute_1) 2025-12-04T11:24:38.1940862Z V1204 11:24:05.540000 101771 site-packages/torch/_inductor/graph.py:1527] [0/0] Force channels last inputs for 0 conv for the current graph with id 2 2025-12-04T11:24:38.1941603Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:24:38.1941761Z warnings.warn( 2025-12-04T11:24:38.1942238Z V1204 11:24:05.540000 101771 site-packages/torch/_inductor/triton_bundler.py:151] [0/0] TritonBundler.end_compile is called 2025-12-04T11:24:38.1943074Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-563749739dc7a57a.xml - 2025-12-04T11:24:38.1943261Z =========================== short test summary info ============================ 2025-12-04T11:24:38.1944151Z FAILED [0.3278s] inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True - AssertionError: Scalars are not equal! 2025-12-04T11:24:38.1944164Z 2025-12-04T11:24:38.1944287Z Expected 1 but got 0. 2025-12-04T11:24:38.1944393Z Absolute difference: 1 2025-12-04T11:24:38.1944503Z Relative difference: 1.0 2025-12-04T11:24:38.1944510Z 2025-12-04T11:24:38.1944737Z To execute this test, run the following from the base repo dir: 2025-12-04T11:24:38.1945585Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_decompose_mem_bound_mm.py TestDecomposeMemMM.test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:38.1945591Z 2025-12-04T11:24:38.1945872Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:24:38.1946077Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:24:38.1946320Z ================== 1 failed, 36 deselected, 2 rerun in 6.63s =================== 2025-12-04T11:24:38.1946431Z --- Logging error --- 2025-12-04T11:24:38.1946589Z Traceback (most recent call last): 2025-12-04T11:24:38.1946914Z File "/opt/conda/envs/py_3.10/lib/python3.10/logging/__init__.py", line 1103, in emit 2025-12-04T11:24:38.1947047Z stream.write(msg + self.terminator) 2025-12-04T11:24:38.1947182Z ValueError: I/O operation on closed file. 2025-12-04T11:24:38.1947290Z Call stack: 2025-12-04T11:24:38.1947810Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/remote_cache.py", line 432, in dump_cache_stats 2025-12-04T11:24:38.1947973Z log.info("Cache Metrics:%s", out.getvalue()) 2025-12-04T11:24:38.1948087Z Message: 'Cache Metrics:%s' 2025-12-04T11:24:38.1948193Z Arguments: (' None\n',) 2025-12-04T11:24:38.1948304Z Got exit code 1 2025-12-04T11:24:38.1949067Z FAILED CONSISTENTLY: test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True 2025-12-04T11:24:38.1949479Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:24:38.1950159Z Test results will be stored in test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-d89a888c67de20be.xml 2025-12-04T11:24:38.1950359Z ============================= test session starts ============================== 2025-12-04T11:24:38.1950727Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:24:38.1950835Z cachedir: .pytest_cache 2025-12-04T11:24:38.1951362Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:24:38.1951505Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:24:38.1951614Z configfile: pytest.ini 2025-12-04T11:24:38.1952242Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:24:38.1953443Z collecting ... /var/lib/jenkins/workspace/test/inductor/test_decompose_mem_bound_mm.py:52: PytestCollectionWarning: cannot collect test class 'TestDecomposeAddMM' because it has a __init__ constructor (from: test/inductor/test_decompose_mem_bound_mm.py) 2025-12-04T11:24:38.1953601Z class TestDecomposeAddMM(torch.nn.Module): 2025-12-04T11:24:38.1953791Z collected 37 items / 31 deselected / 6 selected 2025-12-04T11:24:38.1953933Z stepcurrent: skipping 31 already run items. 2025-12-04T11:24:38.1954047Z Running 6 items in this shard 2025-12-04T11:24:38.1954053Z 2025-12-04T11:24:38.1954791Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_2048_k_2_n_2_should_decompose_False_has_bias_False PASSED [5.8132s] [ 16%] 2025-12-04T11:24:38.1955508Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_2048_k_2_n_2_should_decompose_False_has_bias_True PASSED [0.2893s] [ 33%] 2025-12-04T11:24:38.1956018Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_dynamic_shape_decompose_addmm PASSED [6.3586s] [ 50%] 2025-12-04T11:24:38.1956665Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_dynamic_shape_m_20480_k_5_n_2_should_decompose_True_has_bias_False SKIPPED [0.0003s] [ 66%] 2025-12-04T11:24:38.1957306Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_dynamic_shape_m_20480_k_5_n_2_should_decompose_True_has_bias_True SKIPPED [0.0002s] [ 83%] 2025-12-04T11:24:38.1957734Z inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_realize_input PASSED [0.8071s] [100%] 2025-12-04T11:24:38.1957740Z 2025-12-04T11:24:38.1958586Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-d89a888c67de20be.xml - 2025-12-04T11:24:38.1958810Z ================= 4 passed, 2 skipped, 31 deselected in 13.32s ================= 2025-12-04T11:24:38.1958917Z --- Logging error --- 2025-12-04T11:24:38.1959037Z Traceback (most recent call last): 2025-12-04T11:24:38.1959396Z File "/opt/conda/envs/py_3.10/lib/python3.10/logging/__init__.py", line 1103, in emit 2025-12-04T11:24:38.1959525Z stream.write(msg + self.terminator) 2025-12-04T11:24:38.1959674Z ValueError: I/O operation on closed file. 2025-12-04T11:24:38.1959767Z Call stack: 2025-12-04T11:24:38.1960285Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/remote_cache.py", line 432, in dump_cache_stats 2025-12-04T11:24:38.1960447Z log.info("Cache Metrics:%s", out.getvalue()) 2025-12-04T11:24:38.1960558Z Message: 'Cache Metrics:%s' 2025-12-04T11:24:38.1961174Z Arguments: ('\n LocalAutotuneCache: {hit: 0, miss: 1, put: 2, exception: 0}\n backend:_LocalAutotuneCacheBackend: {hit: 0, miss: 1, put: 2, exception: 0}\n',) 2025-12-04T11:24:38.1964245Z The following tests failed consistently: ['test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False', 'test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True', 'test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False', 'test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True'] 2025-12-04T11:24:38.1964290Z 2025-12-04T11:24:38.1964949Z FINISHED PRINTING LOG FILE of inductor/test_decompose_mem_bound_mm 1/1 (test/test-reports/inductor.test_decompose_mem_bound_mm_1.1_7b3d736361a2b2dc_.log) 2025-12-04T11:24:38.1964954Z 2025-12-04T11:24:38.1965347Z Finished inductor/test_decompose_mem_bound_mm 1/1 ... [2025-12-04 11:24:34.210660][7857.901045769], took 4.57min 2025-12-04T11:24:38.1966291Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-819d138cb6fe4c8f.xml 2025-12-04T11:24:38.1967248Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-3d7372eabf0c20ef.xml 2025-12-04T11:24:38.1968180Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-e6eefab2abf634a5.xml 2025-12-04T11:24:38.1969088Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-fd020a6d441a5738.xml 2025-12-04T11:24:38.1969982Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-b7f552c9fcec58a4.xml 2025-12-04T11:24:38.1970891Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-ba19ebfa7960f6c4.xml 2025-12-04T11:24:38.1971794Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-6aac4defd9d3a3cc.xml 2025-12-04T11:24:38.1972703Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-041aad81ad32b1c4.xml 2025-12-04T11:24:38.1973591Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-043eb8ddb5d84fd4.xml 2025-12-04T11:24:38.1974485Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-555ea3dc15f2a8d0.xml 2025-12-04T11:24:38.1975390Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-01d96e43057cf9b8.xml 2025-12-04T11:24:38.1976315Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-563749739dc7a57a.xml 2025-12-04T11:24:38.1977294Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-d89a888c67de20be.xml 2025-12-04T11:24:38.1977416Z Uploading logs for 57119749282 to S3 2025-12-04T11:24:38.1977546Z Uploading artifacts took 1.45 seconds 2025-12-04T11:24:38.1977716Z inductor/test_decompose_mem_bound_mm 1/1 failed! 2025-12-04T11:24:38.1978028Z Running inductor/test_online_softmax 1/1 ... [2025-12-04 11:24:36.509015][7860.199406064] 2025-12-04T11:24:38.1978168Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T11:24:38.1979195Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_online_softmax.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 11:24:36.509539] 2025-12-04T11:40:52.0605240Z 2025-12-04T11:40:52.0606253Z PRINTING LOG FILE of inductor/test_online_softmax 1/1 (test/test-reports/inductor.test_online_softmax_1.1_d592e7340b8c6a36_.log) 2025-12-04T11:40:52.0607931Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a12b7a446d46a036.xml 2025-12-04T11:40:52.0609584Z ============================= test session starts ============================== 2025-12-04T11:40:52.0610609Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.0611575Z cachedir: .pytest_cache 2025-12-04T11:40:52.0612874Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.0614210Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.0614750Z configfile: pytest.ini 2025-12-04T11:40:52.0616198Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.0617652Z collecting ... collected 31 items 2025-12-04T11:40:52.0618297Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T11:40:52.0661913Z Running 31 items in this shard: test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_3d_tiled_online_softmax, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_causal_mask, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_3pass_softmax_due_to_disable, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_False, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_True, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_False, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_True, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_softmax_persistent_reduction, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_log_softmax, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_no_online_softmax_for_cpu, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_bfloat16, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_float16, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_float32, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_-1, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_0, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_1, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_-1, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_0, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_1, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_perf, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_sdpa, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_bfloat16, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_float16, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_float32, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_bfloat16, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_float16, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_float32, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmin, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_split_reduction, test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_tb_speech_transformer_attn 2025-12-04T11:40:52.0677998Z 2025-12-04T11:40:52.0678454Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_3d_tiled_online_softmax PASSED [6.7367s] [ 3%] 2025-12-04T11:40:52.0679369Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_causal_mask PASSED [0.4220s] [ 6%] 2025-12-04T11:40:52.0680437Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_3pass_softmax_due_to_disable ('RERUN', {'yellow': True}) [0.2594s] [ 9%] 2025-12-04T11:40:52.0681760Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_3pass_softmax_due_to_disable ('RERUN', {'yellow': True}) [0.0944s] [ 9%] 2025-12-04T11:40:52.0682936Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_3pass_softmax_due_to_disable FAILED [0.0896s] [ 9%] 2025-12-04T11:40:52.0683556Z 2025-12-04T11:40:52.0683699Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.0684257Z _________ TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable __________ 2025-12-04T11:40:52.0684821Z Traceback (most recent call last): 2025-12-04T11:40:52.0685605Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 82, in test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0686418Z wrapper_code = self.get_softmax_wrapper() 2025-12-04T11:40:52.0687126Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.0687809Z return source_codes[0] 2025-12-04T11:40:52.0688110Z IndexError: list index out of range 2025-12-04T11:40:52.0688352Z 2025-12-04T11:40:52.0688560Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.0689499Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0690223Z 2025-12-04T11:40:52.0690490Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.0691100Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0691557Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0691873Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0692287Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0692752Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.0693080Z graph_break [] 2025-12-04T11:40:52.0693432Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0694514Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0695521Z warnings.warn( 2025-12-04T11:40:52.0695957Z _________ TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable __________ 2025-12-04T11:40:52.0696691Z Traceback (most recent call last): 2025-12-04T11:40:52.0697534Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 82, in test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0698355Z wrapper_code = self.get_softmax_wrapper() 2025-12-04T11:40:52.0699063Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.0699743Z return source_codes[0] 2025-12-04T11:40:52.0700045Z IndexError: list index out of range 2025-12-04T11:40:52.0700272Z 2025-12-04T11:40:52.0700490Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.0701422Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0702148Z 2025-12-04T11:40:52.0702486Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.0703094Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0703544Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0703937Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0704358Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0704821Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.0705134Z graph_break [] 2025-12-04T11:40:52.0705489Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0706556Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0707549Z warnings.warn( 2025-12-04T11:40:52.0707904Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0708359Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0708673Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0709078Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0709538Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.0709882Z graph_break [] 2025-12-04T11:40:52.0710242Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0711329Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0712299Z warnings.warn( 2025-12-04T11:40:52.0712609Z =================================== FAILURES =================================== 2025-12-04T11:40:52.0713184Z _________ TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable __________ 2025-12-04T11:40:52.0713740Z Traceback (most recent call last): 2025-12-04T11:40:52.0714540Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 82, in test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0715382Z wrapper_code = self.get_softmax_wrapper() 2025-12-04T11:40:52.0716098Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.0716804Z return source_codes[0] 2025-12-04T11:40:52.0717130Z IndexError: list index out of range 2025-12-04T11:40:52.0717367Z 2025-12-04T11:40:52.0717580Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.0718541Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0719283Z 2025-12-04T11:40:52.0719549Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.0720183Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0720695Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0721033Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0721475Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0721946Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.0722291Z graph_break [] 2025-12-04T11:40:52.0722674Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0723764Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0724725Z warnings.warn( 2025-12-04T11:40:52.0725114Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0725587Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0725912Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0726347Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0726834Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.0727205Z graph_break [] 2025-12-04T11:40:52.0727575Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0728717Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0729684Z warnings.warn( 2025-12-04T11:40:52.0730055Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0730531Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0730864Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0731287Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0731765Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.0732104Z graph_break [] 2025-12-04T11:40:52.0732520Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0733599Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0734563Z warnings.warn( 2025-12-04T11:40:52.0735500Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a12b7a446d46a036.xml - 2025-12-04T11:40:52.0736568Z =========================== short test summary info ============================ 2025-12-04T11:40:52.0737608Z FAILED [0.0896s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_3pass_softmax_due_to_disable - IndexError: list index out of range 2025-12-04T11:40:52.0738373Z 2025-12-04T11:40:52.0738590Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.0739542Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0740279Z 2025-12-04T11:40:52.0740560Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.0741141Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.0741654Z ===================== 1 failed, 2 passed, 2 rerun in 7.64s ===================== 2025-12-04T11:40:52.0742087Z Got exit code 1 2025-12-04T11:40:52.0742345Z Retrying single test... 2025-12-04T11:40:52.0743125Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-c354a873b70ad9a4.xml 2025-12-04T11:40:52.0744022Z ============================= test session starts ============================== 2025-12-04T11:40:52.0744685Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.0745277Z cachedir: .pytest_cache 2025-12-04T11:40:52.0745989Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.0746837Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.0747180Z configfile: pytest.ini 2025-12-04T11:40:52.0747962Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.0748914Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.0749946Z stepcurrent: skipping 2 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0750856Z Running 1 items in this shard 2025-12-04T11:40:52.0751077Z 2025-12-04T11:40:52.0751643Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_3pass_softmax_due_to_disable ('RERUN', {'yellow': True}) [4.3549s] [100%] 2025-12-04T11:40:52.0752919Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_3pass_softmax_due_to_disable ('RERUN', {'yellow': True}) [0.0931s] [100%] 2025-12-04T11:40:52.0754108Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_3pass_softmax_due_to_disable FAILED [0.0908s] [100%] 2025-12-04T11:40:52.0754763Z 2025-12-04T11:40:52.0754906Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.0755520Z _________ TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable __________ 2025-12-04T11:40:52.0756075Z Traceback (most recent call last): 2025-12-04T11:40:52.0756872Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 82, in test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0757701Z wrapper_code = self.get_softmax_wrapper() 2025-12-04T11:40:52.0758429Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.0759133Z return source_codes[0] 2025-12-04T11:40:52.0759477Z IndexError: list index out of range 2025-12-04T11:40:52.0759722Z 2025-12-04T11:40:52.0759936Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.0760890Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0761624Z 2025-12-04T11:40:52.0761900Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.0762519Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0762991Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0763325Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0763650Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.0764114Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0764578Z graph_break [] 2025-12-04T11:40:52.0764952Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0766035Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0767021Z warnings.warn( 2025-12-04T11:40:52.0767479Z _________ TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable __________ 2025-12-04T11:40:52.0768023Z Traceback (most recent call last): 2025-12-04T11:40:52.0768821Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 82, in test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0769663Z wrapper_code = self.get_softmax_wrapper() 2025-12-04T11:40:52.0770592Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.0771286Z return source_codes[0] 2025-12-04T11:40:52.0771609Z IndexError: list index out of range 2025-12-04T11:40:52.0771845Z 2025-12-04T11:40:52.0772071Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.0773010Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0773806Z 2025-12-04T11:40:52.0774075Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.0774705Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0775181Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0775507Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0775864Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.0776333Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0776782Z graph_break [] 2025-12-04T11:40:52.0777236Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0778343Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0779323Z warnings.warn( 2025-12-04T11:40:52.0779698Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0780215Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0780555Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0780983Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0781471Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.0781859Z graph_break [] 2025-12-04T11:40:52.0782226Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0783326Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0784307Z warnings.warn( 2025-12-04T11:40:52.0784623Z =================================== FAILURES =================================== 2025-12-04T11:40:52.0785227Z _________ TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable __________ 2025-12-04T11:40:52.0785786Z Traceback (most recent call last): 2025-12-04T11:40:52.0786592Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 82, in test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0787438Z wrapper_code = self.get_softmax_wrapper() 2025-12-04T11:40:52.0788155Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.0788872Z return source_codes[0] 2025-12-04T11:40:52.0789198Z IndexError: list index out of range 2025-12-04T11:40:52.0789437Z 2025-12-04T11:40:52.0789652Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.0790606Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0791348Z 2025-12-04T11:40:52.0791612Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.0792239Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0792699Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0793033Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0793367Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.0793814Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0794275Z graph_break [] 2025-12-04T11:40:52.0794645Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0795746Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0796866Z warnings.warn( 2025-12-04T11:40:52.0797249Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0797721Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0798046Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0798485Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0799043Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.0799375Z graph_break [] 2025-12-04T11:40:52.0799748Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0800842Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0801806Z warnings.warn( 2025-12-04T11:40:52.0802174Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0802645Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0802977Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0803400Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0803884Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.0804223Z graph_break [] 2025-12-04T11:40:52.0804596Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0805678Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0806702Z warnings.warn( 2025-12-04T11:40:52.0807691Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-c354a873b70ad9a4.xml - 2025-12-04T11:40:52.0808758Z =========================== short test summary info ============================ 2025-12-04T11:40:52.0809706Z FAILED [0.0908s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_3pass_softmax_due_to_disable - IndexError: list index out of range 2025-12-04T11:40:52.0810477Z 2025-12-04T11:40:52.0810695Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.0811693Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0812425Z 2025-12-04T11:40:52.0812703Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.0813280Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.0813800Z ================== 1 failed, 30 deselected, 2 rerun in 4.57s =================== 2025-12-04T11:40:52.0814245Z Got exit code 1 2025-12-04T11:40:52.0814502Z Retrying single test... 2025-12-04T11:40:52.0815286Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-72b2d7d92c2f3d43.xml 2025-12-04T11:40:52.0816184Z ============================= test session starts ============================== 2025-12-04T11:40:52.0816854Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.0817513Z cachedir: .pytest_cache 2025-12-04T11:40:52.0818232Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.0819022Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.0819362Z configfile: pytest.ini 2025-12-04T11:40:52.0820143Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.0821091Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.0822121Z stepcurrent: skipping 2 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0823030Z Running 1 items in this shard 2025-12-04T11:40:52.0823251Z 2025-12-04T11:40:52.0823819Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_3pass_softmax_due_to_disable ('RERUN', {'yellow': True}) [4.3585s] [100%] 2025-12-04T11:40:52.0825098Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_3pass_softmax_due_to_disable ('RERUN', {'yellow': True}) [0.0935s] [100%] 2025-12-04T11:40:52.0826327Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_3pass_softmax_due_to_disable FAILED [0.0903s] [100%] 2025-12-04T11:40:52.0826949Z 2025-12-04T11:40:52.0827090Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.0827673Z _________ TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable __________ 2025-12-04T11:40:52.0828227Z Traceback (most recent call last): 2025-12-04T11:40:52.0829027Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 82, in test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0829862Z wrapper_code = self.get_softmax_wrapper() 2025-12-04T11:40:52.0830594Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.0831295Z return source_codes[0] 2025-12-04T11:40:52.0831609Z IndexError: list index out of range 2025-12-04T11:40:52.0831858Z 2025-12-04T11:40:52.0832075Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.0833027Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0833791Z 2025-12-04T11:40:52.0834068Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.0835114Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0835599Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0835937Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0836262Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.0836726Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0837186Z graph_break [] 2025-12-04T11:40:52.0837564Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0838717Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0839695Z warnings.warn( 2025-12-04T11:40:52.0840155Z _________ TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable __________ 2025-12-04T11:40:52.0840702Z Traceback (most recent call last): 2025-12-04T11:40:52.0841511Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 82, in test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0842353Z wrapper_code = self.get_softmax_wrapper() 2025-12-04T11:40:52.0843085Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.0843781Z return source_codes[0] 2025-12-04T11:40:52.0844108Z IndexError: list index out of range 2025-12-04T11:40:52.0844344Z 2025-12-04T11:40:52.0844571Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.0845518Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0846273Z 2025-12-04T11:40:52.0846541Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.0847179Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0847658Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0847985Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0848326Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.0848793Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0849237Z graph_break [] 2025-12-04T11:40:52.0849611Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0850710Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0851691Z warnings.warn( 2025-12-04T11:40:52.0852104Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0852586Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0852924Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0853345Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0853824Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.0854166Z graph_break [] 2025-12-04T11:40:52.0854521Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0855608Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0856572Z warnings.warn( 2025-12-04T11:40:52.0856949Z =================================== FAILURES =================================== 2025-12-04T11:40:52.0857533Z _________ TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable __________ 2025-12-04T11:40:52.0858087Z Traceback (most recent call last): 2025-12-04T11:40:52.0858891Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 82, in test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0859770Z wrapper_code = self.get_softmax_wrapper() 2025-12-04T11:40:52.0860519Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.0861224Z return source_codes[0] 2025-12-04T11:40:52.0861549Z IndexError: list index out of range 2025-12-04T11:40:52.0861783Z 2025-12-04T11:40:52.0861997Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.0862950Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0863695Z 2025-12-04T11:40:52.0863991Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.0864617Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0865079Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0865416Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0865754Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.0866208Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0866671Z graph_break [] 2025-12-04T11:40:52.0867048Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0868148Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0869105Z warnings.warn( 2025-12-04T11:40:52.0869489Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0869963Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0870287Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0870721Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0871203Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.0871532Z graph_break [] 2025-12-04T11:40:52.0871902Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0872995Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0873961Z warnings.warn( 2025-12-04T11:40:52.0874330Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0874803Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0875136Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0875560Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0876040Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.0876381Z graph_break [] 2025-12-04T11:40:52.0876755Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0877868Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0878841Z warnings.warn( 2025-12-04T11:40:52.0879774Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-72b2d7d92c2f3d43.xml - 2025-12-04T11:40:52.0880833Z =========================== short test summary info ============================ 2025-12-04T11:40:52.0881780Z FAILED [0.0903s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_3pass_softmax_due_to_disable - IndexError: list index out of range 2025-12-04T11:40:52.0882544Z 2025-12-04T11:40:52.0882760Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.0883711Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0884444Z 2025-12-04T11:40:52.0884721Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.0885331Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.0885855Z ================== 1 failed, 30 deselected, 2 rerun in 4.57s =================== 2025-12-04T11:40:52.0886329Z Got exit code 1 2025-12-04T11:40:52.0886995Z FAILED CONSISTENTLY: test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_3pass_softmax_due_to_disable 2025-12-04T11:40:52.0888059Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:40:52.0889203Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-c8076b5dd44238e4.xml 2025-12-04T11:40:52.0890133Z ============================= test session starts ============================== 2025-12-04T11:40:52.0890785Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.0891396Z cachedir: .pytest_cache 2025-12-04T11:40:52.0892106Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.0892876Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.0893231Z configfile: pytest.ini 2025-12-04T11:40:52.0894004Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.0894952Z collecting ... collected 31 items / 3 deselected / 28 selected 2025-12-04T11:40:52.0895432Z stepcurrent: skipping 3 already run items. 2025-12-04T11:40:52.0895821Z Running 28 items in this shard 2025-12-04T11:40:52.0896473Z 2025-12-04T11:40:52.0897236Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_False ('RERUN', {'yellow': True}) [4.3737s] [ 3%] 2025-12-04T11:40:52.0898656Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_False ('RERUN', {'yellow': True}) [0.1027s] [ 3%] 2025-12-04T11:40:52.0899968Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_False FAILED [0.1086s] [ 3%] 2025-12-04T11:40:52.0900677Z 2025-12-04T11:40:52.0900825Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.0901423Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False __ 2025-12-04T11:40:52.0901999Z Traceback (most recent call last): 2025-12-04T11:40:52.0902732Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.0903655Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.0904539Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.0905412Z return source_codes[0] 2025-12-04T11:40:52.0905749Z IndexError: list index out of range 2025-12-04T11:40:52.0905997Z 2025-12-04T11:40:52.0906213Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.0907247Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False 2025-12-04T11:40:52.0908046Z 2025-12-04T11:40:52.0908315Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.0908953Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0909435Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0909778Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0910327Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.0911044Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0911515Z graph_break [] 2025-12-04T11:40:52.0911882Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0913038Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0914075Z warnings.warn( 2025-12-04T11:40:52.0914549Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False __ 2025-12-04T11:40:52.0915113Z Traceback (most recent call last): 2025-12-04T11:40:52.0915852Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.0916771Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.0917678Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.0918381Z return source_codes[0] 2025-12-04T11:40:52.0918705Z IndexError: list index out of range 2025-12-04T11:40:52.0918941Z 2025-12-04T11:40:52.0919165Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.0920171Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False 2025-12-04T11:40:52.0920986Z 2025-12-04T11:40:52.0921251Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.0921880Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0922349Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0922673Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0923231Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.0923938Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0924387Z graph_break [] 2025-12-04T11:40:52.0924769Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0925868Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0926839Z warnings.warn( 2025-12-04T11:40:52.0927209Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0927676Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0928009Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0928432Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0929132Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.0929718Z graph_break [] 2025-12-04T11:40:52.0930080Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0931171Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0932182Z warnings.warn( 2025-12-04T11:40:52.0932494Z =================================== FAILURES =================================== 2025-12-04T11:40:52.0933086Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False __ 2025-12-04T11:40:52.0933661Z Traceback (most recent call last): 2025-12-04T11:40:52.0934404Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.0935311Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.0936191Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.0936961Z return source_codes[0] 2025-12-04T11:40:52.0937295Z IndexError: list index out of range 2025-12-04T11:40:52.0937531Z 2025-12-04T11:40:52.0937748Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.0938818Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False 2025-12-04T11:40:52.0939631Z 2025-12-04T11:40:52.0939900Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.0940566Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0941032Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0941367Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0941925Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.0942617Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0943073Z graph_break [] 2025-12-04T11:40:52.0943479Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0944576Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0945531Z warnings.warn( 2025-12-04T11:40:52.0945917Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0946392Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0946713Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0947148Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0947848Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.0948432Z graph_break [] 2025-12-04T11:40:52.0948792Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0949882Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0950847Z warnings.warn( 2025-12-04T11:40:52.0951217Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0951693Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0952030Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0952451Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0953154Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.0953739Z graph_break [] 2025-12-04T11:40:52.0954114Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0955188Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0956152Z warnings.warn( 2025-12-04T11:40:52.0957085Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-c8076b5dd44238e4.xml - 2025-12-04T11:40:52.0958197Z =========================== short test summary info ============================ 2025-12-04T11:40:52.0959189Z FAILED [0.1086s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_False - IndexError: list index out of range 2025-12-04T11:40:52.0960016Z 2025-12-04T11:40:52.0960232Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.0961249Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False 2025-12-04T11:40:52.0962052Z 2025-12-04T11:40:52.0962330Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.0962908Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.0963434Z =================== 1 failed, 3 deselected, 2 rerun in 4.62s =================== 2025-12-04T11:40:52.0963875Z Got exit code 1 2025-12-04T11:40:52.0964132Z Retrying single test... 2025-12-04T11:40:52.0964948Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-d2d17ca8c68c28e7.xml 2025-12-04T11:40:52.0965848Z ============================= test session starts ============================== 2025-12-04T11:40:52.0966549Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.0967139Z cachedir: .pytest_cache 2025-12-04T11:40:52.0967854Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.0968653Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.0968998Z configfile: pytest.ini 2025-12-04T11:40:52.0969813Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.0970768Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.0971874Z stepcurrent: skipping 3 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_False 2025-12-04T11:40:52.0972860Z Running 1 items in this shard 2025-12-04T11:40:52.0973086Z 2025-12-04T11:40:52.0973728Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_False ('RERUN', {'yellow': True}) [4.3763s] [100%] 2025-12-04T11:40:52.0975138Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_False ('RERUN', {'yellow': True}) [0.1027s] [100%] 2025-12-04T11:40:52.0976466Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_False FAILED [0.1001s] [100%] 2025-12-04T11:40:52.0977217Z 2025-12-04T11:40:52.0977376Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.0977970Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False __ 2025-12-04T11:40:52.0978547Z Traceback (most recent call last): 2025-12-04T11:40:52.0979283Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.0980192Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.0981075Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.0981783Z return source_codes[0] 2025-12-04T11:40:52.0982110Z IndexError: list index out of range 2025-12-04T11:40:52.0982343Z 2025-12-04T11:40:52.0982556Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.0983583Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False 2025-12-04T11:40:52.0984419Z 2025-12-04T11:40:52.0984697Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.0985326Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0985785Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0986116Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0986675Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.0987363Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.0987817Z graph_break [] 2025-12-04T11:40:52.0988198Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.0989284Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.0990260Z warnings.warn( 2025-12-04T11:40:52.0990734Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False __ 2025-12-04T11:40:52.0991338Z Traceback (most recent call last): 2025-12-04T11:40:52.0992059Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.0992976Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.0993907Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.0994613Z return source_codes[0] 2025-12-04T11:40:52.0994924Z IndexError: list index out of range 2025-12-04T11:40:52.0995172Z 2025-12-04T11:40:52.0995385Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.0996692Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False 2025-12-04T11:40:52.0997501Z 2025-12-04T11:40:52.0997774Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.0998414Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.0998895Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.0999239Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.0999794Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1000506Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1000978Z graph_break [] 2025-12-04T11:40:52.1001341Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1002435Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1003402Z warnings.warn( 2025-12-04T11:40:52.1003785Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1004246Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1004584Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1005021Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1005711Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1006299Z graph_break [] 2025-12-04T11:40:52.1006672Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1007765Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1008724Z warnings.warn( 2025-12-04T11:40:52.1009032Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1009639Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False __ 2025-12-04T11:40:52.1010194Z Traceback (most recent call last): 2025-12-04T11:40:52.1010975Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1011890Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1012768Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1013457Z return source_codes[0] 2025-12-04T11:40:52.1013783Z IndexError: list index out of range 2025-12-04T11:40:52.1014014Z 2025-12-04T11:40:52.1014240Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1015248Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False 2025-12-04T11:40:52.1016060Z 2025-12-04T11:40:52.1016327Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1017015Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1017500Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1017876Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1018437Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1019197Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1019666Z graph_break [] 2025-12-04T11:40:52.1020030Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1021135Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1022111Z warnings.warn( 2025-12-04T11:40:52.1022481Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1022988Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1023327Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1023753Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1024456Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1025038Z graph_break [] 2025-12-04T11:40:52.1025414Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1026496Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1027470Z warnings.warn( 2025-12-04T11:40:52.1027857Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1028315Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1028656Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1029098Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1029804Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1030384Z graph_break [] 2025-12-04T11:40:52.1030765Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1031862Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1032825Z warnings.warn( 2025-12-04T11:40:52.1033770Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-d2d17ca8c68c28e7.xml - 2025-12-04T11:40:52.1034864Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1035870Z FAILED [0.1001s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_False - IndexError: list index out of range 2025-12-04T11:40:52.1036687Z 2025-12-04T11:40:52.1036921Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1037967Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False 2025-12-04T11:40:52.1038782Z 2025-12-04T11:40:52.1039050Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1039646Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1040153Z ================== 1 failed, 30 deselected, 2 rerun in 4.61s =================== 2025-12-04T11:40:52.1040600Z Got exit code 1 2025-12-04T11:40:52.1040874Z Retrying single test... 2025-12-04T11:40:52.1041663Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-b56c2dd706b17a20.xml 2025-12-04T11:40:52.1042548Z ============================= test session starts ============================== 2025-12-04T11:40:52.1043210Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1043841Z cachedir: .pytest_cache 2025-12-04T11:40:52.1044541Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1045330Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1045717Z configfile: pytest.ini 2025-12-04T11:40:52.1046499Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1047432Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.1048531Z stepcurrent: skipping 3 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_False 2025-12-04T11:40:52.1049529Z Running 1 items in this shard 2025-12-04T11:40:52.1049775Z 2025-12-04T11:40:52.1050422Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_False ('RERUN', {'yellow': True}) [4.3552s] [100%] 2025-12-04T11:40:52.1051832Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_False ('RERUN', {'yellow': True}) [0.1013s] [100%] 2025-12-04T11:40:52.1053165Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_False FAILED [0.0991s] [100%] 2025-12-04T11:40:52.1053858Z 2025-12-04T11:40:52.1053998Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1054592Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False __ 2025-12-04T11:40:52.1055149Z Traceback (most recent call last): 2025-12-04T11:40:52.1055895Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1056817Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1057770Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1058465Z return source_codes[0] 2025-12-04T11:40:52.1058787Z IndexError: list index out of range 2025-12-04T11:40:52.1059020Z 2025-12-04T11:40:52.1059246Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1060260Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False 2025-12-04T11:40:52.1061070Z 2025-12-04T11:40:52.1061337Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1061965Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1062438Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1062761Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1063326Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1064084Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1064539Z graph_break [] 2025-12-04T11:40:52.1064919Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1066022Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1067006Z warnings.warn( 2025-12-04T11:40:52.1067467Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False __ 2025-12-04T11:40:52.1068041Z Traceback (most recent call last): 2025-12-04T11:40:52.1068772Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1069679Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1070558Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1071307Z return source_codes[0] 2025-12-04T11:40:52.1071630Z IndexError: list index out of range 2025-12-04T11:40:52.1071864Z 2025-12-04T11:40:52.1072078Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1073137Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False 2025-12-04T11:40:52.1073951Z 2025-12-04T11:40:52.1074216Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1074844Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1075304Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1075641Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1076233Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1076935Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1077398Z graph_break [] 2025-12-04T11:40:52.1077772Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1078861Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1079818Z warnings.warn( 2025-12-04T11:40:52.1080197Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1080667Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1080995Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1081426Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1082127Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1082712Z graph_break [] 2025-12-04T11:40:52.1083072Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1084170Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1085141Z warnings.warn( 2025-12-04T11:40:52.1085447Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1086051Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False __ 2025-12-04T11:40:52.1086621Z Traceback (most recent call last): 2025-12-04T11:40:52.1087354Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1088258Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1089139Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1089837Z return source_codes[0] 2025-12-04T11:40:52.1090188Z IndexError: list index out of range 2025-12-04T11:40:52.1090434Z 2025-12-04T11:40:52.1090645Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1091660Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False 2025-12-04T11:40:52.1092460Z 2025-12-04T11:40:52.1092736Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1093355Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1093830Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1094167Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1094713Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1095426Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1095891Z graph_break [] 2025-12-04T11:40:52.1096486Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1097736Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1098709Z warnings.warn( 2025-12-04T11:40:52.1099158Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1099642Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1099971Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1100414Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1101125Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1101698Z graph_break [] 2025-12-04T11:40:52.1102122Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1103215Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1104180Z warnings.warn( 2025-12-04T11:40:52.1104565Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1105045Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1105384Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1105806Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1106510Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1107098Z graph_break [] 2025-12-04T11:40:52.1107455Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1108547Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1109510Z warnings.warn( 2025-12-04T11:40:52.1110451Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-b56c2dd706b17a20.xml - 2025-12-04T11:40:52.1111516Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1112522Z FAILED [0.0991s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_False - IndexError: list index out of range 2025-12-04T11:40:52.1113348Z 2025-12-04T11:40:52.1113564Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1114586Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_False 2025-12-04T11:40:52.1115387Z 2025-12-04T11:40:52.1115657Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1116249Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1116820Z ================== 1 failed, 30 deselected, 2 rerun in 4.59s =================== 2025-12-04T11:40:52.1117265Z Got exit code 1 2025-12-04T11:40:52.1117999Z FAILED CONSISTENTLY: test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_False 2025-12-04T11:40:52.1119131Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:40:52.1120288Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-fa3c7a9046cc8737.xml 2025-12-04T11:40:52.1121183Z ============================= test session starts ============================== 2025-12-04T11:40:52.1121833Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1122435Z cachedir: .pytest_cache 2025-12-04T11:40:52.1123156Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1123964Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1124315Z configfile: pytest.ini 2025-12-04T11:40:52.1125089Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1126084Z collecting ... collected 31 items / 4 deselected / 27 selected 2025-12-04T11:40:52.1126570Z stepcurrent: skipping 4 already run items. 2025-12-04T11:40:52.1126952Z Running 27 items in this shard 2025-12-04T11:40:52.1127163Z 2025-12-04T11:40:52.1127817Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_True ('RERUN', {'yellow': True}) [4.3637s] [ 3%] 2025-12-04T11:40:52.1129241Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_True ('RERUN', {'yellow': True}) [0.1035s] [ 3%] 2025-12-04T11:40:52.1130566Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_True FAILED [0.1022s] [ 3%] 2025-12-04T11:40:52.1131261Z 2025-12-04T11:40:52.1131404Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1131999Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True ___ 2025-12-04T11:40:52.1132563Z Traceback (most recent call last): 2025-12-04T11:40:52.1133299Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1134217Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1135092Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1135781Z return source_codes[0] 2025-12-04T11:40:52.1136101Z IndexError: list index out of range 2025-12-04T11:40:52.1136333Z 2025-12-04T11:40:52.1136559Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1137653Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True 2025-12-04T11:40:52.1138450Z 2025-12-04T11:40:52.1138719Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1139355Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1139831Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1140152Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1140710Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1141421Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1141885Z graph_break [] 2025-12-04T11:40:52.1142247Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1143351Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1144380Z warnings.warn( 2025-12-04T11:40:52.1144836Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True ___ 2025-12-04T11:40:52.1145405Z Traceback (most recent call last): 2025-12-04T11:40:52.1146139Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1147058Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1147917Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1148621Z return source_codes[0] 2025-12-04T11:40:52.1148942Z IndexError: list index out of range 2025-12-04T11:40:52.1149177Z 2025-12-04T11:40:52.1149392Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1150407Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True 2025-12-04T11:40:52.1151251Z 2025-12-04T11:40:52.1151518Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1153362Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1153822Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1154159Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1154720Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1155425Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1155871Z graph_break [] 2025-12-04T11:40:52.1156245Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1157387Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1158356Z warnings.warn( 2025-12-04T11:40:52.1158741Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1159220Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1159561Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1159991Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1160697Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1161287Z graph_break [] 2025-12-04T11:40:52.1161651Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1162755Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1163743Z warnings.warn( 2025-12-04T11:40:52.1164062Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1164653Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True ___ 2025-12-04T11:40:52.1165224Z Traceback (most recent call last): 2025-12-04T11:40:52.1165966Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1166877Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1167761Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1168471Z return source_codes[0] 2025-12-04T11:40:52.1168801Z IndexError: list index out of range 2025-12-04T11:40:52.1169038Z 2025-12-04T11:40:52.1169254Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1170276Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True 2025-12-04T11:40:52.1171114Z 2025-12-04T11:40:52.1171397Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1172023Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1172490Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1172828Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1173384Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1174079Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1174536Z graph_break [] 2025-12-04T11:40:52.1174908Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1175992Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1177040Z warnings.warn( 2025-12-04T11:40:52.1177423Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1177945Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1178266Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1178704Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1179440Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1180016Z graph_break [] 2025-12-04T11:40:52.1180391Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1181485Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1182456Z warnings.warn( 2025-12-04T11:40:52.1182855Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1183329Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1183665Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1184090Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1184788Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1185379Z graph_break [] 2025-12-04T11:40:52.1185755Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1186832Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1187802Z warnings.warn( 2025-12-04T11:40:52.1188741Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-fa3c7a9046cc8737.xml - 2025-12-04T11:40:52.1189815Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1190820Z FAILED [0.1022s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_True - IndexError: list index out of range 2025-12-04T11:40:52.1191647Z 2025-12-04T11:40:52.1191864Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1192883Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True 2025-12-04T11:40:52.1193681Z 2025-12-04T11:40:52.1193960Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1194536Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1195055Z =================== 1 failed, 4 deselected, 2 rerun in 4.60s =================== 2025-12-04T11:40:52.1195494Z Got exit code 1 2025-12-04T11:40:52.1195749Z Retrying single test... 2025-12-04T11:40:52.1196759Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3da5ad44ee5d80af.xml 2025-12-04T11:40:52.1197767Z ============================= test session starts ============================== 2025-12-04T11:40:52.1198436Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1199026Z cachedir: .pytest_cache 2025-12-04T11:40:52.1199740Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1200530Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1200869Z configfile: pytest.ini 2025-12-04T11:40:52.1201649Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1202591Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.1203683Z stepcurrent: skipping 4 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_True 2025-12-04T11:40:52.1204713Z Running 1 items in this shard 2025-12-04T11:40:52.1204936Z 2025-12-04T11:40:52.1205565Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_True ('RERUN', {'yellow': True}) [4.3857s] [100%] 2025-12-04T11:40:52.1207012Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_True ('RERUN', {'yellow': True}) [0.1021s] [100%] 2025-12-04T11:40:52.1208328Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_True FAILED [0.1032s] [100%] 2025-12-04T11:40:52.1209016Z 2025-12-04T11:40:52.1209155Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1209792Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True ___ 2025-12-04T11:40:52.1210364Z Traceback (most recent call last): 2025-12-04T11:40:52.1211102Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1212009Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1212885Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1213592Z return source_codes[0] 2025-12-04T11:40:52.1213906Z IndexError: list index out of range 2025-12-04T11:40:52.1214157Z 2025-12-04T11:40:52.1214371Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1215404Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True 2025-12-04T11:40:52.1216199Z 2025-12-04T11:40:52.1216477Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1217190Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1217669Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1218009Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1218567Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1233798Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1234399Z graph_break [] 2025-12-04T11:40:52.1234783Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1235881Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1236850Z warnings.warn( 2025-12-04T11:40:52.1237315Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True ___ 2025-12-04T11:40:52.1237869Z Traceback (most recent call last): 2025-12-04T11:40:52.1238601Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1239617Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1240491Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1241169Z return source_codes[0] 2025-12-04T11:40:52.1241483Z IndexError: list index out of range 2025-12-04T11:40:52.1241717Z 2025-12-04T11:40:52.1241936Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1242937Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True 2025-12-04T11:40:52.1243725Z 2025-12-04T11:40:52.1243984Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1244605Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1245069Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1245388Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1245974Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1246666Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1247113Z graph_break [] 2025-12-04T11:40:52.1247503Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1248601Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1249559Z warnings.warn( 2025-12-04T11:40:52.1249922Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1250378Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1250700Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1251155Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1251836Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1252406Z graph_break [] 2025-12-04T11:40:52.1252760Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1253824Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1254775Z warnings.warn( 2025-12-04T11:40:52.1255070Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1255656Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True ___ 2025-12-04T11:40:52.1256197Z Traceback (most recent call last): 2025-12-04T11:40:52.1257017Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1257925Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1258773Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1259469Z return source_codes[0] 2025-12-04T11:40:52.1259778Z IndexError: list index out of range 2025-12-04T11:40:52.1260008Z 2025-12-04T11:40:52.1260230Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1261227Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True 2025-12-04T11:40:52.1262020Z 2025-12-04T11:40:52.1262280Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1262896Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1263355Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1263669Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1264215Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1264956Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1265395Z graph_break [] 2025-12-04T11:40:52.1265754Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1266833Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1267796Z warnings.warn( 2025-12-04T11:40:52.1268159Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1268619Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1268943Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1269358Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1270048Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1270619Z graph_break [] 2025-12-04T11:40:52.1270976Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1272102Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1273087Z warnings.warn( 2025-12-04T11:40:52.1273458Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1273906Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1274231Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1274659Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1275335Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1275902Z graph_break [] 2025-12-04T11:40:52.1276294Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1277380Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1278327Z warnings.warn( 2025-12-04T11:40:52.1279254Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3da5ad44ee5d80af.xml - 2025-12-04T11:40:52.1280327Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1281312Z FAILED [0.1032s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_True - IndexError: list index out of range 2025-12-04T11:40:52.1282114Z 2025-12-04T11:40:52.1282327Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1283328Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True 2025-12-04T11:40:52.1284121Z 2025-12-04T11:40:52.1284388Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1284969Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1285473Z ================== 1 failed, 30 deselected, 2 rerun in 4.62s =================== 2025-12-04T11:40:52.1285904Z Got exit code 1 2025-12-04T11:40:52.1286153Z Retrying single test... 2025-12-04T11:40:52.1286912Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2586fe9efcb2d9c1.xml 2025-12-04T11:40:52.1287801Z ============================= test session starts ============================== 2025-12-04T11:40:52.1288445Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1289029Z cachedir: .pytest_cache 2025-12-04T11:40:52.1289714Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1290511Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1290849Z configfile: pytest.ini 2025-12-04T11:40:52.1291611Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1292557Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.1293650Z stepcurrent: skipping 4 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_True 2025-12-04T11:40:52.1294635Z Running 1 items in this shard 2025-12-04T11:40:52.1294846Z 2025-12-04T11:40:52.1295478Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_True ('RERUN', {'yellow': True}) [4.3494s] [100%] 2025-12-04T11:40:52.1297158Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_True ('RERUN', {'yellow': True}) [0.1017s] [100%] 2025-12-04T11:40:52.1298658Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_True FAILED [0.0990s] [100%] 2025-12-04T11:40:52.1299340Z 2025-12-04T11:40:52.1299499Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1300134Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True ___ 2025-12-04T11:40:52.1300710Z Traceback (most recent call last): 2025-12-04T11:40:52.1301451Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1302377Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1303241Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1303989Z return source_codes[0] 2025-12-04T11:40:52.1304320Z IndexError: list index out of range 2025-12-04T11:40:52.1304557Z 2025-12-04T11:40:52.1304774Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1305805Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True 2025-12-04T11:40:52.1306619Z 2025-12-04T11:40:52.1306889Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1307526Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1307995Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1308338Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1308902Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1309619Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1310070Z graph_break [] 2025-12-04T11:40:52.1310451Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1311563Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1312528Z warnings.warn( 2025-12-04T11:40:52.1313009Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True ___ 2025-12-04T11:40:52.1313585Z Traceback (most recent call last): 2025-12-04T11:40:52.1314316Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1315219Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1316091Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1316791Z return source_codes[0] 2025-12-04T11:40:52.1317103Z IndexError: list index out of range 2025-12-04T11:40:52.1317395Z 2025-12-04T11:40:52.1317606Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1318628Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True 2025-12-04T11:40:52.1319420Z 2025-12-04T11:40:52.1319696Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1320310Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1320782Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1321123Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1321680Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1322375Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1322834Z graph_break [] 2025-12-04T11:40:52.1323210Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1324297Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1325301Z warnings.warn( 2025-12-04T11:40:52.1325685Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1326184Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1326507Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1326945Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1327646Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1328217Z graph_break [] 2025-12-04T11:40:52.1328592Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1329720Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1330694Z warnings.warn( 2025-12-04T11:40:52.1330995Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1331594Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True ___ 2025-12-04T11:40:52.1332158Z Traceback (most recent call last): 2025-12-04T11:40:52.1332886Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1333802Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1334676Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1335376Z return source_codes[0] 2025-12-04T11:40:52.1335685Z IndexError: list index out of range 2025-12-04T11:40:52.1335931Z 2025-12-04T11:40:52.1336146Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1337238Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True 2025-12-04T11:40:52.1338039Z 2025-12-04T11:40:52.1338305Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1338937Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1339416Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1339757Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1340302Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1341011Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1341475Z graph_break [] 2025-12-04T11:40:52.1341835Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1342936Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1343951Z warnings.warn( 2025-12-04T11:40:52.1344331Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1344790Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1345130Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1345573Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1346260Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1346846Z graph_break [] 2025-12-04T11:40:52.1347219Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1348306Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1349261Z warnings.warn( 2025-12-04T11:40:52.1349642Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1350117Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1350471Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1350913Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1351650Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1352238Z graph_break [] 2025-12-04T11:40:52.1352601Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1353692Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1354662Z warnings.warn( 2025-12-04T11:40:52.1355620Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2586fe9efcb2d9c1.xml - 2025-12-04T11:40:52.1356706Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1357708Z FAILED [0.0990s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_True - IndexError: list index out of range 2025-12-04T11:40:52.1358516Z 2025-12-04T11:40:52.1358744Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1359748Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_2048_use_log_softmax_True 2025-12-04T11:40:52.1360559Z 2025-12-04T11:40:52.1360826Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1361419Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1361943Z ================== 1 failed, 30 deselected, 2 rerun in 4.58s =================== 2025-12-04T11:40:52.1362370Z Got exit code 1 2025-12-04T11:40:52.1363103Z FAILED CONSISTENTLY: test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_True 2025-12-04T11:40:52.1364230Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:40:52.1365373Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-4d9d69bfeda71662.xml 2025-12-04T11:40:52.1366258Z ============================= test session starts ============================== 2025-12-04T11:40:52.1366912Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1367514Z cachedir: .pytest_cache 2025-12-04T11:40:52.1368202Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1368985Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1369334Z configfile: pytest.ini 2025-12-04T11:40:52.1370110Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1371090Z collecting ... collected 31 items / 5 deselected / 26 selected 2025-12-04T11:40:52.1371582Z stepcurrent: skipping 5 already run items. 2025-12-04T11:40:52.1371969Z Running 26 items in this shard 2025-12-04T11:40:52.1372175Z 2025-12-04T11:40:52.1372831Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_False ('RERUN', {'yellow': True}) [4.3641s] [ 3%] 2025-12-04T11:40:52.1374234Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_False ('RERUN', {'yellow': True}) [0.1016s] [ 3%] 2025-12-04T11:40:52.1375556Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_False FAILED [0.0991s] [ 3%] 2025-12-04T11:40:52.1376239Z 2025-12-04T11:40:52.1376391Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1377052Z _ TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False __ 2025-12-04T11:40:52.1377653Z Traceback (most recent call last): 2025-12-04T11:40:52.1378383Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1379324Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1380182Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1380879Z return source_codes[0] 2025-12-04T11:40:52.1381195Z IndexError: list index out of range 2025-12-04T11:40:52.1381424Z 2025-12-04T11:40:52.1381646Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1382685Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False 2025-12-04T11:40:52.1383498Z 2025-12-04T11:40:52.1383760Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1384381Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1384852Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1385164Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1385713Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1385935Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1386044Z graph_break [] 2025-12-04T11:40:52.1386260Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1386996Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1387104Z warnings.warn( 2025-12-04T11:40:52.1387412Z _ TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False __ 2025-12-04T11:40:52.1387540Z Traceback (most recent call last): 2025-12-04T11:40:52.1388030Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1388305Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1388764Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1388867Z return source_codes[0] 2025-12-04T11:40:52.1388983Z IndexError: list index out of range 2025-12-04T11:40:52.1389000Z 2025-12-04T11:40:52.1389213Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1389881Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False 2025-12-04T11:40:52.1389887Z 2025-12-04T11:40:52.1390192Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1390409Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1390515Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1390636Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1390980Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1391206Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1391302Z graph_break [] 2025-12-04T11:40:52.1391513Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1392257Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1392355Z warnings.warn( 2025-12-04T11:40:52.1392568Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1392688Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1392832Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1393057Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1393393Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1393522Z graph_break [] 2025-12-04T11:40:52.1393747Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1394472Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1394570Z warnings.warn( 2025-12-04T11:40:52.1394723Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1395052Z _ TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False __ 2025-12-04T11:40:52.1395179Z Traceback (most recent call last): 2025-12-04T11:40:52.1395674Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1395951Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1396637Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1396749Z return source_codes[0] 2025-12-04T11:40:52.1396870Z IndexError: list index out of range 2025-12-04T11:40:52.1396887Z 2025-12-04T11:40:52.1397104Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1397771Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False 2025-12-04T11:40:52.1397776Z 2025-12-04T11:40:52.1398055Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1398271Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1398383Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1398509Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1398849Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1399081Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1399177Z graph_break [] 2025-12-04T11:40:52.1399389Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1400134Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1400235Z warnings.warn( 2025-12-04T11:40:52.1400447Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1400565Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1400676Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1400973Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1401315Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1401411Z graph_break [] 2025-12-04T11:40:52.1401635Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1402359Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1402459Z warnings.warn( 2025-12-04T11:40:52.1402682Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1402795Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1402925Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1403139Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1403475Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1403623Z graph_break [] 2025-12-04T11:40:52.1403832Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1404597Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1404703Z warnings.warn( 2025-12-04T11:40:52.1405474Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-4d9d69bfeda71662.xml - 2025-12-04T11:40:52.1405651Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1406379Z FAILED [0.0991s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_False - IndexError: list index out of range 2025-12-04T11:40:52.1406385Z 2025-12-04T11:40:52.1406600Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1407281Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False 2025-12-04T11:40:52.1407289Z 2025-12-04T11:40:52.1407551Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1407738Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1407931Z =================== 1 failed, 5 deselected, 2 rerun in 4.60s =================== 2025-12-04T11:40:52.1408025Z Got exit code 1 2025-12-04T11:40:52.1408139Z Retrying single test... 2025-12-04T11:40:52.1408726Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0045cf67af4f7e36.xml 2025-12-04T11:40:52.1408896Z ============================= test session starts ============================== 2025-12-04T11:40:52.1409242Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1409351Z cachedir: .pytest_cache 2025-12-04T11:40:52.1409879Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1410001Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1410106Z configfile: pytest.ini 2025-12-04T11:40:52.1410702Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1410917Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.1411671Z stepcurrent: skipping 5 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_False 2025-12-04T11:40:52.1411786Z Running 1 items in this shard 2025-12-04T11:40:52.1411791Z 2025-12-04T11:40:52.1412430Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_False ('RERUN', {'yellow': True}) [4.3683s] [100%] 2025-12-04T11:40:52.1413101Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_False ('RERUN', {'yellow': True}) [0.1038s] [100%] 2025-12-04T11:40:52.1413653Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_False FAILED [0.0984s] [100%] 2025-12-04T11:40:52.1413659Z 2025-12-04T11:40:52.1413803Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1414111Z _ TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False __ 2025-12-04T11:40:52.1414229Z Traceback (most recent call last): 2025-12-04T11:40:52.1414733Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1415010Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1415471Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1415610Z return source_codes[0] 2025-12-04T11:40:52.1415731Z IndexError: list index out of range 2025-12-04T11:40:52.1415736Z 2025-12-04T11:40:52.1415988Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1416656Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False 2025-12-04T11:40:52.1416662Z 2025-12-04T11:40:52.1416998Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1417216Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1417324Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1417497Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1417838Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1418059Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1418168Z graph_break [] 2025-12-04T11:40:52.1418377Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1419124Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1419220Z warnings.warn( 2025-12-04T11:40:52.1419527Z _ TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False __ 2025-12-04T11:40:52.1419654Z Traceback (most recent call last): 2025-12-04T11:40:52.1420150Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1420433Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1420897Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1421009Z return source_codes[0] 2025-12-04T11:40:52.1421140Z IndexError: list index out of range 2025-12-04T11:40:52.1421145Z 2025-12-04T11:40:52.1421355Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1422024Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False 2025-12-04T11:40:52.1422030Z 2025-12-04T11:40:52.1422308Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1422523Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1422647Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1422760Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1423099Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1423360Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1423458Z graph_break [] 2025-12-04T11:40:52.1423673Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1424419Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1424521Z warnings.warn( 2025-12-04T11:40:52.1424743Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1424855Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1424965Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1425192Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1425534Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1425627Z graph_break [] 2025-12-04T11:40:52.1425850Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1426604Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1426713Z warnings.warn( 2025-12-04T11:40:52.1426886Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1427190Z _ TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False __ 2025-12-04T11:40:52.1427317Z Traceback (most recent call last): 2025-12-04T11:40:52.1427811Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1428084Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1428582Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1428697Z return source_codes[0] 2025-12-04T11:40:52.1428826Z IndexError: list index out of range 2025-12-04T11:40:52.1428833Z 2025-12-04T11:40:52.1429045Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1429714Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False 2025-12-04T11:40:52.1429720Z 2025-12-04T11:40:52.1429991Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1430205Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1430319Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1430431Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1430769Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1430998Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1431097Z graph_break [] 2025-12-04T11:40:52.1431307Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1432046Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1432147Z warnings.warn( 2025-12-04T11:40:52.1432370Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1432475Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1432587Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1432812Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1433151Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1433246Z graph_break [] 2025-12-04T11:40:52.1433466Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1434189Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1434331Z warnings.warn( 2025-12-04T11:40:52.1434539Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1434645Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1434764Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1434980Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1435315Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1435417Z graph_break [] 2025-12-04T11:40:52.1435625Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1436349Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1436453Z warnings.warn( 2025-12-04T11:40:52.1437222Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0045cf67af4f7e36.xml - 2025-12-04T11:40:52.1437429Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1438144Z FAILED [0.0984s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_False - IndexError: list index out of range 2025-12-04T11:40:52.1438150Z 2025-12-04T11:40:52.1438379Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1439043Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False 2025-12-04T11:40:52.1439048Z 2025-12-04T11:40:52.1439340Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1439534Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1439729Z ================== 1 failed, 30 deselected, 2 rerun in 4.60s =================== 2025-12-04T11:40:52.1439826Z Got exit code 1 2025-12-04T11:40:52.1439942Z Retrying single test... 2025-12-04T11:40:52.1440534Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-38fbcefce9021f2a.xml 2025-12-04T11:40:52.1440702Z ============================= test session starts ============================== 2025-12-04T11:40:52.1441044Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1441150Z cachedir: .pytest_cache 2025-12-04T11:40:52.1441680Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1441799Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1441906Z configfile: pytest.ini 2025-12-04T11:40:52.1442509Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1442721Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.1443478Z stepcurrent: skipping 5 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_False 2025-12-04T11:40:52.1443587Z Running 1 items in this shard 2025-12-04T11:40:52.1443592Z 2025-12-04T11:40:52.1444222Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_False ('RERUN', {'yellow': True}) [4.3722s] [100%] 2025-12-04T11:40:52.1444865Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_False ('RERUN', {'yellow': True}) [0.1031s] [100%] 2025-12-04T11:40:52.1445416Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_False FAILED [0.0990s] [100%] 2025-12-04T11:40:52.1445452Z 2025-12-04T11:40:52.1445601Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1445908Z _ TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False __ 2025-12-04T11:40:52.1446023Z Traceback (most recent call last): 2025-12-04T11:40:52.1446528Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1446804Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1447258Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1447368Z return source_codes[0] 2025-12-04T11:40:52.1447488Z IndexError: list index out of range 2025-12-04T11:40:52.1447493Z 2025-12-04T11:40:52.1447718Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1448390Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False 2025-12-04T11:40:52.1448424Z 2025-12-04T11:40:52.1448695Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1448939Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1449049Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1449173Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1449513Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1449731Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1449835Z graph_break [] 2025-12-04T11:40:52.1450049Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1450814Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1450918Z warnings.warn( 2025-12-04T11:40:52.1451225Z _ TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False __ 2025-12-04T11:40:52.1451353Z Traceback (most recent call last): 2025-12-04T11:40:52.1451847Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1452132Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1452578Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1452687Z return source_codes[0] 2025-12-04T11:40:52.1452815Z IndexError: list index out of range 2025-12-04T11:40:52.1452820Z 2025-12-04T11:40:52.1453034Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1453705Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False 2025-12-04T11:40:52.1453724Z 2025-12-04T11:40:52.1453987Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1454202Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1454320Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1454428Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1454764Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1454987Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1455079Z graph_break [] 2025-12-04T11:40:52.1455289Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1456028Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1456153Z warnings.warn( 2025-12-04T11:40:52.1456373Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1456479Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1456591Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1456812Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1457222Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1457322Z graph_break [] 2025-12-04T11:40:52.1457545Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1458267Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1458377Z warnings.warn( 2025-12-04T11:40:52.1458520Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1458822Z _ TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False __ 2025-12-04T11:40:52.1458991Z Traceback (most recent call last): 2025-12-04T11:40:52.1459485Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1459815Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1460276Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1460382Z return source_codes[0] 2025-12-04T11:40:52.1460510Z IndexError: list index out of range 2025-12-04T11:40:52.1460515Z 2025-12-04T11:40:52.1460728Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1461421Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False 2025-12-04T11:40:52.1461428Z 2025-12-04T11:40:52.1461702Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1461915Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1462031Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1462143Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1462478Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1462706Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1462801Z graph_break [] 2025-12-04T11:40:52.1463012Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1463750Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1463853Z warnings.warn( 2025-12-04T11:40:52.1464071Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1464182Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1464293Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1464514Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1464848Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1464946Z graph_break [] 2025-12-04T11:40:52.1465159Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1465880Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1465978Z warnings.warn( 2025-12-04T11:40:52.1466182Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1466287Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1466398Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1466643Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1466974Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1467067Z graph_break [] 2025-12-04T11:40:52.1467272Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1468001Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1468094Z warnings.warn( 2025-12-04T11:40:52.1468861Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-38fbcefce9021f2a.xml - 2025-12-04T11:40:52.1469036Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1469722Z FAILED [0.0990s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_False - IndexError: list index out of range 2025-12-04T11:40:52.1469760Z 2025-12-04T11:40:52.1469988Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1470654Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_False 2025-12-04T11:40:52.1470688Z 2025-12-04T11:40:52.1470960Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1471147Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1471338Z ================== 1 failed, 30 deselected, 2 rerun in 4.60s =================== 2025-12-04T11:40:52.1471439Z Got exit code 1 2025-12-04T11:40:52.1472057Z FAILED CONSISTENTLY: test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_False 2025-12-04T11:40:52.1472462Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:40:52.1473057Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0625eb2c44495ce0.xml 2025-12-04T11:40:52.1473216Z ============================= test session starts ============================== 2025-12-04T11:40:52.1473581Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1473691Z cachedir: .pytest_cache 2025-12-04T11:40:52.1474208Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1474344Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1474452Z configfile: pytest.ini 2025-12-04T11:40:52.1475044Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1475270Z collecting ... collected 31 items / 6 deselected / 25 selected 2025-12-04T11:40:52.1475415Z stepcurrent: skipping 6 already run items. 2025-12-04T11:40:52.1475539Z Running 25 items in this shard 2025-12-04T11:40:52.1475545Z 2025-12-04T11:40:52.1476182Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_True ('RERUN', {'yellow': True}) [4.3973s] [ 4%] 2025-12-04T11:40:52.1476822Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_True ('RERUN', {'yellow': True}) [0.1055s] [ 4%] 2025-12-04T11:40:52.1477382Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_True FAILED [0.1003s] [ 4%] 2025-12-04T11:40:52.1477387Z 2025-12-04T11:40:52.1477528Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1477841Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True __ 2025-12-04T11:40:52.1477993Z Traceback (most recent call last): 2025-12-04T11:40:52.1478488Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1478779Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1479232Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1479342Z return source_codes[0] 2025-12-04T11:40:52.1479475Z IndexError: list index out of range 2025-12-04T11:40:52.1479481Z 2025-12-04T11:40:52.1479696Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1480366Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True 2025-12-04T11:40:52.1480371Z 2025-12-04T11:40:52.1480642Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1480862Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1481016Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1481132Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1481489Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1481739Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1481841Z graph_break [] 2025-12-04T11:40:52.1482065Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1482793Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1482893Z warnings.warn( 2025-12-04T11:40:52.1483213Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True __ 2025-12-04T11:40:52.1483367Z Traceback (most recent call last): 2025-12-04T11:40:52.1483874Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1484152Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1484603Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1484723Z return source_codes[0] 2025-12-04T11:40:52.1484841Z IndexError: list index out of range 2025-12-04T11:40:52.1484846Z 2025-12-04T11:40:52.1485060Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1485738Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True 2025-12-04T11:40:52.1485743Z 2025-12-04T11:40:52.1486009Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1486240Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1486350Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1486468Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1486819Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1487034Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1487146Z graph_break [] 2025-12-04T11:40:52.1487356Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1488083Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1488196Z warnings.warn( 2025-12-04T11:40:52.1488404Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1488508Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1488634Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1488850Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1489230Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1489327Z graph_break [] 2025-12-04T11:40:52.1489536Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1490271Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1490370Z warnings.warn( 2025-12-04T11:40:52.1490508Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1490824Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True __ 2025-12-04T11:40:52.1490942Z Traceback (most recent call last): 2025-12-04T11:40:52.1491450Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1491728Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1492206Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1492328Z return source_codes[0] 2025-12-04T11:40:52.1492448Z IndexError: list index out of range 2025-12-04T11:40:52.1492453Z 2025-12-04T11:40:52.1492696Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1493375Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True 2025-12-04T11:40:52.1493381Z 2025-12-04T11:40:52.1493647Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1493873Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1494012Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1494124Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1494476Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1494697Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1494800Z graph_break [] 2025-12-04T11:40:52.1495014Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1495742Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1495849Z warnings.warn( 2025-12-04T11:40:52.1496252Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1496365Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1496487Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1496710Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1497110Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1497210Z graph_break [] 2025-12-04T11:40:52.1497427Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1498167Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1498264Z warnings.warn( 2025-12-04T11:40:52.1498477Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1498599Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1498714Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1498943Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1499279Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1499383Z graph_break [] 2025-12-04T11:40:52.1499610Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1500410Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1500515Z warnings.warn( 2025-12-04T11:40:52.1501295Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0625eb2c44495ce0.xml - 2025-12-04T11:40:52.1501464Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1502165Z FAILED [0.1003s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_True - IndexError: list index out of range 2025-12-04T11:40:52.1502171Z 2025-12-04T11:40:52.1502387Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1503060Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True 2025-12-04T11:40:52.1503081Z 2025-12-04T11:40:52.1503404Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1503580Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1503826Z =================== 1 failed, 6 deselected, 2 rerun in 4.63s =================== 2025-12-04T11:40:52.1503924Z Got exit code 1 2025-12-04T11:40:52.1504029Z Retrying single test... 2025-12-04T11:40:52.1504646Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-ea417fd0de1ef597.xml 2025-12-04T11:40:52.1504806Z ============================= test session starts ============================== 2025-12-04T11:40:52.1505164Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1505311Z cachedir: .pytest_cache 2025-12-04T11:40:52.1505831Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1505970Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1506078Z configfile: pytest.ini 2025-12-04T11:40:52.1506670Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1506896Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.1507635Z stepcurrent: skipping 6 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_True 2025-12-04T11:40:52.1507763Z Running 1 items in this shard 2025-12-04T11:40:52.1507768Z 2025-12-04T11:40:52.1508401Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_True ('RERUN', {'yellow': True}) [4.3831s] [100%] 2025-12-04T11:40:52.1509025Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_True ('RERUN', {'yellow': True}) [0.1044s] [100%] 2025-12-04T11:40:52.1509592Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_True FAILED [0.1027s] [100%] 2025-12-04T11:40:52.1509597Z 2025-12-04T11:40:52.1509737Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1510051Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True __ 2025-12-04T11:40:52.1510170Z Traceback (most recent call last): 2025-12-04T11:40:52.1510662Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1510948Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1511402Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1511551Z return source_codes[0] 2025-12-04T11:40:52.1511673Z IndexError: list index out of range 2025-12-04T11:40:52.1511680Z 2025-12-04T11:40:52.1511895Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1512567Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True 2025-12-04T11:40:52.1512572Z 2025-12-04T11:40:52.1512837Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1513061Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1513169Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1513280Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1513631Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1513851Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1513950Z graph_break [] 2025-12-04T11:40:52.1514173Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1514941Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1515049Z warnings.warn( 2025-12-04T11:40:52.1515381Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True __ 2025-12-04T11:40:52.1515502Z Traceback (most recent call last): 2025-12-04T11:40:52.1516003Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1516281Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1516728Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1516873Z return source_codes[0] 2025-12-04T11:40:52.1516991Z IndexError: list index out of range 2025-12-04T11:40:52.1516998Z 2025-12-04T11:40:52.1517222Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1517881Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True 2025-12-04T11:40:52.1517886Z 2025-12-04T11:40:52.1518157Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1518381Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1518490Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1518601Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1518951Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1519169Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1519277Z graph_break [] 2025-12-04T11:40:52.1519488Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1520225Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1520337Z warnings.warn( 2025-12-04T11:40:52.1520550Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1520668Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1520779Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1520993Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1521341Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1521437Z graph_break [] 2025-12-04T11:40:52.1521648Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1522392Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1522597Z warnings.warn( 2025-12-04T11:40:52.1522734Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1523048Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True __ 2025-12-04T11:40:52.1523167Z Traceback (most recent call last): 2025-12-04T11:40:52.1523671Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1523946Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1524393Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1524513Z return source_codes[0] 2025-12-04T11:40:52.1524630Z IndexError: list index out of range 2025-12-04T11:40:52.1524637Z 2025-12-04T11:40:52.1524864Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1525528Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True 2025-12-04T11:40:52.1525568Z 2025-12-04T11:40:52.1525837Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1526091Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1526199Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1526309Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1526660Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1526876Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1526983Z graph_break [] 2025-12-04T11:40:52.1527226Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1527954Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1528069Z warnings.warn( 2025-12-04T11:40:52.1528284Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1528392Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1528521Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1528739Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1529089Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1529188Z graph_break [] 2025-12-04T11:40:52.1529402Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1530145Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1530245Z warnings.warn( 2025-12-04T11:40:52.1530459Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1530586Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1530698Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1530924Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1531265Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1531362Z graph_break [] 2025-12-04T11:40:52.1531586Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1532307Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1532408Z warnings.warn( 2025-12-04T11:40:52.1533193Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-ea417fd0de1ef597.xml - 2025-12-04T11:40:52.1533392Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1534093Z FAILED [0.1027s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_True - IndexError: list index out of range 2025-12-04T11:40:52.1534098Z 2025-12-04T11:40:52.1534318Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1534993Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True 2025-12-04T11:40:52.1534998Z 2025-12-04T11:40:52.1535264Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1535443Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1535652Z ================== 1 failed, 30 deselected, 2 rerun in 4.62s =================== 2025-12-04T11:40:52.1535752Z Got exit code 1 2025-12-04T11:40:52.1535859Z Retrying single test... 2025-12-04T11:40:52.1536464Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-e58bcfb06df4ce4d.xml 2025-12-04T11:40:52.1536656Z ============================= test session starts ============================== 2025-12-04T11:40:52.1537122Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1537233Z cachedir: .pytest_cache 2025-12-04T11:40:52.1537754Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1537883Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1537990Z configfile: pytest.ini 2025-12-04T11:40:52.1538609Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1538833Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.1539579Z stepcurrent: skipping 6 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_True 2025-12-04T11:40:52.1539705Z Running 1 items in this shard 2025-12-04T11:40:52.1539710Z 2025-12-04T11:40:52.1540342Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_True ('RERUN', {'yellow': True}) [4.3639s] [100%] 2025-12-04T11:40:52.1540974Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_True ('RERUN', {'yellow': True}) [0.1058s] [100%] 2025-12-04T11:40:52.1541525Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_True FAILED [0.0996s] [100%] 2025-12-04T11:40:52.1541530Z 2025-12-04T11:40:52.1541668Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1541984Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True __ 2025-12-04T11:40:52.1542104Z Traceback (most recent call last): 2025-12-04T11:40:52.1542594Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1542888Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1543340Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1543458Z return source_codes[0] 2025-12-04T11:40:52.1543578Z IndexError: list index out of range 2025-12-04T11:40:52.1543583Z 2025-12-04T11:40:52.1543793Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1544476Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True 2025-12-04T11:40:52.1544481Z 2025-12-04T11:40:52.1544789Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1545016Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1545126Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1545237Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1545592Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1545807Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1545906Z graph_break [] 2025-12-04T11:40:52.1546129Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1546865Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1546979Z warnings.warn( 2025-12-04T11:40:52.1547284Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True __ 2025-12-04T11:40:52.1547404Z Traceback (most recent call last): 2025-12-04T11:40:52.1547942Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1548218Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1548694Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1548811Z return source_codes[0] 2025-12-04T11:40:52.1548933Z IndexError: list index out of range 2025-12-04T11:40:52.1548938Z 2025-12-04T11:40:52.1549162Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1549828Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True 2025-12-04T11:40:52.1549834Z 2025-12-04T11:40:52.1550123Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1550350Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1550462Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1550582Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1550923Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1551143Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1551247Z graph_break [] 2025-12-04T11:40:52.1551462Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1552196Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1552303Z warnings.warn( 2025-12-04T11:40:52.1552519Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1552639Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1552751Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1552971Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1553319Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1553416Z graph_break [] 2025-12-04T11:40:52.1553632Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1554367Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1554465Z warnings.warn( 2025-12-04T11:40:52.1554616Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1554915Z __ TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True __ 2025-12-04T11:40:52.1555039Z Traceback (most recent call last): 2025-12-04T11:40:52.1555574Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 89, in test_codegen_online_softmax 2025-12-04T11:40:52.1555852Z wrapper_code = self.get_softmax_wrapper(use_log_softmax=use_log_softmax, V=V) 2025-12-04T11:40:52.1556307Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1556427Z return source_codes[0] 2025-12-04T11:40:52.1556545Z IndexError: list index out of range 2025-12-04T11:40:52.1556550Z 2025-12-04T11:40:52.1556772Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1557431Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True 2025-12-04T11:40:52.1557436Z 2025-12-04T11:40:52.1557699Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1557929Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1558041Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1558167Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1558536Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1558750Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1558857Z graph_break [] 2025-12-04T11:40:52.1559099Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1559828Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1559936Z warnings.warn( 2025-12-04T11:40:52.1560147Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1560268Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1560410Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1560630Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1560981Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1561077Z graph_break [] 2025-12-04T11:40:52.1561287Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1562022Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1562123Z warnings.warn( 2025-12-04T11:40:52.1562344Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1562449Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1562564Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1562794Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1563126Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1563223Z graph_break [] 2025-12-04T11:40:52.1563444Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1564169Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1564282Z warnings.warn( 2025-12-04T11:40:52.1565054Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-e58bcfb06df4ce4d.xml - 2025-12-04T11:40:52.1565220Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1565913Z FAILED [0.0996s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_True - IndexError: list index out of range 2025-12-04T11:40:52.1565918Z 2025-12-04T11:40:52.1566129Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1566840Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_online_softmax_V_50304_use_log_softmax_True 2025-12-04T11:40:52.1566847Z 2025-12-04T11:40:52.1567113Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1567289Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1567496Z ================== 1 failed, 30 deselected, 2 rerun in 4.60s =================== 2025-12-04T11:40:52.1567592Z Got exit code 1 2025-12-04T11:40:52.1568182Z FAILED CONSISTENTLY: test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_True 2025-12-04T11:40:52.1568591Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:40:52.1569179Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0c1484308d51b00a.xml 2025-12-04T11:40:52.1569353Z ============================= test session starts ============================== 2025-12-04T11:40:52.1569734Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1569837Z cachedir: .pytest_cache 2025-12-04T11:40:52.1570402Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1570529Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1570640Z configfile: pytest.ini 2025-12-04T11:40:52.1571230Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1571443Z collecting ... collected 31 items / 7 deselected / 24 selected 2025-12-04T11:40:52.1571592Z stepcurrent: skipping 7 already run items. 2025-12-04T11:40:52.1571736Z Running 24 items in this shard 2025-12-04T11:40:52.1571744Z 2025-12-04T11:40:52.1572341Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_softmax_persistent_reduction ('RERUN', {'yellow': True}) [4.3710s] [ 4%] 2025-12-04T11:40:52.1572929Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_softmax_persistent_reduction ('RERUN', {'yellow': True}) [0.0998s] [ 4%] 2025-12-04T11:40:52.1573416Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_softmax_persistent_reduction FAILED [0.0977s] [ 4%] 2025-12-04T11:40:52.1573422Z 2025-12-04T11:40:52.1573572Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1573864Z _________ TestOnlineSoftmax.test_codegen_softmax_persistent_reduction __________ 2025-12-04T11:40:52.1573994Z Traceback (most recent call last): 2025-12-04T11:40:52.1574563Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 104, in test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1574720Z wrapper_code = self.get_softmax_wrapper(1024) 2025-12-04T11:40:52.1575183Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1575294Z return source_codes[0] 2025-12-04T11:40:52.1575417Z IndexError: list index out of range 2025-12-04T11:40:52.1575423Z 2025-12-04T11:40:52.1575651Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1576253Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1576258Z 2025-12-04T11:40:52.1576535Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1576751Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1576858Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1577053Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1577405Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1577676Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1577774Z graph_break [] 2025-12-04T11:40:52.1577989Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1578737Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1578836Z warnings.warn( 2025-12-04T11:40:52.1579125Z _________ TestOnlineSoftmax.test_codegen_softmax_persistent_reduction __________ 2025-12-04T11:40:52.1579254Z Traceback (most recent call last): 2025-12-04T11:40:52.1579812Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 104, in test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1579985Z wrapper_code = self.get_softmax_wrapper(1024) 2025-12-04T11:40:52.1580445Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1580564Z return source_codes[0] 2025-12-04T11:40:52.1580731Z IndexError: list index out of range 2025-12-04T11:40:52.1580736Z 2025-12-04T11:40:52.1580952Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1581583Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1581603Z 2025-12-04T11:40:52.1581872Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1582089Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1582217Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1582336Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1582727Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1582964Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1583065Z graph_break [] 2025-12-04T11:40:52.1583281Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1584031Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1584136Z warnings.warn( 2025-12-04T11:40:52.1584367Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1584480Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1584595Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1584829Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1585168Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1585266Z graph_break [] 2025-12-04T11:40:52.1585497Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1586227Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1586345Z warnings.warn( 2025-12-04T11:40:52.1586489Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1586783Z _________ TestOnlineSoftmax.test_codegen_softmax_persistent_reduction __________ 2025-12-04T11:40:52.1586921Z Traceback (most recent call last): 2025-12-04T11:40:52.1587486Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 104, in test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1587645Z wrapper_code = self.get_softmax_wrapper(1024) 2025-12-04T11:40:52.1588111Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1588225Z return source_codes[0] 2025-12-04T11:40:52.1588358Z IndexError: list index out of range 2025-12-04T11:40:52.1588394Z 2025-12-04T11:40:52.1588615Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1589217Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1589222Z 2025-12-04T11:40:52.1589507Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1589725Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1589850Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1589965Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1590307Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1590541Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1590643Z graph_break [] 2025-12-04T11:40:52.1590859Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1591608Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1591739Z warnings.warn( 2025-12-04T11:40:52.1591965Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1592108Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1592224Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1592457Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1592794Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1592890Z graph_break [] 2025-12-04T11:40:52.1593112Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1593865Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1593981Z warnings.warn( 2025-12-04T11:40:52.1594195Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1594303Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1594427Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1594648Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1594985Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1595092Z graph_break [] 2025-12-04T11:40:52.1595303Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1596225Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1596333Z warnings.warn( 2025-12-04T11:40:52.1597095Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0c1484308d51b00a.xml - 2025-12-04T11:40:52.1597279Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1597922Z FAILED [0.0977s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_softmax_persistent_reduction - IndexError: list index out of range 2025-12-04T11:40:52.1597928Z 2025-12-04T11:40:52.1598156Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1598752Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1598757Z 2025-12-04T11:40:52.1599022Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1599214Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1599409Z =================== 1 failed, 7 deselected, 2 rerun in 4.60s =================== 2025-12-04T11:40:52.1599589Z Got exit code 1 2025-12-04T11:40:52.1599704Z Retrying single test... 2025-12-04T11:40:52.1600295Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-9682f9212923ea0d.xml 2025-12-04T11:40:52.1600470Z ============================= test session starts ============================== 2025-12-04T11:40:52.1600821Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1600928Z cachedir: .pytest_cache 2025-12-04T11:40:52.1601464Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1601587Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1601705Z configfile: pytest.ini 2025-12-04T11:40:52.1602298Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1602516Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.1603250Z stepcurrent: skipping 7 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1603365Z Running 1 items in this shard 2025-12-04T11:40:52.1603371Z 2025-12-04T11:40:52.1604006Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_softmax_persistent_reduction ('RERUN', {'yellow': True}) [4.3675s] [100%] 2025-12-04T11:40:52.1604586Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_softmax_persistent_reduction ('RERUN', {'yellow': True}) [0.0993s] [100%] 2025-12-04T11:40:52.1605072Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_softmax_persistent_reduction FAILED [0.0971s] [100%] 2025-12-04T11:40:52.1605078Z 2025-12-04T11:40:52.1605267Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1605564Z _________ TestOnlineSoftmax.test_codegen_softmax_persistent_reduction __________ 2025-12-04T11:40:52.1605686Z Traceback (most recent call last): 2025-12-04T11:40:52.1606258Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 104, in test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1606416Z wrapper_code = self.get_softmax_wrapper(1024) 2025-12-04T11:40:52.1606878Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1606987Z return source_codes[0] 2025-12-04T11:40:52.1607109Z IndexError: list index out of range 2025-12-04T11:40:52.1607114Z 2025-12-04T11:40:52.1607344Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1607949Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1607955Z 2025-12-04T11:40:52.1608233Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1608453Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1608568Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1608695Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1609042Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1609261Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1609373Z graph_break [] 2025-12-04T11:40:52.1609587Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1610333Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1610433Z warnings.warn( 2025-12-04T11:40:52.1610726Z _________ TestOnlineSoftmax.test_codegen_softmax_persistent_reduction __________ 2025-12-04T11:40:52.1610890Z Traceback (most recent call last): 2025-12-04T11:40:52.1611458Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 104, in test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1611616Z wrapper_code = self.get_softmax_wrapper(1024) 2025-12-04T11:40:52.1612080Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1612189Z return source_codes[0] 2025-12-04T11:40:52.1612324Z IndexError: list index out of range 2025-12-04T11:40:52.1612329Z 2025-12-04T11:40:52.1612542Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1613142Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1613148Z 2025-12-04T11:40:52.1613430Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1613646Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1613799Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1613914Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1614256Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1614514Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1614613Z graph_break [] 2025-12-04T11:40:52.1614825Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1615570Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1615672Z warnings.warn( 2025-12-04T11:40:52.1615897Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1616035Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1616151Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1616384Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1616721Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1616817Z graph_break [] 2025-12-04T11:40:52.1617110Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1617844Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1617958Z warnings.warn( 2025-12-04T11:40:52.1618101Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1618390Z _________ TestOnlineSoftmax.test_codegen_softmax_persistent_reduction __________ 2025-12-04T11:40:52.1618526Z Traceback (most recent call last): 2025-12-04T11:40:52.1619093Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 104, in test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1619251Z wrapper_code = self.get_softmax_wrapper(1024) 2025-12-04T11:40:52.1619714Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1619822Z return source_codes[0] 2025-12-04T11:40:52.1619959Z IndexError: list index out of range 2025-12-04T11:40:52.1619965Z 2025-12-04T11:40:52.1620180Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1620776Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1620782Z 2025-12-04T11:40:52.1621059Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1621274Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1621401Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1621554Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1621894Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1622127Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1622227Z graph_break [] 2025-12-04T11:40:52.1622444Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1623190Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1623290Z warnings.warn( 2025-12-04T11:40:52.1623514Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1623625Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1623737Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1623973Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1624314Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1624443Z graph_break [] 2025-12-04T11:40:52.1624669Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1625446Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1625560Z warnings.warn( 2025-12-04T11:40:52.1625772Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1625880Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1626006Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1626225Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1626596Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1626708Z graph_break [] 2025-12-04T11:40:52.1626925Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1627659Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1627758Z warnings.warn( 2025-12-04T11:40:52.1628521Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-9682f9212923ea0d.xml - 2025-12-04T11:40:52.1628700Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1629341Z FAILED [0.0971s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_softmax_persistent_reduction - IndexError: list index out of range 2025-12-04T11:40:52.1629346Z 2025-12-04T11:40:52.1629572Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1630175Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1630185Z 2025-12-04T11:40:52.1630454Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1630642Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1630842Z ================== 1 failed, 30 deselected, 2 rerun in 4.59s =================== 2025-12-04T11:40:52.1630940Z Got exit code 1 2025-12-04T11:40:52.1631056Z Retrying single test... 2025-12-04T11:40:52.1631651Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-ef1f71bad0a01b52.xml 2025-12-04T11:40:52.1631826Z ============================= test session starts ============================== 2025-12-04T11:40:52.1632172Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1632284Z cachedir: .pytest_cache 2025-12-04T11:40:52.1632823Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1632982Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1633090Z configfile: pytest.ini 2025-12-04T11:40:52.1633694Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1633910Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.1634603Z stepcurrent: skipping 7 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1634715Z Running 1 items in this shard 2025-12-04T11:40:52.1634721Z 2025-12-04T11:40:52.1635306Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_softmax_persistent_reduction ('RERUN', {'yellow': True}) [4.3876s] [100%] 2025-12-04T11:40:52.1635908Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_softmax_persistent_reduction ('RERUN', {'yellow': True}) [0.1009s] [100%] 2025-12-04T11:40:52.1636426Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_softmax_persistent_reduction FAILED [0.1010s] [100%] 2025-12-04T11:40:52.1636431Z 2025-12-04T11:40:52.1636588Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1636918Z _________ TestOnlineSoftmax.test_codegen_softmax_persistent_reduction __________ 2025-12-04T11:40:52.1637043Z Traceback (most recent call last): 2025-12-04T11:40:52.1637619Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 104, in test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1637777Z wrapper_code = self.get_softmax_wrapper(1024) 2025-12-04T11:40:52.1638247Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1638388Z return source_codes[0] 2025-12-04T11:40:52.1638514Z IndexError: list index out of range 2025-12-04T11:40:52.1638521Z 2025-12-04T11:40:52.1638753Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1639358Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1639363Z 2025-12-04T11:40:52.1639647Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1639865Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1639975Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1640105Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1640447Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1640666Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1640781Z graph_break [] 2025-12-04T11:40:52.1640998Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1641748Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1641853Z warnings.warn( 2025-12-04T11:40:52.1642146Z _________ TestOnlineSoftmax.test_codegen_softmax_persistent_reduction __________ 2025-12-04T11:40:52.1642286Z Traceback (most recent call last): 2025-12-04T11:40:52.1642848Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 104, in test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1643005Z wrapper_code = self.get_softmax_wrapper(1024) 2025-12-04T11:40:52.1643468Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1643580Z return source_codes[0] 2025-12-04T11:40:52.1643725Z IndexError: list index out of range 2025-12-04T11:40:52.1643730Z 2025-12-04T11:40:52.1643947Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1644580Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1644588Z 2025-12-04T11:40:52.1644870Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1645091Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1645221Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1645335Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1645675Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1645905Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1646002Z graph_break [] 2025-12-04T11:40:52.1646220Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1646963Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1647099Z warnings.warn( 2025-12-04T11:40:52.1647319Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1647426Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1647569Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1647801Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1648139Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1648233Z graph_break [] 2025-12-04T11:40:52.1648456Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1649208Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1649323Z warnings.warn( 2025-12-04T11:40:52.1649467Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1649756Z _________ TestOnlineSoftmax.test_codegen_softmax_persistent_reduction __________ 2025-12-04T11:40:52.1649891Z Traceback (most recent call last): 2025-12-04T11:40:52.1650460Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 104, in test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1650616Z wrapper_code = self.get_softmax_wrapper(1024) 2025-12-04T11:40:52.1651075Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 78, in get_softmax_wrapper 2025-12-04T11:40:52.1651187Z return source_codes[0] 2025-12-04T11:40:52.1651318Z IndexError: list index out of range 2025-12-04T11:40:52.1651323Z 2025-12-04T11:40:52.1651538Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1652136Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1652143Z 2025-12-04T11:40:52.1652424Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1652643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1652768Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1652884Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1653218Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1653447Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1653544Z graph_break [] 2025-12-04T11:40:52.1653757Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1654502Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1654632Z warnings.warn( 2025-12-04T11:40:52.1654855Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1654970Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1655082Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1655310Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1655651Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1655746Z graph_break [] 2025-12-04T11:40:52.1655971Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1656700Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1656811Z warnings.warn( 2025-12-04T11:40:52.1657105Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1657219Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1657346Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.1657603Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1657943Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1658056Z graph_break [] 2025-12-04T11:40:52.1658299Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1659035Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1659146Z warnings.warn( 2025-12-04T11:40:52.1659918Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-ef1f71bad0a01b52.xml - 2025-12-04T11:40:52.1660126Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1660770Z FAILED [0.1010s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_softmax_persistent_reduction - IndexError: list index out of range 2025-12-04T11:40:52.1660778Z 2025-12-04T11:40:52.1661005Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1661608Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1661614Z 2025-12-04T11:40:52.1661881Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1662069Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1662266Z ================== 1 failed, 30 deselected, 2 rerun in 4.62s =================== 2025-12-04T11:40:52.1662364Z Got exit code 1 2025-12-04T11:40:52.1662899Z FAILED CONSISTENTLY: test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_softmax_persistent_reduction 2025-12-04T11:40:52.1663312Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:40:52.1663912Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3d18fa53804224f4.xml 2025-12-04T11:40:52.1664070Z ============================= test session starts ============================== 2025-12-04T11:40:52.1664419Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1664540Z cachedir: .pytest_cache 2025-12-04T11:40:52.1665062Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1665199Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1665305Z configfile: pytest.ini 2025-12-04T11:40:52.1665898Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1666155Z collecting ... collected 31 items / 8 deselected / 23 selected 2025-12-04T11:40:52.1666299Z stepcurrent: skipping 8 already run items. 2025-12-04T11:40:52.1666411Z Running 23 items in this shard 2025-12-04T11:40:52.1666416Z 2025-12-04T11:40:52.1666796Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_log_softmax PASSED [4.3671s] [ 4%] 2025-12-04T11:40:52.1667231Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_no_online_softmax_for_cpu PASSED [4.0844s] [ 8%] 2025-12-04T11:40:52.1667823Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_bfloat16 ('RERUN', {'yellow': True}) [0.1290s] [ 13%] 2025-12-04T11:40:52.1668403Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_bfloat16 ('RERUN', {'yellow': True}) [0.0949s] [ 13%] 2025-12-04T11:40:52.1668891Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_bfloat16 FAILED [0.0950s] [ 13%] 2025-12-04T11:40:52.1668899Z 2025-12-04T11:40:52.1669053Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1669391Z ________ TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 _________ 2025-12-04T11:40:52.1669524Z Traceback (most recent call last): 2025-12-04T11:40:52.1670083Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 181, in test_prepare_softmax_acc_with_fp64 2025-12-04T11:40:52.1670329Z res, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim=-1) 2025-12-04T11:40:52.1670552Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1670557Z 2025-12-04T11:40:52.1670772Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1671420Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 2025-12-04T11:40:52.1671426Z 2025-12-04T11:40:52.1671692Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1671908Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1672035Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1672152Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1672372Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1672727Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1672824Z graph_break [] 2025-12-04T11:40:52.1673051Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1673778Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1673879Z warnings.warn( 2025-12-04T11:40:52.1674188Z ________ TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 _________ 2025-12-04T11:40:52.1674310Z Traceback (most recent call last): 2025-12-04T11:40:52.1674842Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 181, in test_prepare_softmax_acc_with_fp64 2025-12-04T11:40:52.1675096Z res, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim=-1) 2025-12-04T11:40:52.1675308Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1675313Z 2025-12-04T11:40:52.1675536Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1676140Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 2025-12-04T11:40:52.1676146Z 2025-12-04T11:40:52.1676419Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1676634Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1676744Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1676921Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1677140Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1677482Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1677587Z graph_break [] 2025-12-04T11:40:52.1677801Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1678527Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1678637Z warnings.warn( 2025-12-04T11:40:52.1678851Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1678978Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1679091Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1679314Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1679665Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1679797Z graph_break [] 2025-12-04T11:40:52.1680012Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1680784Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1680885Z warnings.warn( 2025-12-04T11:40:52.1681039Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1681327Z ________ TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 _________ 2025-12-04T11:40:52.1681446Z Traceback (most recent call last): 2025-12-04T11:40:52.1681991Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 181, in test_prepare_softmax_acc_with_fp64 2025-12-04T11:40:52.1682257Z res, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim=-1) 2025-12-04T11:40:52.1682484Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1682491Z 2025-12-04T11:40:52.1682702Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1683308Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 2025-12-04T11:40:52.1683313Z 2025-12-04T11:40:52.1683589Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1683803Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1683913Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1684035Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1684253Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1684605Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1684704Z graph_break [] 2025-12-04T11:40:52.1684918Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1685658Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1685761Z warnings.warn( 2025-12-04T11:40:52.1685976Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1686090Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1686202Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1686424Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1686758Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1686852Z graph_break [] 2025-12-04T11:40:52.1687074Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1687802Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1687933Z warnings.warn( 2025-12-04T11:40:52.1688150Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1688255Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1688372Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1688588Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1688921Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1689025Z graph_break [] 2025-12-04T11:40:52.1689235Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1689956Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1690066Z warnings.warn( 2025-12-04T11:40:52.1690823Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3d18fa53804224f4.xml - 2025-12-04T11:40:52.1691027Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1691802Z FAILED [0.0950s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_bfloat16 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1691808Z 2025-12-04T11:40:52.1692023Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1692635Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 2025-12-04T11:40:52.1692641Z 2025-12-04T11:40:52.1693003Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1693189Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1693397Z ============== 1 failed, 2 passed, 8 deselected, 2 rerun in 8.80s ============== 2025-12-04T11:40:52.1693493Z Got exit code 1 2025-12-04T11:40:52.1693608Z Retrying single test... 2025-12-04T11:40:52.1694200Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2256fbb35262b330.xml 2025-12-04T11:40:52.1694363Z ============================= test session starts ============================== 2025-12-04T11:40:52.1694709Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1694815Z cachedir: .pytest_cache 2025-12-04T11:40:52.1695338Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1695457Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1695566Z configfile: pytest.ini 2025-12-04T11:40:52.1696350Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1696572Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.1697328Z stepcurrent: skipping 10 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_bfloat16 2025-12-04T11:40:52.1697444Z Running 1 items in this shard 2025-12-04T11:40:52.1697450Z 2025-12-04T11:40:52.1698029Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_bfloat16 ('RERUN', {'yellow': True}) [4.2447s] [100%] 2025-12-04T11:40:52.1698614Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_bfloat16 ('RERUN', {'yellow': True}) [0.0943s] [100%] 2025-12-04T11:40:52.1699105Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_bfloat16 FAILED [0.0926s] [100%] 2025-12-04T11:40:52.1699187Z 2025-12-04T11:40:52.1699332Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1699622Z ________ TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 _________ 2025-12-04T11:40:52.1699741Z Traceback (most recent call last): 2025-12-04T11:40:52.1700283Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 181, in test_prepare_softmax_acc_with_fp64 2025-12-04T11:40:52.1700518Z res, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim=-1) 2025-12-04T11:40:52.1700733Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1700738Z 2025-12-04T11:40:52.1700945Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1701551Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 2025-12-04T11:40:52.1701556Z 2025-12-04T11:40:52.1701820Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1702033Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1702196Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1702302Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1702678Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1702900Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1702994Z graph_break [] 2025-12-04T11:40:52.1703202Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1703938Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1704033Z warnings.warn( 2025-12-04T11:40:52.1704364Z ________ TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 _________ 2025-12-04T11:40:52.1704483Z Traceback (most recent call last): 2025-12-04T11:40:52.1705017Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 181, in test_prepare_softmax_acc_with_fp64 2025-12-04T11:40:52.1705261Z res, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim=-1) 2025-12-04T11:40:52.1705468Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1705473Z 2025-12-04T11:40:52.1705679Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1706289Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 2025-12-04T11:40:52.1706294Z 2025-12-04T11:40:52.1706552Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1706768Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1706874Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1706985Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1707327Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1707542Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1707651Z graph_break [] 2025-12-04T11:40:52.1707866Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1708609Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1708726Z warnings.warn( 2025-12-04T11:40:52.1708941Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1709050Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1709175Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1709397Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1709793Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1709892Z graph_break [] 2025-12-04T11:40:52.1710106Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1710847Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1710947Z warnings.warn( 2025-12-04T11:40:52.1711090Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1711390Z ________ TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 _________ 2025-12-04T11:40:52.1711512Z Traceback (most recent call last): 2025-12-04T11:40:52.1712066Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 181, in test_prepare_softmax_acc_with_fp64 2025-12-04T11:40:52.1712303Z res, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim=-1) 2025-12-04T11:40:52.1712516Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1712549Z 2025-12-04T11:40:52.1712777Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1713419Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 2025-12-04T11:40:52.1713424Z 2025-12-04T11:40:52.1713702Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1713918Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1714030Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1714157Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1714498Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1714742Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1714855Z graph_break [] 2025-12-04T11:40:52.1715070Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1715816Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1715916Z warnings.warn( 2025-12-04T11:40:52.1716131Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1716251Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1716363Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1716582Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1716933Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1717030Z graph_break [] 2025-12-04T11:40:52.1717256Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1717982Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1718085Z warnings.warn( 2025-12-04T11:40:52.1718309Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1718418Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1718531Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1718763Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1719099Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1719207Z graph_break [] 2025-12-04T11:40:52.1719420Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1720144Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1720286Z warnings.warn( 2025-12-04T11:40:52.1721055Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2256fbb35262b330.xml - 2025-12-04T11:40:52.1721227Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1721987Z FAILED [0.0926s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_bfloat16 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1721993Z 2025-12-04T11:40:52.1722208Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1722821Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 2025-12-04T11:40:52.1722826Z 2025-12-04T11:40:52.1723091Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1723273Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1723483Z ================== 1 failed, 30 deselected, 2 rerun in 4.46s =================== 2025-12-04T11:40:52.1723612Z Got exit code 1 2025-12-04T11:40:52.1723732Z Retrying single test... 2025-12-04T11:40:52.1724350Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a993cfee9c5b5572.xml 2025-12-04T11:40:52.1724514Z ============================= test session starts ============================== 2025-12-04T11:40:52.1724870Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1724978Z cachedir: .pytest_cache 2025-12-04T11:40:52.1725507Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1725630Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1725765Z configfile: pytest.ini 2025-12-04T11:40:52.1726371Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1726585Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.1727272Z stepcurrent: skipping 10 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_bfloat16 2025-12-04T11:40:52.1727394Z Running 1 items in this shard 2025-12-04T11:40:52.1727399Z 2025-12-04T11:40:52.1727976Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_bfloat16 ('RERUN', {'yellow': True}) [4.2842s] [100%] 2025-12-04T11:40:52.1728570Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_bfloat16 ('RERUN', {'yellow': True}) [0.0946s] [100%] 2025-12-04T11:40:52.1729060Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_bfloat16 FAILED [0.0931s] [100%] 2025-12-04T11:40:52.1729068Z 2025-12-04T11:40:52.1729208Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1729511Z ________ TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 _________ 2025-12-04T11:40:52.1729631Z Traceback (most recent call last): 2025-12-04T11:40:52.1730181Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 181, in test_prepare_softmax_acc_with_fp64 2025-12-04T11:40:52.1730419Z res, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim=-1) 2025-12-04T11:40:52.1730628Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1730633Z 2025-12-04T11:40:52.1730859Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1731465Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 2025-12-04T11:40:52.1731473Z 2025-12-04T11:40:52.1731750Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1731996Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1732110Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1732236Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1732580Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1732798Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1732911Z graph_break [] 2025-12-04T11:40:52.1733127Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1733874Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1733975Z warnings.warn( 2025-12-04T11:40:52.1734273Z ________ TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 _________ 2025-12-04T11:40:52.1734411Z Traceback (most recent call last): 2025-12-04T11:40:52.1734946Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 181, in test_prepare_softmax_acc_with_fp64 2025-12-04T11:40:52.1735231Z res, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim=-1) 2025-12-04T11:40:52.1735471Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1735476Z 2025-12-04T11:40:52.1735690Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1736304Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 2025-12-04T11:40:52.1736309Z 2025-12-04T11:40:52.1736572Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1743101Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1743279Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1743402Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1743766Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1743990Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1744089Z graph_break [] 2025-12-04T11:40:52.1744319Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1745065Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1745169Z warnings.warn( 2025-12-04T11:40:52.1745396Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1745504Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1745615Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1745846Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1746185Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1746294Z graph_break [] 2025-12-04T11:40:52.1746508Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1747243Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1747349Z warnings.warn( 2025-12-04T11:40:52.1747492Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1747789Z ________ TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 _________ 2025-12-04T11:40:52.1747910Z Traceback (most recent call last): 2025-12-04T11:40:52.1748451Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 181, in test_prepare_softmax_acc_with_fp64 2025-12-04T11:40:52.1748705Z res, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim=-1) 2025-12-04T11:40:52.1748958Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1748968Z 2025-12-04T11:40:52.1749185Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1749802Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 2025-12-04T11:40:52.1749808Z 2025-12-04T11:40:52.1750071Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1750292Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1750402Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1750512Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1750859Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1751076Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1751173Z graph_break [] 2025-12-04T11:40:52.1751391Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1752160Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1752268Z warnings.warn( 2025-12-04T11:40:52.1752518Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1752623Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1752742Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1752957Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1753291Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1753393Z graph_break [] 2025-12-04T11:40:52.1753655Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1754397Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1754501Z warnings.warn( 2025-12-04T11:40:52.1754711Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1754828Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1754939Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1755156Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1755500Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1755594Z graph_break [] 2025-12-04T11:40:52.1755817Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1756539Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1756638Z warnings.warn( 2025-12-04T11:40:52.1757415Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a993cfee9c5b5572.xml - 2025-12-04T11:40:52.1757580Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1758350Z FAILED [0.0931s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_bfloat16 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1758356Z 2025-12-04T11:40:52.1758569Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1759173Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_acc_with_fp64_bfloat16 2025-12-04T11:40:52.1759179Z 2025-12-04T11:40:52.1759451Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1759625Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1759860Z ================== 1 failed, 30 deselected, 2 rerun in 4.50s =================== 2025-12-04T11:40:52.1759957Z Got exit code 1 2025-12-04T11:40:52.1760484Z FAILED CONSISTENTLY: test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_bfloat16 2025-12-04T11:40:52.1760902Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:40:52.1761485Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-4097f4a492ce8bb6.xml 2025-12-04T11:40:52.1761656Z ============================= test session starts ============================== 2025-12-04T11:40:52.1762002Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1762110Z cachedir: .pytest_cache 2025-12-04T11:40:52.1762638Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1762793Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1762900Z configfile: pytest.ini 2025-12-04T11:40:52.1763500Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1763749Z collecting ... collected 31 items / 11 deselected / 20 selected 2025-12-04T11:40:52.1763898Z stepcurrent: skipping 11 already run items. 2025-12-04T11:40:52.1764010Z Running 20 items in this shard 2025-12-04T11:40:52.1764015Z 2025-12-04T11:40:52.1764506Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_float16 PASSED [4.7811s] [ 5%] 2025-12-04T11:40:52.1764996Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_float32 PASSED [0.5663s] [ 10%] 2025-12-04T11:40:52.1765569Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_-1 ('RERUN', {'yellow': True}) [0.1328s] [ 15%] 2025-12-04T11:40:52.1766132Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_-1 ('RERUN', {'yellow': True}) [0.0964s] [ 15%] 2025-12-04T11:40:52.1766592Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_-1 FAILED [0.0945s] [ 15%] 2025-12-04T11:40:52.1766600Z 2025-12-04T11:40:52.1766738Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1767024Z ___________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 ____________ 2025-12-04T11:40:52.1767140Z Traceback (most recent call last): 2025-12-04T11:40:52.1767609Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1767847Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1768058Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1768066Z 2025-12-04T11:40:52.1768290Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1768867Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 2025-12-04T11:40:52.1768872Z 2025-12-04T11:40:52.1769145Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1769358Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1769466Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1769585Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1769801Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1770141Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1770248Z graph_break [] 2025-12-04T11:40:52.1770462Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1771233Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1771343Z warnings.warn( 2025-12-04T11:40:52.1771617Z ___________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 ____________ 2025-12-04T11:40:52.1771751Z Traceback (most recent call last): 2025-12-04T11:40:52.1772217Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1772440Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1772659Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1772664Z 2025-12-04T11:40:52.1772874Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1773460Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 2025-12-04T11:40:52.1773468Z 2025-12-04T11:40:52.1773730Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1773973Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1774091Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1774237Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1774456Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1774805Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1774901Z graph_break [] 2025-12-04T11:40:52.1775121Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1775880Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1775979Z warnings.warn( 2025-12-04T11:40:52.1776202Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1776311Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1776419Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1776644Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1777067Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1777180Z graph_break [] 2025-12-04T11:40:52.1777392Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1778118Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1778227Z warnings.warn( 2025-12-04T11:40:52.1778368Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1778643Z ___________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 ____________ 2025-12-04T11:40:52.1778775Z Traceback (most recent call last): 2025-12-04T11:40:52.1779245Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1779477Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1779688Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1779693Z 2025-12-04T11:40:52.1779903Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1780478Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 2025-12-04T11:40:52.1780483Z 2025-12-04T11:40:52.1780744Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1780972Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1781080Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1781230Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1781459Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1781793Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1781888Z graph_break [] 2025-12-04T11:40:52.1782112Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1782845Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1782951Z warnings.warn( 2025-12-04T11:40:52.1783163Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1783271Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1783395Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1783609Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1783947Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1784083Z graph_break [] 2025-12-04T11:40:52.1784294Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1785061Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1785162Z warnings.warn( 2025-12-04T11:40:52.1785373Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1785489Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1785600Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1785815Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1786199Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1786293Z graph_break [] 2025-12-04T11:40:52.1786518Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1787247Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1787344Z warnings.warn( 2025-12-04T11:40:52.1788122Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-4097f4a492ce8bb6.xml - 2025-12-04T11:40:52.1788289Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1789004Z FAILED [0.0945s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_-1 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1789010Z 2025-12-04T11:40:52.1789223Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1789791Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 2025-12-04T11:40:52.1789801Z 2025-12-04T11:40:52.1790074Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1790249Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1790471Z ============= 1 failed, 2 passed, 11 deselected, 2 rerun in 5.71s ============== 2025-12-04T11:40:52.1790566Z Got exit code 1 2025-12-04T11:40:52.1790671Z Retrying single test... 2025-12-04T11:40:52.1791269Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-9df84d07ff6a448b.xml 2025-12-04T11:40:52.1791426Z ============================= test session starts ============================== 2025-12-04T11:40:52.1791776Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1791894Z cachedir: .pytest_cache 2025-12-04T11:40:52.1792447Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1792585Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1792692Z configfile: pytest.ini 2025-12-04T11:40:52.1793283Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1793509Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.1794168Z stepcurrent: skipping 13 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_-1 2025-12-04T11:40:52.1794280Z Running 1 items in this shard 2025-12-04T11:40:52.1794295Z 2025-12-04T11:40:52.1794840Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_-1 ('RERUN', {'yellow': True}) [4.2994s] [100%] 2025-12-04T11:40:52.1795382Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_-1 ('RERUN', {'yellow': True}) [0.1011s] [100%] 2025-12-04T11:40:52.1795892Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_-1 FAILED [0.1028s] [100%] 2025-12-04T11:40:52.1795897Z 2025-12-04T11:40:52.1796401Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1796704Z ___________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 ____________ 2025-12-04T11:40:52.1796826Z Traceback (most recent call last): 2025-12-04T11:40:52.1797296Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1797539Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1797752Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1797801Z 2025-12-04T11:40:52.1798016Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1798602Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 2025-12-04T11:40:52.1798610Z 2025-12-04T11:40:52.1798879Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1799107Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1799217Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1799333Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1799684Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1799902Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1799999Z graph_break [] 2025-12-04T11:40:52.1800223Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1800961Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1801076Z warnings.warn( 2025-12-04T11:40:52.1801351Z ___________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 ____________ 2025-12-04T11:40:52.1801470Z Traceback (most recent call last): 2025-12-04T11:40:52.1801953Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1802185Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1802405Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1802411Z 2025-12-04T11:40:52.1802624Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1803203Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 2025-12-04T11:40:52.1803208Z 2025-12-04T11:40:52.1803481Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1803738Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1803852Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1803977Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1804317Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1804542Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1804636Z graph_break [] 2025-12-04T11:40:52.1804848Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1805590Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1805688Z warnings.warn( 2025-12-04T11:40:52.1805902Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1806021Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1806133Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1806401Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1806733Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1806828Z graph_break [] 2025-12-04T11:40:52.1807079Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1807805Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1807904Z warnings.warn( 2025-12-04T11:40:52.1808053Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1808356Z ___________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 ____________ 2025-12-04T11:40:52.1808492Z Traceback (most recent call last): 2025-12-04T11:40:52.1808960Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1809188Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1809406Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1809411Z 2025-12-04T11:40:52.1809624Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1810205Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 2025-12-04T11:40:52.1810210Z 2025-12-04T11:40:52.1810474Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1810688Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1810807Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1810920Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1811260Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1811487Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1811583Z graph_break [] 2025-12-04T11:40:52.1811807Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1812536Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1812634Z warnings.warn( 2025-12-04T11:40:52.1812857Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1812964Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1813076Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1813303Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1813638Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1813776Z graph_break [] 2025-12-04T11:40:52.1813988Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1814721Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1814832Z warnings.warn( 2025-12-04T11:40:52.1815045Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1815151Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1815272Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1815489Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1815833Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1815927Z graph_break [] 2025-12-04T11:40:52.1816140Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1816965Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1817112Z warnings.warn( 2025-12-04T11:40:52.1817929Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-9df84d07ff6a448b.xml - 2025-12-04T11:40:52.1818109Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1818824Z FAILED [0.1028s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_-1 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1818829Z 2025-12-04T11:40:52.1819055Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1819660Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 2025-12-04T11:40:52.1819668Z 2025-12-04T11:40:52.1819947Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1820129Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1820328Z ================== 1 failed, 30 deselected, 2 rerun in 4.53s =================== 2025-12-04T11:40:52.1820438Z Got exit code 1 2025-12-04T11:40:52.1820543Z Retrying single test... 2025-12-04T11:40:52.1821133Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-b20a3959da2014da.xml 2025-12-04T11:40:52.1821302Z ============================= test session starts ============================== 2025-12-04T11:40:52.1821646Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1821765Z cachedir: .pytest_cache 2025-12-04T11:40:52.1822281Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1822406Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1822526Z configfile: pytest.ini 2025-12-04T11:40:52.1823117Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1823333Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.1824000Z stepcurrent: skipping 13 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_-1 2025-12-04T11:40:52.1824112Z Running 1 items in this shard 2025-12-04T11:40:52.1824117Z 2025-12-04T11:40:52.1824682Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_-1 ('RERUN', {'yellow': True}) [4.2666s] [100%] 2025-12-04T11:40:52.1825223Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_-1 ('RERUN', {'yellow': True}) [0.0948s] [100%] 2025-12-04T11:40:52.1825712Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_-1 FAILED [0.0928s] [100%] 2025-12-04T11:40:52.1825730Z 2025-12-04T11:40:52.1825866Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1826144Z ___________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 ____________ 2025-12-04T11:40:52.1826276Z Traceback (most recent call last): 2025-12-04T11:40:52.1826743Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1826970Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1827187Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1827193Z 2025-12-04T11:40:52.1827404Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1827991Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 2025-12-04T11:40:52.1828032Z 2025-12-04T11:40:52.1828301Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1828514Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1828637Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1828778Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1829121Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1829349Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1829445Z graph_break [] 2025-12-04T11:40:52.1829668Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1830437Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1830539Z warnings.warn( 2025-12-04T11:40:52.1830824Z ___________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 ____________ 2025-12-04T11:40:52.1830943Z Traceback (most recent call last): 2025-12-04T11:40:52.1831422Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1831655Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1831864Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1831869Z 2025-12-04T11:40:52.1832091Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1832664Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 2025-12-04T11:40:52.1832669Z 2025-12-04T11:40:52.1832943Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1833155Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1833269Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1833400Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1833741Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1833956Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1834061Z graph_break [] 2025-12-04T11:40:52.1834275Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1835013Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1835112Z warnings.warn( 2025-12-04T11:40:52.1835323Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1835441Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1835553Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1835803Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1836150Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1836244Z graph_break [] 2025-12-04T11:40:52.1836457Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1837197Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1837295Z warnings.warn( 2025-12-04T11:40:52.1837449Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1837721Z ___________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 ____________ 2025-12-04T11:40:52.1837840Z Traceback (most recent call last): 2025-12-04T11:40:52.1838321Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1838549Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1838815Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1838820Z 2025-12-04T11:40:52.1839030Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1839628Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 2025-12-04T11:40:52.1839633Z 2025-12-04T11:40:52.1839907Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1840117Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1840227Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1840349Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1840714Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1840943Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1841042Z graph_break [] 2025-12-04T11:40:52.1841253Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1841996Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1842094Z warnings.warn( 2025-12-04T11:40:52.1842304Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1842423Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1842537Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1842764Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1843101Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1843198Z graph_break [] 2025-12-04T11:40:52.1843422Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1844152Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1844250Z warnings.warn( 2025-12-04T11:40:52.1844474Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1844584Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1844705Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1844919Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1845254Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1845361Z graph_break [] 2025-12-04T11:40:52.1845572Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1846300Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1846435Z warnings.warn( 2025-12-04T11:40:52.1847201Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-b20a3959da2014da.xml - 2025-12-04T11:40:52.1847367Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1848075Z FAILED [0.0928s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_-1 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1848080Z 2025-12-04T11:40:52.1848290Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1848857Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_-1 2025-12-04T11:40:52.1848872Z 2025-12-04T11:40:52.1849133Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1849310Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1849541Z ================== 1 failed, 30 deselected, 2 rerun in 4.48s =================== 2025-12-04T11:40:52.1849631Z Got exit code 1 2025-12-04T11:40:52.1850144Z FAILED CONSISTENTLY: test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_-1 2025-12-04T11:40:52.1850557Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:40:52.1851137Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-89205029d9c04efd.xml 2025-12-04T11:40:52.1851301Z ============================= test session starts ============================== 2025-12-04T11:40:52.1851675Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1851781Z cachedir: .pytest_cache 2025-12-04T11:40:52.1852309Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1852429Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1852534Z configfile: pytest.ini 2025-12-04T11:40:52.1853131Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1853344Z collecting ... collected 31 items / 14 deselected / 17 selected 2025-12-04T11:40:52.1853488Z stepcurrent: skipping 14 already run items. 2025-12-04T11:40:52.1853596Z Running 17 items in this shard 2025-12-04T11:40:52.1853602Z 2025-12-04T11:40:52.1854142Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_0 ('RERUN', {'yellow': True}) [4.2524s] [ 5%] 2025-12-04T11:40:52.1854684Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_0 ('RERUN', {'yellow': True}) [0.0996s] [ 5%] 2025-12-04T11:40:52.1855134Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_0 FAILED [0.0985s] [ 5%] 2025-12-04T11:40:52.1855141Z 2025-12-04T11:40:52.1855281Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1855554Z ____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 ____________ 2025-12-04T11:40:52.1855669Z Traceback (most recent call last): 2025-12-04T11:40:52.1856138Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1856364Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1856572Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1856577Z 2025-12-04T11:40:52.1856786Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1857425Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 2025-12-04T11:40:52.1857473Z 2025-12-04T11:40:52.1857737Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1857947Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1858064Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1858174Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1858508Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1858728Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1858820Z graph_break [] 2025-12-04T11:40:52.1859029Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1859445Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.1859663Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.1859859Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.1860092Z important use case and you want to speed it up with online 2025-12-04T11:40:52.1860178Z softmax. 2025-12-04T11:40:52.1860183Z 2025-12-04T11:40:52.1860284Z warnings.warn( 2025-12-04T11:40:52.1861040Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1861133Z warnings.warn( 2025-12-04T11:40:52.1861411Z ____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 ____________ 2025-12-04T11:40:52.1861523Z Traceback (most recent call last): 2025-12-04T11:40:52.1861984Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1862243Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1862446Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1862453Z 2025-12-04T11:40:52.1862673Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1863229Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 2025-12-04T11:40:52.1863233Z 2025-12-04T11:40:52.1863494Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1863711Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1863815Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1863929Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1864266Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1864478Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1864580Z graph_break [] 2025-12-04T11:40:52.1864788Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1865196Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.1865423Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.1865617Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.1865818Z important use case and you want to speed it up with online 2025-12-04T11:40:52.1865903Z softmax. 2025-12-04T11:40:52.1865909Z 2025-12-04T11:40:52.1866001Z warnings.warn( 2025-12-04T11:40:52.1866733Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1866826Z warnings.warn( 2025-12-04T11:40:52.1867033Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1867146Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1867286Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1867510Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1867849Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1867943Z graph_break [] 2025-12-04T11:40:52.1868160Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1868562Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.1868783Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.1868986Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.1869180Z important use case and you want to speed it up with online 2025-12-04T11:40:52.1869274Z softmax. 2025-12-04T11:40:52.1869279Z 2025-12-04T11:40:52.1869374Z warnings.warn( 2025-12-04T11:40:52.1870098Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1870230Z warnings.warn( 2025-12-04T11:40:52.1870372Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1870645Z ____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 ____________ 2025-12-04T11:40:52.1870867Z Traceback (most recent call last): 2025-12-04T11:40:52.1871337Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1871577Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1871786Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1871791Z 2025-12-04T11:40:52.1872004Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1872612Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 2025-12-04T11:40:52.1872619Z 2025-12-04T11:40:52.1872885Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1873109Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1873218Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1873332Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1873682Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1873898Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1873995Z graph_break [] 2025-12-04T11:40:52.1874212Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1874618Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.1874851Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.1875047Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.1875246Z important use case and you want to speed it up with online 2025-12-04T11:40:52.1875342Z softmax. 2025-12-04T11:40:52.1875348Z 2025-12-04T11:40:52.1875442Z warnings.warn( 2025-12-04T11:40:52.1876163Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1876266Z warnings.warn( 2025-12-04T11:40:52.1876477Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1876590Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1876700Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1876914Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1877259Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1877356Z graph_break [] 2025-12-04T11:40:52.1877610Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1878023Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.1878244Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.1878447Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.1878644Z important use case and you want to speed it up with online 2025-12-04T11:40:52.1878733Z softmax. 2025-12-04T11:40:52.1878738Z 2025-12-04T11:40:52.1878837Z warnings.warn( 2025-12-04T11:40:52.1879560Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1879658Z warnings.warn( 2025-12-04T11:40:52.1879877Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1879986Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1880105Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1880322Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1880692Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1880794Z graph_break [] 2025-12-04T11:40:52.1881033Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1881435Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.1881664Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.1881862Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.1882067Z important use case and you want to speed it up with online 2025-12-04T11:40:52.1882156Z softmax. 2025-12-04T11:40:52.1882160Z 2025-12-04T11:40:52.1882255Z warnings.warn( 2025-12-04T11:40:52.1883015Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1883117Z warnings.warn( 2025-12-04T11:40:52.1883877Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-89205029d9c04efd.xml - 2025-12-04T11:40:52.1884056Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1884763Z FAILED [0.0985s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_0 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1884768Z 2025-12-04T11:40:52.1884988Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1885555Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 2025-12-04T11:40:52.1885563Z 2025-12-04T11:40:52.1885823Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1886011Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1886208Z ================== 1 failed, 14 deselected, 2 rerun in 4.48s =================== 2025-12-04T11:40:52.1886306Z Got exit code 1 2025-12-04T11:40:52.1886408Z Retrying single test... 2025-12-04T11:40:52.1886994Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-476b174d4b08ad56.xml 2025-12-04T11:40:52.1887163Z ============================= test session starts ============================== 2025-12-04T11:40:52.1887508Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1887610Z cachedir: .pytest_cache 2025-12-04T11:40:52.1888133Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1888286Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1888397Z configfile: pytest.ini 2025-12-04T11:40:52.1888985Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1889194Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.1889848Z stepcurrent: skipping 14 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_0 2025-12-04T11:40:52.1889956Z Running 1 items in this shard 2025-12-04T11:40:52.1889961Z 2025-12-04T11:40:52.1890509Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_0 ('RERUN', {'yellow': True}) [4.2428s] [100%] 2025-12-04T11:40:52.1891048Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_0 ('RERUN', {'yellow': True}) [0.1000s] [100%] 2025-12-04T11:40:52.1891498Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_0 FAILED [0.0987s] [100%] 2025-12-04T11:40:52.1891535Z 2025-12-04T11:40:52.1891677Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1891948Z ____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 ____________ 2025-12-04T11:40:52.1892074Z Traceback (most recent call last): 2025-12-04T11:40:52.1892570Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1892796Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1893010Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1893015Z 2025-12-04T11:40:52.1893228Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1893834Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 2025-12-04T11:40:52.1893843Z 2025-12-04T11:40:52.1894110Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1894327Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1894450Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1894564Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1894908Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1895137Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1895236Z graph_break [] 2025-12-04T11:40:52.1895461Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1895875Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.1896280Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.1896496Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.1896700Z important use case and you want to speed it up with online 2025-12-04T11:40:52.1896796Z softmax. 2025-12-04T11:40:52.1896801Z 2025-12-04T11:40:52.1896987Z warnings.warn( 2025-12-04T11:40:52.1897724Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1897845Z warnings.warn( 2025-12-04T11:40:52.1898124Z ____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 ____________ 2025-12-04T11:40:52.1898249Z Traceback (most recent call last): 2025-12-04T11:40:52.1898739Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1898972Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1899187Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1899207Z 2025-12-04T11:40:52.1899498Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1900066Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 2025-12-04T11:40:52.1900073Z 2025-12-04T11:40:52.1900353Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1900574Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1900686Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1900818Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1901160Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1901395Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1901495Z graph_break [] 2025-12-04T11:40:52.1901714Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1902150Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.1902420Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.1902618Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.1902834Z important use case and you want to speed it up with online 2025-12-04T11:40:52.1902928Z softmax. 2025-12-04T11:40:52.1902974Z 2025-12-04T11:40:52.1903091Z warnings.warn( 2025-12-04T11:40:52.1903823Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1903924Z warnings.warn( 2025-12-04T11:40:52.1904157Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1904272Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1904388Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1904666Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1905009Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1905125Z graph_break [] 2025-12-04T11:40:52.1905341Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1905757Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.1905995Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.1906196Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.1906398Z important use case and you want to speed it up with online 2025-12-04T11:40:52.1906504Z softmax. 2025-12-04T11:40:52.1906509Z 2025-12-04T11:40:52.1906614Z warnings.warn( 2025-12-04T11:40:52.1907367Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1907469Z warnings.warn( 2025-12-04T11:40:52.1907614Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1907902Z ____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 ____________ 2025-12-04T11:40:52.1908025Z Traceback (most recent call last): 2025-12-04T11:40:52.1908497Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1908737Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1908946Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1908951Z 2025-12-04T11:40:52.1909173Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1909737Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 2025-12-04T11:40:52.1909746Z 2025-12-04T11:40:52.1910010Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1910272Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1910385Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1910513Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1910862Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1911079Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1911189Z graph_break [] 2025-12-04T11:40:52.1911403Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1911820Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.1912059Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.1912263Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.1912474Z important use case and you want to speed it up with online 2025-12-04T11:40:52.1912570Z softmax. 2025-12-04T11:40:52.1912575Z 2025-12-04T11:40:52.1912710Z warnings.warn( 2025-12-04T11:40:52.1913445Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1913544Z warnings.warn( 2025-12-04T11:40:52.1913787Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1913912Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1914026Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1914248Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1914603Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1914701Z graph_break [] 2025-12-04T11:40:52.1914975Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1915390Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.1915619Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.1915834Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.1916034Z important use case and you want to speed it up with online 2025-12-04T11:40:52.1916128Z softmax. 2025-12-04T11:40:52.1916146Z 2025-12-04T11:40:52.1916246Z warnings.warn( 2025-12-04T11:40:52.1916970Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1917083Z warnings.warn( 2025-12-04T11:40:52.1917299Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1917407Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1917537Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1917757Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1918094Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1918206Z graph_break [] 2025-12-04T11:40:52.1918418Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1918842Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.1919068Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.1919269Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.1919483Z important use case and you want to speed it up with online 2025-12-04T11:40:52.1919574Z softmax. 2025-12-04T11:40:52.1919579Z 2025-12-04T11:40:52.1919680Z warnings.warn( 2025-12-04T11:40:52.1920416Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1920548Z warnings.warn( 2025-12-04T11:40:52.1921332Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-476b174d4b08ad56.xml - 2025-12-04T11:40:52.1921509Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1922220Z FAILED [0.0987s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_0 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1922240Z 2025-12-04T11:40:52.1922453Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1923021Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 2025-12-04T11:40:52.1923026Z 2025-12-04T11:40:52.1923306Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1923488Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1923690Z ================== 1 failed, 30 deselected, 2 rerun in 4.47s =================== 2025-12-04T11:40:52.1923835Z Got exit code 1 2025-12-04T11:40:52.1923945Z Retrying single test... 2025-12-04T11:40:52.1924580Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-99551b4e3863e77a.xml 2025-12-04T11:40:52.1924744Z ============================= test session starts ============================== 2025-12-04T11:40:52.1925092Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1925212Z cachedir: .pytest_cache 2025-12-04T11:40:52.1925730Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1925856Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1926010Z configfile: pytest.ini 2025-12-04T11:40:52.1926606Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1926832Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.1927487Z stepcurrent: skipping 14 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_0 2025-12-04T11:40:52.1927601Z Running 1 items in this shard 2025-12-04T11:40:52.1927606Z 2025-12-04T11:40:52.1928159Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_0 ('RERUN', {'yellow': True}) [4.3066s] [100%] 2025-12-04T11:40:52.1928701Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_0 ('RERUN', {'yellow': True}) [0.1011s] [100%] 2025-12-04T11:40:52.1929173Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_0 FAILED [0.0997s] [100%] 2025-12-04T11:40:52.1929178Z 2025-12-04T11:40:52.1929321Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1929596Z ____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 ____________ 2025-12-04T11:40:52.1929732Z Traceback (most recent call last): 2025-12-04T11:40:52.1930207Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1930451Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1930661Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1930666Z 2025-12-04T11:40:52.1930880Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1931458Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 2025-12-04T11:40:52.1931463Z 2025-12-04T11:40:52.1931731Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1931998Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1932112Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1932225Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1932579Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1932800Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1932898Z graph_break [] 2025-12-04T11:40:52.1933123Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1933537Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.1933773Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.1933971Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.1934175Z important use case and you want to speed it up with online 2025-12-04T11:40:52.1934281Z softmax. 2025-12-04T11:40:52.1934286Z 2025-12-04T11:40:52.1934386Z warnings.warn( 2025-12-04T11:40:52.1935151Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1935263Z warnings.warn( 2025-12-04T11:40:52.1935567Z ____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 ____________ 2025-12-04T11:40:52.1935699Z Traceback (most recent call last): 2025-12-04T11:40:52.1936173Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1936404Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1936625Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1936630Z 2025-12-04T11:40:52.1936876Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1937516Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 2025-12-04T11:40:52.1937538Z 2025-12-04T11:40:52.1937803Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1938019Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1938146Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1938261Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1938600Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1938834Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1938933Z graph_break [] 2025-12-04T11:40:52.1939148Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1939581Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.1939813Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.1940026Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.1940230Z important use case and you want to speed it up with online 2025-12-04T11:40:52.1940322Z softmax. 2025-12-04T11:40:52.1940326Z 2025-12-04T11:40:52.1940442Z warnings.warn( 2025-12-04T11:40:52.1941170Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1941283Z warnings.warn( 2025-12-04T11:40:52.1941499Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1941609Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1941736Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1941955Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1942294Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1942449Z graph_break [] 2025-12-04T11:40:52.1942664Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1943070Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.1943310Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.1943507Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.1943718Z important use case and you want to speed it up with online 2025-12-04T11:40:52.1943809Z softmax. 2025-12-04T11:40:52.1943814Z 2025-12-04T11:40:52.1943911Z warnings.warn( 2025-12-04T11:40:52.1944647Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1944748Z warnings.warn( 2025-12-04T11:40:52.1944900Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1945188Z ____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 ____________ 2025-12-04T11:40:52.1945339Z Traceback (most recent call last): 2025-12-04T11:40:52.1945816Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1946077Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1946288Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1946293Z 2025-12-04T11:40:52.1946519Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1947088Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 2025-12-04T11:40:52.1947093Z 2025-12-04T11:40:52.1947402Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1947616Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1947729Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1947855Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1948193Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1948411Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1948523Z graph_break [] 2025-12-04T11:40:52.1948738Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1949157Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.1949381Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.1949577Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.1949784Z important use case and you want to speed it up with online 2025-12-04T11:40:52.1949875Z softmax. 2025-12-04T11:40:52.1949880Z 2025-12-04T11:40:52.1949980Z warnings.warn( 2025-12-04T11:40:52.1950721Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1950821Z warnings.warn( 2025-12-04T11:40:52.1951047Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1951156Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1951269Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1951498Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1951835Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1951933Z graph_break [] 2025-12-04T11:40:52.1952156Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1952566Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.1952832Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.1953034Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.1953234Z important use case and you want to speed it up with online 2025-12-04T11:40:52.1953335Z softmax. 2025-12-04T11:40:52.1953340Z 2025-12-04T11:40:52.1953439Z warnings.warn( 2025-12-04T11:40:52.1954168Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1954277Z warnings.warn( 2025-12-04T11:40:52.1954492Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1954609Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1954716Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1954933Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1955283Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1955382Z graph_break [] 2025-12-04T11:40:52.1955631Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1956046Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.1956310Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.1956515Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.1956710Z important use case and you want to speed it up with online 2025-12-04T11:40:52.1956800Z softmax. 2025-12-04T11:40:52.1956805Z 2025-12-04T11:40:52.1956906Z warnings.warn( 2025-12-04T11:40:52.1957628Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1957757Z warnings.warn( 2025-12-04T11:40:52.1958536Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-99551b4e3863e77a.xml - 2025-12-04T11:40:52.1958708Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1959427Z FAILED [0.0997s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_0 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1959432Z 2025-12-04T11:40:52.1959642Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1960210Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_0 2025-12-04T11:40:52.1960222Z 2025-12-04T11:40:52.1960484Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1960666Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1960877Z ================== 1 failed, 30 deselected, 2 rerun in 4.54s =================== 2025-12-04T11:40:52.1960972Z Got exit code 1 2025-12-04T11:40:52.1961458Z FAILED CONSISTENTLY: test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_0 2025-12-04T11:40:52.1961877Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:40:52.1962469Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-498b08ff107b8914.xml 2025-12-04T11:40:52.1962637Z ============================= test session starts ============================== 2025-12-04T11:40:52.1962979Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1963085Z cachedir: .pytest_cache 2025-12-04T11:40:52.1963614Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1964409Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1964517Z configfile: pytest.ini 2025-12-04T11:40:52.1965117Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1965334Z collecting ... collected 31 items / 15 deselected / 16 selected 2025-12-04T11:40:52.1965486Z stepcurrent: skipping 15 already run items. 2025-12-04T11:40:52.1965596Z Running 16 items in this shard 2025-12-04T11:40:52.1965603Z 2025-12-04T11:40:52.1966149Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_1 ('RERUN', {'yellow': True}) [4.2595s] [ 6%] 2025-12-04T11:40:52.1966690Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_1 ('RERUN', {'yellow': True}) [0.0936s] [ 6%] 2025-12-04T11:40:52.1967143Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_1 FAILED [0.0931s] [ 6%] 2025-12-04T11:40:52.1967151Z 2025-12-04T11:40:52.1967301Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1967605Z ____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 ____________ 2025-12-04T11:40:52.1967726Z Traceback (most recent call last): 2025-12-04T11:40:52.1968235Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1968466Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1968686Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1968691Z 2025-12-04T11:40:52.1968900Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1969469Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 2025-12-04T11:40:52.1969474Z 2025-12-04T11:40:52.1969776Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1969994Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1970120Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1970232Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1970575Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1970808Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1970905Z graph_break [] 2025-12-04T11:40:52.1971119Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1971866Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1971966Z warnings.warn( 2025-12-04T11:40:52.1972255Z ____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 ____________ 2025-12-04T11:40:52.1972373Z Traceback (most recent call last): 2025-12-04T11:40:52.1972840Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1973082Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1973287Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1973292Z 2025-12-04T11:40:52.1973505Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1974083Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 2025-12-04T11:40:52.1974088Z 2025-12-04T11:40:52.1974354Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1974581Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1974691Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1974808Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1975190Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1975409Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1975508Z graph_break [] 2025-12-04T11:40:52.1975732Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1976470Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1976579Z warnings.warn( 2025-12-04T11:40:52.1976789Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1976969Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1977099Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1977315Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1977658Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1977766Z graph_break [] 2025-12-04T11:40:52.1978019Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1978755Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1978887Z warnings.warn( 2025-12-04T11:40:52.1979028Z =================================== FAILURES =================================== 2025-12-04T11:40:52.1979312Z ____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 ____________ 2025-12-04T11:40:52.1979432Z Traceback (most recent call last): 2025-12-04T11:40:52.1979901Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1980165Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1980373Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1980380Z 2025-12-04T11:40:52.1980600Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1981167Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 2025-12-04T11:40:52.1981172Z 2025-12-04T11:40:52.1981438Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1981659Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1981766Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1981894Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1982228Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1982444Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1982548Z graph_break [] 2025-12-04T11:40:52.1982757Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1983488Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1983598Z warnings.warn( 2025-12-04T11:40:52.1983808Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1983929Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1984040Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1984258Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1984600Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1984695Z graph_break [] 2025-12-04T11:40:52.1984905Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1985645Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1985778Z warnings.warn( 2025-12-04T11:40:52.1986004Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1986110Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1986222Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.1986455Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.1986793Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.1986888Z graph_break [] 2025-12-04T11:40:52.1987106Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.1987828Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.1987938Z warnings.warn( 2025-12-04T11:40:52.1988701Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-498b08ff107b8914.xml - 2025-12-04T11:40:52.1988905Z =========================== short test summary info ============================ 2025-12-04T11:40:52.1989667Z FAILED [0.0931s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_1 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1989673Z 2025-12-04T11:40:52.1989887Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1990466Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 2025-12-04T11:40:52.1990472Z 2025-12-04T11:40:52.1990734Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1990940Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.1991148Z ================== 1 failed, 15 deselected, 2 rerun in 4.48s =================== 2025-12-04T11:40:52.1991245Z Got exit code 1 2025-12-04T11:40:52.1991365Z Retrying single test... 2025-12-04T11:40:52.1991950Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0e343f505ff50c95.xml 2025-12-04T11:40:52.1992108Z ============================= test session starts ============================== 2025-12-04T11:40:52.1992465Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.1992570Z cachedir: .pytest_cache 2025-12-04T11:40:52.1993094Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.1993225Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.1993330Z configfile: pytest.ini 2025-12-04T11:40:52.1993932Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.1994145Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.1994795Z stepcurrent: skipping 15 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_1 2025-12-04T11:40:52.1994918Z Running 1 items in this shard 2025-12-04T11:40:52.1994926Z 2025-12-04T11:40:52.1995465Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_1 ('RERUN', {'yellow': True}) [4.2283s] [100%] 2025-12-04T11:40:52.1996185Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_1 ('RERUN', {'yellow': True}) [0.0946s] [100%] 2025-12-04T11:40:52.1996638Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_1 FAILED [0.0937s] [100%] 2025-12-04T11:40:52.1996644Z 2025-12-04T11:40:52.1996786Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.1997146Z ____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 ____________ 2025-12-04T11:40:52.1997269Z Traceback (most recent call last): 2025-12-04T11:40:52.1997753Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.1997984Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.1998195Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.1998201Z 2025-12-04T11:40:52.1998428Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.1998991Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 2025-12-04T11:40:52.1998996Z 2025-12-04T11:40:52.1999273Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.1999490Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.1999603Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.1999784Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2000122Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2000339Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2000488Z graph_break [] 2025-12-04T11:40:52.2000705Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2001452Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2001551Z warnings.warn( 2025-12-04T11:40:52.2001822Z ____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 ____________ 2025-12-04T11:40:52.2001955Z Traceback (most recent call last): 2025-12-04T11:40:52.2002467Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2002702Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2002922Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2002927Z 2025-12-04T11:40:52.2003141Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2003723Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 2025-12-04T11:40:52.2003728Z 2025-12-04T11:40:52.2003989Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2004206Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2004331Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2004443Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2004786Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2005018Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2005118Z graph_break [] 2025-12-04T11:40:52.2005346Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2006080Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2006181Z warnings.warn( 2025-12-04T11:40:52.2006404Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2006512Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2006625Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2006848Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2007187Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2007293Z graph_break [] 2025-12-04T11:40:52.2007536Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2008265Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2008374Z warnings.warn( 2025-12-04T11:40:52.2008517Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2008791Z ____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 ____________ 2025-12-04T11:40:52.2008922Z Traceback (most recent call last): 2025-12-04T11:40:52.2009390Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2009627Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2009830Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2009838Z 2025-12-04T11:40:52.2010046Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2010624Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 2025-12-04T11:40:52.2010664Z 2025-12-04T11:40:52.2010926Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2011182Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2011291Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2011403Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2011748Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2011962Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2012055Z graph_break [] 2025-12-04T11:40:52.2012307Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2013037Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2013151Z warnings.warn( 2025-12-04T11:40:52.2013356Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2013463Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2013590Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2013805Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2014140Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2014244Z graph_break [] 2025-12-04T11:40:52.2014455Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2015189Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2015285Z warnings.warn( 2025-12-04T11:40:52.2015498Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2015620Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2015730Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2015949Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2016297Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2016393Z graph_break [] 2025-12-04T11:40:52.2016614Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2017402Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2017503Z warnings.warn( 2025-12-04T11:40:52.2018278Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0e343f505ff50c95.xml - 2025-12-04T11:40:52.2018492Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2019211Z FAILED [0.0937s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_1 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2019216Z 2025-12-04T11:40:52.2019430Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2019993Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 2025-12-04T11:40:52.2019998Z 2025-12-04T11:40:52.2020273Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2020451Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2020662Z ================== 1 failed, 30 deselected, 2 rerun in 4.45s =================== 2025-12-04T11:40:52.2020754Z Got exit code 1 2025-12-04T11:40:52.2020859Z Retrying single test... 2025-12-04T11:40:52.2021462Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-486e74c9978d3a7f.xml 2025-12-04T11:40:52.2021648Z ============================= test session starts ============================== 2025-12-04T11:40:52.2022029Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2022150Z cachedir: .pytest_cache 2025-12-04T11:40:52.2022675Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2022814Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2022921Z configfile: pytest.ini 2025-12-04T11:40:52.2023542Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2023769Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.2024425Z stepcurrent: skipping 15 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_1 2025-12-04T11:40:52.2024541Z Running 1 items in this shard 2025-12-04T11:40:52.2024560Z 2025-12-04T11:40:52.2025103Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_1 ('RERUN', {'yellow': True}) [4.2409s] [100%] 2025-12-04T11:40:52.2025641Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_1 ('RERUN', {'yellow': True}) [0.0936s] [100%] 2025-12-04T11:40:52.2026113Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_1 FAILED [0.0918s] [100%] 2025-12-04T11:40:52.2026119Z 2025-12-04T11:40:52.2026257Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2026547Z ____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 ____________ 2025-12-04T11:40:52.2026669Z Traceback (most recent call last): 2025-12-04T11:40:52.2027142Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2027385Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2027593Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2027600Z 2025-12-04T11:40:52.2027810Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2028391Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 2025-12-04T11:40:52.2028397Z 2025-12-04T11:40:52.2028658Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2028886Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2028997Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2029107Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2029487Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2029705Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2029811Z graph_break [] 2025-12-04T11:40:52.2030023Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2030754Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2030866Z warnings.warn( 2025-12-04T11:40:52.2031137Z ____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 ____________ 2025-12-04T11:40:52.2031258Z Traceback (most recent call last): 2025-12-04T11:40:52.2031734Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2031959Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2032180Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2032215Z 2025-12-04T11:40:52.2032425Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2033022Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 2025-12-04T11:40:52.2033028Z 2025-12-04T11:40:52.2033302Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2033513Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2033638Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2033748Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2034089Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2034348Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2034450Z graph_break [] 2025-12-04T11:40:52.2034662Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2035410Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2035512Z warnings.warn( 2025-12-04T11:40:52.2035740Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2035850Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2035961Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2036187Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2036527Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2036623Z graph_break [] 2025-12-04T11:40:52.2036847Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2037577Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2037689Z warnings.warn( 2025-12-04T11:40:52.2037828Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2038103Z ____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 ____________ 2025-12-04T11:40:52.2038230Z Traceback (most recent call last): 2025-12-04T11:40:52.2038698Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2038928Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2039141Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2039147Z 2025-12-04T11:40:52.2039360Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2039937Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 2025-12-04T11:40:52.2039993Z 2025-12-04T11:40:52.2040255Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2040467Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2040590Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2040704Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2041043Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2041276Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2041374Z graph_break [] 2025-12-04T11:40:52.2041600Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2042332Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2042437Z warnings.warn( 2025-12-04T11:40:52.2042664Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2042884Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2042998Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2043226Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2043592Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2043704Z graph_break [] 2025-12-04T11:40:52.2043913Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2044640Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2044747Z warnings.warn( 2025-12-04T11:40:52.2044989Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2045102Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2045224Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2045443Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2045789Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2045882Z graph_break [] 2025-12-04T11:40:52.2046096Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2046837Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2046934Z warnings.warn( 2025-12-04T11:40:52.2047697Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-486e74c9978d3a7f.xml - 2025-12-04T11:40:52.2047879Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2048592Z FAILED [0.0918s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_1 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2048600Z 2025-12-04T11:40:52.2048825Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2049392Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2048_dim_1 2025-12-04T11:40:52.2049397Z 2025-12-04T11:40:52.2049675Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2049857Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2050053Z ================== 1 failed, 30 deselected, 2 rerun in 4.46s =================== 2025-12-04T11:40:52.2050164Z Got exit code 1 2025-12-04T11:40:52.2050655Z FAILED CONSISTENTLY: test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_1 2025-12-04T11:40:52.2051100Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:40:52.2051701Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a0676d6303da3615.xml 2025-12-04T11:40:52.2051868Z ============================= test session starts ============================== 2025-12-04T11:40:52.2052225Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2052336Z cachedir: .pytest_cache 2025-12-04T11:40:52.2052856Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2052993Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2053103Z configfile: pytest.ini 2025-12-04T11:40:52.2053713Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2053934Z collecting ... collected 31 items / 16 deselected / 15 selected 2025-12-04T11:40:52.2054117Z stepcurrent: skipping 16 already run items. 2025-12-04T11:40:52.2054244Z Running 15 items in this shard 2025-12-04T11:40:52.2054249Z 2025-12-04T11:40:52.2054817Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_-1 ('RERUN', {'yellow': True}) [4.2593s] [ 6%] 2025-12-04T11:40:52.2055348Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_-1 ('RERUN', {'yellow': True}) [0.0941s] [ 6%] 2025-12-04T11:40:52.2055808Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_-1 FAILED [0.0939s] [ 6%] 2025-12-04T11:40:52.2055814Z 2025-12-04T11:40:52.2055955Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2056272Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 _____________ 2025-12-04T11:40:52.2056399Z Traceback (most recent call last): 2025-12-04T11:40:52.2056869Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2057189Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2057401Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2057409Z 2025-12-04T11:40:52.2057635Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2058191Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 2025-12-04T11:40:52.2058197Z 2025-12-04T11:40:52.2058463Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2058695Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2058809Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2058937Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2059282Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2059503Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2059613Z graph_break [] 2025-12-04T11:40:52.2059828Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2060569Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2060683Z warnings.warn( 2025-12-04T11:40:52.2060957Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 _____________ 2025-12-04T11:40:52.2061092Z Traceback (most recent call last): 2025-12-04T11:40:52.2061564Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2061794Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2062051Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2062059Z 2025-12-04T11:40:52.2062269Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2062825Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 2025-12-04T11:40:52.2062840Z 2025-12-04T11:40:52.2063104Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2063316Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2063438Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2063549Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2063888Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2064117Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2064215Z graph_break [] 2025-12-04T11:40:52.2064424Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2065193Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2065294Z warnings.warn( 2025-12-04T11:40:52.2065546Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2065655Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2065767Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2065995Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2066332Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2066428Z graph_break [] 2025-12-04T11:40:52.2066677Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2067409Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2067523Z warnings.warn( 2025-12-04T11:40:52.2067663Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2067936Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 _____________ 2025-12-04T11:40:52.2068069Z Traceback (most recent call last): 2025-12-04T11:40:52.2068539Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2068780Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2068988Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2068994Z 2025-12-04T11:40:52.2069207Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2069771Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 2025-12-04T11:40:52.2069780Z 2025-12-04T11:40:52.2070045Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2070268Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2070376Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2070491Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2070842Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2071060Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2071157Z graph_break [] 2025-12-04T11:40:52.2071375Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2072109Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2072249Z warnings.warn( 2025-12-04T11:40:52.2072461Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2072573Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2072695Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2072911Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2073251Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2073357Z graph_break [] 2025-12-04T11:40:52.2073568Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2074294Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2074401Z warnings.warn( 2025-12-04T11:40:52.2074618Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2074739Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2074851Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2075102Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2075453Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2075550Z graph_break [] 2025-12-04T11:40:52.2075792Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2076533Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2076631Z warnings.warn( 2025-12-04T11:40:52.2077405Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a0676d6303da3615.xml - 2025-12-04T11:40:52.2077618Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2078313Z FAILED [0.0939s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_-1 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2078334Z 2025-12-04T11:40:52.2078550Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2079106Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 2025-12-04T11:40:52.2079110Z 2025-12-04T11:40:52.2079387Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2079562Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2079761Z ================== 1 failed, 16 deselected, 2 rerun in 4.48s =================== 2025-12-04T11:40:52.2079867Z Got exit code 1 2025-12-04T11:40:52.2079973Z Retrying single test... 2025-12-04T11:40:52.2080575Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-e27b1d19486b6954.xml 2025-12-04T11:40:52.2080742Z ============================= test session starts ============================== 2025-12-04T11:40:52.2081089Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2081209Z cachedir: .pytest_cache 2025-12-04T11:40:52.2081732Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2081858Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2081977Z configfile: pytest.ini 2025-12-04T11:40:52.2082571Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2082791Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.2083433Z stepcurrent: skipping 16 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_-1 2025-12-04T11:40:52.2083582Z Running 1 items in this shard 2025-12-04T11:40:52.2083587Z 2025-12-04T11:40:52.2084131Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_-1 ('RERUN', {'yellow': True}) [4.2333s] [100%] 2025-12-04T11:40:52.2084663Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_-1 ('RERUN', {'yellow': True}) [0.0930s] [100%] 2025-12-04T11:40:52.2085122Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_-1 FAILED [0.0935s] [100%] 2025-12-04T11:40:52.2085127Z 2025-12-04T11:40:52.2085265Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2085539Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 _____________ 2025-12-04T11:40:52.2085672Z Traceback (most recent call last): 2025-12-04T11:40:52.2086138Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2086383Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2086628Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2086633Z 2025-12-04T11:40:52.2086846Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2087445Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 2025-12-04T11:40:52.2087451Z 2025-12-04T11:40:52.2087715Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2087944Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2088059Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2088172Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2088551Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2088776Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2088877Z graph_break [] 2025-12-04T11:40:52.2089103Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2089839Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2089951Z warnings.warn( 2025-12-04T11:40:52.2090223Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 _____________ 2025-12-04T11:40:52.2090344Z Traceback (most recent call last): 2025-12-04T11:40:52.2090827Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2091057Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2091267Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2091275Z 2025-12-04T11:40:52.2091494Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2092052Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 2025-12-04T11:40:52.2092057Z 2025-12-04T11:40:52.2092333Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2092547Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2092658Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2092780Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2093118Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2093345Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2093443Z graph_break [] 2025-12-04T11:40:52.2093657Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2094428Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2094532Z warnings.warn( 2025-12-04T11:40:52.2094746Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2094869Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2094982Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2095202Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2095553Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2095650Z graph_break [] 2025-12-04T11:40:52.2095876Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2096759Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2096863Z warnings.warn( 2025-12-04T11:40:52.2097146Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2097416Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 _____________ 2025-12-04T11:40:52.2097546Z Traceback (most recent call last): 2025-12-04T11:40:52.2098061Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2098293Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2098513Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2098519Z 2025-12-04T11:40:52.2098731Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2099328Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 2025-12-04T11:40:52.2099346Z 2025-12-04T11:40:52.2099613Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2099831Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2099952Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2100066Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2100407Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2100636Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2100733Z graph_break [] 2025-12-04T11:40:52.2100947Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2101685Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2101785Z warnings.warn( 2025-12-04T11:40:52.2102011Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2102123Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2102239Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2102471Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2102812Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2102913Z graph_break [] 2025-12-04T11:40:52.2103139Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2103866Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2103981Z warnings.warn( 2025-12-04T11:40:52.2104197Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2104309Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2104439Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2104701Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2105039Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2105148Z graph_break [] 2025-12-04T11:40:52.2105362Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2106108Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2106209Z warnings.warn( 2025-12-04T11:40:52.2106982Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-e27b1d19486b6954.xml - 2025-12-04T11:40:52.2107165Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2107865Z FAILED [0.0935s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_-1 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2107905Z 2025-12-04T11:40:52.2108132Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2108690Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 2025-12-04T11:40:52.2108695Z 2025-12-04T11:40:52.2108990Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2109178Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2109379Z ================== 1 failed, 30 deselected, 2 rerun in 4.45s =================== 2025-12-04T11:40:52.2109489Z Got exit code 1 2025-12-04T11:40:52.2109595Z Retrying single test... 2025-12-04T11:40:52.2110217Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3f1ba8060f3abc48.xml 2025-12-04T11:40:52.2110393Z ============================= test session starts ============================== 2025-12-04T11:40:52.2110741Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2110855Z cachedir: .pytest_cache 2025-12-04T11:40:52.2111394Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2111520Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2111640Z configfile: pytest.ini 2025-12-04T11:40:52.2112229Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2112442Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.2113091Z stepcurrent: skipping 16 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_-1 2025-12-04T11:40:52.2113205Z Running 1 items in this shard 2025-12-04T11:40:52.2113212Z 2025-12-04T11:40:52.2113755Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_-1 ('RERUN', {'yellow': True}) [4.2408s] [100%] 2025-12-04T11:40:52.2114282Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_-1 ('RERUN', {'yellow': True}) [0.0938s] [100%] 2025-12-04T11:40:52.2114732Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_-1 FAILED [0.0937s] [100%] 2025-12-04T11:40:52.2114738Z 2025-12-04T11:40:52.2114885Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2115158Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 _____________ 2025-12-04T11:40:52.2115289Z Traceback (most recent call last): 2025-12-04T11:40:52.2115756Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2115992Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2116246Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2116254Z 2025-12-04T11:40:52.2116469Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2117028Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 2025-12-04T11:40:52.2117042Z 2025-12-04T11:40:52.2117302Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2117516Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2117632Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2117746Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2118086Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2118316Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2118415Z graph_break [] 2025-12-04T11:40:52.2118628Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2119413Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2119515Z warnings.warn( 2025-12-04T11:40:52.2119842Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 _____________ 2025-12-04T11:40:52.2119962Z Traceback (most recent call last): 2025-12-04T11:40:52.2120432Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2120670Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2120878Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2120883Z 2025-12-04T11:40:52.2121132Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2121690Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 2025-12-04T11:40:52.2121697Z 2025-12-04T11:40:52.2121960Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2122181Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2122293Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2122419Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2122759Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2122978Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2123086Z graph_break [] 2025-12-04T11:40:52.2123299Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2124027Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2124142Z warnings.warn( 2025-12-04T11:40:52.2124360Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2124480Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2124594Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2124816Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2125163Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2125260Z graph_break [] 2025-12-04T11:40:52.2125471Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2126209Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2126316Z warnings.warn( 2025-12-04T11:40:52.2126459Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2126773Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 _____________ 2025-12-04T11:40:52.2126898Z Traceback (most recent call last): 2025-12-04T11:40:52.2127374Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2127606Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2127817Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2127823Z 2025-12-04T11:40:52.2128045Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2128600Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 2025-12-04T11:40:52.2128605Z 2025-12-04T11:40:52.2128882Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2129094Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2129208Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2129363Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2129698Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2129944Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2130052Z graph_break [] 2025-12-04T11:40:52.2130265Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2131005Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2131104Z warnings.warn( 2025-12-04T11:40:52.2131315Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2131462Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2131573Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2131793Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2132146Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2132242Z graph_break [] 2025-12-04T11:40:52.2132466Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2133190Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2133292Z warnings.warn( 2025-12-04T11:40:52.2133518Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2133628Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2133741Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2133970Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2134306Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2134414Z graph_break [] 2025-12-04T11:40:52.2134625Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2135352Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2135461Z warnings.warn( 2025-12-04T11:40:52.2136234Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3f1ba8060f3abc48.xml - 2025-12-04T11:40:52.2136413Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2137179Z FAILED [0.0937s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_-1 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2137186Z 2025-12-04T11:40:52.2137442Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2138011Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_-1 2025-12-04T11:40:52.2138019Z 2025-12-04T11:40:52.2138283Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2138478Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2138673Z ================== 1 failed, 30 deselected, 2 rerun in 4.46s =================== 2025-12-04T11:40:52.2138772Z Got exit code 1 2025-12-04T11:40:52.2139265Z FAILED CONSISTENTLY: test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_-1 2025-12-04T11:40:52.2139668Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:40:52.2140262Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0209a84b5ca28602.xml 2025-12-04T11:40:52.2140436Z ============================= test session starts ============================== 2025-12-04T11:40:52.2140817Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2140934Z cachedir: .pytest_cache 2025-12-04T11:40:52.2141484Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2141607Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2141722Z configfile: pytest.ini 2025-12-04T11:40:52.2142310Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2142541Z collecting ... collected 31 items / 17 deselected / 14 selected 2025-12-04T11:40:52.2142712Z stepcurrent: skipping 17 already run items. 2025-12-04T11:40:52.2142825Z Running 14 items in this shard 2025-12-04T11:40:52.2142833Z 2025-12-04T11:40:52.2143378Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_0 ('RERUN', {'yellow': True}) [4.2352s] [ 7%] 2025-12-04T11:40:52.2143905Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_0 ('RERUN', {'yellow': True}) [0.0964s] [ 7%] 2025-12-04T11:40:52.2144352Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_0 FAILED [0.0962s] [ 7%] 2025-12-04T11:40:52.2144370Z 2025-12-04T11:40:52.2144507Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2144779Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 ______________ 2025-12-04T11:40:52.2144910Z Traceback (most recent call last): 2025-12-04T11:40:52.2145378Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2145611Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2145832Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2145840Z 2025-12-04T11:40:52.2146051Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2146619Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 2025-12-04T11:40:52.2146627Z 2025-12-04T11:40:52.2146891Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2147106Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2147229Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2147344Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2147695Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2147914Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2148013Z graph_break [] 2025-12-04T11:40:52.2148269Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2148689Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2148912Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2149125Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2149323Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2149414Z softmax. 2025-12-04T11:40:52.2149429Z 2025-12-04T11:40:52.2149528Z warnings.warn( 2025-12-04T11:40:52.2150257Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2150367Z warnings.warn( 2025-12-04T11:40:52.2150642Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 ______________ 2025-12-04T11:40:52.2150764Z Traceback (most recent call last): 2025-12-04T11:40:52.2151243Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2151506Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2151725Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2151730Z 2025-12-04T11:40:52.2151966Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2152516Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 2025-12-04T11:40:52.2152521Z 2025-12-04T11:40:52.2152794Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2153009Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2153139Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2153281Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2153624Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2153860Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2153960Z graph_break [] 2025-12-04T11:40:52.2154176Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2154607Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2154834Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2155034Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2155246Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2155341Z softmax. 2025-12-04T11:40:52.2155346Z 2025-12-04T11:40:52.2155461Z warnings.warn( 2025-12-04T11:40:52.2156191Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2156293Z warnings.warn( 2025-12-04T11:40:52.2156525Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2156637Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2156753Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2156989Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2157327Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2157440Z graph_break [] 2025-12-04T11:40:52.2157658Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2158072Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2158311Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2158513Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2158749Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2158857Z softmax. 2025-12-04T11:40:52.2158862Z 2025-12-04T11:40:52.2158963Z warnings.warn( 2025-12-04T11:40:52.2159703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2159806Z warnings.warn( 2025-12-04T11:40:52.2159947Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2160235Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 ______________ 2025-12-04T11:40:52.2160356Z Traceback (most recent call last): 2025-12-04T11:40:52.2160823Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2161071Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2161281Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2161289Z 2025-12-04T11:40:52.2161516Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2162115Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 2025-12-04T11:40:52.2162121Z 2025-12-04T11:40:52.2162426Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2162652Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2162767Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2162894Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2163233Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2163449Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2163564Z graph_break [] 2025-12-04T11:40:52.2163804Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2164221Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2164458Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2164656Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2164869Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2164959Z softmax. 2025-12-04T11:40:52.2164964Z 2025-12-04T11:40:52.2165063Z warnings.warn( 2025-12-04T11:40:52.2165798Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2165897Z warnings.warn( 2025-12-04T11:40:52.2166111Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2166235Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2166349Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2166581Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2166920Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2167017Z graph_break [] 2025-12-04T11:40:52.2167241Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2167652Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2167875Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2168085Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2168284Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2168386Z softmax. 2025-12-04T11:40:52.2168391Z 2025-12-04T11:40:52.2168491Z warnings.warn( 2025-12-04T11:40:52.2169222Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2169369Z warnings.warn( 2025-12-04T11:40:52.2169587Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2169696Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2169826Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2170047Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2170398Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2170499Z graph_break [] 2025-12-04T11:40:52.2170712Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2171133Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2171357Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2171556Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2171769Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2171888Z softmax. 2025-12-04T11:40:52.2171893Z 2025-12-04T11:40:52.2172003Z warnings.warn( 2025-12-04T11:40:52.2172756Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2172856Z warnings.warn( 2025-12-04T11:40:52.2173628Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0209a84b5ca28602.xml - 2025-12-04T11:40:52.2173800Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2174487Z FAILED [0.0962s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_0 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2174534Z 2025-12-04T11:40:52.2174749Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2175305Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 2025-12-04T11:40:52.2175312Z 2025-12-04T11:40:52.2175586Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2175768Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2175978Z ================== 1 failed, 17 deselected, 2 rerun in 4.46s =================== 2025-12-04T11:40:52.2176077Z Got exit code 1 2025-12-04T11:40:52.2176184Z Retrying single test... 2025-12-04T11:40:52.2176788Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3d838b8cb24fd196.xml 2025-12-04T11:40:52.2177025Z ============================= test session starts ============================== 2025-12-04T11:40:52.2177382Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2177510Z cachedir: .pytest_cache 2025-12-04T11:40:52.2178037Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2178176Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2178285Z configfile: pytest.ini 2025-12-04T11:40:52.2178882Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2179112Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.2179749Z stepcurrent: skipping 17 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_0 2025-12-04T11:40:52.2179864Z Running 1 items in this shard 2025-12-04T11:40:52.2179869Z 2025-12-04T11:40:52.2180415Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_0 ('RERUN', {'yellow': True}) [4.2339s] [100%] 2025-12-04T11:40:52.2180979Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_0 ('RERUN', {'yellow': True}) [0.0946s] [100%] 2025-12-04T11:40:52.2181433Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_0 FAILED [0.0946s] [100%] 2025-12-04T11:40:52.2181439Z 2025-12-04T11:40:52.2181579Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2181852Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 ______________ 2025-12-04T11:40:52.2181984Z Traceback (most recent call last): 2025-12-04T11:40:52.2182455Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2182695Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2182910Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2182917Z 2025-12-04T11:40:52.2183131Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2183721Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 2025-12-04T11:40:52.2183726Z 2025-12-04T11:40:52.2183992Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2184251Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2184367Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2184481Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2184835Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2185054Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2185151Z graph_break [] 2025-12-04T11:40:52.2185405Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2185819Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2186061Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2186261Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2186462Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2186569Z softmax. 2025-12-04T11:40:52.2186573Z 2025-12-04T11:40:52.2186672Z warnings.warn( 2025-12-04T11:40:52.2187400Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2187510Z warnings.warn( 2025-12-04T11:40:52.2187784Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 ______________ 2025-12-04T11:40:52.2187919Z Traceback (most recent call last): 2025-12-04T11:40:52.2188391Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2188622Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2188845Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2188850Z 2025-12-04T11:40:52.2189064Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2189626Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 2025-12-04T11:40:52.2189631Z 2025-12-04T11:40:52.2189894Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2190109Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2190232Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2190346Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2190687Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2190948Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2191047Z graph_break [] 2025-12-04T11:40:52.2191271Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2191682Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2191913Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2192123Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2192322Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2192412Z softmax. 2025-12-04T11:40:52.2192416Z 2025-12-04T11:40:52.2192527Z warnings.warn( 2025-12-04T11:40:52.2193257Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2193371Z warnings.warn( 2025-12-04T11:40:52.2193589Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2193701Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2193857Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2194077Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2194446Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2194556Z graph_break [] 2025-12-04T11:40:52.2194769Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2195194Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2195420Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2195618Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2195858Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2195950Z softmax. 2025-12-04T11:40:52.2195957Z 2025-12-04T11:40:52.2196231Z warnings.warn( 2025-12-04T11:40:52.2196975Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2197076Z warnings.warn( 2025-12-04T11:40:52.2197235Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2197509Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 ______________ 2025-12-04T11:40:52.2197629Z Traceback (most recent call last): 2025-12-04T11:40:52.2198111Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2198343Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2212459Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2212484Z 2025-12-04T11:40:52.2212790Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2213358Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 2025-12-04T11:40:52.2213368Z 2025-12-04T11:40:52.2213634Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2213860Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2213977Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2214090Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2214431Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2214659Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2214753Z graph_break [] 2025-12-04T11:40:52.2214975Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2215391Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2215761Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2215965Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2216163Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2216251Z softmax. 2025-12-04T11:40:52.2216256Z 2025-12-04T11:40:52.2216365Z warnings.warn( 2025-12-04T11:40:52.2217195Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2217299Z warnings.warn( 2025-12-04T11:40:52.2217516Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2217620Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2217737Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2217957Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2218293Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2218458Z graph_break [] 2025-12-04T11:40:52.2218668Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2219082Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2219350Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2219544Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2219749Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2219838Z softmax. 2025-12-04T11:40:52.2219844Z 2025-12-04T11:40:52.2219936Z warnings.warn( 2025-12-04T11:40:52.2220714Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2220810Z warnings.warn( 2025-12-04T11:40:52.2221031Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2221140Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2221246Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2221475Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2221808Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2221902Z graph_break [] 2025-12-04T11:40:52.2222124Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2222528Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2222752Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2222947Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2223146Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2223241Z softmax. 2025-12-04T11:40:52.2223248Z 2025-12-04T11:40:52.2223344Z warnings.warn( 2025-12-04T11:40:52.2224067Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2224170Z warnings.warn( 2025-12-04T11:40:52.2224936Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3d838b8cb24fd196.xml - 2025-12-04T11:40:52.2225110Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2225805Z FAILED [0.0946s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_0 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2225810Z 2025-12-04T11:40:52.2226018Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2226573Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 2025-12-04T11:40:52.2226686Z 2025-12-04T11:40:52.2226948Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2227129Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2227327Z ================== 1 failed, 30 deselected, 2 rerun in 4.45s =================== 2025-12-04T11:40:52.2227417Z Got exit code 1 2025-12-04T11:40:52.2227525Z Retrying single test... 2025-12-04T11:40:52.2228109Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-6a6a2484271011a1.xml 2025-12-04T11:40:52.2228273Z ============================= test session starts ============================== 2025-12-04T11:40:52.2228615Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2228720Z cachedir: .pytest_cache 2025-12-04T11:40:52.2229248Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2229399Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2229502Z configfile: pytest.ini 2025-12-04T11:40:52.2230126Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2230335Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.2230973Z stepcurrent: skipping 17 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_0 2025-12-04T11:40:52.2231082Z Running 1 items in this shard 2025-12-04T11:40:52.2231087Z 2025-12-04T11:40:52.2231611Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_0 ('RERUN', {'yellow': True}) [4.2481s] [100%] 2025-12-04T11:40:52.2232168Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_0 ('RERUN', {'yellow': True}) [0.0963s] [100%] 2025-12-04T11:40:52.2232610Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_0 FAILED [0.0959s] [100%] 2025-12-04T11:40:52.2232616Z 2025-12-04T11:40:52.2232755Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2233025Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 ______________ 2025-12-04T11:40:52.2233143Z Traceback (most recent call last): 2025-12-04T11:40:52.2233614Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2233841Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2234048Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2234059Z 2025-12-04T11:40:52.2234269Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2234822Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 2025-12-04T11:40:52.2234832Z 2025-12-04T11:40:52.2235097Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2235308Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2235416Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2235531Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2235863Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2236082Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2236176Z graph_break [] 2025-12-04T11:40:52.2236384Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2236799Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2237051Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2237243Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2237447Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2237536Z softmax. 2025-12-04T11:40:52.2237541Z 2025-12-04T11:40:52.2237642Z warnings.warn( 2025-12-04T11:40:52.2238374Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2238468Z warnings.warn( 2025-12-04T11:40:52.2238744Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 ______________ 2025-12-04T11:40:52.2238861Z Traceback (most recent call last): 2025-12-04T11:40:52.2239323Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2239555Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2239758Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2239798Z 2025-12-04T11:40:52.2240011Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2240560Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 2025-12-04T11:40:52.2240565Z 2025-12-04T11:40:52.2240851Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2241070Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2241178Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2241294Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2241630Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2241870Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2241976Z graph_break [] 2025-12-04T11:40:52.2242186Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2242596Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2242821Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2243016Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2243215Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2243303Z softmax. 2025-12-04T11:40:52.2243308Z 2025-12-04T11:40:52.2243399Z warnings.warn( 2025-12-04T11:40:52.2244136Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2244231Z warnings.warn( 2025-12-04T11:40:52.2244440Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2244555Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2244665Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2244885Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2245218Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2245309Z graph_break [] 2025-12-04T11:40:52.2245528Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2245939Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2246157Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2246357Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2246553Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2246648Z softmax. 2025-12-04T11:40:52.2246653Z 2025-12-04T11:40:52.2246748Z warnings.warn( 2025-12-04T11:40:52.2247469Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2247610Z warnings.warn( 2025-12-04T11:40:52.2247751Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2248019Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 ______________ 2025-12-04T11:40:52.2248149Z Traceback (most recent call last): 2025-12-04T11:40:52.2248614Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2248851Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2249056Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2249061Z 2025-12-04T11:40:52.2249268Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2249827Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 2025-12-04T11:40:52.2249835Z 2025-12-04T11:40:52.2250121Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2250340Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2250445Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2250584Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2250929Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2251144Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2251232Z graph_break [] 2025-12-04T11:40:52.2251450Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2251854Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2252123Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2252320Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2252513Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2252606Z softmax. 2025-12-04T11:40:52.2252611Z 2025-12-04T11:40:52.2252705Z warnings.warn( 2025-12-04T11:40:52.2253429Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2253532Z warnings.warn( 2025-12-04T11:40:52.2253744Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2253857Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2253963Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2254174Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2254517Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2254607Z graph_break [] 2025-12-04T11:40:52.2254817Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2255227Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2255444Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2255649Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2255840Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2255924Z softmax. 2025-12-04T11:40:52.2255929Z 2025-12-04T11:40:52.2256025Z warnings.warn( 2025-12-04T11:40:52.2256748Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2256838Z warnings.warn( 2025-12-04T11:40:52.2257161Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2257273Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2257425Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2257660Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2257998Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2258106Z graph_break [] 2025-12-04T11:40:52.2258323Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2258732Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2258966Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2259166Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2259366Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2259473Z softmax. 2025-12-04T11:40:52.2259478Z 2025-12-04T11:40:52.2259580Z warnings.warn( 2025-12-04T11:40:52.2260321Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2260454Z warnings.warn( 2025-12-04T11:40:52.2261220Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-6a6a2484271011a1.xml - 2025-12-04T11:40:52.2261436Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2262131Z FAILED [0.0959s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_0 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2262136Z 2025-12-04T11:40:52.2262364Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2262944Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_0 2025-12-04T11:40:52.2262950Z 2025-12-04T11:40:52.2263217Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2263410Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2263609Z ================== 1 failed, 30 deselected, 2 rerun in 4.47s =================== 2025-12-04T11:40:52.2263721Z Got exit code 1 2025-12-04T11:40:52.2264196Z FAILED CONSISTENTLY: test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_0 2025-12-04T11:40:52.2264605Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:40:52.2265209Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-9356c0308b0c66f7.xml 2025-12-04T11:40:52.2265372Z ============================= test session starts ============================== 2025-12-04T11:40:52.2265737Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2265848Z cachedir: .pytest_cache 2025-12-04T11:40:52.2266369Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2266510Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2266619Z configfile: pytest.ini 2025-12-04T11:40:52.2267218Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2267453Z collecting ... collected 31 items / 18 deselected / 13 selected 2025-12-04T11:40:52.2267597Z stepcurrent: skipping 18 already run items. 2025-12-04T11:40:52.2267730Z Running 13 items in this shard 2025-12-04T11:40:52.2267735Z 2025-12-04T11:40:52.2268271Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_1 ('RERUN', {'yellow': True}) [4.2653s] [ 7%] 2025-12-04T11:40:52.2268799Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_1 ('RERUN', {'yellow': True}) [0.0947s] [ 7%] 2025-12-04T11:40:52.2269288Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_1 FAILED [0.0940s] [ 7%] 2025-12-04T11:40:52.2269296Z 2025-12-04T11:40:52.2269437Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2269727Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 ______________ 2025-12-04T11:40:52.2269849Z Traceback (most recent call last): 2025-12-04T11:40:52.2270320Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2270562Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2270770Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2270775Z 2025-12-04T11:40:52.2270992Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2271563Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 2025-12-04T11:40:52.2271601Z 2025-12-04T11:40:52.2271869Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2272097Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2272238Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2272354Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2272707Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2272924Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2273035Z graph_break [] 2025-12-04T11:40:52.2273249Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2274012Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2274125Z warnings.warn( 2025-12-04T11:40:52.2274395Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 ______________ 2025-12-04T11:40:52.2274520Z Traceback (most recent call last): 2025-12-04T11:40:52.2274999Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2275227Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2275447Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2275452Z 2025-12-04T11:40:52.2275663Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2276215Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 2025-12-04T11:40:52.2276221Z 2025-12-04T11:40:52.2276496Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2276714Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2276840Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2276954Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2277294Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2277528Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2277621Z graph_break [] 2025-12-04T11:40:52.2277835Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2278582Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2278680Z warnings.warn( 2025-12-04T11:40:52.2278902Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2279013Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2279155Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2279386Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2279725Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2279821Z graph_break [] 2025-12-04T11:40:52.2280052Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2280783Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2280896Z warnings.warn( 2025-12-04T11:40:52.2281034Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2281302Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 ______________ 2025-12-04T11:40:52.2281436Z Traceback (most recent call last): 2025-12-04T11:40:52.2281909Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2282139Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2282393Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2282398Z 2025-12-04T11:40:52.2282610Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2283200Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 2025-12-04T11:40:52.2283206Z 2025-12-04T11:40:52.2283470Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2283684Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2283804Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2283917Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2284282Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2284516Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2284618Z graph_break [] 2025-12-04T11:40:52.2284842Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2285575Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2285674Z warnings.warn( 2025-12-04T11:40:52.2285895Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2286004Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2286119Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2286348Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2286689Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2286797Z graph_break [] 2025-12-04T11:40:52.2287012Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2287738Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2287848Z warnings.warn( 2025-12-04T11:40:52.2288063Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2288173Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2288298Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2288517Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2288864Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2288961Z graph_break [] 2025-12-04T11:40:52.2289173Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2289911Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2290041Z warnings.warn( 2025-12-04T11:40:52.2290819Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-9356c0308b0c66f7.xml - 2025-12-04T11:40:52.2290990Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2291681Z FAILED [0.0940s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_1 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2291686Z 2025-12-04T11:40:52.2291912Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2292466Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 2025-12-04T11:40:52.2292474Z 2025-12-04T11:40:52.2292751Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2292928Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2293152Z ================== 1 failed, 18 deselected, 2 rerun in 4.49s =================== 2025-12-04T11:40:52.2293259Z Got exit code 1 2025-12-04T11:40:52.2293367Z Retrying single test... 2025-12-04T11:40:52.2294004Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-51758daa3a3dc599.xml 2025-12-04T11:40:52.2294178Z ============================= test session starts ============================== 2025-12-04T11:40:52.2294529Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2294647Z cachedir: .pytest_cache 2025-12-04T11:40:52.2295196Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2295321Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2295442Z configfile: pytest.ini 2025-12-04T11:40:52.2296229Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2296444Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.2297154Z stepcurrent: skipping 18 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_1 2025-12-04T11:40:52.2297271Z Running 1 items in this shard 2025-12-04T11:40:52.2297276Z 2025-12-04T11:40:52.2297820Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_1 ('RERUN', {'yellow': True}) [4.2623s] [100%] 2025-12-04T11:40:52.2298346Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_1 ('RERUN', {'yellow': True}) [0.0959s] [100%] 2025-12-04T11:40:52.2298785Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_1 FAILED [0.0950s] [100%] 2025-12-04T11:40:52.2298808Z 2025-12-04T11:40:52.2298949Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2299222Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 ______________ 2025-12-04T11:40:52.2299356Z Traceback (most recent call last): 2025-12-04T11:40:52.2299828Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2300055Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2300275Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2300281Z 2025-12-04T11:40:52.2300490Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2301050Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 2025-12-04T11:40:52.2301055Z 2025-12-04T11:40:52.2301395Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2301614Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2301742Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2301858Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2302211Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2302429Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2302527Z graph_break [] 2025-12-04T11:40:52.2302753Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2303486Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2303589Z warnings.warn( 2025-12-04T11:40:52.2303874Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 ______________ 2025-12-04T11:40:52.2303997Z Traceback (most recent call last): 2025-12-04T11:40:52.2304520Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2304748Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2304994Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2304999Z 2025-12-04T11:40:52.2305228Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2305775Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 2025-12-04T11:40:52.2305780Z 2025-12-04T11:40:52.2306057Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2306314Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2306428Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2306561Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2306903Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2307121Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2307231Z graph_break [] 2025-12-04T11:40:52.2307447Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2308190Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2308289Z warnings.warn( 2025-12-04T11:40:52.2308502Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2308625Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2308739Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2308960Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2309313Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2309412Z graph_break [] 2025-12-04T11:40:52.2309624Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2310366Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2310466Z warnings.warn( 2025-12-04T11:40:52.2310621Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2310896Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 ______________ 2025-12-04T11:40:52.2311019Z Traceback (most recent call last): 2025-12-04T11:40:52.2311500Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2311730Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2311982Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2311990Z 2025-12-04T11:40:52.2312201Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2312746Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 2025-12-04T11:40:52.2312753Z 2025-12-04T11:40:52.2313031Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2313240Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2313354Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2313462Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2313796Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2314025Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2314117Z graph_break [] 2025-12-04T11:40:52.2314331Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2315099Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2315195Z warnings.warn( 2025-12-04T11:40:52.2315445Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2315551Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2315660Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2315888Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2316221Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2316316Z graph_break [] 2025-12-04T11:40:52.2316565Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2317288Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2317400Z warnings.warn( 2025-12-04T11:40:52.2317607Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2317707Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2317829Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2318043Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2318378Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2318478Z graph_break [] 2025-12-04T11:40:52.2318688Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2319416Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2319522Z warnings.warn( 2025-12-04T11:40:52.2320290Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-51758daa3a3dc599.xml - 2025-12-04T11:40:52.2320467Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2321156Z FAILED [0.0950s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_1 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2321161Z 2025-12-04T11:40:52.2321371Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2321928Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 2025-12-04T11:40:52.2321933Z 2025-12-04T11:40:52.2322193Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2322376Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2322597Z ================== 1 failed, 30 deselected, 2 rerun in 4.48s =================== 2025-12-04T11:40:52.2322694Z Got exit code 1 2025-12-04T11:40:52.2322806Z Retrying single test... 2025-12-04T11:40:52.2323394Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a0d42e3d123a44e8.xml 2025-12-04T11:40:52.2323563Z ============================= test session starts ============================== 2025-12-04T11:40:52.2323907Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2324012Z cachedir: .pytest_cache 2025-12-04T11:40:52.2324537Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2324656Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2324763Z configfile: pytest.ini 2025-12-04T11:40:52.2325354Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2325596Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.2326239Z stepcurrent: skipping 18 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_1 2025-12-04T11:40:52.2326378Z Running 1 items in this shard 2025-12-04T11:40:52.2326384Z 2025-12-04T11:40:52.2326910Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_1 ('RERUN', {'yellow': True}) [4.2236s] [100%] 2025-12-04T11:40:52.2327439Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_1 ('RERUN', {'yellow': True}) [0.0939s] [100%] 2025-12-04T11:40:52.2327874Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_1 FAILED [0.0920s] [100%] 2025-12-04T11:40:52.2327908Z 2025-12-04T11:40:52.2328054Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2328324Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 ______________ 2025-12-04T11:40:52.2328444Z Traceback (most recent call last): 2025-12-04T11:40:52.2328912Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2329137Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2329355Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2329360Z 2025-12-04T11:40:52.2329566Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2330110Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 2025-12-04T11:40:52.2330115Z 2025-12-04T11:40:52.2330381Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2330591Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2330710Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2330820Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2331153Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2331382Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2331474Z graph_break [] 2025-12-04T11:40:52.2331682Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2332423Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2332519Z warnings.warn( 2025-12-04T11:40:52.2332800Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 ______________ 2025-12-04T11:40:52.2332917Z Traceback (most recent call last): 2025-12-04T11:40:52.2333380Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2333646Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2333848Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2333853Z 2025-12-04T11:40:52.2334064Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2334618Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 2025-12-04T11:40:52.2334623Z 2025-12-04T11:40:52.2334879Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2335101Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2335205Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2335315Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2335664Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2335880Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2336029Z graph_break [] 2025-12-04T11:40:52.2336236Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2337070Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2337184Z warnings.warn( 2025-12-04T11:40:52.2337392Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2337496Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2337620Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2337834Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2338198Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2338306Z graph_break [] 2025-12-04T11:40:52.2338514Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2339252Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2339348Z warnings.warn( 2025-12-04T11:40:52.2339491Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2339770Z _____________ TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 ______________ 2025-12-04T11:40:52.2339888Z Traceback (most recent call last): 2025-12-04T11:40:52.2340364Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 136, in test_prepare_softmax 2025-12-04T11:40:52.2340586Z act, (code,) = run_and_get_code(torch.compile(_prepare_softmax), x, dim) 2025-12-04T11:40:52.2340793Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2340798Z 2025-12-04T11:40:52.2341021Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2341571Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 2025-12-04T11:40:52.2341577Z 2025-12-04T11:40:52.2341846Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2342058Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2342165Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2342284Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2342618Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2342831Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2342935Z graph_break [] 2025-12-04T11:40:52.2343143Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2343903Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2343998Z warnings.warn( 2025-12-04T11:40:52.2344208Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2344324Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2344432Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2344649Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2344986Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2345081Z graph_break [] 2025-12-04T11:40:52.2345287Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2346018Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2346116Z warnings.warn( 2025-12-04T11:40:52.2346335Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2346473Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2346582Z stats [('calls_captured', 4)] 2025-12-04T11:40:52.2346804Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2347177Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2347272Z graph_break [] 2025-12-04T11:40:52.2347491Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2348209Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2348311Z warnings.warn( 2025-12-04T11:40:52.2349106Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a0d42e3d123a44e8.xml - 2025-12-04T11:40:52.2349277Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2349976Z FAILED [0.0920s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_1 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2349981Z 2025-12-04T11:40:52.2350194Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2350755Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_prepare_softmax_nrow_2_dim_1 2025-12-04T11:40:52.2350760Z 2025-12-04T11:40:52.2351021Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2351199Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2351402Z ================== 1 failed, 30 deselected, 2 rerun in 4.44s =================== 2025-12-04T11:40:52.2351501Z Got exit code 1 2025-12-04T11:40:52.2351977Z FAILED CONSISTENTLY: test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_1 2025-12-04T11:40:52.2352382Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:40:52.2352971Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-20d10a20187a2cb9.xml 2025-12-04T11:40:52.2353139Z ============================= test session starts ============================== 2025-12-04T11:40:52.2353480Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2353596Z cachedir: .pytest_cache 2025-12-04T11:40:52.2354112Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2354234Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2354345Z configfile: pytest.ini 2025-12-04T11:40:52.2354962Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2355177Z collecting ... collected 31 items / 19 deselected / 12 selected 2025-12-04T11:40:52.2355324Z stepcurrent: skipping 19 already run items. 2025-12-04T11:40:52.2355437Z Running 12 items in this shard 2025-12-04T11:40:52.2355442Z 2025-12-04T11:40:52.2355861Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_perf PASSED [4.4185s] [ 8%] 2025-12-04T11:40:52.2356636Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_sdpa W1204 11:37:00.246000 109120 site-packages/torch/_inductor/utils.py:1703] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:40:52.2356764Z ('RERUN', {'yellow': True}) [1.3725s] [ 16%] 2025-12-04T11:40:52.2357190Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_sdpa ('RERUN', {'yellow': True}) [0.2195s] [ 16%] 2025-12-04T11:40:52.2357514Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_sdpa FAILED [0.2191s] [ 16%] 2025-12-04T11:40:52.2357550Z 2025-12-04T11:40:52.2357697Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2357923Z _________________________ TestOnlineSoftmax.test_sdpa __________________________ 2025-12-04T11:40:52.2358067Z Traceback (most recent call last): 2025-12-04T11:40:52.2358477Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 128, in test_sdpa 2025-12-04T11:40:52.2358621Z act, (code,) = run_and_get_code(opt_f, q, k, v) 2025-12-04T11:40:52.2358825Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2358830Z 2025-12-04T11:40:52.2359045Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2359500Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_sdpa 2025-12-04T11:40:52.2359506Z 2025-12-04T11:40:52.2359782Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2359996Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2360100Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2360219Z stats [('calls_captured', 5)] 2025-12-04T11:40:52.2360434Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2360865Z inductor [('pattern_matcher_nodes', 26), ('pattern_matcher_count', 3), ('fuse_attention', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.2360958Z graph_break [] 2025-12-04T11:40:52.2361158Z aten_mm_info [('aten.bmm_8_16_16_32', 1), ('aten.bmm_8_16_32_16', 1)] 2025-12-04T11:40:52.2361378Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2362114Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2362212Z warnings.warn( 2025-12-04T11:40:52.2362445Z _________________________ TestOnlineSoftmax.test_sdpa __________________________ 2025-12-04T11:40:52.2362566Z Traceback (most recent call last): 2025-12-04T11:40:52.2362965Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 128, in test_sdpa 2025-12-04T11:40:52.2363110Z act, (code,) = run_and_get_code(opt_f, q, k, v) 2025-12-04T11:40:52.2363315Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2363320Z 2025-12-04T11:40:52.2363537Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2363965Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_sdpa 2025-12-04T11:40:52.2363970Z 2025-12-04T11:40:52.2364231Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2364448Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2364585Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2364704Z stats [('calls_captured', 5)] 2025-12-04T11:40:52.2364923Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2365349Z inductor [('pattern_matcher_nodes', 26), ('pattern_matcher_count', 3), ('fuse_attention', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.2365458Z graph_break [] 2025-12-04T11:40:52.2365660Z aten_mm_info [('aten.bmm_8_16_16_32', 1), ('aten.bmm_8_16_32_16', 1)] 2025-12-04T11:40:52.2365871Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2366608Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2366704Z warnings.warn( 2025-12-04T11:40:52.2366926Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2367031Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2367147Z stats [('calls_captured', 5)] 2025-12-04T11:40:52.2367373Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2367823Z inductor [('pattern_matcher_nodes', 26), ('pattern_matcher_count', 3), ('fuse_attention', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.2367914Z graph_break [] 2025-12-04T11:40:52.2368151Z aten_mm_info [('aten.bmm_8_16_16_32', 1), ('aten.bmm_8_16_32_16', 1)] 2025-12-04T11:40:52.2368361Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2369094Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2369189Z warnings.warn( 2025-12-04T11:40:52.2369328Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2369593Z _________________________ TestOnlineSoftmax.test_sdpa __________________________ 2025-12-04T11:40:52.2369714Z Traceback (most recent call last): 2025-12-04T11:40:52.2370119Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 128, in test_sdpa 2025-12-04T11:40:52.2370265Z act, (code,) = run_and_get_code(opt_f, q, k, v) 2025-12-04T11:40:52.2370472Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2370477Z 2025-12-04T11:40:52.2370696Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2371123Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_sdpa 2025-12-04T11:40:52.2371128Z 2025-12-04T11:40:52.2371389Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2371610Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2371713Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2371834Z stats [('calls_captured', 5)] 2025-12-04T11:40:52.2372053Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2372478Z inductor [('pattern_matcher_nodes', 26), ('pattern_matcher_count', 3), ('fuse_attention', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.2372586Z graph_break [] 2025-12-04T11:40:52.2372786Z aten_mm_info [('aten.bmm_8_16_16_32', 1), ('aten.bmm_8_16_32_16', 1)] 2025-12-04T11:40:52.2373000Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2373736Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2373832Z warnings.warn( 2025-12-04T11:40:52.2374052Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2374156Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2374267Z stats [('calls_captured', 5)] 2025-12-04T11:40:52.2374494Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2374962Z inductor [('pattern_matcher_nodes', 26), ('pattern_matcher_count', 3), ('fuse_attention', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.2375060Z graph_break [] 2025-12-04T11:40:52.2375264Z aten_mm_info [('aten.bmm_8_16_16_32', 1), ('aten.bmm_8_16_32_16', 1)] 2025-12-04T11:40:52.2375478Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2376218Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2376317Z warnings.warn( 2025-12-04T11:40:52.2376523Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2376636Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2376743Z stats [('calls_captured', 5)] 2025-12-04T11:40:52.2377032Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2377464Z inductor [('pattern_matcher_nodes', 26), ('pattern_matcher_count', 3), ('fuse_attention', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.2377593Z graph_break [] 2025-12-04T11:40:52.2377798Z aten_mm_info [('aten.bmm_8_16_16_32', 1), ('aten.bmm_8_16_32_16', 1)] 2025-12-04T11:40:52.2378007Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2378768Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2378870Z warnings.warn( 2025-12-04T11:40:52.2379634Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-20d10a20187a2cb9.xml - 2025-12-04T11:40:52.2379805Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2380414Z FAILED [0.2191s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_sdpa - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2380422Z 2025-12-04T11:40:52.2380639Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2381074Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_sdpa 2025-12-04T11:40:52.2381079Z 2025-12-04T11:40:52.2381343Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2381529Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2381736Z ============= 1 failed, 1 passed, 19 deselected, 2 rerun in 6.26s ============== 2025-12-04T11:40:52.2381832Z Got exit code 1 2025-12-04T11:40:52.2381945Z Retrying single test... 2025-12-04T11:40:52.2382535Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-5f3be1bf80a1ce40.xml 2025-12-04T11:40:52.2382692Z ============================= test session starts ============================== 2025-12-04T11:40:52.2383043Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2383149Z cachedir: .pytest_cache 2025-12-04T11:40:52.2383671Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2383793Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2383898Z configfile: pytest.ini 2025-12-04T11:40:52.2384492Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2384703Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.2385220Z stepcurrent: skipping 20 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_sdpa 2025-12-04T11:40:52.2385339Z Running 1 items in this shard 2025-12-04T11:40:52.2385344Z 2025-12-04T11:40:52.2386165Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_sdpa W1204 11:37:18.903000 109289 site-packages/torch/_inductor/utils.py:1703] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:40:52.2386300Z ('RERUN', {'yellow': True}) [5.5624s] [100%] 2025-12-04T11:40:52.2386720Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_sdpa ('RERUN', {'yellow': True}) [0.2206s] [100%] 2025-12-04T11:40:52.2387040Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_sdpa FAILED [0.2209s] [100%] 2025-12-04T11:40:52.2387056Z 2025-12-04T11:40:52.2387186Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2387414Z _________________________ TestOnlineSoftmax.test_sdpa __________________________ 2025-12-04T11:40:52.2387542Z Traceback (most recent call last): 2025-12-04T11:40:52.2387938Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 128, in test_sdpa 2025-12-04T11:40:52.2388083Z act, (code,) = run_and_get_code(opt_f, q, k, v) 2025-12-04T11:40:52.2388300Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2388334Z 2025-12-04T11:40:52.2388545Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2388987Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_sdpa 2025-12-04T11:40:52.2389020Z 2025-12-04T11:40:52.2389285Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2389498Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2389618Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2389728Z stats [('calls_captured', 5)] 2025-12-04T11:40:52.2390161Z inductor [('pattern_matcher_nodes', 26), ('pattern_matcher_count', 3), ('fuse_attention', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.2390488Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2390587Z graph_break [] 2025-12-04T11:40:52.2390798Z aten_mm_info [('aten.bmm_8_16_16_32', 1), ('aten.bmm_8_16_32_16', 1)] 2025-12-04T11:40:52.2391013Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2391752Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2391863Z warnings.warn( 2025-12-04T11:40:52.2392092Z _________________________ TestOnlineSoftmax.test_sdpa __________________________ 2025-12-04T11:40:52.2392211Z Traceback (most recent call last): 2025-12-04T11:40:52.2392618Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 128, in test_sdpa 2025-12-04T11:40:52.2392766Z act, (code,) = run_and_get_code(opt_f, q, k, v) 2025-12-04T11:40:52.2392986Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2392993Z 2025-12-04T11:40:52.2393200Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2393632Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_sdpa 2025-12-04T11:40:52.2393640Z 2025-12-04T11:40:52.2393911Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2394123Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2394241Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2394349Z stats [('calls_captured', 5)] 2025-12-04T11:40:52.2394780Z inductor [('pattern_matcher_nodes', 26), ('pattern_matcher_count', 3), ('fuse_attention', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.2395009Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2395101Z graph_break [] 2025-12-04T11:40:52.2395306Z aten_mm_info [('aten.bmm_8_16_16_32', 1), ('aten.bmm_8_16_32_16', 1)] 2025-12-04T11:40:52.2395525Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2396486Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2396604Z warnings.warn( 2025-12-04T11:40:52.2396814Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2396920Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2397043Z stats [('calls_captured', 5)] 2025-12-04T11:40:52.2397253Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2397681Z inductor [('pattern_matcher_nodes', 26), ('pattern_matcher_count', 3), ('fuse_attention', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.2397781Z graph_break [] 2025-12-04T11:40:52.2397983Z aten_mm_info [('aten.bmm_8_16_16_32', 1), ('aten.bmm_8_16_32_16', 1)] 2025-12-04T11:40:52.2398205Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2398931Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2399083Z warnings.warn( 2025-12-04T11:40:52.2399227Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2399493Z _________________________ TestOnlineSoftmax.test_sdpa __________________________ 2025-12-04T11:40:52.2399614Z Traceback (most recent call last): 2025-12-04T11:40:52.2400014Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 128, in test_sdpa 2025-12-04T11:40:52.2400161Z act, (code,) = run_and_get_code(opt_f, q, k, v) 2025-12-04T11:40:52.2400374Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2400380Z 2025-12-04T11:40:52.2400587Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2401052Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_sdpa 2025-12-04T11:40:52.2401060Z 2025-12-04T11:40:52.2401329Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2401540Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2401655Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2401760Z stats [('calls_captured', 5)] 2025-12-04T11:40:52.2402192Z inductor [('pattern_matcher_nodes', 26), ('pattern_matcher_count', 3), ('fuse_attention', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.2402412Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2402502Z graph_break [] 2025-12-04T11:40:52.2402705Z aten_mm_info [('aten.bmm_8_16_16_32', 1), ('aten.bmm_8_16_32_16', 1)] 2025-12-04T11:40:52.2402920Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2403654Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2403767Z warnings.warn( 2025-12-04T11:40:52.2403973Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2404079Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2404195Z stats [('calls_captured', 5)] 2025-12-04T11:40:52.2404412Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2404837Z inductor [('pattern_matcher_nodes', 26), ('pattern_matcher_count', 3), ('fuse_attention', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.2404936Z graph_break [] 2025-12-04T11:40:52.2405138Z aten_mm_info [('aten.bmm_8_16_16_32', 1), ('aten.bmm_8_16_32_16', 1)] 2025-12-04T11:40:52.2405356Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2406079Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2406213Z warnings.warn( 2025-12-04T11:40:52.2406427Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2406535Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2406645Z stats [('calls_captured', 5)] 2025-12-04T11:40:52.2406866Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2407287Z inductor [('pattern_matcher_nodes', 26), ('pattern_matcher_count', 3), ('fuse_attention', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.2407391Z graph_break [] 2025-12-04T11:40:52.2407591Z aten_mm_info [('aten.bmm_8_16_16_32', 1), ('aten.bmm_8_16_32_16', 1)] 2025-12-04T11:40:52.2407801Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2408537Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2408635Z warnings.warn( 2025-12-04T11:40:52.2409404Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-5f3be1bf80a1ce40.xml - 2025-12-04T11:40:52.2409615Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2410223Z FAILED [0.2209s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_sdpa - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2410229Z 2025-12-04T11:40:52.2410450Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2410879Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_sdpa 2025-12-04T11:40:52.2410884Z 2025-12-04T11:40:52.2411145Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2411366Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2411562Z ================== 1 failed, 30 deselected, 2 rerun in 6.03s =================== 2025-12-04T11:40:52.2411667Z Got exit code 1 2025-12-04T11:40:52.2411766Z Retrying single test... 2025-12-04T11:40:52.2412355Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-baeb09a4e62b9d77.xml 2025-12-04T11:40:52.2412524Z ============================= test session starts ============================== 2025-12-04T11:40:52.2412868Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2412988Z cachedir: .pytest_cache 2025-12-04T11:40:52.2413507Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2413628Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2413746Z configfile: pytest.ini 2025-12-04T11:40:52.2414339Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2414559Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.2415093Z stepcurrent: skipping 20 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_sdpa 2025-12-04T11:40:52.2415207Z Running 1 items in this shard 2025-12-04T11:40:52.2415214Z 2025-12-04T11:40:52.2416007Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_sdpa W1204 11:37:38.693000 109458 site-packages/torch/_inductor/utils.py:1703] [0/0] Not enough SMs to use max_autotune_gemm mode 2025-12-04T11:40:52.2416137Z ('RERUN', {'yellow': True}) [5.6131s] [100%] 2025-12-04T11:40:52.2416560Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_sdpa ('RERUN', {'yellow': True}) [0.2218s] [100%] 2025-12-04T11:40:52.2416982Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_sdpa FAILED [0.2180s] [100%] 2025-12-04T11:40:52.2416989Z 2025-12-04T11:40:52.2417130Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2417409Z _________________________ TestOnlineSoftmax.test_sdpa __________________________ 2025-12-04T11:40:52.2417531Z Traceback (most recent call last): 2025-12-04T11:40:52.2417932Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 128, in test_sdpa 2025-12-04T11:40:52.2418091Z act, (code,) = run_and_get_code(opt_f, q, k, v) 2025-12-04T11:40:52.2418298Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2418304Z 2025-12-04T11:40:52.2418515Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2418957Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_sdpa 2025-12-04T11:40:52.2418963Z 2025-12-04T11:40:52.2419227Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2419460Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2419572Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2419717Z stats [('calls_captured', 5)] 2025-12-04T11:40:52.2420158Z inductor [('pattern_matcher_nodes', 26), ('pattern_matcher_count', 3), ('fuse_attention', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.2420373Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2420511Z graph_break [] 2025-12-04T11:40:52.2420714Z aten_mm_info [('aten.bmm_8_16_16_32', 1), ('aten.bmm_8_16_32_16', 1)] 2025-12-04T11:40:52.2420924Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2421666Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2421765Z warnings.warn( 2025-12-04T11:40:52.2422023Z _________________________ TestOnlineSoftmax.test_sdpa __________________________ 2025-12-04T11:40:52.2422158Z Traceback (most recent call last): 2025-12-04T11:40:52.2422553Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 128, in test_sdpa 2025-12-04T11:40:52.2422715Z act, (code,) = run_and_get_code(opt_f, q, k, v) 2025-12-04T11:40:52.2422921Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2422926Z 2025-12-04T11:40:52.2423139Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2423579Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_sdpa 2025-12-04T11:40:52.2423584Z 2025-12-04T11:40:52.2423846Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2424068Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2424175Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2424288Z stats [('calls_captured', 5)] 2025-12-04T11:40:52.2424731Z inductor [('pattern_matcher_nodes', 26), ('pattern_matcher_count', 3), ('fuse_attention', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.2424952Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2425047Z graph_break [] 2025-12-04T11:40:52.2425263Z aten_mm_info [('aten.bmm_8_16_16_32', 1), ('aten.bmm_8_16_32_16', 1)] 2025-12-04T11:40:52.2425473Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2426217Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2426317Z warnings.warn( 2025-12-04T11:40:52.2426526Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2426638Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2426745Z stats [('calls_captured', 5)] 2025-12-04T11:40:52.2426962Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2427431Z inductor [('pattern_matcher_nodes', 26), ('pattern_matcher_count', 3), ('fuse_attention', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.2427526Z graph_break [] 2025-12-04T11:40:52.2427726Z aten_mm_info [('aten.bmm_8_16_16_32', 1), ('aten.bmm_8_16_32_16', 1)] 2025-12-04T11:40:52.2427951Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2428676Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2428784Z warnings.warn( 2025-12-04T11:40:52.2428924Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2429151Z _________________________ TestOnlineSoftmax.test_sdpa __________________________ 2025-12-04T11:40:52.2429279Z Traceback (most recent call last): 2025-12-04T11:40:52.2429671Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 128, in test_sdpa 2025-12-04T11:40:52.2429826Z act, (code,) = run_and_get_code(opt_f, q, k, v) 2025-12-04T11:40:52.2430060Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2430065Z 2025-12-04T11:40:52.2430274Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2430738Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_sdpa 2025-12-04T11:40:52.2430743Z 2025-12-04T11:40:52.2431004Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2431216Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2431329Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2431438Z stats [('calls_captured', 5)] 2025-12-04T11:40:52.2431904Z inductor [('pattern_matcher_nodes', 26), ('pattern_matcher_count', 3), ('fuse_attention', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.2432118Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2432215Z graph_break [] 2025-12-04T11:40:52.2432424Z aten_mm_info [('aten.bmm_8_16_16_32', 1), ('aten.bmm_8_16_32_16', 1)] 2025-12-04T11:40:52.2432632Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2433362Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2433465Z warnings.warn( 2025-12-04T11:40:52.2433677Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2433790Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2433900Z stats [('calls_captured', 5)] 2025-12-04T11:40:52.2434115Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2434553Z inductor [('pattern_matcher_nodes', 26), ('pattern_matcher_count', 3), ('fuse_attention', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.2434650Z graph_break [] 2025-12-04T11:40:52.2434854Z aten_mm_info [('aten.bmm_8_16_16_32', 1), ('aten.bmm_8_16_32_16', 1)] 2025-12-04T11:40:52.2435071Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2435801Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2435908Z warnings.warn( 2025-12-04T11:40:52.2436118Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2436226Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2436344Z stats [('calls_captured', 5)] 2025-12-04T11:40:52.2436561Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2436988Z inductor [('pattern_matcher_nodes', 26), ('pattern_matcher_count', 3), ('fuse_attention', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T11:40:52.2437092Z graph_break [] 2025-12-04T11:40:52.2437327Z aten_mm_info [('aten.bmm_8_16_16_32', 1), ('aten.bmm_8_16_32_16', 1)] 2025-12-04T11:40:52.2437551Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2438277Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2438374Z warnings.warn( 2025-12-04T11:40:52.2439152Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-baeb09a4e62b9d77.xml - 2025-12-04T11:40:52.2439322Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2439915Z FAILED [0.2180s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_sdpa - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2439923Z 2025-12-04T11:40:52.2440137Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2440568Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_sdpa 2025-12-04T11:40:52.2440602Z 2025-12-04T11:40:52.2440872Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2441078Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2441287Z ================== 1 failed, 30 deselected, 2 rerun in 6.08s =================== 2025-12-04T11:40:52.2441383Z Got exit code 1 2025-12-04T11:40:52.2441736Z FAILED CONSISTENTLY: test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_sdpa 2025-12-04T11:40:52.2442149Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:40:52.2442768Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2cdd67c45b570140.xml 2025-12-04T11:40:52.2442929Z ============================= test session starts ============================== 2025-12-04T11:40:52.2443283Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2443393Z cachedir: .pytest_cache 2025-12-04T11:40:52.2443923Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2444043Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2444146Z configfile: pytest.ini 2025-12-04T11:40:52.2444740Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2444955Z collecting ... collected 31 items / 21 deselected / 10 selected 2025-12-04T11:40:52.2445106Z stepcurrent: skipping 21 already run items. 2025-12-04T11:40:52.2445213Z Running 10 items in this shard 2025-12-04T11:40:52.2445220Z 2025-12-04T11:40:52.2445561Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax PASSED [4.3549s] [ 10%] 2025-12-04T11:40:52.2446124Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_bfloat16 ('RERUN', {'yellow': True}) [0.1330s] [ 20%] 2025-12-04T11:40:52.2446680Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_bfloat16 ('RERUN', {'yellow': True}) [0.1023s] [ 20%] 2025-12-04T11:40:52.2447142Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_bfloat16 FAILED [0.1013s] [ 20%] 2025-12-04T11:40:52.2447158Z 2025-12-04T11:40:52.2447292Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2447571Z __________ TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 ___________ 2025-12-04T11:40:52.2447699Z Traceback (most recent call last): 2025-12-04T11:40:52.2448192Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 216, in test_softmax_acc_with_fp64 2025-12-04T11:40:52.2448385Z res, (code,) = run_and_get_code(torch.compile(fn), x, dim=-1) 2025-12-04T11:40:52.2448633Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2448641Z 2025-12-04T11:40:52.2448849Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2449445Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 2025-12-04T11:40:52.2449451Z 2025-12-04T11:40:52.2449717Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2449931Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2450053Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2450163Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2450395Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2450731Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2450830Z graph_break [] 2025-12-04T11:40:52.2451051Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2451810Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2451953Z warnings.warn( 2025-12-04T11:40:52.2452241Z __________ TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 ___________ 2025-12-04T11:40:52.2452359Z Traceback (most recent call last): 2025-12-04T11:40:52.2452863Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 216, in test_softmax_acc_with_fp64 2025-12-04T11:40:52.2453051Z res, (code,) = run_and_get_code(torch.compile(fn), x, dim=-1) 2025-12-04T11:40:52.2453261Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2453266Z 2025-12-04T11:40:52.2453513Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2454093Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 2025-12-04T11:40:52.2454100Z 2025-12-04T11:40:52.2454378Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2454591Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2454699Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2454826Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2455040Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2455377Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2455483Z graph_break [] 2025-12-04T11:40:52.2455692Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2456438Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2456536Z warnings.warn( 2025-12-04T11:40:52.2456743Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2456859Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2457039Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2457262Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2457607Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2457700Z graph_break [] 2025-12-04T11:40:52.2457911Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2458650Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2458750Z warnings.warn( 2025-12-04T11:40:52.2458898Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2459217Z __________ TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 ___________ 2025-12-04T11:40:52.2459335Z Traceback (most recent call last): 2025-12-04T11:40:52.2459841Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 216, in test_softmax_acc_with_fp64 2025-12-04T11:40:52.2460034Z res, (code,) = run_and_get_code(torch.compile(fn), x, dim=-1) 2025-12-04T11:40:52.2460252Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2460257Z 2025-12-04T11:40:52.2460466Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2461050Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 2025-12-04T11:40:52.2461055Z 2025-12-04T11:40:52.2461328Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2461542Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2461694Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2461802Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2462017Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2462389Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2462485Z graph_break [] 2025-12-04T11:40:52.2462697Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2463444Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2463542Z warnings.warn( 2025-12-04T11:40:52.2463762Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2463898Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2464018Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2464244Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2464581Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2464678Z graph_break [] 2025-12-04T11:40:52.2464898Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2465623Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2465731Z warnings.warn( 2025-12-04T11:40:52.2465941Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2466048Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2466170Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2466392Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2466729Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2466837Z graph_break [] 2025-12-04T11:40:52.2467044Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2467772Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2467877Z warnings.warn( 2025-12-04T11:40:52.2468645Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2cdd67c45b570140.xml - 2025-12-04T11:40:52.2468818Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2469539Z FAILED [0.1013s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_bfloat16 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2469575Z 2025-12-04T11:40:52.2469788Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2470378Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 2025-12-04T11:40:52.2470383Z 2025-12-04T11:40:52.2470644Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2470832Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2471041Z ============= 1 failed, 1 passed, 21 deselected, 2 rerun in 4.72s ============== 2025-12-04T11:40:52.2471140Z Got exit code 1 2025-12-04T11:40:52.2471256Z Retrying single test... 2025-12-04T11:40:52.2471842Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2423c3b275573349.xml 2025-12-04T11:40:52.2472016Z ============================= test session starts ============================== 2025-12-04T11:40:52.2472363Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2472502Z cachedir: .pytest_cache 2025-12-04T11:40:52.2473029Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2473148Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2473285Z configfile: pytest.ini 2025-12-04T11:40:52.2473879Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2474091Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.2474764Z stepcurrent: skipping 22 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_bfloat16 2025-12-04T11:40:52.2474875Z Running 1 items in this shard 2025-12-04T11:40:52.2474909Z 2025-12-04T11:40:52.2475468Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_bfloat16 ('RERUN', {'yellow': True}) [4.3917s] [100%] 2025-12-04T11:40:52.2476029Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_bfloat16 ('RERUN', {'yellow': True}) [0.1025s] [100%] 2025-12-04T11:40:52.2476492Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_bfloat16 FAILED [0.1012s] [100%] 2025-12-04T11:40:52.2476497Z 2025-12-04T11:40:52.2476645Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2476921Z __________ TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 ___________ 2025-12-04T11:40:52.2477042Z Traceback (most recent call last): 2025-12-04T11:40:52.2477544Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 216, in test_softmax_acc_with_fp64 2025-12-04T11:40:52.2477735Z res, (code,) = run_and_get_code(torch.compile(fn), x, dim=-1) 2025-12-04T11:40:52.2477951Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2477958Z 2025-12-04T11:40:52.2478166Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2478745Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 2025-12-04T11:40:52.2478750Z 2025-12-04T11:40:52.2479026Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2479240Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2479360Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2479468Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2479805Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2480029Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2480122Z graph_break [] 2025-12-04T11:40:52.2480335Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2481114Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2481212Z warnings.warn( 2025-12-04T11:40:52.2481505Z __________ TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 ___________ 2025-12-04T11:40:52.2481624Z Traceback (most recent call last): 2025-12-04T11:40:52.2482117Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 216, in test_softmax_acc_with_fp64 2025-12-04T11:40:52.2482314Z res, (code,) = run_and_get_code(torch.compile(fn), x, dim=-1) 2025-12-04T11:40:52.2482516Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2482521Z 2025-12-04T11:40:52.2482732Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2483323Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 2025-12-04T11:40:52.2483330Z 2025-12-04T11:40:52.2483624Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2483850Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2483958Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2484100Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2484448Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2484661Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2484767Z graph_break [] 2025-12-04T11:40:52.2484975Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2485756Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2485869Z warnings.warn( 2025-12-04T11:40:52.2486077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2486185Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2486302Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2486516Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2486863Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2486957Z graph_break [] 2025-12-04T11:40:52.2487168Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2487906Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2488001Z warnings.warn( 2025-12-04T11:40:52.2488141Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2488431Z __________ TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 ___________ 2025-12-04T11:40:52.2488553Z Traceback (most recent call last): 2025-12-04T11:40:52.2489058Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 216, in test_softmax_acc_with_fp64 2025-12-04T11:40:52.2489247Z res, (code,) = run_and_get_code(torch.compile(fn), x, dim=-1) 2025-12-04T11:40:52.2489458Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2489463Z 2025-12-04T11:40:52.2489685Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2490264Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 2025-12-04T11:40:52.2490270Z 2025-12-04T11:40:52.2490542Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2490754Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2490892Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2491016Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2491353Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2491569Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2491674Z graph_break [] 2025-12-04T11:40:52.2491887Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2492621Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2492718Z warnings.warn( 2025-12-04T11:40:52.2492927Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2493043Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2493155Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2493370Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2493713Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2493860Z graph_break [] 2025-12-04T11:40:52.2494072Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2494833Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2494932Z warnings.warn( 2025-12-04T11:40:52.2495149Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2495256Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2495366Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2495590Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2495954Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2496343Z graph_break [] 2025-12-04T11:40:52.2496567Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2497366Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2497476Z warnings.warn( 2025-12-04T11:40:52.2498238Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2423c3b275573349.xml - 2025-12-04T11:40:52.2498407Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2499142Z FAILED [0.1012s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_bfloat16 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2499148Z 2025-12-04T11:40:52.2499363Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2499955Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 2025-12-04T11:40:52.2499963Z 2025-12-04T11:40:52.2500225Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2500403Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2500612Z ================== 1 failed, 30 deselected, 2 rerun in 4.63s =================== 2025-12-04T11:40:52.2500708Z Got exit code 1 2025-12-04T11:40:52.2500822Z Retrying single test... 2025-12-04T11:40:52.2501408Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-66e69110dfb99f31.xml 2025-12-04T11:40:52.2501568Z ============================= test session starts ============================== 2025-12-04T11:40:52.2501927Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2502104Z cachedir: .pytest_cache 2025-12-04T11:40:52.2502625Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2502758Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2502862Z configfile: pytest.ini 2025-12-04T11:40:52.2503471Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2503685Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.2504349Z stepcurrent: skipping 22 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_bfloat16 2025-12-04T11:40:52.2504469Z Running 1 items in this shard 2025-12-04T11:40:52.2504474Z 2025-12-04T11:40:52.2505028Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_bfloat16 ('RERUN', {'yellow': True}) [4.3623s] [100%] 2025-12-04T11:40:52.2505589Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_bfloat16 ('RERUN', {'yellow': True}) [0.1032s] [100%] 2025-12-04T11:40:52.2506095Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_bfloat16 FAILED [0.1012s] [100%] 2025-12-04T11:40:52.2506100Z 2025-12-04T11:40:52.2506280Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2506568Z __________ TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 ___________ 2025-12-04T11:40:52.2506686Z Traceback (most recent call last): 2025-12-04T11:40:52.2507195Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 216, in test_softmax_acc_with_fp64 2025-12-04T11:40:52.2507384Z res, (code,) = run_and_get_code(torch.compile(fn), x, dim=-1) 2025-12-04T11:40:52.2507632Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2507638Z 2025-12-04T11:40:52.2507861Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2508445Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 2025-12-04T11:40:52.2508452Z 2025-12-04T11:40:52.2508726Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2508940Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2509047Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2509170Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2509507Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2509724Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2509830Z graph_break [] 2025-12-04T11:40:52.2510043Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2510790Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2510892Z warnings.warn( 2025-12-04T11:40:52.2511170Z __________ TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 ___________ 2025-12-04T11:40:52.2511296Z Traceback (most recent call last): 2025-12-04T11:40:52.2511785Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 216, in test_softmax_acc_with_fp64 2025-12-04T11:40:52.2511974Z res, (code,) = run_and_get_code(torch.compile(fn), x, dim=-1) 2025-12-04T11:40:52.2512191Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2512197Z 2025-12-04T11:40:52.2512404Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2512995Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 2025-12-04T11:40:52.2513001Z 2025-12-04T11:40:52.2513292Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2513506Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2513625Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2513736Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2514086Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2514300Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2514394Z graph_break [] 2025-12-04T11:40:52.2514617Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2515341Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2515441Z warnings.warn( 2025-12-04T11:40:52.2515664Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2515771Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2515930Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2516148Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2516485Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2516643Z graph_break [] 2025-12-04T11:40:52.2516858Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2517585Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2517698Z warnings.warn( 2025-12-04T11:40:52.2517841Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2518163Z __________ TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 ___________ 2025-12-04T11:40:52.2518285Z Traceback (most recent call last): 2025-12-04T11:40:52.2518780Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 216, in test_softmax_acc_with_fp64 2025-12-04T11:40:52.2518987Z res, (code,) = run_and_get_code(torch.compile(fn), x, dim=-1) 2025-12-04T11:40:52.2519195Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2519202Z 2025-12-04T11:40:52.2519415Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2520013Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 2025-12-04T11:40:52.2520018Z 2025-12-04T11:40:52.2520284Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2520515Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2520629Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2520743Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2521099Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2521319Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2521430Z graph_break [] 2025-12-04T11:40:52.2521645Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2522382Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2522497Z warnings.warn( 2025-12-04T11:40:52.2522709Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2522820Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2522949Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2523170Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2523515Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2523647Z graph_break [] 2025-12-04T11:40:52.2523864Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2524610Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2524711Z warnings.warn( 2025-12-04T11:40:52.2524925Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2525046Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2525161Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2525381Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2525726Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2525826Z graph_break [] 2025-12-04T11:40:52.2526053Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2526780Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2526909Z warnings.warn( 2025-12-04T11:40:52.2527714Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-66e69110dfb99f31.xml - 2025-12-04T11:40:52.2527885Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2528618Z FAILED [0.1012s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_bfloat16 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2528624Z 2025-12-04T11:40:52.2528835Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2529444Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn0_bfloat16 2025-12-04T11:40:52.2529452Z 2025-12-04T11:40:52.2529727Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2529902Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2530108Z ================== 1 failed, 30 deselected, 2 rerun in 4.60s =================== 2025-12-04T11:40:52.2530203Z Got exit code 1 2025-12-04T11:40:52.2530705Z FAILED CONSISTENTLY: test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_bfloat16 2025-12-04T11:40:52.2531125Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:40:52.2531714Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-4e57fa3cbc487129.xml 2025-12-04T11:40:52.2531892Z ============================= test session starts ============================== 2025-12-04T11:40:52.2532235Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2532343Z cachedir: .pytest_cache 2025-12-04T11:40:52.2532875Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2532997Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2533105Z configfile: pytest.ini 2025-12-04T11:40:52.2533706Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2533917Z collecting ... collected 31 items / 23 deselected / 8 selected 2025-12-04T11:40:52.2534070Z stepcurrent: skipping 23 already run items. 2025-12-04T11:40:52.2534182Z Running 8 items in this shard 2025-12-04T11:40:52.2534187Z 2025-12-04T11:40:52.2534660Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_float16 PASSED [5.1437s] [ 12%] 2025-12-04T11:40:52.2535183Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_float32 PASSED [0.4028s] [ 25%] 2025-12-04T11:40:52.2535737Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_bfloat16 ('RERUN', {'yellow': True}) [0.1317s] [ 37%] 2025-12-04T11:40:52.2536303Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_bfloat16 ('RERUN', {'yellow': True}) [0.1011s] [ 37%] 2025-12-04T11:40:52.2536765Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_bfloat16 FAILED [0.0994s] [ 37%] 2025-12-04T11:40:52.2536770Z 2025-12-04T11:40:52.2536977Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2537271Z __________ TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 ___________ 2025-12-04T11:40:52.2537390Z Traceback (most recent call last): 2025-12-04T11:40:52.2537904Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 216, in test_softmax_acc_with_fp64 2025-12-04T11:40:52.2538100Z res, (code,) = run_and_get_code(torch.compile(fn), x, dim=-1) 2025-12-04T11:40:52.2538348Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2538353Z 2025-12-04T11:40:52.2538578Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2539188Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 2025-12-04T11:40:52.2539194Z 2025-12-04T11:40:52.2539471Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2539686Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2539796Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2539919Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2540164Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2540505Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2540615Z graph_break [] 2025-12-04T11:40:52.2540827Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2541573Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2541673Z warnings.warn( 2025-12-04T11:40:52.2541950Z __________ TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 ___________ 2025-12-04T11:40:52.2542081Z Traceback (most recent call last): 2025-12-04T11:40:52.2542572Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 216, in test_softmax_acc_with_fp64 2025-12-04T11:40:52.2542764Z res, (code,) = run_and_get_code(torch.compile(fn), x, dim=-1) 2025-12-04T11:40:52.2542990Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2542997Z 2025-12-04T11:40:52.2543209Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2543809Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 2025-12-04T11:40:52.2543814Z 2025-12-04T11:40:52.2544080Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2544292Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2544414Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2544527Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2544749Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2545102Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2545202Z graph_break [] 2025-12-04T11:40:52.2545432Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2546192Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2546294Z warnings.warn( 2025-12-04T11:40:52.2546515Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2546625Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2546738Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2546967Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2547302Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2547410Z graph_break [] 2025-12-04T11:40:52.2547623Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2548345Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2548455Z warnings.warn( 2025-12-04T11:40:52.2548597Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2548917Z __________ TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 ___________ 2025-12-04T11:40:52.2549035Z Traceback (most recent call last): 2025-12-04T11:40:52.2549562Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 216, in test_softmax_acc_with_fp64 2025-12-04T11:40:52.2549765Z res, (code,) = run_and_get_code(torch.compile(fn), x, dim=-1) 2025-12-04T11:40:52.2549974Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2549979Z 2025-12-04T11:40:52.2550189Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2550808Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 2025-12-04T11:40:52.2550813Z 2025-12-04T11:40:52.2551080Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2551306Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2551413Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2551528Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2551759Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2552094Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2552190Z graph_break [] 2025-12-04T11:40:52.2552413Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2553138Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2553249Z warnings.warn( 2025-12-04T11:40:52.2553464Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2553576Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2553698Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2553915Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2554247Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2554355Z graph_break [] 2025-12-04T11:40:52.2554565Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2555302Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2555397Z warnings.warn( 2025-12-04T11:40:52.2555606Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2555722Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2555835Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2556087Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2556434Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2556532Z graph_break [] 2025-12-04T11:40:52.2556753Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2557480Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2557580Z warnings.warn( 2025-12-04T11:40:52.2558350Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-4e57fa3cbc487129.xml - 2025-12-04T11:40:52.2558518Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2559251Z FAILED [0.0994s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_bfloat16 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2559259Z 2025-12-04T11:40:52.2559503Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2560087Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 2025-12-04T11:40:52.2560092Z 2025-12-04T11:40:52.2560477Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2560654Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2560883Z ============= 1 failed, 2 passed, 23 deselected, 2 rerun in 5.91s ============== 2025-12-04T11:40:52.2560978Z Got exit code 1 2025-12-04T11:40:52.2561084Z Retrying single test... 2025-12-04T11:40:52.2561711Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-7331926cdb107464.xml 2025-12-04T11:40:52.2561873Z ============================= test session starts ============================== 2025-12-04T11:40:52.2562221Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2562340Z cachedir: .pytest_cache 2025-12-04T11:40:52.2562864Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2563000Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2563106Z configfile: pytest.ini 2025-12-04T11:40:52.2563692Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2563917Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.2564582Z stepcurrent: skipping 25 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_bfloat16 2025-12-04T11:40:52.2564693Z Running 1 items in this shard 2025-12-04T11:40:52.2564710Z 2025-12-04T11:40:52.2565264Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_bfloat16 ('RERUN', {'yellow': True}) [4.3901s] [100%] 2025-12-04T11:40:52.2565815Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_bfloat16 ('RERUN', {'yellow': True}) [0.1021s] [100%] 2025-12-04T11:40:52.2566292Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_bfloat16 FAILED [0.0987s] [100%] 2025-12-04T11:40:52.2566297Z 2025-12-04T11:40:52.2566436Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2566728Z __________ TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 ___________ 2025-12-04T11:40:52.2566847Z Traceback (most recent call last): 2025-12-04T11:40:52.2567343Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 216, in test_softmax_acc_with_fp64 2025-12-04T11:40:52.2567549Z res, (code,) = run_and_get_code(torch.compile(fn), x, dim=-1) 2025-12-04T11:40:52.2567792Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2567799Z 2025-12-04T11:40:52.2568012Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2568612Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 2025-12-04T11:40:52.2568617Z 2025-12-04T11:40:52.2568883Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2569110Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2569220Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2569334Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2569686Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2569906Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2570020Z graph_break [] 2025-12-04T11:40:52.2570234Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2571000Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2571114Z warnings.warn( 2025-12-04T11:40:52.2571426Z __________ TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 ___________ 2025-12-04T11:40:52.2571550Z Traceback (most recent call last): 2025-12-04T11:40:52.2572060Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 216, in test_softmax_acc_with_fp64 2025-12-04T11:40:52.2572251Z res, (code,) = run_and_get_code(torch.compile(fn), x, dim=-1) 2025-12-04T11:40:52.2572474Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2572480Z 2025-12-04T11:40:52.2572736Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2573321Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 2025-12-04T11:40:52.2573328Z 2025-12-04T11:40:52.2573609Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2573826Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2573953Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2574069Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2574410Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2574646Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2574743Z graph_break [] 2025-12-04T11:40:52.2574959Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2575710Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2575815Z warnings.warn( 2025-12-04T11:40:52.2576043Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2576152Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2576267Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2576503Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2576840Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2577003Z graph_break [] 2025-12-04T11:40:52.2577235Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2577964Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2578081Z warnings.warn( 2025-12-04T11:40:52.2578222Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2578539Z __________ TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 ___________ 2025-12-04T11:40:52.2578675Z Traceback (most recent call last): 2025-12-04T11:40:52.2579168Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 216, in test_softmax_acc_with_fp64 2025-12-04T11:40:52.2579362Z res, (code,) = run_and_get_code(torch.compile(fn), x, dim=-1) 2025-12-04T11:40:52.2579583Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2579588Z 2025-12-04T11:40:52.2579802Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2580397Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 2025-12-04T11:40:52.2580403Z 2025-12-04T11:40:52.2580668Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2580888Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2581038Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2581149Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2581490Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2581744Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2581841Z graph_break [] 2025-12-04T11:40:52.2582062Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2582788Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2582886Z warnings.warn( 2025-12-04T11:40:52.2583109Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2583251Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2583367Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2583594Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2583931Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2584045Z graph_break [] 2025-12-04T11:40:52.2584257Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2584987Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2585099Z warnings.warn( 2025-12-04T11:40:52.2585310Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2585416Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2585538Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2585756Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2586101Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2586199Z graph_break [] 2025-12-04T11:40:52.2586411Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2587148Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2587248Z warnings.warn( 2025-12-04T11:40:52.2588012Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-7331926cdb107464.xml - 2025-12-04T11:40:52.2588188Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2588911Z FAILED [0.0987s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_bfloat16 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2588917Z 2025-12-04T11:40:52.2589180Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2589764Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 2025-12-04T11:40:52.2589770Z 2025-12-04T11:40:52.2590049Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2590229Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2590424Z ================== 1 failed, 30 deselected, 2 rerun in 4.62s =================== 2025-12-04T11:40:52.2590533Z Got exit code 1 2025-12-04T11:40:52.2590637Z Retrying single test... 2025-12-04T11:40:52.2591227Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-7d66ea27b5ac2c65.xml 2025-12-04T11:40:52.2591403Z ============================= test session starts ============================== 2025-12-04T11:40:52.2591745Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2591897Z cachedir: .pytest_cache 2025-12-04T11:40:52.2592417Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2592538Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2592690Z configfile: pytest.ini 2025-12-04T11:40:52.2593281Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2593494Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.2594172Z stepcurrent: skipping 25 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_bfloat16 2025-12-04T11:40:52.2594286Z Running 1 items in this shard 2025-12-04T11:40:52.2594321Z 2025-12-04T11:40:52.2594889Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_bfloat16 ('RERUN', {'yellow': True}) [4.3673s] [100%] 2025-12-04T11:40:52.2595444Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_bfloat16 ('RERUN', {'yellow': True}) [0.1042s] [100%] 2025-12-04T11:40:52.2595912Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_bfloat16 FAILED [0.1025s] [100%] 2025-12-04T11:40:52.2595930Z 2025-12-04T11:40:52.2596236Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2596518Z __________ TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 ___________ 2025-12-04T11:40:52.2596654Z Traceback (most recent call last): 2025-12-04T11:40:52.2597152Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 216, in test_softmax_acc_with_fp64 2025-12-04T11:40:52.2597347Z res, (code,) = run_and_get_code(torch.compile(fn), x, dim=-1) 2025-12-04T11:40:52.2597573Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2597581Z 2025-12-04T11:40:52.2597793Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2598394Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 2025-12-04T11:40:52.2598399Z 2025-12-04T11:40:52.2598669Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2598883Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2599007Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2599119Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2599471Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2599688Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2599783Z graph_break [] 2025-12-04T11:40:52.2600007Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2600801Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2600904Z warnings.warn( 2025-12-04T11:40:52.2601204Z __________ TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 ___________ 2025-12-04T11:40:52.2601325Z Traceback (most recent call last): 2025-12-04T11:40:52.2601830Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 216, in test_softmax_acc_with_fp64 2025-12-04T11:40:52.2602019Z res, (code,) = run_and_get_code(torch.compile(fn), x, dim=-1) 2025-12-04T11:40:52.2602227Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2602233Z 2025-12-04T11:40:52.2602459Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2603045Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 2025-12-04T11:40:52.2603052Z 2025-12-04T11:40:52.2603372Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2603584Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2603691Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2603858Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2604198Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2604413Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2604520Z graph_break [] 2025-12-04T11:40:52.2604732Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2605516Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2605618Z warnings.warn( 2025-12-04T11:40:52.2605830Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2605955Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2606068Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2606286Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2606634Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2606729Z graph_break [] 2025-12-04T11:40:52.2606941Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2607678Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2607776Z warnings.warn( 2025-12-04T11:40:52.2607931Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2608208Z __________ TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 ___________ 2025-12-04T11:40:52.2608331Z Traceback (most recent call last): 2025-12-04T11:40:52.2608842Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 216, in test_softmax_acc_with_fp64 2025-12-04T11:40:52.2609032Z res, (code,) = run_and_get_code(torch.compile(fn), x, dim=-1) 2025-12-04T11:40:52.2609251Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2609256Z 2025-12-04T11:40:52.2609469Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2610050Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 2025-12-04T11:40:52.2610056Z 2025-12-04T11:40:52.2610329Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2610543Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2610695Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2610807Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2611148Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2611377Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2611474Z graph_break [] 2025-12-04T11:40:52.2611683Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2612428Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2612529Z warnings.warn( 2025-12-04T11:40:52.2612755Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2612868Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2612981Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2613210Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2613549Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2613675Z graph_break [] 2025-12-04T11:40:52.2613903Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2614675Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2614789Z warnings.warn( 2025-12-04T11:40:52.2615003Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2615114Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2615242Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2615460Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2615824Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2615936Z graph_break [] 2025-12-04T11:40:52.2616147Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2616871Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2617057Z warnings.warn( 2025-12-04T11:40:52.2617826Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-7d66ea27b5ac2c65.xml - 2025-12-04T11:40:52.2618005Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2618733Z FAILED [0.1025s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_bfloat16 - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2618738Z 2025-12-04T11:40:52.2618967Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2619550Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_softmax_acc_with_fp64_fn1_bfloat16 2025-12-04T11:40:52.2619557Z 2025-12-04T11:40:52.2619821Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2620012Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2620213Z ================== 1 failed, 30 deselected, 2 rerun in 4.60s =================== 2025-12-04T11:40:52.2620312Z Got exit code 1 2025-12-04T11:40:52.2620826Z FAILED CONSISTENTLY: test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_bfloat16 2025-12-04T11:40:52.2621539Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:40:52.2622193Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-356c1a001f2d20d8.xml 2025-12-04T11:40:52.2622491Z ============================= test session starts ============================== 2025-12-04T11:40:52.2622883Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2623043Z cachedir: .pytest_cache 2025-12-04T11:40:52.2623633Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2623840Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2624005Z configfile: pytest.ini 2025-12-04T11:40:52.2624688Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2624951Z collecting ... collected 31 items / 26 deselected / 5 selected 2025-12-04T11:40:52.2625180Z stepcurrent: skipping 26 already run items. 2025-12-04T11:40:52.2625309Z Running 5 items in this shard 2025-12-04T11:40:52.2625316Z 2025-12-04T11:40:52.2625863Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_float16 PASSED [5.0973s] [ 20%] 2025-12-04T11:40:52.2626541Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_float32 PASSED [0.3982s] [ 40%] 2025-12-04T11:40:52.2626931Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmin PASSED [0.2433s] [ 60%] 2025-12-04T11:40:52.2627527Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_split_reduction ('RERUN', {'yellow': True}) [0.1445s] [ 80%] 2025-12-04T11:40:52.2628035Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_split_reduction ('RERUN', {'yellow': True}) [0.1048s] [ 80%] 2025-12-04T11:40:52.2628432Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_split_reduction FAILED [0.1040s] [ 80%] 2025-12-04T11:40:52.2628438Z 2025-12-04T11:40:52.2628722Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2629054Z ____________________ TestOnlineSoftmax.test_split_reduction ____________________ 2025-12-04T11:40:52.2629266Z Traceback (most recent call last): 2025-12-04T11:40:52.2629771Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 165, in test_split_reduction 2025-12-04T11:40:52.2630039Z act, (code,) = run_and_get_code(torch.compile(torch.softmax), x, dim=-1) 2025-12-04T11:40:52.2630356Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2630362Z 2025-12-04T11:40:52.2630642Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2631219Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_split_reduction 2025-12-04T11:40:52.2631225Z 2025-12-04T11:40:52.2631526Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2631780Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2631957Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2632149Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2632437Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2632867Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2633027Z graph_break [] 2025-12-04T11:40:52.2633705Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2634692Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2635549Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2636229Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2636965Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2637450Z softmax. 2025-12-04T11:40:52.2637693Z 2025-12-04T11:40:52.2637842Z warnings.warn( 2025-12-04T11:40:52.2638873Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2640030Z warnings.warn( 2025-12-04T11:40:52.2640553Z ____________________ TestOnlineSoftmax.test_split_reduction ____________________ 2025-12-04T11:40:52.2641283Z Traceback (most recent call last): 2025-12-04T11:40:52.2642091Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 165, in test_split_reduction 2025-12-04T11:40:52.2643167Z act, (code,) = run_and_get_code(torch.compile(torch.softmax), x, dim=-1) 2025-12-04T11:40:52.2643916Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2644343Z 2025-12-04T11:40:52.2644596Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2645596Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_split_reduction 2025-12-04T11:40:52.2646263Z 2025-12-04T11:40:52.2646618Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2647312Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2647983Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2648431Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2648944Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2649837Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2650536Z graph_break [] 2025-12-04T11:40:52.2651015Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2651985Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2652883Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2653552Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2654261Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2654821Z softmax. 2025-12-04T11:40:52.2655040Z 2025-12-04T11:40:52.2655158Z warnings.warn( 2025-12-04T11:40:52.2656238Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2657386Z warnings.warn( 2025-12-04T11:40:52.2657864Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2658512Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2658917Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2659446Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2660396Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2661102Z graph_break [] 2025-12-04T11:40:52.2661520Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2662473Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2663372Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2664056Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2664689Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2665225Z softmax. 2025-12-04T11:40:52.2665396Z 2025-12-04T11:40:52.2665591Z warnings.warn( 2025-12-04T11:40:52.2666574Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2667654Z warnings.warn( 2025-12-04T11:40:52.2668143Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2668841Z ____________________ TestOnlineSoftmax.test_split_reduction ____________________ 2025-12-04T11:40:52.2669429Z Traceback (most recent call last): 2025-12-04T11:40:52.2670274Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 165, in test_split_reduction 2025-12-04T11:40:52.2671287Z act, (code,) = run_and_get_code(torch.compile(torch.softmax), x, dim=-1) 2025-12-04T11:40:52.2671981Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2672346Z 2025-12-04T11:40:52.2678104Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2678991Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_split_reduction 2025-12-04T11:40:52.2679631Z 2025-12-04T11:40:52.2679898Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2680522Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2680989Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2681307Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2681733Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2682429Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2683084Z graph_break [] 2025-12-04T11:40:52.2683450Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2684257Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2685029Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2685577Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2686103Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2686525Z softmax. 2025-12-04T11:40:52.2686652Z 2025-12-04T11:40:52.2686745Z warnings.warn( 2025-12-04T11:40:52.2687680Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2688642Z warnings.warn( 2025-12-04T11:40:52.2689013Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2689471Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2689795Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2690218Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2690900Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2691469Z graph_break [] 2025-12-04T11:40:52.2691832Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2692591Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2693352Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2693905Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2694430Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2694836Z softmax. 2025-12-04T11:40:52.2694971Z 2025-12-04T11:40:52.2695068Z warnings.warn( 2025-12-04T11:40:52.2695949Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2697180Z warnings.warn( 2025-12-04T11:40:52.2697545Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2698005Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2698325Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2698739Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2699421Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2699986Z graph_break [] 2025-12-04T11:40:52.2700341Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2701099Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2701948Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2702496Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2703012Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2703425Z softmax. 2025-12-04T11:40:52.2703551Z 2025-12-04T11:40:52.2703655Z warnings.warn( 2025-12-04T11:40:52.2704521Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2705476Z warnings.warn( 2025-12-04T11:40:52.2706400Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-356c1a001f2d20d8.xml - 2025-12-04T11:40:52.2707470Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2708431Z FAILED [0.1040s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_split_reduction - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2709266Z 2025-12-04T11:40:52.2709478Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2710357Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_split_reduction 2025-12-04T11:40:52.2710976Z 2025-12-04T11:40:52.2711250Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2711819Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2712344Z ============= 1 failed, 3 passed, 26 deselected, 2 rerun in 6.13s ============== 2025-12-04T11:40:52.2712788Z Got exit code 1 2025-12-04T11:40:52.2713048Z Retrying single test... 2025-12-04T11:40:52.2713847Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-426f4e961e373829.xml 2025-12-04T11:40:52.2714732Z ============================= test session starts ============================== 2025-12-04T11:40:52.2715387Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2715965Z cachedir: .pytest_cache 2025-12-04T11:40:52.2716665Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2717437Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2717783Z configfile: pytest.ini 2025-12-04T11:40:52.2718540Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2719479Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.2720388Z stepcurrent: skipping 29 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_split_reduction 2025-12-04T11:40:52.2721198Z Running 1 items in this shard 2025-12-04T11:40:52.2721415Z 2025-12-04T11:40:52.2721894Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_split_reduction ('RERUN', {'yellow': True}) [4.3984s] [100%] 2025-12-04T11:40:52.2722975Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_split_reduction ('RERUN', {'yellow': True}) [0.1076s] [100%] 2025-12-04T11:40:52.2723964Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_split_reduction FAILED [0.1062s] [100%] 2025-12-04T11:40:52.2724481Z 2025-12-04T11:40:52.2724618Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2725149Z ____________________ TestOnlineSoftmax.test_split_reduction ____________________ 2025-12-04T11:40:52.2725655Z Traceback (most recent call last): 2025-12-04T11:40:52.2726355Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 165, in test_split_reduction 2025-12-04T11:40:52.2727175Z act, (code,) = run_and_get_code(torch.compile(torch.softmax), x, dim=-1) 2025-12-04T11:40:52.2727780Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2728121Z 2025-12-04T11:40:52.2728336Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2729163Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_split_reduction 2025-12-04T11:40:52.2729778Z 2025-12-04T11:40:52.2730039Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2730651Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2731107Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2731417Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2731959Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2732648Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2733096Z graph_break [] 2025-12-04T11:40:52.2733447Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2734247Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2735015Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2735584Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2736115Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2736530Z softmax. 2025-12-04T11:40:52.2736655Z 2025-12-04T11:40:52.2736763Z warnings.warn( 2025-12-04T11:40:52.2737717Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2738682Z warnings.warn( 2025-12-04T11:40:52.2739140Z ____________________ TestOnlineSoftmax.test_split_reduction ____________________ 2025-12-04T11:40:52.2739652Z Traceback (most recent call last): 2025-12-04T11:40:52.2740350Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 165, in test_split_reduction 2025-12-04T11:40:52.2741181Z act, (code,) = run_and_get_code(torch.compile(torch.softmax), x, dim=-1) 2025-12-04T11:40:52.2741753Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2742087Z 2025-12-04T11:40:52.2742298Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2743127Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_split_reduction 2025-12-04T11:40:52.2743748Z 2025-12-04T11:40:52.2744016Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2744631Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2745088Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2745411Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2745961Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2746652Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2747098Z graph_break [] 2025-12-04T11:40:52.2747459Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2748212Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2748980Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2749521Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2750043Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2750449Z softmax. 2025-12-04T11:40:52.2750585Z 2025-12-04T11:40:52.2750681Z warnings.warn( 2025-12-04T11:40:52.2751576Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2752558Z warnings.warn( 2025-12-04T11:40:52.2752927Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2753380Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2753697Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2754112Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2754797Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2755365Z graph_break [] 2025-12-04T11:40:52.2755714Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2756467Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2757231Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2757783Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2758301Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2758749Z softmax. 2025-12-04T11:40:52.2758871Z 2025-12-04T11:40:52.2758973Z warnings.warn( 2025-12-04T11:40:52.2759883Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2760837Z warnings.warn( 2025-12-04T11:40:52.2761132Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2761668Z ____________________ TestOnlineSoftmax.test_split_reduction ____________________ 2025-12-04T11:40:52.2762161Z Traceback (most recent call last): 2025-12-04T11:40:52.2762846Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 165, in test_split_reduction 2025-12-04T11:40:52.2763717Z act, (code,) = run_and_get_code(torch.compile(torch.softmax), x, dim=-1) 2025-12-04T11:40:52.2764276Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2764626Z 2025-12-04T11:40:52.2764838Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2765662Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_split_reduction 2025-12-04T11:40:52.2766281Z 2025-12-04T11:40:52.2766552Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2767159Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2767621Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2767948Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2768486Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2769175Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2769621Z graph_break [] 2025-12-04T11:40:52.2769983Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2770744Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2771507Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2772060Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2772580Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2772993Z softmax. 2025-12-04T11:40:52.2773120Z 2025-12-04T11:40:52.2773218Z warnings.warn( 2025-12-04T11:40:52.2774101Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2775046Z warnings.warn( 2025-12-04T11:40:52.2775414Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2775878Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2776188Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2776642Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2777406Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2777979Z graph_break [] 2025-12-04T11:40:52.2778328Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2779090Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2779856Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2780397Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2780927Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2781342Z softmax. 2025-12-04T11:40:52.2781467Z 2025-12-04T11:40:52.2781559Z warnings.warn( 2025-12-04T11:40:52.2782444Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2783402Z warnings.warn( 2025-12-04T11:40:52.2783807Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2784260Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2784583Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2785040Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2785719Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2786286Z graph_break [] 2025-12-04T11:40:52.2786649Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2787411Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2788167Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2788743Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2789266Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2789675Z softmax. 2025-12-04T11:40:52.2789805Z 2025-12-04T11:40:52.2789900Z warnings.warn( 2025-12-04T11:40:52.2790790Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2791747Z warnings.warn( 2025-12-04T11:40:52.2792668Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-426f4e961e373829.xml - 2025-12-04T11:40:52.2793738Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2794704Z FAILED [0.1062s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_split_reduction - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2795484Z 2025-12-04T11:40:52.2795713Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2796747Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_split_reduction 2025-12-04T11:40:52.2797383Z 2025-12-04T11:40:52.2797648Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2798236Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2798755Z ================== 1 failed, 30 deselected, 2 rerun in 4.64s =================== 2025-12-04T11:40:52.2799181Z Got exit code 1 2025-12-04T11:40:52.2799442Z Retrying single test... 2025-12-04T11:40:52.2800220Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-1a1453b5e01c9120.xml 2025-12-04T11:40:52.2801096Z ============================= test session starts ============================== 2025-12-04T11:40:52.2801761Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2802420Z cachedir: .pytest_cache 2025-12-04T11:40:52.2803131Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2803899Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2804244Z configfile: pytest.ini 2025-12-04T11:40:52.2805021Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2805946Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.2806863Z stepcurrent: skipping 29 already run items. Running only test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_split_reduction 2025-12-04T11:40:52.2807666Z Running 1 items in this shard 2025-12-04T11:40:52.2807874Z 2025-12-04T11:40:52.2808360Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_split_reduction ('RERUN', {'yellow': True}) [4.3757s] [100%] 2025-12-04T11:40:52.2809444Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_split_reduction ('RERUN', {'yellow': True}) [0.1053s] [100%] 2025-12-04T11:40:52.2810474Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_split_reduction FAILED [0.1034s] [100%] 2025-12-04T11:40:52.2810997Z 2025-12-04T11:40:52.2811171Z ==================================== RERUNS ==================================== 2025-12-04T11:40:52.2811709Z ____________________ TestOnlineSoftmax.test_split_reduction ____________________ 2025-12-04T11:40:52.2812205Z Traceback (most recent call last): 2025-12-04T11:40:52.2812908Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 165, in test_split_reduction 2025-12-04T11:40:52.2813734Z act, (code,) = run_and_get_code(torch.compile(torch.softmax), x, dim=-1) 2025-12-04T11:40:52.2814302Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2814687Z 2025-12-04T11:40:52.2814903Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2815726Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_split_reduction 2025-12-04T11:40:52.2816351Z 2025-12-04T11:40:52.2816620Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2817301Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2817759Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2818083Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2818633Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2819325Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2819771Z graph_break [] 2025-12-04T11:40:52.2820135Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2820900Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2821682Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2822241Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2822768Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2823177Z softmax. 2025-12-04T11:40:52.2823317Z 2025-12-04T11:40:52.2823413Z warnings.warn( 2025-12-04T11:40:52.2824305Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2825264Z warnings.warn( 2025-12-04T11:40:52.2825678Z ____________________ TestOnlineSoftmax.test_split_reduction ____________________ 2025-12-04T11:40:52.2826190Z Traceback (most recent call last): 2025-12-04T11:40:52.2826891Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 165, in test_split_reduction 2025-12-04T11:40:52.2827720Z act, (code,) = run_and_get_code(torch.compile(torch.softmax), x, dim=-1) 2025-12-04T11:40:52.2828328Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2828669Z 2025-12-04T11:40:52.2828892Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2829717Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_split_reduction 2025-12-04T11:40:52.2830348Z 2025-12-04T11:40:52.2830612Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2831234Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2831699Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2832015Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2832567Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2833266Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2833710Z graph_break [] 2025-12-04T11:40:52.2834069Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2834873Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2835640Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2836219Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2836761Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2837177Z softmax. 2025-12-04T11:40:52.2837302Z 2025-12-04T11:40:52.2837398Z warnings.warn( 2025-12-04T11:40:52.2838289Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2839248Z warnings.warn( 2025-12-04T11:40:52.2839653Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2840109Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2840437Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2840870Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2841550Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2842129Z graph_break [] 2025-12-04T11:40:52.2842499Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2843258Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2844027Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2844573Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2845102Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2845515Z softmax. 2025-12-04T11:40:52.2845647Z 2025-12-04T11:40:52.2845746Z warnings.warn( 2025-12-04T11:40:52.2846636Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2847592Z warnings.warn( 2025-12-04T11:40:52.2847893Z =================================== FAILURES =================================== 2025-12-04T11:40:52.2848442Z ____________________ TestOnlineSoftmax.test_split_reduction ____________________ 2025-12-04T11:40:52.2848946Z Traceback (most recent call last): 2025-12-04T11:40:52.2849637Z File "/var/lib/jenkins/workspace/test/inductor/test_online_softmax.py", line 165, in test_split_reduction 2025-12-04T11:40:52.2850467Z act, (code,) = run_and_get_code(torch.compile(torch.softmax), x, dim=-1) 2025-12-04T11:40:52.2851043Z ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2851378Z 2025-12-04T11:40:52.2851598Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2852424Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_split_reduction 2025-12-04T11:40:52.2853092Z 2025-12-04T11:40:52.2853362Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2853980Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2854436Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2854766Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2855324Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2856021Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2856464Z graph_break [] 2025-12-04T11:40:52.2856836Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2857673Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2858439Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2859000Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2859567Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2859995Z softmax. 2025-12-04T11:40:52.2860121Z 2025-12-04T11:40:52.2860217Z warnings.warn( 2025-12-04T11:40:52.2861151Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2862121Z warnings.warn( 2025-12-04T11:40:52.2862486Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2862954Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2863276Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2863697Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2864486Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2865070Z graph_break [] 2025-12-04T11:40:52.2865436Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2866194Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2866969Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2867523Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2868044Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2868461Z softmax. 2025-12-04T11:40:52.2868596Z 2025-12-04T11:40:52.2868689Z warnings.warn( 2025-12-04T11:40:52.2869572Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2870526Z warnings.warn( 2025-12-04T11:40:52.2870898Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T11:40:52.2871364Z frames [('total', 1), ('ok', 1)] 2025-12-04T11:40:52.2871680Z stats [('calls_captured', 1)] 2025-12-04T11:40:52.2872107Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T11:40:52.2872800Z inductor [('pattern_matcher_nodes', 4), ('fxgraph_cache_miss', 1), ('pattern_matcher_count', 1)] 2025-12-04T11:40:52.2873374Z graph_break [] 2025-12-04T11:40:52.2873733Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T11:40:52.2874497Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/lowering.py:7581: UserWarning: 2025-12-04T11:40:52.2875261Z Online softmax is disabled on the fly since Inductor decides to 2025-12-04T11:40:52.2875805Z split the reduction. Cut an issue to PyTorch if this is an 2025-12-04T11:40:52.2876332Z important use case and you want to speed it up with online 2025-12-04T11:40:52.2876753Z softmax. 2025-12-04T11:40:52.2876880Z 2025-12-04T11:40:52.2876978Z warnings.warn( 2025-12-04T11:40:52.2877869Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T11:40:52.2878871Z warnings.warn( 2025-12-04T11:40:52.2879800Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-1a1453b5e01c9120.xml - 2025-12-04T11:40:52.2880865Z =========================== short test summary info ============================ 2025-12-04T11:40:52.2881835Z FAILED [0.1034s] inductor/test_online_softmax.py::TestOnlineSoftmax::test_split_reduction - ValueError: not enough values to unpack (expected 1, got 0) 2025-12-04T11:40:52.2882618Z 2025-12-04T11:40:52.2882832Z To execute this test, run the following from the base repo dir: 2025-12-04T11:40:52.2883661Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_online_softmax.py TestOnlineSoftmax.test_split_reduction 2025-12-04T11:40:52.2884288Z 2025-12-04T11:40:52.2884553Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T11:40:52.2885137Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T11:40:52.2885685Z ================== 1 failed, 30 deselected, 2 rerun in 4.61s =================== 2025-12-04T11:40:52.2886118Z Got exit code 1 2025-12-04T11:40:52.2886696Z FAILED CONSISTENTLY: test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_split_reduction 2025-12-04T11:40:52.2887654Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T11:40:52.2888788Z Test results will be stored in test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-09c678f485469fcb.xml 2025-12-04T11:40:52.2889670Z ============================= test session starts ============================== 2025-12-04T11:40:52.2890334Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T11:40:52.2890930Z cachedir: .pytest_cache 2025-12-04T11:40:52.2891636Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T11:40:52.2892402Z rootdir: /var/lib/jenkins/workspace 2025-12-04T11:40:52.2892741Z configfile: pytest.ini 2025-12-04T11:40:52.2893516Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T11:40:52.2894456Z collecting ... collected 31 items / 30 deselected / 1 selected 2025-12-04T11:40:52.2894937Z stepcurrent: skipping 30 already run items. 2025-12-04T11:40:52.2895314Z Running 1 items in this shard 2025-12-04T11:40:52.2895519Z 2025-12-04T11:40:52.2896124Z inductor/test_online_softmax.py::TestOnlineSoftmax::test_tb_speech_transformer_attn PASSED [5.2290s] [100%] 2025-12-04T11:40:52.2896699Z 2025-12-04T11:40:52.2897539Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-09c678f485469fcb.xml - 2025-12-04T11:40:52.2898619Z ======================= 1 passed, 30 deselected in 5.26s ======================= 2025-12-04T11:40:52.2907073Z The following tests failed consistently: ['test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_3pass_softmax_due_to_disable', 'test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_False', 'test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_2048_use_log_softmax_True', 'test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_False', 'test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_online_softmax_V_50304_use_log_softmax_True', 'test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_codegen_softmax_persistent_reduction', 'test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_acc_with_fp64_bfloat16', 'test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_-1', 'test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_0', 'test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2048_dim_1', 'test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_-1', 'test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_0', 'test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_prepare_softmax_nrow_2_dim_1', 'test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_sdpa', 'test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn0_bfloat16', 'test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_softmax_acc_with_fp64_fn1_bfloat16', 'test/inductor/test_online_softmax.py::TestOnlineSoftmax::test_split_reduction'] 2025-12-04T11:40:52.2915665Z 2025-12-04T11:40:52.2916234Z FINISHED PRINTING LOG FILE of inductor/test_online_softmax 1/1 (test/test-reports/inductor.test_online_softmax_1.1_d592e7340b8c6a36_.log) 2025-12-04T11:40:52.2916999Z 2025-12-04T11:40:52.2917362Z Finished inductor/test_online_softmax 1/1 ... [2025-12-04 11:40:52.066565][8835.756951501], took 16.26min 2025-12-04T11:40:52.2918740Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a12b7a446d46a036.xml 2025-12-04T11:40:52.2920560Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-c354a873b70ad9a4.xml 2025-12-04T11:40:52.2922329Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-72b2d7d92c2f3d43.xml 2025-12-04T11:40:52.2924141Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-c8076b5dd44238e4.xml 2025-12-04T11:40:52.2925919Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-d2d17ca8c68c28e7.xml 2025-12-04T11:40:52.2927710Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-b56c2dd706b17a20.xml 2025-12-04T11:40:52.3177311Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-fa3c7a9046cc8737.xml 2025-12-04T11:40:52.3455340Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3da5ad44ee5d80af.xml 2025-12-04T11:40:52.3736566Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2586fe9efcb2d9c1.xml 2025-12-04T11:40:52.4036561Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-4d9d69bfeda71662.xml 2025-12-04T11:40:52.4327353Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0045cf67af4f7e36.xml 2025-12-04T11:40:52.4664180Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-38fbcefce9021f2a.xml 2025-12-04T11:40:52.4967386Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0625eb2c44495ce0.xml 2025-12-04T11:40:52.5229897Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-ea417fd0de1ef597.xml 2025-12-04T11:40:52.5501462Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-e58bcfb06df4ce4d.xml 2025-12-04T11:40:52.5831558Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0c1484308d51b00a.xml 2025-12-04T11:40:52.6199516Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-9682f9212923ea0d.xml 2025-12-04T11:40:52.6522015Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-ef1f71bad0a01b52.xml 2025-12-04T11:40:52.6874292Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3d18fa53804224f4.xml 2025-12-04T11:40:52.7228821Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2256fbb35262b330.xml 2025-12-04T11:40:52.7532682Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a993cfee9c5b5572.xml 2025-12-04T11:40:52.7854583Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-4097f4a492ce8bb6.xml 2025-12-04T11:40:52.8155859Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-9df84d07ff6a448b.xml 2025-12-04T11:40:52.8477075Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-b20a3959da2014da.xml 2025-12-04T11:40:52.8756046Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-89205029d9c04efd.xml 2025-12-04T11:40:52.9223165Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-476b174d4b08ad56.xml 2025-12-04T11:40:52.9527416Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-99551b4e3863e77a.xml 2025-12-04T11:40:52.9873515Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-498b08ff107b8914.xml 2025-12-04T11:40:53.0216418Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0e343f505ff50c95.xml 2025-12-04T11:40:53.0549495Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-486e74c9978d3a7f.xml 2025-12-04T11:40:53.1032637Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a0676d6303da3615.xml 2025-12-04T11:40:53.1369377Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-e27b1d19486b6954.xml 2025-12-04T11:40:53.1832456Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3f1ba8060f3abc48.xml 2025-12-04T11:40:53.2494158Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0209a84b5ca28602.xml 2025-12-04T11:40:53.2825517Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3d838b8cb24fd196.xml 2025-12-04T11:40:53.3182398Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-6a6a2484271011a1.xml 2025-12-04T11:40:53.3517273Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-9356c0308b0c66f7.xml 2025-12-04T11:40:53.3824402Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-51758daa3a3dc599.xml 2025-12-04T11:40:53.4177932Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a0d42e3d123a44e8.xml 2025-12-04T11:40:53.4503702Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-20d10a20187a2cb9.xml 2025-12-04T11:40:53.4857902Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-5f3be1bf80a1ce40.xml 2025-12-04T11:40:53.5745628Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-baeb09a4e62b9d77.xml 2025-12-04T11:40:53.6071414Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2cdd67c45b570140.xml 2025-12-04T11:40:53.6519456Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2423c3b275573349.xml 2025-12-04T11:40:53.6859729Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-66e69110dfb99f31.xml 2025-12-04T11:40:53.7240027Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-4e57fa3cbc487129.xml 2025-12-04T11:40:53.7662842Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-7331926cdb107464.xml 2025-12-04T11:40:53.7955040Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-7d66ea27b5ac2c65.xml 2025-12-04T11:40:53.8244971Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-356c1a001f2d20d8.xml 2025-12-04T11:40:53.8531773Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-426f4e961e373829.xml 2025-12-04T11:40:53.8848753Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-1a1453b5e01c9120.xml 2025-12-04T11:40:53.9135515Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-09c678f485469fcb.xml 2025-12-04T11:40:55.4542927Z Uploading logs for 57119749282 to S3 2025-12-04T11:40:55.7229354Z Uploading artifacts took 1.78 seconds 2025-12-04T11:40:55.7230080Z inductor/test_online_softmax 1/1 failed! 2025-12-04T11:40:55.7235632Z Running inductor/test_mix_order_reduction 2/2 ... [2025-12-04 11:40:55.723324][8839.413714071] 2025-12-04T11:40:55.7236573Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T11:40:55.7242620Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_mix_order_reduction.py', '--shard-id=2', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 11:40:55.723888] 2025-12-04T12:25:16.4415304Z 2025-12-04T12:25:16.4416568Z PRINTING LOG FILE of inductor/test_mix_order_reduction 2/2 (test/test-reports/inductor.test_mix_order_reduction_2.2_7b65ea7a55b42f88_.log) 2025-12-04T12:25:16.4418695Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-fb2ec7b0e0c81fea.xml 2025-12-04T12:25:16.4419905Z ============================= test session starts ============================== 2025-12-04T12:25:16.4420717Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.4421423Z cachedir: .pytest_cache 2025-12-04T12:25:16.4422447Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.4423624Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.4424475Z configfile: pytest.ini 2025-12-04T12:25:16.4425803Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.4427295Z collecting ... collected 380 items 2025-12-04T12:25:16.4428174Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T12:25:16.4663354Z Running 205 items in this shard: test/inductor/test_mix_order_reduction.py::SkipPatternTest::test_dimension_too_close, test/inductor/test_mix_order_reduction.py::SkipPatternTest::test_skip_if_outer_reduction_followed_by_full_pointwise, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_XBLOCK_coordest_tuning, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_avoid_non_coalesced_access, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_independent_split_size, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_no_bias_split_reductions_True_shape0, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_no_bias_split_reductions_True_shape1, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_dynamic_shape_dynamic_dims0, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_dynamic_shape_dynamic_dims1, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_mean_swap_False_split_reductions_False_shape1, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_mean_swap_False_split_reductions_False_shape2, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_mean_swap_False_split_reductions_True_shape2, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_mean_swap_True_split_reductions_False_shape1, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_mean_swap_True_split_reductions_False_shape2, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_mean_swap_True_split_reductions_True_shape0, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_mean_swap_True_split_reductions_True_shape1, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_prod_swap_False_split_reductions_False_shape1, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_prod_swap_False_split_reductions_True_shape1, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_prod_swap_True_split_reductions_False_shape0, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_prod_swap_True_split_reductions_True_shape0, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_prod_swap_True_split_reductions_True_shape1, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_prod_swap_True_split_reductions_True_shape2, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_sum_swap_False_split_reductions_False_shape2, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_sum_swap_False_split_reductions_True_shape1, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_sum_swap_True_split_reductions_False_shape0, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_sum_swap_True_split_reductions_False_shape1, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_sum_swap_True_split_reductions_False_shape2, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_sum_swap_True_split_reductions_True_shape1, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_sum_swap_True_split_reductions_True_shape2, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_True_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_True_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_True_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_True_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_True_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_True_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_True_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_True_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_True_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_True_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_True_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_True_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_True_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_True_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_True_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_True_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_True_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_True_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_True_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_True_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_True_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_True_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_True_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape1_max_autotune_True_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape1_max_autotune_True_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape1_max_autotune_True_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_True_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_True_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_True_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_True_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_True_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_True_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_True_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_True_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape2_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape2_max_autotune_True_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_True_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_with_dynamic_shape_dynamic_dims1, test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_xmask, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_avoid_non_coalesced_access, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_independent_split_size, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_no_bias_split_reductions_False_shape0, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_no_bias_split_reductions_False_shape1, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_float32_split_reductions_False_shape1, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_float32_split_reductions_True_shape1, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_dynamic_shape_dynamic_dims1, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_mean_swap_False_split_reductions_False_shape1, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_mean_swap_False_split_reductions_False_shape2, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_mean_swap_False_split_reductions_True_shape1, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_mean_swap_False_split_reductions_True_shape2, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_mean_swap_True_split_reductions_False_shape0, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_mean_swap_True_split_reductions_False_shape1, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_mean_swap_True_split_reductions_False_shape2, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_mean_swap_True_split_reductions_True_shape0, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_mean_swap_True_split_reductions_True_shape2, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_prod_swap_False_split_reductions_False_shape0, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_prod_swap_False_split_reductions_False_shape2, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_prod_swap_False_split_reductions_True_shape0, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_prod_swap_False_split_reductions_True_shape2, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_prod_swap_True_split_reductions_False_shape1, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_prod_swap_True_split_reductions_True_shape1, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_prod_swap_True_split_reductions_True_shape2, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_sum_swap_False_split_reductions_False_shape1, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_sum_swap_False_split_reductions_False_shape2, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_sum_swap_False_split_reductions_True_shape0, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_sum_swap_False_split_reductions_True_shape1, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_sum_swap_True_split_reductions_False_shape0, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_sum_swap_True_split_reductions_True_shape2, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_True_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_True_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_True_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_True_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_True_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_True_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_True_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_True_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_True_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_True_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_True_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_True_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_True_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_True_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_True_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape1_max_autotune_True_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape1_max_autotune_True_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape1_max_autotune_True_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_True_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_True_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_True_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_True_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_True_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_True_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_True_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_True_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape2_max_autotune_True_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape2_max_autotune_True_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape2_max_autotune_True_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape2_max_autotune_True_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_True_initial_xblock_2_add_1dim_False, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_True_initial_xblock_2_add_1dim_True, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_with_dynamic_shape_dynamic_dims0, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_with_dynamic_shape_dynamic_dims2, test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_xmask 2025-12-04T12:25:16.4820713Z 2025-12-04T12:25:16.4821188Z inductor/test_mix_order_reduction.py::SkipPatternTest::test_dimension_too_close PASSED [4.9301s] [ 0%] 2025-12-04T12:25:16.4822333Z inductor/test_mix_order_reduction.py::SkipPatternTest::test_skip_if_outer_reduction_followed_by_full_pointwise PASSED [0.3758s] [ 0%] 2025-12-04T12:25:16.4823511Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_3layer_split_reduction PASSED [1.0183s] [ 1%] 2025-12-04T12:25:16.4824689Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_XBLOCK_coordest_tuning PASSED [5.5717s] [ 1%] 2025-12-04T12:25:16.4825772Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_avoid_non_coalesced_access PASSED [1.4232s] [ 2%] 2025-12-04T12:25:16.4826874Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_independent_split_size PASSED [1.3206s] [ 2%] 2025-12-04T12:25:16.4828083Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_no_bias_split_reductions_True_shape0 PASSED [1.5411s] [ 3%] 2025-12-04T12:25:16.4829416Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_no_bias_split_reductions_True_shape1 PASSED [1.4345s] [ 3%] 2025-12-04T12:25:16.4830890Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 ('RERUN', {'yellow': True}) [0.1789s] [ 4%] 2025-12-04T12:25:16.4832524Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 ('RERUN', {'yellow': True}) [0.1430s] [ 4%] 2025-12-04T12:25:16.4834063Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 FAILED [0.1419s] [ 4%] 2025-12-04T12:25:16.4834852Z 2025-12-04T12:25:16.4835006Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.4835669Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 _ 2025-12-04T12:25:16.4836319Z Traceback (most recent call last): 2025-12-04T12:25:16.4837092Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.4837964Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.4838501Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.4838855Z 2025-12-04T12:25:16.4839074Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.4840259Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.4841164Z 2025-12-04T12:25:16.4841446Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.4842081Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.4842561Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.4842899Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.4843323Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.4843812Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.4844156Z graph_break [] 2025-12-04T12:25:16.4844531Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.4845624Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.4846654Z warnings.warn( 2025-12-04T12:25:16.4847204Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 _ 2025-12-04T12:25:16.4847840Z Traceback (most recent call last): 2025-12-04T12:25:16.4848656Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.4849529Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.4850079Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.4850420Z 2025-12-04T12:25:16.4850635Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.4851798Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.4852710Z 2025-12-04T12:25:16.4852984Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.4853620Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.4854089Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.4854429Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.4854869Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.4855351Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.4855697Z graph_break [] 2025-12-04T12:25:16.4856075Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.4857262Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.4858233Z warnings.warn( 2025-12-04T12:25:16.4858633Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.4859123Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.4859451Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.4859898Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.4860388Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.4860719Z graph_break [] 2025-12-04T12:25:16.4861093Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.4862192Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.4863163Z warnings.warn( 2025-12-04T12:25:16.4863460Z =================================== FAILURES =================================== 2025-12-04T12:25:16.4864134Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 _ 2025-12-04T12:25:16.4864779Z Traceback (most recent call last): 2025-12-04T12:25:16.4865537Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.4866469Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.4867019Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.4867359Z 2025-12-04T12:25:16.4867586Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.4868701Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.4869615Z 2025-12-04T12:25:16.4869881Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.4870506Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.4870978Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.4871306Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.4871742Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.4872226Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.4872597Z graph_break [] 2025-12-04T12:25:16.4872966Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.4874091Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.4875070Z warnings.warn( 2025-12-04T12:25:16.4875443Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.4875985Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.4876327Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.4876747Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.4877231Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.4877573Z graph_break [] 2025-12-04T12:25:16.4877965Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.4879061Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.4880030Z warnings.warn( 2025-12-04T12:25:16.4880410Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.4880872Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.4881207Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.4881639Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.4882109Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.4882451Z graph_break [] 2025-12-04T12:25:16.4882822Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.4883914Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.4884875Z warnings.warn( 2025-12-04T12:25:16.4885866Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-fb2ec7b0e0c81fea.xml - 2025-12-04T12:25:16.4887005Z =========================== short test summary info ============================ 2025-12-04T12:25:16.4888243Z FAILED [0.1419s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.4889281Z 2025-12-04T12:25:16.4889498Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.4890622Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.4891539Z 2025-12-04T12:25:16.4891806Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.4892434Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.4892938Z ==================== 1 failed, 8 passed, 2 rerun in 18.15s ===================== 2025-12-04T12:25:16.4893370Z Got exit code 1 2025-12-04T12:25:16.4893642Z Retrying single test... 2025-12-04T12:25:16.4894466Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6ae6fdb299843870.xml 2025-12-04T12:25:16.4895410Z ============================= test session starts ============================== 2025-12-04T12:25:16.4896266Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.4896935Z cachedir: .pytest_cache 2025-12-04T12:25:16.4897645Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.4898437Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.4898794Z configfile: pytest.ini 2025-12-04T12:25:16.4899559Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.4900631Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.4901896Z stepcurrent: skipping 8 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.4903001Z Running 1 items in this shard 2025-12-04T12:25:16.4903214Z 2025-12-04T12:25:16.4903961Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 ('RERUN', {'yellow': True}) [4.6234s] [100%] 2025-12-04T12:25:16.4905642Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 ('RERUN', {'yellow': True}) [0.1408s] [100%] 2025-12-04T12:25:16.4907180Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 FAILED [0.1392s] [100%] 2025-12-04T12:25:16.4907978Z 2025-12-04T12:25:16.4908120Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.4908793Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 _ 2025-12-04T12:25:16.4909424Z Traceback (most recent call last): 2025-12-04T12:25:16.4910198Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.4911066Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.4911602Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.4911956Z 2025-12-04T12:25:16.4912173Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.4913296Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.4914195Z 2025-12-04T12:25:16.4914473Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.4915113Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.4915591Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.4915933Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.4916268Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.4916735Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.4917201Z graph_break [] 2025-12-04T12:25:16.4917585Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.4918676Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.4919708Z warnings.warn( 2025-12-04T12:25:16.4920271Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 _ 2025-12-04T12:25:16.4920911Z Traceback (most recent call last): 2025-12-04T12:25:16.4921692Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.4922568Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.4923121Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.4923461Z 2025-12-04T12:25:16.4923675Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.4924804Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.4925716Z 2025-12-04T12:25:16.4925985Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.4926624Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.4927125Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.4927466Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.4927809Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.4928295Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.4928761Z graph_break [] 2025-12-04T12:25:16.4929140Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.4930236Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.4931194Z warnings.warn( 2025-12-04T12:25:16.4931606Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.4932080Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.4932402Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.4932835Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.4933316Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.4933642Z graph_break [] 2025-12-04T12:25:16.4934018Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.4935109Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.4936073Z warnings.warn( 2025-12-04T12:25:16.4936371Z =================================== FAILURES =================================== 2025-12-04T12:25:16.4937128Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 _ 2025-12-04T12:25:16.4937780Z Traceback (most recent call last): 2025-12-04T12:25:16.4938547Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.4939425Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.4939975Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.4940319Z 2025-12-04T12:25:16.4940544Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.4941661Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.4942575Z 2025-12-04T12:25:16.4942841Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.4943470Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.4943947Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.4944271Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.4944615Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.4945120Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.4945572Z graph_break [] 2025-12-04T12:25:16.4945948Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.4947039Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.4948010Z warnings.warn( 2025-12-04T12:25:16.4948384Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.4948858Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.4949191Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.4949615Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.4950095Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.4950437Z graph_break [] 2025-12-04T12:25:16.4950799Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.4951888Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.4952893Z warnings.warn( 2025-12-04T12:25:16.4953272Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.4953861Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.4954198Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.4954634Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.4955102Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.4955442Z graph_break [] 2025-12-04T12:25:16.4955814Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.4956938Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.4957899Z warnings.warn( 2025-12-04T12:25:16.4958891Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6ae6fdb299843870.xml - 2025-12-04T12:25:16.4960014Z =========================== short test summary info ============================ 2025-12-04T12:25:16.4961250Z FAILED [0.1392s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.4962287Z 2025-12-04T12:25:16.4962506Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.4963634Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.4964548Z 2025-12-04T12:25:16.4964818Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.4965415Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.4965934Z ================== 1 failed, 204 deselected, 2 rerun in 4.96s ================== 2025-12-04T12:25:16.4966380Z Got exit code 1 2025-12-04T12:25:16.4966649Z Retrying single test... 2025-12-04T12:25:16.4967473Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2362ba98ed8b8900.xml 2025-12-04T12:25:16.4968427Z ============================= test session starts ============================== 2025-12-04T12:25:16.4969091Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.4969691Z cachedir: .pytest_cache 2025-12-04T12:25:16.4970390Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.4971179Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.4971574Z configfile: pytest.ini 2025-12-04T12:25:16.5011981Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.5013118Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.5014333Z stepcurrent: skipping 8 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.5015417Z Running 1 items in this shard 2025-12-04T12:25:16.5015625Z 2025-12-04T12:25:16.5016365Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 ('RERUN', {'yellow': True}) [4.6198s] [100%] 2025-12-04T12:25:16.5018050Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 ('RERUN', {'yellow': True}) [0.1415s] [100%] 2025-12-04T12:25:16.5019568Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 FAILED [0.1397s] [100%] 2025-12-04T12:25:16.5020487Z 2025-12-04T12:25:16.5020630Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.5021330Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 _ 2025-12-04T12:25:16.5021955Z Traceback (most recent call last): 2025-12-04T12:25:16.5022705Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5023550Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5024071Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5024412Z 2025-12-04T12:25:16.5024669Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5025772Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.5026670Z 2025-12-04T12:25:16.5026938Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5027551Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5028013Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5028329Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5028640Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5029079Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5029522Z graph_break [] 2025-12-04T12:25:16.5029874Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5030954Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5031904Z warnings.warn( 2025-12-04T12:25:16.5032425Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 _ 2025-12-04T12:25:16.5033045Z Traceback (most recent call last): 2025-12-04T12:25:16.5033795Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5034636Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5035159Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5035492Z 2025-12-04T12:25:16.5035698Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5036794Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.5037689Z 2025-12-04T12:25:16.5037948Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5038610Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5039062Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5039376Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5039693Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5040128Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5040566Z graph_break [] 2025-12-04T12:25:16.5040917Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5041989Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5042932Z warnings.warn( 2025-12-04T12:25:16.5043290Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5043742Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5044051Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5044463Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5044959Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5045278Z graph_break [] 2025-12-04T12:25:16.5045625Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5046722Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5047670Z warnings.warn( 2025-12-04T12:25:16.5047953Z =================================== FAILURES =================================== 2025-12-04T12:25:16.5048620Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 _ 2025-12-04T12:25:16.5049249Z Traceback (most recent call last): 2025-12-04T12:25:16.5050027Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5050887Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5051420Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5051755Z 2025-12-04T12:25:16.5051970Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5053064Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.5053963Z 2025-12-04T12:25:16.5054225Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5054838Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5055292Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5055608Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5055940Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5056398Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5056904Z graph_break [] 2025-12-04T12:25:16.5057282Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5058373Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5059344Z warnings.warn( 2025-12-04T12:25:16.5059703Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5060164Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5060488Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5060905Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5061377Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5061713Z graph_break [] 2025-12-04T12:25:16.5062071Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5063193Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5064155Z warnings.warn( 2025-12-04T12:25:16.5064529Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5064982Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5065311Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5065734Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5066203Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5066534Z graph_break [] 2025-12-04T12:25:16.5066890Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5067975Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5068929Z warnings.warn( 2025-12-04T12:25:16.5069913Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2362ba98ed8b8900.xml - 2025-12-04T12:25:16.5071065Z =========================== short test summary info ============================ 2025-12-04T12:25:16.5072320Z FAILED [0.1397s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5073359Z 2025-12-04T12:25:16.5073574Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5074692Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.5075602Z 2025-12-04T12:25:16.5075895Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5076479Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.5076987Z ================== 1 failed, 204 deselected, 2 rerun in 4.96s ================== 2025-12-04T12:25:16.5077425Z Got exit code 1 2025-12-04T12:25:16.5078271Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.5079498Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.5080667Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-52334a8157857cd5.xml 2025-12-04T12:25:16.5081602Z ============================= test session starts ============================== 2025-12-04T12:25:16.5082262Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.5082854Z cachedir: .pytest_cache 2025-12-04T12:25:16.5083565Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.5084353Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.5084706Z configfile: pytest.ini 2025-12-04T12:25:16.5085471Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.5086419Z collecting ... collected 380 items / 9 deselected / 371 selected 2025-12-04T12:25:16.5086919Z stepcurrent: skipping 9 already run items. 2025-12-04T12:25:16.5087288Z Running 196 items in this shard 2025-12-04T12:25:16.5087507Z 2025-12-04T12:25:16.5088249Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 ('RERUN', {'yellow': True}) [4.6123s] [ 0%] 2025-12-04T12:25:16.5089873Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 ('RERUN', {'yellow': True}) [0.1408s] [ 0%] 2025-12-04T12:25:16.5091439Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 FAILED [0.1403s] [ 0%] 2025-12-04T12:25:16.5092215Z 2025-12-04T12:25:16.5092366Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.5093021Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 _ 2025-12-04T12:25:16.5093664Z Traceback (most recent call last): 2025-12-04T12:25:16.5094430Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5095274Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5095818Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5096373Z 2025-12-04T12:25:16.5096587Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5097756Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 2025-12-04T12:25:16.5098726Z 2025-12-04T12:25:16.5098987Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5099673Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5100145Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5100476Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5100797Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5101252Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5101707Z graph_break [] 2025-12-04T12:25:16.5102061Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5103205Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5104181Z warnings.warn( 2025-12-04T12:25:16.5104717Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 _ 2025-12-04T12:25:16.5105344Z Traceback (most recent call last): 2025-12-04T12:25:16.5106113Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5106977Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5107501Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5107846Z 2025-12-04T12:25:16.5108053Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5109171Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 2025-12-04T12:25:16.5110072Z 2025-12-04T12:25:16.5110345Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5110954Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5111419Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5111747Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5112082Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5112528Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5112980Z graph_break [] 2025-12-04T12:25:16.5113342Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5114416Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5115384Z warnings.warn( 2025-12-04T12:25:16.5115756Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5116260Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5116578Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5117011Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5117486Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5117806Z graph_break [] 2025-12-04T12:25:16.5118176Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5119256Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5120199Z warnings.warn( 2025-12-04T12:25:16.5120501Z =================================== FAILURES =================================== 2025-12-04T12:25:16.5121171Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 _ 2025-12-04T12:25:16.5121814Z Traceback (most recent call last): 2025-12-04T12:25:16.5122567Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5123468Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5124010Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5124342Z 2025-12-04T12:25:16.5124584Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5125700Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 2025-12-04T12:25:16.5126616Z 2025-12-04T12:25:16.5126876Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5127497Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5127947Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5128303Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5128638Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5129087Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5129530Z graph_break [] 2025-12-04T12:25:16.5129890Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5130974Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5131919Z warnings.warn( 2025-12-04T12:25:16.5132289Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5132749Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5133078Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5133495Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5133966Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5134301Z graph_break [] 2025-12-04T12:25:16.5134715Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5135827Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5136789Z warnings.warn( 2025-12-04T12:25:16.5137230Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5137690Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5138018Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5138441Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5138917Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5139261Z graph_break [] 2025-12-04T12:25:16.5139637Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5140722Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5141749Z warnings.warn( 2025-12-04T12:25:16.5142741Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-52334a8157857cd5.xml - 2025-12-04T12:25:16.5143878Z =========================== short test summary info ============================ 2025-12-04T12:25:16.5145097Z FAILED [0.1403s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5146159Z 2025-12-04T12:25:16.5146377Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5147512Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 2025-12-04T12:25:16.5148416Z 2025-12-04T12:25:16.5148700Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5149284Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.5149848Z =================== 1 failed, 9 deselected, 2 rerun in 4.95s =================== 2025-12-04T12:25:16.5150294Z Got exit code 1 2025-12-04T12:25:16.5150569Z Retrying single test... 2025-12-04T12:25:16.5151429Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0b589120e3bef198.xml 2025-12-04T12:25:16.5152389Z ============================= test session starts ============================== 2025-12-04T12:25:16.5153052Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.5153638Z cachedir: .pytest_cache 2025-12-04T12:25:16.5154382Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.5155175Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.5155542Z configfile: pytest.ini 2025-12-04T12:25:16.5156313Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.5157270Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.5158465Z stepcurrent: skipping 9 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 2025-12-04T12:25:16.5159538Z Running 1 items in this shard 2025-12-04T12:25:16.5159741Z 2025-12-04T12:25:16.5160479Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 ('RERUN', {'yellow': True}) [4.6712s] [100%] 2025-12-04T12:25:16.5162087Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 ('RERUN', {'yellow': True}) [0.1424s] [100%] 2025-12-04T12:25:16.5163600Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 FAILED [0.1399s] [100%] 2025-12-04T12:25:16.5164380Z 2025-12-04T12:25:16.5164523Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.5165168Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 _ 2025-12-04T12:25:16.5165792Z Traceback (most recent call last): 2025-12-04T12:25:16.5166540Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5167384Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5167912Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5168248Z 2025-12-04T12:25:16.5168456Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5169598Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 2025-12-04T12:25:16.5170498Z 2025-12-04T12:25:16.5170772Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5171386Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5171859Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5172184Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5172503Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5172949Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5173408Z graph_break [] 2025-12-04T12:25:16.5173769Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5174860Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5175859Z warnings.warn( 2025-12-04T12:25:16.5176393Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 _ 2025-12-04T12:25:16.5177108Z Traceback (most recent call last): 2025-12-04T12:25:16.5177907Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5178774Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5179311Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5179644Z 2025-12-04T12:25:16.5179854Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5180996Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 2025-12-04T12:25:16.5181892Z 2025-12-04T12:25:16.5182161Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5182776Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5183223Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5183541Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5183861Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5184301Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5184747Z graph_break [] 2025-12-04T12:25:16.5185109Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5186185Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5187135Z warnings.warn( 2025-12-04T12:25:16.5187498Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5187947Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5188254Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5188672Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5189137Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5189456Z graph_break [] 2025-12-04T12:25:16.5189813Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5190884Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5191835Z warnings.warn( 2025-12-04T12:25:16.5192121Z =================================== FAILURES =================================== 2025-12-04T12:25:16.5192775Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 _ 2025-12-04T12:25:16.5193403Z Traceback (most recent call last): 2025-12-04T12:25:16.5194149Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5195029Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5195556Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5195890Z 2025-12-04T12:25:16.5196318Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5197429Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 2025-12-04T12:25:16.5198324Z 2025-12-04T12:25:16.5198582Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5199195Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5199654Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5199966Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5200285Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5200732Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5201269Z graph_break [] 2025-12-04T12:25:16.5201622Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5202745Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5203693Z warnings.warn( 2025-12-04T12:25:16.5204047Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5204494Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5204807Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5205216Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5205677Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5205998Z graph_break [] 2025-12-04T12:25:16.5206386Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5207459Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5208404Z warnings.warn( 2025-12-04T12:25:16.5208770Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5209222Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5209544Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5209966Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5210424Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5210754Z graph_break [] 2025-12-04T12:25:16.5211110Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5212177Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5213121Z warnings.warn( 2025-12-04T12:25:16.5214089Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0b589120e3bef198.xml - 2025-12-04T12:25:16.5215202Z =========================== short test summary info ============================ 2025-12-04T12:25:16.5216406Z FAILED [0.1399s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5217509Z 2025-12-04T12:25:16.5217720Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5218823Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 2025-12-04T12:25:16.5219715Z 2025-12-04T12:25:16.5219981Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5220604Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.5221105Z ================== 1 failed, 204 deselected, 2 rerun in 5.01s ================== 2025-12-04T12:25:16.5221527Z Got exit code 1 2025-12-04T12:25:16.5221777Z Retrying single test... 2025-12-04T12:25:16.5222588Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f9457e8d87a8d0e5.xml 2025-12-04T12:25:16.5223518Z ============================= test session starts ============================== 2025-12-04T12:25:16.5224153Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.5224731Z cachedir: .pytest_cache 2025-12-04T12:25:16.5225417Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.5226180Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.5226515Z configfile: pytest.ini 2025-12-04T12:25:16.5227263Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.5228227Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.5229446Z stepcurrent: skipping 9 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 2025-12-04T12:25:16.5230518Z Running 1 items in this shard 2025-12-04T12:25:16.5230720Z 2025-12-04T12:25:16.5231454Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 ('RERUN', {'yellow': True}) [4.6262s] [100%] 2025-12-04T12:25:16.5233077Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 ('RERUN', {'yellow': True}) [0.1413s] [100%] 2025-12-04T12:25:16.5234587Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 FAILED [0.1372s] [100%] 2025-12-04T12:25:16.5235367Z 2025-12-04T12:25:16.5235506Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.5236157Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 _ 2025-12-04T12:25:16.5236777Z Traceback (most recent call last): 2025-12-04T12:25:16.5237528Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5238379Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5238901Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5239240Z 2025-12-04T12:25:16.5239449Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5240552Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 2025-12-04T12:25:16.5241449Z 2025-12-04T12:25:16.5241713Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5242315Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5242770Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5243086Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5243406Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5243843Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5244286Z graph_break [] 2025-12-04T12:25:16.5244638Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5245712Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5246693Z warnings.warn( 2025-12-04T12:25:16.5247223Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 _ 2025-12-04T12:25:16.5247847Z Traceback (most recent call last): 2025-12-04T12:25:16.5248601Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5249446Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5249970Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5250300Z 2025-12-04T12:25:16.5250505Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5251598Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 2025-12-04T12:25:16.5252497Z 2025-12-04T12:25:16.5252757Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5253368Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5253850Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5254162Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5254478Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5254942Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5255381Z graph_break [] 2025-12-04T12:25:16.5255734Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5256803Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5257814Z warnings.warn( 2025-12-04T12:25:16.5258208Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5258658Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5258968Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5259378Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5259839Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5260160Z graph_break [] 2025-12-04T12:25:16.5260505Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5261574Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5262523Z warnings.warn( 2025-12-04T12:25:16.5262806Z =================================== FAILURES =================================== 2025-12-04T12:25:16.5263459Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 _ 2025-12-04T12:25:16.5264086Z Traceback (most recent call last): 2025-12-04T12:25:16.5264839Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5265681Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5266203Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5266538Z 2025-12-04T12:25:16.5266748Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5267852Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 2025-12-04T12:25:16.5268753Z 2025-12-04T12:25:16.5269013Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5269624Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5270090Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5270417Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5270764Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5271263Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5271717Z graph_break [] 2025-12-04T12:25:16.5272097Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5273199Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5274178Z warnings.warn( 2025-12-04T12:25:16.5274555Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5275037Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5275380Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5275813Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5276302Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5276651Z graph_break [] 2025-12-04T12:25:16.5277030Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5278121Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5279127Z warnings.warn( 2025-12-04T12:25:16.5279516Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5280014Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5280361Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5280806Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5281277Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5281626Z graph_break [] 2025-12-04T12:25:16.5281999Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5283119Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5284079Z warnings.warn( 2025-12-04T12:25:16.5285069Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f9457e8d87a8d0e5.xml - 2025-12-04T12:25:16.5286201Z =========================== short test summary info ============================ 2025-12-04T12:25:16.5287440Z FAILED [0.1372s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5288480Z 2025-12-04T12:25:16.5288696Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5289819Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 2025-12-04T12:25:16.5290729Z 2025-12-04T12:25:16.5290996Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5291591Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.5292105Z ================== 1 failed, 204 deselected, 2 rerun in 4.96s ================== 2025-12-04T12:25:16.5292544Z Got exit code 1 2025-12-04T12:25:16.5293401Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1 2025-12-04T12:25:16.5294641Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.5296181Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e4f574242b541525.xml 2025-12-04T12:25:16.5297222Z ============================= test session starts ============================== 2025-12-04T12:25:16.5297892Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.5298611Z cachedir: .pytest_cache 2025-12-04T12:25:16.5299318Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.5300115Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.5300468Z configfile: pytest.ini 2025-12-04T12:25:16.5301233Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.5302185Z collecting ... collected 380 items / 10 deselected / 370 selected 2025-12-04T12:25:16.5302687Z stepcurrent: skipping 10 already run items. 2025-12-04T12:25:16.5303074Z Running 195 items in this shard 2025-12-04T12:25:16.5303287Z 2025-12-04T12:25:16.5304039Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 ('RERUN', {'yellow': True}) [4.6374s] [ 0%] 2025-12-04T12:25:16.5305661Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 ('RERUN', {'yellow': True}) [0.1426s] [ 0%] 2025-12-04T12:25:16.5307473Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 FAILED [0.1393s] [ 0%] 2025-12-04T12:25:16.5308259Z 2025-12-04T12:25:16.5308475Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.5309134Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 _ 2025-12-04T12:25:16.5309778Z Traceback (most recent call last): 2025-12-04T12:25:16.5310553Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5311430Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5312009Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5312365Z 2025-12-04T12:25:16.5312582Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5313707Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.5314608Z 2025-12-04T12:25:16.5314888Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5315509Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5315990Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5316330Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5316659Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5317128Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5317589Z graph_break [] 2025-12-04T12:25:16.5317951Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5319052Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5320030Z warnings.warn( 2025-12-04T12:25:16.5320572Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 _ 2025-12-04T12:25:16.5321200Z Traceback (most recent call last): 2025-12-04T12:25:16.5321971Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5322833Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5323369Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5323722Z 2025-12-04T12:25:16.5323934Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5325050Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.5325981Z 2025-12-04T12:25:16.5326259Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5326889Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5327354Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5327692Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5328032Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5328490Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5328955Z graph_break [] 2025-12-04T12:25:16.5329332Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5330420Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5331389Z warnings.warn( 2025-12-04T12:25:16.5331776Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5332258Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5332617Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5333053Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5333540Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5333869Z graph_break [] 2025-12-04T12:25:16.5334273Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5335364Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5336337Z warnings.warn( 2025-12-04T12:25:16.5336639Z =================================== FAILURES =================================== 2025-12-04T12:25:16.5337405Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 _ 2025-12-04T12:25:16.5339093Z Traceback (most recent call last): 2025-12-04T12:25:16.5339877Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5340759Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5341312Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5341656Z 2025-12-04T12:25:16.5341890Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5343001Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.5343917Z 2025-12-04T12:25:16.5344186Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5344826Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5345312Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5345639Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5345988Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5346497Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5346950Z graph_break [] 2025-12-04T12:25:16.5347331Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5348437Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5349404Z warnings.warn( 2025-12-04T12:25:16.5349791Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5350271Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5350614Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5351042Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5351533Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5351877Z graph_break [] 2025-12-04T12:25:16.5352276Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5353370Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5354340Z warnings.warn( 2025-12-04T12:25:16.5354723Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5355181Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5355515Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5355949Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5356420Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5356764Z graph_break [] 2025-12-04T12:25:16.5357136Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5358215Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5359189Z warnings.warn( 2025-12-04T12:25:16.5360207Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e4f574242b541525.xml - 2025-12-04T12:25:16.5361366Z =========================== short test summary info ============================ 2025-12-04T12:25:16.5362585Z FAILED [0.1393s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5363627Z 2025-12-04T12:25:16.5363845Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5364991Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.5365889Z 2025-12-04T12:25:16.5366170Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5366764Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.5367274Z ================== 1 failed, 10 deselected, 2 rerun in 4.97s =================== 2025-12-04T12:25:16.5367717Z Got exit code 1 2025-12-04T12:25:16.5367989Z Retrying single test... 2025-12-04T12:25:16.5368813Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-19baa3ae62072b2f.xml 2025-12-04T12:25:16.5369763Z ============================= test session starts ============================== 2025-12-04T12:25:16.5370423Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.5371023Z cachedir: .pytest_cache 2025-12-04T12:25:16.5371727Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.5372519Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.5372872Z configfile: pytest.ini 2025-12-04T12:25:16.5373636Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.5374590Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.5375811Z stepcurrent: skipping 10 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.5376975Z Running 1 items in this shard 2025-12-04T12:25:16.5377193Z 2025-12-04T12:25:16.5377935Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 ('RERUN', {'yellow': True}) [4.6492s] [100%] 2025-12-04T12:25:16.5379556Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 ('RERUN', {'yellow': True}) [0.1439s] [100%] 2025-12-04T12:25:16.5381138Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 FAILED [0.1412s] [100%] 2025-12-04T12:25:16.5381922Z 2025-12-04T12:25:16.5382081Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.5382736Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 _ 2025-12-04T12:25:16.5383381Z Traceback (most recent call last): 2025-12-04T12:25:16.5384159Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5385027Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5385562Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5385916Z 2025-12-04T12:25:16.5386129Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5387246Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.5388181Z 2025-12-04T12:25:16.5388463Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5389105Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5389581Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5389917Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5390243Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5390705Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5391166Z graph_break [] 2025-12-04T12:25:16.5391540Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5392675Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5393655Z warnings.warn( 2025-12-04T12:25:16.5394200Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 _ 2025-12-04T12:25:16.5394828Z Traceback (most recent call last): 2025-12-04T12:25:16.5395609Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5396662Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5397210Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5397554Z 2025-12-04T12:25:16.5397767Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5398895Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.5399807Z 2025-12-04T12:25:16.5400073Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5400706Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5401167Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5401501Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5401843Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5402296Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5402756Z graph_break [] 2025-12-04T12:25:16.5403129Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5404227Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5405191Z warnings.warn( 2025-12-04T12:25:16.5405581Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5406130Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5406459Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5406902Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5407391Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5407723Z graph_break [] 2025-12-04T12:25:16.5408108Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5409201Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5410179Z warnings.warn( 2025-12-04T12:25:16.5410482Z =================================== FAILURES =================================== 2025-12-04T12:25:16.5411162Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 _ 2025-12-04T12:25:16.5411817Z Traceback (most recent call last): 2025-12-04T12:25:16.5412588Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5413507Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5414055Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5414395Z 2025-12-04T12:25:16.5414662Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5415770Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.5416679Z 2025-12-04T12:25:16.5417003Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5417637Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5418114Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5418491Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5418832Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5419298Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5419748Z graph_break [] 2025-12-04T12:25:16.5420122Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5421225Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5422200Z warnings.warn( 2025-12-04T12:25:16.5422573Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5423048Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5423409Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5423834Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5424316Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5424661Z graph_break [] 2025-12-04T12:25:16.5425024Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5426116Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5427083Z warnings.warn( 2025-12-04T12:25:16.5427464Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5427928Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5428261Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5428695Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5429165Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5429507Z graph_break [] 2025-12-04T12:25:16.5429880Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5430961Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5431954Z warnings.warn( 2025-12-04T12:25:16.5432947Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-19baa3ae62072b2f.xml - 2025-12-04T12:25:16.5434087Z =========================== short test summary info ============================ 2025-12-04T12:25:16.5435318Z FAILED [0.1412s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5436353Z 2025-12-04T12:25:16.5436570Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5437695Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.5438606Z 2025-12-04T12:25:16.5438873Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5439468Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.5440014Z ================== 1 failed, 204 deselected, 2 rerun in 4.99s ================== 2025-12-04T12:25:16.5440458Z Got exit code 1 2025-12-04T12:25:16.5440729Z Retrying single test... 2025-12-04T12:25:16.5441586Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-debf0ef50304edda.xml 2025-12-04T12:25:16.5442537Z ============================= test session starts ============================== 2025-12-04T12:25:16.5443199Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.5443803Z cachedir: .pytest_cache 2025-12-04T12:25:16.5444537Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.5445325Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.5445682Z configfile: pytest.ini 2025-12-04T12:25:16.5446446Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.5447525Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.5448749Z stepcurrent: skipping 10 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.5449845Z Running 1 items in this shard 2025-12-04T12:25:16.5450057Z 2025-12-04T12:25:16.5450795Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 ('RERUN', {'yellow': True}) [4.6321s] [100%] 2025-12-04T12:25:16.5452413Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 ('RERUN', {'yellow': True}) [0.1436s] [100%] 2025-12-04T12:25:16.5453929Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 FAILED [0.1414s] [100%] 2025-12-04T12:25:16.5454715Z 2025-12-04T12:25:16.5454872Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.5455540Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 _ 2025-12-04T12:25:16.5456174Z Traceback (most recent call last): 2025-12-04T12:25:16.5457007Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5457934Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5458531Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5458887Z 2025-12-04T12:25:16.5459104Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5460284Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.5461184Z 2025-12-04T12:25:16.5461468Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5462092Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5462571Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5462915Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5463256Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5463707Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5464166Z graph_break [] 2025-12-04T12:25:16.5464540Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5465632Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5466610Z warnings.warn( 2025-12-04T12:25:16.5467195Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 _ 2025-12-04T12:25:16.5467896Z Traceback (most recent call last): 2025-12-04T12:25:16.5468688Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5469562Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5470109Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5470449Z 2025-12-04T12:25:16.5470662Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5471822Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.5472737Z 2025-12-04T12:25:16.5473004Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5473639Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5474112Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5474454Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5474798Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5475255Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5475719Z graph_break [] 2025-12-04T12:25:16.5476099Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5477201Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5478163Z warnings.warn( 2025-12-04T12:25:16.5478557Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5479041Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5479369Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5479807Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5480297Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5480645Z graph_break [] 2025-12-04T12:25:16.5481007Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5482107Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5483082Z warnings.warn( 2025-12-04T12:25:16.5483381Z =================================== FAILURES =================================== 2025-12-04T12:25:16.5484048Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 _ 2025-12-04T12:25:16.5484686Z Traceback (most recent call last): 2025-12-04T12:25:16.5485459Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5486355Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5486908Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5487249Z 2025-12-04T12:25:16.5487474Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5488590Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.5489482Z 2025-12-04T12:25:16.5489749Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5490377Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5490856Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5491182Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5491524Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5491989Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5492490Z graph_break [] 2025-12-04T12:25:16.5492853Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5494007Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5494984Z warnings.warn( 2025-12-04T12:25:16.5495354Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5495825Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5496331Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5496758Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5497305Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5497647Z graph_break [] 2025-12-04T12:25:16.5498093Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5499191Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5500170Z warnings.warn( 2025-12-04T12:25:16.5500558Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5501024Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5501368Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5501809Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5502295Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5502623Z graph_break [] 2025-12-04T12:25:16.5502995Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5504087Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5505043Z warnings.warn( 2025-12-04T12:25:16.5506033Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-debf0ef50304edda.xml - 2025-12-04T12:25:16.5507167Z =========================== short test summary info ============================ 2025-12-04T12:25:16.5508397Z FAILED [0.1414s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5509437Z 2025-12-04T12:25:16.5509653Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5510777Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.5511690Z 2025-12-04T12:25:16.5511959Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5512598Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.5513195Z ================== 1 failed, 204 deselected, 2 rerun in 4.97s ================== 2025-12-04T12:25:16.5513646Z Got exit code 1 2025-12-04T12:25:16.5514496Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.5515729Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.5516917Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a0af7423e85581ef.xml 2025-12-04T12:25:16.5517858Z ============================= test session starts ============================== 2025-12-04T12:25:16.5518515Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.5519119Z cachedir: .pytest_cache 2025-12-04T12:25:16.5519815Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.5520644Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.5520994Z configfile: pytest.ini 2025-12-04T12:25:16.5521801Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.5522754Z collecting ... collected 380 items / 11 deselected / 369 selected 2025-12-04T12:25:16.5523259Z stepcurrent: skipping 11 already run items. 2025-12-04T12:25:16.5523644Z Running 194 items in this shard 2025-12-04T12:25:16.5523856Z 2025-12-04T12:25:16.5524596Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 ('RERUN', {'yellow': True}) [4.6128s] [ 0%] 2025-12-04T12:25:16.5526260Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 ('RERUN', {'yellow': True}) [0.1437s] [ 0%] 2025-12-04T12:25:16.5527797Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 FAILED [0.1408s] [ 0%] 2025-12-04T12:25:16.5528582Z 2025-12-04T12:25:16.5528742Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.5529395Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 _ 2025-12-04T12:25:16.5530042Z Traceback (most recent call last): 2025-12-04T12:25:16.5530818Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5531686Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5532225Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5532575Z 2025-12-04T12:25:16.5532790Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5533911Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.5534808Z 2025-12-04T12:25:16.5535089Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5535704Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5536181Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5536521Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5536904Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5537371Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5537833Z graph_break [] 2025-12-04T12:25:16.5538196Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5539298Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5540311Z warnings.warn( 2025-12-04T12:25:16.5540847Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 _ 2025-12-04T12:25:16.5541473Z Traceback (most recent call last): 2025-12-04T12:25:16.5542247Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5543112Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5543658Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5543996Z 2025-12-04T12:25:16.5544207Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5545319Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.5546222Z 2025-12-04T12:25:16.5546497Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5547159Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5547613Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5547948Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5548318Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5548772Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5549231Z graph_break [] 2025-12-04T12:25:16.5549595Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5550678Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5551650Z warnings.warn( 2025-12-04T12:25:16.5552055Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5552523Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5552846Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5553272Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5553749Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5554071Z graph_break [] 2025-12-04T12:25:16.5554436Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5555521Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5556484Z warnings.warn( 2025-12-04T12:25:16.5556776Z =================================== FAILURES =================================== 2025-12-04T12:25:16.5557438Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 _ 2025-12-04T12:25:16.5558081Z Traceback (most recent call last): 2025-12-04T12:25:16.5558831Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5559694Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5560230Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5560565Z 2025-12-04T12:25:16.5560789Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5561889Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.5562791Z 2025-12-04T12:25:16.5563055Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5563673Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5564145Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5564465Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5564834Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5565291Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5565739Z graph_break [] 2025-12-04T12:25:16.5566107Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5567200Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5568164Z warnings.warn( 2025-12-04T12:25:16.5568526Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5568647Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5568757Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5568975Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5569104Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5569200Z graph_break [] 2025-12-04T12:25:16.5569413Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5570183Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5570281Z warnings.warn( 2025-12-04T12:25:16.5570533Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5570639Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5570748Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5570978Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5571099Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5571191Z graph_break [] 2025-12-04T12:25:16.5571414Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5572169Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5572281Z warnings.warn( 2025-12-04T12:25:16.5573102Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a0af7423e85581ef.xml - 2025-12-04T12:25:16.5573274Z =========================== short test summary info ============================ 2025-12-04T12:25:16.5574197Z FAILED [0.1408s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5574203Z 2025-12-04T12:25:16.5574419Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5575191Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.5575196Z 2025-12-04T12:25:16.5575462Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5575641Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.5575845Z ================== 1 failed, 11 deselected, 2 rerun in 4.95s =================== 2025-12-04T12:25:16.5575943Z Got exit code 1 2025-12-04T12:25:16.5576049Z Retrying single test... 2025-12-04T12:25:16.5576706Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6e37efe6d0d7b5f7.xml 2025-12-04T12:25:16.5576938Z ============================= test session starts ============================== 2025-12-04T12:25:16.5577301Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.5577411Z cachedir: .pytest_cache 2025-12-04T12:25:16.5577936Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.5578105Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.5578213Z configfile: pytest.ini 2025-12-04T12:25:16.5578817Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.5579041Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.5579895Z stepcurrent: skipping 11 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.5580014Z Running 1 items in this shard 2025-12-04T12:25:16.5580019Z 2025-12-04T12:25:16.5580755Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 ('RERUN', {'yellow': True}) [4.6146s] [100%] 2025-12-04T12:25:16.5581501Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 ('RERUN', {'yellow': True}) [0.1422s] [100%] 2025-12-04T12:25:16.5582189Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 FAILED [0.1396s] [100%] 2025-12-04T12:25:16.5582194Z 2025-12-04T12:25:16.5582365Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.5582750Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 _ 2025-12-04T12:25:16.5582873Z Traceback (most recent call last): 2025-12-04T12:25:16.5583416Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5583614Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5583822Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5583857Z 2025-12-04T12:25:16.5584081Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5584847Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.5584854Z 2025-12-04T12:25:16.5585131Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5585347Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5585458Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5585583Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5585701Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5585919Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5586033Z graph_break [] 2025-12-04T12:25:16.5586246Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5586993Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5587098Z warnings.warn( 2025-12-04T12:25:16.5587466Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 _ 2025-12-04T12:25:16.5587597Z Traceback (most recent call last): 2025-12-04T12:25:16.5588123Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5588317Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5588542Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5588547Z 2025-12-04T12:25:16.5588758Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5589528Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.5589565Z 2025-12-04T12:25:16.5589829Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5590043Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5590164Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5590276Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5590405Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5601969Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5602163Z graph_break [] 2025-12-04T12:25:16.5602416Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5603169Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5603267Z warnings.warn( 2025-12-04T12:25:16.5603505Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5603619Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5603730Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5604105Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5604223Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5604329Z graph_break [] 2025-12-04T12:25:16.5604590Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5605324Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5605429Z warnings.warn( 2025-12-04T12:25:16.5605571Z =================================== FAILURES =================================== 2025-12-04T12:25:16.5605943Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 _ 2025-12-04T12:25:16.5606135Z Traceback (most recent call last): 2025-12-04T12:25:16.5606665Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5606876Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5607086Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5607095Z 2025-12-04T12:25:16.5607305Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5608067Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.5608074Z 2025-12-04T12:25:16.5608335Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5608555Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5608663Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5608774Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5608897Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5609116Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5609212Z graph_break [] 2025-12-04T12:25:16.5609430Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5610166Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5610272Z warnings.warn( 2025-12-04T12:25:16.5610482Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5610587Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5610707Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5610924Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5611041Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5611144Z graph_break [] 2025-12-04T12:25:16.5611349Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5612127Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5612227Z warnings.warn( 2025-12-04T12:25:16.5612432Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5615361Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5615473Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5615689Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5615809Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5615898Z graph_break [] 2025-12-04T12:25:16.5616105Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5616833Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5617036Z warnings.warn( 2025-12-04T12:25:16.5617861Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6e37efe6d0d7b5f7.xml - 2025-12-04T12:25:16.5618091Z =========================== short test summary info ============================ 2025-12-04T12:25:16.5619030Z FAILED [0.1396s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5619046Z 2025-12-04T12:25:16.5619259Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5620021Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.5620059Z 2025-12-04T12:25:16.5620331Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5620509Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.5620712Z ================== 1 failed, 204 deselected, 2 rerun in 4.95s ================== 2025-12-04T12:25:16.5620808Z Got exit code 1 2025-12-04T12:25:16.5620911Z Retrying single test... 2025-12-04T12:25:16.5621562Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-64b9c8b908c45f51.xml 2025-12-04T12:25:16.5621719Z ============================= test session starts ============================== 2025-12-04T12:25:16.5622066Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.5622179Z cachedir: .pytest_cache 2025-12-04T12:25:16.5622695Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.5622823Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.5622931Z configfile: pytest.ini 2025-12-04T12:25:16.5623523Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.5623752Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.5624597Z stepcurrent: skipping 11 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.5624711Z Running 1 items in this shard 2025-12-04T12:25:16.5624723Z 2025-12-04T12:25:16.5625456Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 ('RERUN', {'yellow': True}) [4.6283s] [100%] 2025-12-04T12:25:16.5626183Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 ('RERUN', {'yellow': True}) [0.1428s] [100%] 2025-12-04T12:25:16.5626864Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 FAILED [0.1400s] [100%] 2025-12-04T12:25:16.5626872Z 2025-12-04T12:25:16.5627009Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.5627390Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 _ 2025-12-04T12:25:16.5627508Z Traceback (most recent call last): 2025-12-04T12:25:16.5628038Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5628246Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5628452Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5628458Z 2025-12-04T12:25:16.5628676Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5629434Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.5629473Z 2025-12-04T12:25:16.5629734Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5629989Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5630099Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5630206Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5630334Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5630549Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5630652Z graph_break [] 2025-12-04T12:25:16.5630864Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5631627Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5631737Z warnings.warn( 2025-12-04T12:25:16.5632109Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 _ 2025-12-04T12:25:16.5632226Z Traceback (most recent call last): 2025-12-04T12:25:16.5632767Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5632957Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5633169Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5633175Z 2025-12-04T12:25:16.5633386Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5634149Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.5634154Z 2025-12-04T12:25:16.5634420Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5634633Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5634752Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5634862Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5634976Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5635206Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5635300Z graph_break [] 2025-12-04T12:25:16.5635510Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5636247Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5636344Z warnings.warn( 2025-12-04T12:25:16.5636564Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5636669Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5636821Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5637045Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5637163Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5637255Z graph_break [] 2025-12-04T12:25:16.5637473Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5638202Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5638308Z warnings.warn( 2025-12-04T12:25:16.5638449Z =================================== FAILURES =================================== 2025-12-04T12:25:16.5638821Z _ MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 _ 2025-12-04T12:25:16.5638946Z Traceback (most recent call last): 2025-12-04T12:25:16.5639475Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.5639673Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5639917Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5639922Z 2025-12-04T12:25:16.5640129Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5640936Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.5640941Z 2025-12-04T12:25:16.5641205Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5641418Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5641534Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5641648Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5641803Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5642023Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5642123Z graph_break [] 2025-12-04T12:25:16.5642344Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5643068Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5643167Z warnings.warn( 2025-12-04T12:25:16.5643385Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5643491Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5643601Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5643825Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5643940Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5644045Z graph_break [] 2025-12-04T12:25:16.5644257Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5644980Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5645086Z warnings.warn( 2025-12-04T12:25:16.5645294Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5645399Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5645517Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.5645733Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5645857Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5645949Z graph_break [] 2025-12-04T12:25:16.5646159Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5646882Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5646976Z warnings.warn( 2025-12-04T12:25:16.5647831Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-64b9c8b908c45f51.xml - 2025-12-04T12:25:16.5648009Z =========================== short test summary info ============================ 2025-12-04T12:25:16.5648910Z FAILED [0.1400s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5648916Z 2025-12-04T12:25:16.5649138Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5649900Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.5649905Z 2025-12-04T12:25:16.5650175Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5650352Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.5650578Z ================== 1 failed, 204 deselected, 2 rerun in 4.97s ================== 2025-12-04T12:25:16.5650685Z Got exit code 1 2025-12-04T12:25:16.5651394Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.5651801Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.5652442Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3343e9ff31c5255a.xml 2025-12-04T12:25:16.5652604Z ============================= test session starts ============================== 2025-12-04T12:25:16.5652986Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.5653099Z cachedir: .pytest_cache 2025-12-04T12:25:16.5653621Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.5653753Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.5653860Z configfile: pytest.ini 2025-12-04T12:25:16.5654455Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.5654677Z collecting ... collected 380 items / 12 deselected / 368 selected 2025-12-04T12:25:16.5654821Z stepcurrent: skipping 12 already run items. 2025-12-04T12:25:16.5654942Z Running 193 items in this shard 2025-12-04T12:25:16.5654947Z 2025-12-04T12:25:16.5655524Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_dynamic_shape_dynamic_dims0 PASSED [6.8098s] [ 0%] 2025-12-04T12:25:16.5656107Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_dynamic_shape_dynamic_dims1 PASSED [2.2328s] [ 1%] 2025-12-04T12:25:16.5656793Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_mean_swap_False_split_reductions_False_shape1 PASSED [0.8517s] [ 1%] 2025-12-04T12:25:16.5657571Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_mean_swap_False_split_reductions_False_shape2 PASSED [0.8490s] [ 2%] 2025-12-04T12:25:16.5658259Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_mean_swap_False_split_reductions_True_shape2 PASSED [0.9018s] [ 2%] 2025-12-04T12:25:16.5658932Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_mean_swap_True_split_reductions_False_shape1 PASSED [0.8304s] [ 3%] 2025-12-04T12:25:16.5659617Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_mean_swap_True_split_reductions_False_shape2 PASSED [0.8450s] [ 3%] 2025-12-04T12:25:16.5660323Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_mean_swap_True_split_reductions_True_shape0 PASSED [0.8901s] [ 4%] 2025-12-04T12:25:16.5661000Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_mean_swap_True_split_reductions_True_shape1 PASSED [0.8910s] [ 4%] 2025-12-04T12:25:16.5661692Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_prod_swap_False_split_reductions_False_shape1 PASSED [1.9286s] [ 5%] 2025-12-04T12:25:16.5662366Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_prod_swap_False_split_reductions_True_shape1 PASSED [0.4436s] [ 5%] 2025-12-04T12:25:16.5663054Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_prod_swap_True_split_reductions_False_shape0 PASSED [0.4281s] [ 6%] 2025-12-04T12:25:16.5663722Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_prod_swap_True_split_reductions_True_shape0 PASSED [0.4465s] [ 6%] 2025-12-04T12:25:16.5664437Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_prod_swap_True_split_reductions_True_shape1 PASSED [0.4524s] [ 7%] 2025-12-04T12:25:16.5665251Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_prod_swap_True_split_reductions_True_shape2 SKIPPED [0.0032s] (Invalid combination) [ 7%] 2025-12-04T12:25:16.5665926Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_sum_swap_False_split_reductions_False_shape2 PASSED [0.4805s] [ 8%] 2025-12-04T12:25:16.5666604Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_sum_swap_False_split_reductions_True_shape1 PASSED [0.4996s] [ 8%] 2025-12-04T12:25:16.5667295Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_sum_swap_True_split_reductions_False_shape0 PASSED [0.4728s] [ 9%] 2025-12-04T12:25:16.5667974Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_sum_swap_True_split_reductions_False_shape1 PASSED [0.7060s] [ 9%] 2025-12-04T12:25:16.5668651Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_sum_swap_True_split_reductions_False_shape2 PASSED [0.4664s] [ 10%] 2025-12-04T12:25:16.5669311Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_sum_swap_True_split_reductions_True_shape1 PASSED [0.4883s] [ 10%] 2025-12-04T12:25:16.5669987Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_mix_order_reduction_name_sum_swap_True_split_reductions_True_shape2 PASSED [0.4853s] [ 11%] 2025-12-04T12:25:16.5670903Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [0.2356s] [ 11%] 2025-12-04T12:25:16.5671822Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [0.1681s] [ 11%] 2025-12-04T12:25:16.5672652Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True FAILED [0.1638s] [ 11%] 2025-12-04T12:25:16.5672659Z 2025-12-04T12:25:16.5672803Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.5673357Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.5673476Z Traceback (most recent call last): 2025-12-04T12:25:16.5673952Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5674146Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5674392Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5674400Z 2025-12-04T12:25:16.5674612Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5675567Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.5675573Z 2025-12-04T12:25:16.5675843Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5676055Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5676173Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5676284Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5676500Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5676843Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5676939Z graph_break [] 2025-12-04T12:25:16.5677183Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5679942Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5680048Z return x.grad, w.grad 2025-12-04T12:25:16.5680823Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5680924Z warnings.warn( 2025-12-04T12:25:16.5683651Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5683754Z return x.grad, w.grad 2025-12-04T12:25:16.5684323Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.5684444Z Traceback (most recent call last): 2025-12-04T12:25:16.5684914Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5685119Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5685329Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5685335Z 2025-12-04T12:25:16.5685559Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5686502Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.5686507Z 2025-12-04T12:25:16.5686770Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5686994Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5687107Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5687248Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5687470Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5687807Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5687909Z graph_break [] 2025-12-04T12:25:16.5688123Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5690812Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5690926Z return x.grad, w.grad 2025-12-04T12:25:16.5691654Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5691790Z warnings.warn( 2025-12-04T12:25:16.5694505Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5694649Z return x.grad, w.grad 2025-12-04T12:25:16.5694865Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5694977Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5695097Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5695317Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5695666Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5695759Z graph_break [] 2025-12-04T12:25:16.5696154Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5698920Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5699033Z return x.grad, w.grad 2025-12-04T12:25:16.5699772Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5699871Z warnings.warn( 2025-12-04T12:25:16.5702581Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5702754Z return x.grad, w.grad 2025-12-04T12:25:16.5702895Z =================================== FAILURES =================================== 2025-12-04T12:25:16.5703468Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.5703589Z Traceback (most recent call last): 2025-12-04T12:25:16.5704067Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5704264Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5704475Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5704481Z 2025-12-04T12:25:16.5704701Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5705647Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.5705694Z 2025-12-04T12:25:16.5705971Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5706184Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5706334Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5706457Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5706675Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5707012Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5707120Z graph_break [] 2025-12-04T12:25:16.5707335Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5710099Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5710206Z return x.grad, w.grad 2025-12-04T12:25:16.5710946Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5711043Z warnings.warn( 2025-12-04T12:25:16.5713736Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5713855Z return x.grad, w.grad 2025-12-04T12:25:16.5714068Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5714185Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5714298Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5714513Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5714862Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5714957Z graph_break [] 2025-12-04T12:25:16.5715182Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5717966Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5718129Z return x.grad, w.grad 2025-12-04T12:25:16.5718957Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5719057Z warnings.warn( 2025-12-04T12:25:16.5721790Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5721928Z return x.grad, w.grad 2025-12-04T12:25:16.5722151Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5722256Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5722367Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5722626Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5722965Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5723070Z graph_break [] 2025-12-04T12:25:16.5723287Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5724019Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5724125Z warnings.warn( 2025-12-04T12:25:16.5726824Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5726938Z return x.grad, w.grad 2025-12-04T12:25:16.5727754Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3343e9ff31c5255a.xml - 2025-12-04T12:25:16.5727924Z =========================== short test summary info ============================ 2025-12-04T12:25:16.5729017Z FAILED [0.1638s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5729023Z 2025-12-04T12:25:16.5729235Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5730189Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.5730226Z 2025-12-04T12:25:16.5730494Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5730680Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.5730917Z ======= 1 failed, 21 passed, 1 skipped, 12 deselected, 2 rerun in 23.07s ======= 2025-12-04T12:25:16.5731015Z Got exit code 1 2025-12-04T12:25:16.5731130Z Retrying single test... 2025-12-04T12:25:16.5731767Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ae334ed847078f28.xml 2025-12-04T12:25:16.5731927Z ============================= test session starts ============================== 2025-12-04T12:25:16.5732280Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.5732387Z cachedir: .pytest_cache 2025-12-04T12:25:16.5732913Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.5733069Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.5733173Z configfile: pytest.ini 2025-12-04T12:25:16.5733774Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.5734042Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.5735089Z stepcurrent: skipping 34 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.5735197Z Running 1 items in this shard 2025-12-04T12:25:16.5735202Z 2025-12-04T12:25:16.5736145Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [4.6643s] [100%] 2025-12-04T12:25:16.5737140Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [0.1715s] [100%] 2025-12-04T12:25:16.5737981Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True FAILED [0.1658s] [100%] 2025-12-04T12:25:16.5737987Z 2025-12-04T12:25:16.5738134Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.5738696Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.5738814Z Traceback (most recent call last): 2025-12-04T12:25:16.5739293Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5739489Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5739706Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5739711Z 2025-12-04T12:25:16.5739924Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5740873Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.5740889Z 2025-12-04T12:25:16.5741153Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5741370Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5741490Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5741602Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5741945Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5742208Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5742306Z graph_break [] 2025-12-04T12:25:16.5742517Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5745220Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5745326Z return x.grad, w.grad 2025-12-04T12:25:16.5746066Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5746199Z warnings.warn( 2025-12-04T12:25:16.5748918Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5749024Z return x.grad, w.grad 2025-12-04T12:25:16.5749617Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.5749755Z Traceback (most recent call last): 2025-12-04T12:25:16.5750218Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5750413Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5750633Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5750641Z 2025-12-04T12:25:16.5750853Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5751802Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.5751808Z 2025-12-04T12:25:16.5752076Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5752293Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5752413Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5752525Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5752876Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5753093Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5753188Z graph_break [] 2025-12-04T12:25:16.5753413Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5756126Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5756271Z return x.grad, w.grad 2025-12-04T12:25:16.5756999Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5757097Z warnings.warn( 2025-12-04T12:25:16.5759790Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5759897Z return x.grad, w.grad 2025-12-04T12:25:16.5760123Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5760263Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5760374Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5760601Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5760968Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5761075Z graph_break [] 2025-12-04T12:25:16.5761287Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5764009Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5764125Z return x.grad, w.grad 2025-12-04T12:25:16.5764853Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5764963Z warnings.warn( 2025-12-04T12:25:16.5767647Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5767763Z return x.grad, w.grad 2025-12-04T12:25:16.5767908Z =================================== FAILURES =================================== 2025-12-04T12:25:16.5768470Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.5768598Z Traceback (most recent call last): 2025-12-04T12:25:16.5769063Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5769270Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5769481Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5769486Z 2025-12-04T12:25:16.5769697Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5770686Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.5770694Z 2025-12-04T12:25:16.5770957Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5771181Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5771288Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5771399Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5771748Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5771963Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5772058Z graph_break [] 2025-12-04T12:25:16.5772280Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5775023Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5775166Z return x.grad, w.grad 2025-12-04T12:25:16.5775891Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5775998Z warnings.warn( 2025-12-04T12:25:16.5778792Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5778912Z return x.grad, w.grad 2025-12-04T12:25:16.5779123Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5779230Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5779354Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5779569Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5779907Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5780008Z graph_break [] 2025-12-04T12:25:16.5780220Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5782922Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5783023Z return x.grad, w.grad 2025-12-04T12:25:16.5783760Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5783858Z warnings.warn( 2025-12-04T12:25:16.5786573Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5786682Z return x.grad, w.grad 2025-12-04T12:25:16.5786897Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5787010Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5787121Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5787344Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5787693Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5787835Z graph_break [] 2025-12-04T12:25:16.5788058Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5788817Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5788918Z warnings.warn( 2025-12-04T12:25:16.5791653Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5791762Z return x.grad, w.grad 2025-12-04T12:25:16.5792593Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ae334ed847078f28.xml - 2025-12-04T12:25:16.5792759Z =========================== short test summary info ============================ 2025-12-04T12:25:16.5793849Z FAILED [0.1658s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5793856Z 2025-12-04T12:25:16.5794065Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5795009Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.5795018Z 2025-12-04T12:25:16.5795290Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5795464Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.5795671Z ================== 1 failed, 204 deselected, 2 rerun in 5.06s ================== 2025-12-04T12:25:16.5795766Z Got exit code 1 2025-12-04T12:25:16.5795865Z Retrying single test... 2025-12-04T12:25:16.5796691Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0eab24e125a2cf7d.xml 2025-12-04T12:25:16.5796854Z ============================= test session starts ============================== 2025-12-04T12:25:16.5797202Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.5797320Z cachedir: .pytest_cache 2025-12-04T12:25:16.5797906Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.5798040Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.5798144Z configfile: pytest.ini 2025-12-04T12:25:16.5798734Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.5798969Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.5800004Z stepcurrent: skipping 34 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.5800128Z Running 1 items in this shard 2025-12-04T12:25:16.5800133Z 2025-12-04T12:25:16.5801058Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [4.6569s] [100%] 2025-12-04T12:25:16.5802007Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [0.1696s] [100%] 2025-12-04T12:25:16.5802880Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True FAILED [0.1638s] [100%] 2025-12-04T12:25:16.5802886Z 2025-12-04T12:25:16.5803024Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.5803585Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.5803743Z Traceback (most recent call last): 2025-12-04T12:25:16.5804219Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5804417Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5804626Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5804631Z 2025-12-04T12:25:16.5804848Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5805799Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.5805805Z 2025-12-04T12:25:16.5806076Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5806290Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5806402Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5806528Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5806874Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5807093Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5807199Z graph_break [] 2025-12-04T12:25:16.5807417Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5810156Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5810258Z return x.grad, w.grad 2025-12-04T12:25:16.5811028Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5811130Z warnings.warn( 2025-12-04T12:25:16.5813843Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5813959Z return x.grad, w.grad 2025-12-04T12:25:16.5814514Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.5814640Z Traceback (most recent call last): 2025-12-04T12:25:16.5815139Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5815332Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5815579Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5815585Z 2025-12-04T12:25:16.5815796Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5816751Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.5816756Z 2025-12-04T12:25:16.5817112Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5817332Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5817456Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5817570Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5817921Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5818135Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5818229Z graph_break [] 2025-12-04T12:25:16.5818460Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5821158Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5821277Z return x.grad, w.grad 2025-12-04T12:25:16.5822005Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5822104Z warnings.warn( 2025-12-04T12:25:16.5824801Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5824936Z return x.grad, w.grad 2025-12-04T12:25:16.5825156Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5825265Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5825383Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5825599Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5825937Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5826041Z graph_break [] 2025-12-04T12:25:16.5826252Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5828940Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5829080Z return x.grad, w.grad 2025-12-04T12:25:16.5829836Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5829942Z warnings.warn( 2025-12-04T12:25:16.5832677Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5832791Z return x.grad, w.grad 2025-12-04T12:25:16.5832931Z =================================== FAILURES =================================== 2025-12-04T12:25:16.5833504Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.5833621Z Traceback (most recent call last): 2025-12-04T12:25:16.5834081Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5834284Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5834495Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5834503Z 2025-12-04T12:25:16.5834709Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5835669Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.5835677Z 2025-12-04T12:25:16.5835946Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5836165Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5836274Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5836382Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5836726Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5836943Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5837044Z graph_break [] 2025-12-04T12:25:16.5837258Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5840003Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5840112Z return x.grad, w.grad 2025-12-04T12:25:16.5840840Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5840945Z warnings.warn( 2025-12-04T12:25:16.5843678Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5843828Z return x.grad, w.grad 2025-12-04T12:25:16.5844041Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5844149Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5844268Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5844482Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5844851Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5844955Z graph_break [] 2025-12-04T12:25:16.5845165Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5847875Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5847975Z return x.grad, w.grad 2025-12-04T12:25:16.5848710Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5848808Z warnings.warn( 2025-12-04T12:25:16.5851517Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5851623Z return x.grad, w.grad 2025-12-04T12:25:16.5851838Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5851954Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5852068Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5852317Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5852665Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5852767Z graph_break [] 2025-12-04T12:25:16.5852991Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5853724Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5853823Z warnings.warn( 2025-12-04T12:25:16.5856534Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.5856670Z return x.grad, w.grad 2025-12-04T12:25:16.5857587Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0eab24e125a2cf7d.xml - 2025-12-04T12:25:16.5857760Z =========================== short test summary info ============================ 2025-12-04T12:25:16.5858854Z FAILED [0.1638s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5858860Z 2025-12-04T12:25:16.5859101Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5860047Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.5860068Z 2025-12-04T12:25:16.5860329Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5860506Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.5860716Z ================== 1 failed, 204 deselected, 2 rerun in 5.04s ================== 2025-12-04T12:25:16.5860810Z Got exit code 1 2025-12-04T12:25:16.5861674Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.5862094Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.5862735Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6060ecdf9a2577f9.xml 2025-12-04T12:25:16.5862909Z ============================= test session starts ============================== 2025-12-04T12:25:16.5863256Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.5863363Z cachedir: .pytest_cache 2025-12-04T12:25:16.5863892Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.5864015Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.5864124Z configfile: pytest.ini 2025-12-04T12:25:16.5864726Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.5864946Z collecting ... collected 380 items / 35 deselected / 345 selected 2025-12-04T12:25:16.5865100Z stepcurrent: skipping 35 already run items. 2025-12-04T12:25:16.5865242Z Running 170 items in this shard 2025-12-04T12:25:16.5865247Z 2025-12-04T12:25:16.5866283Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_True_initial_xblock_1_add_1dim_False SKIPPED [0.0040s] (Skip non-critical tests to save resources.) [ 0%] 2025-12-04T12:25:16.5867321Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_True_initial_xblock_1_add_1dim_True SKIPPED [0.0029s] (Skip non-critical tests to save resources.) [ 1%] 2025-12-04T12:25:16.5868234Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6713s] [ 1%] 2025-12-04T12:25:16.5869156Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1655s] [ 1%] 2025-12-04T12:25:16.5869997Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1616s] [ 1%] 2025-12-04T12:25:16.5870039Z 2025-12-04T12:25:16.5870213Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.5870783Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.5870904Z Traceback (most recent call last): 2025-12-04T12:25:16.5871378Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5871571Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5871811Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5871817Z 2025-12-04T12:25:16.5872028Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5872981Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.5872996Z 2025-12-04T12:25:16.5873257Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5873472Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5873593Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5873701Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5874036Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5874260Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5874353Z graph_break [] 2025-12-04T12:25:16.5874564Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5875307Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5875404Z warnings.warn( 2025-12-04T12:25:16.5875983Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.5876099Z Traceback (most recent call last): 2025-12-04T12:25:16.5876565Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5876767Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5876973Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5876978Z 2025-12-04T12:25:16.5877197Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5878149Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.5878188Z 2025-12-04T12:25:16.5878451Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5878674Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5878781Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5878900Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5879238Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5879451Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5879550Z graph_break [] 2025-12-04T12:25:16.5879759Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5880490Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5880627Z warnings.warn( 2025-12-04T12:25:16.5880838Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5880949Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5881061Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5881304Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5881643Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5881741Z graph_break [] 2025-12-04T12:25:16.5881951Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5882727Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5882828Z warnings.warn( 2025-12-04T12:25:16.5882977Z =================================== FAILURES =================================== 2025-12-04T12:25:16.5883539Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.5883660Z Traceback (most recent call last): 2025-12-04T12:25:16.5884134Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5884329Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5884536Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5884549Z 2025-12-04T12:25:16.5884759Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5885715Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.5885722Z 2025-12-04T12:25:16.5885995Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5886209Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5886318Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5886439Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5886776Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5887003Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5887097Z graph_break [] 2025-12-04T12:25:16.5887311Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5888058Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5888159Z warnings.warn( 2025-12-04T12:25:16.5888371Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5888517Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5888630Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5888858Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5889192Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5889290Z graph_break [] 2025-12-04T12:25:16.5889507Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5890231Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5890333Z warnings.warn( 2025-12-04T12:25:16.5890551Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5890662Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5890783Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5890996Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5891365Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5891466Z graph_break [] 2025-12-04T12:25:16.5891677Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5892427Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5892535Z warnings.warn( 2025-12-04T12:25:16.5893350Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6060ecdf9a2577f9.xml - 2025-12-04T12:25:16.5893559Z =========================== short test summary info ============================ 2025-12-04T12:25:16.5894646Z FAILED [0.1616s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5894657Z 2025-12-04T12:25:16.5894879Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5895834Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.5895840Z 2025-12-04T12:25:16.5896258Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5896447Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.5896658Z ============= 1 failed, 2 skipped, 35 deselected, 2 rerun in 5.06s ============= 2025-12-04T12:25:16.5896765Z Got exit code 1 2025-12-04T12:25:16.5896955Z Retrying single test... 2025-12-04T12:25:16.5897603Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0eb661cc20714db1.xml 2025-12-04T12:25:16.5897776Z ============================= test session starts ============================== 2025-12-04T12:25:16.5898124Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.5898227Z cachedir: .pytest_cache 2025-12-04T12:25:16.5898763Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.5898886Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.5899001Z configfile: pytest.ini 2025-12-04T12:25:16.5899588Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.5899814Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.5900939Z stepcurrent: skipping 37 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.5901055Z Running 1 items in this shard 2025-12-04T12:25:16.5901060Z 2025-12-04T12:25:16.5901990Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6504s] [100%] 2025-12-04T12:25:16.5902907Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1651s] [100%] 2025-12-04T12:25:16.5903739Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1613s] [100%] 2025-12-04T12:25:16.5903757Z 2025-12-04T12:25:16.5903898Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.5904508Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.5904673Z Traceback (most recent call last): 2025-12-04T12:25:16.5905146Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5905343Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5905561Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5905566Z 2025-12-04T12:25:16.5905777Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5906779Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.5906789Z 2025-12-04T12:25:16.5907051Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5907274Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5907393Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5907508Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5907858Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5908078Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5908175Z graph_break [] 2025-12-04T12:25:16.5908397Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5909135Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5909236Z warnings.warn( 2025-12-04T12:25:16.5909811Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.5909930Z Traceback (most recent call last): 2025-12-04T12:25:16.5910413Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5910607Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5910814Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5910819Z 2025-12-04T12:25:16.5911035Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5911987Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.5911993Z 2025-12-04T12:25:16.5912271Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5912523Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5912633Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5912764Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5913102Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5913329Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5913424Z graph_break [] 2025-12-04T12:25:16.5913638Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5914383Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5914480Z warnings.warn( 2025-12-04T12:25:16.5914694Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5914816Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5914930Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5915182Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5915531Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5915655Z graph_break [] 2025-12-04T12:25:16.5915884Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5916611Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5916711Z warnings.warn( 2025-12-04T12:25:16.5916868Z =================================== FAILURES =================================== 2025-12-04T12:25:16.5917466Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.5917602Z Traceback (most recent call last): 2025-12-04T12:25:16.5918073Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5918271Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5918495Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5918503Z 2025-12-04T12:25:16.5918709Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5919666Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.5919684Z 2025-12-04T12:25:16.5919950Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5920164Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5920291Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5920407Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5920745Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5920976Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5921074Z graph_break [] 2025-12-04T12:25:16.5921300Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5922028Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5922130Z warnings.warn( 2025-12-04T12:25:16.5922359Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5922467Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5922584Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5922812Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5923180Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5923288Z graph_break [] 2025-12-04T12:25:16.5923500Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5924234Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5924350Z warnings.warn( 2025-12-04T12:25:16.5924565Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5924676Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5924795Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5925012Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5925357Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5925453Z graph_break [] 2025-12-04T12:25:16.5925665Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5926437Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5926538Z warnings.warn( 2025-12-04T12:25:16.5927397Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0eb661cc20714db1.xml - 2025-12-04T12:25:16.5927573Z =========================== short test summary info ============================ 2025-12-04T12:25:16.5928686Z FAILED [0.1613s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5928693Z 2025-12-04T12:25:16.5928917Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5929871Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.5929876Z 2025-12-04T12:25:16.5930152Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5930325Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.5930521Z ================== 1 failed, 204 deselected, 2 rerun in 5.03s ================== 2025-12-04T12:25:16.5930629Z Got exit code 1 2025-12-04T12:25:16.5930732Z Retrying single test... 2025-12-04T12:25:16.5931378Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-5774b4beb84c0975.xml 2025-12-04T12:25:16.5931547Z ============================= test session starts ============================== 2025-12-04T12:25:16.5931896Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.5932016Z cachedir: .pytest_cache 2025-12-04T12:25:16.5932536Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.5932662Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.5932778Z configfile: pytest.ini 2025-12-04T12:25:16.5933368Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.5933595Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.5934639Z stepcurrent: skipping 37 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.5934784Z Running 1 items in this shard 2025-12-04T12:25:16.5934789Z 2025-12-04T12:25:16.5935714Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6400s] [100%] 2025-12-04T12:25:16.5936628Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1635s] [100%] 2025-12-04T12:25:16.5937564Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1612s] [100%] 2025-12-04T12:25:16.5937572Z 2025-12-04T12:25:16.5937709Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.5938275Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.5938453Z Traceback (most recent call last): 2025-12-04T12:25:16.5938916Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5939124Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5939365Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5939371Z 2025-12-04T12:25:16.5939582Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5940547Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.5940553Z 2025-12-04T12:25:16.5940849Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5941074Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5941190Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5941302Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5941650Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5941872Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5941977Z graph_break [] 2025-12-04T12:25:16.5942191Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5942930Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5943040Z warnings.warn( 2025-12-04T12:25:16.5943611Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.5943732Z Traceback (most recent call last): 2025-12-04T12:25:16.5944210Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5944408Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5944628Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5944633Z 2025-12-04T12:25:16.5944845Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5945802Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.5945807Z 2025-12-04T12:25:16.5946083Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5946298Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5946419Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5946565Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5946905Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5947138Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5947235Z graph_break [] 2025-12-04T12:25:16.5947450Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5948188Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5948287Z warnings.warn( 2025-12-04T12:25:16.5948516Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5948673Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5948846Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5949093Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5949437Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5949632Z graph_break [] 2025-12-04T12:25:16.5949855Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5950621Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5950732Z warnings.warn( 2025-12-04T12:25:16.5950874Z =================================== FAILURES =================================== 2025-12-04T12:25:16.5951437Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.5951568Z Traceback (most recent call last): 2025-12-04T12:25:16.5952072Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5952269Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5952491Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5952499Z 2025-12-04T12:25:16.5952713Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5953681Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.5953687Z 2025-12-04T12:25:16.5953952Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5954180Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5954286Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5954395Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5954749Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5954972Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5955069Z graph_break [] 2025-12-04T12:25:16.5955292Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5956022Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5956133Z warnings.warn( 2025-12-04T12:25:16.5956344Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5956449Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5956575Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5956792Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5957128Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5957242Z graph_break [] 2025-12-04T12:25:16.5957454Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5958212Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5958327Z warnings.warn( 2025-12-04T12:25:16.5958534Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5958653Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5958762Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5958974Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5959363Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5959508Z graph_break [] 2025-12-04T12:25:16.5959842Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5960579Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5960725Z warnings.warn( 2025-12-04T12:25:16.5961560Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-5774b4beb84c0975.xml - 2025-12-04T12:25:16.5961761Z =========================== short test summary info ============================ 2025-12-04T12:25:16.5962851Z FAILED [0.1612s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5962870Z 2025-12-04T12:25:16.5963085Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5964068Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.5964077Z 2025-12-04T12:25:16.5964354Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5964532Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.5964743Z ================== 1 failed, 204 deselected, 2 rerun in 5.02s ================== 2025-12-04T12:25:16.5964840Z Got exit code 1 2025-12-04T12:25:16.5965718Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.5966139Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.5966782Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-74651aae35dd9844.xml 2025-12-04T12:25:16.5966957Z ============================= test session starts ============================== 2025-12-04T12:25:16.5967305Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.5967415Z cachedir: .pytest_cache 2025-12-04T12:25:16.5967947Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.5968068Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.5968177Z configfile: pytest.ini 2025-12-04T12:25:16.5968775Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.5969004Z collecting ... collected 380 items / 38 deselected / 342 selected 2025-12-04T12:25:16.5969157Z stepcurrent: skipping 38 already run items. 2025-12-04T12:25:16.5969271Z Running 167 items in this shard 2025-12-04T12:25:16.5969276Z 2025-12-04T12:25:16.5970192Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6930s] [ 0%] 2025-12-04T12:25:16.5971154Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1684s] [ 0%] 2025-12-04T12:25:16.5971990Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1649s] [ 0%] 2025-12-04T12:25:16.5971995Z 2025-12-04T12:25:16.5972143Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.5972711Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.5972830Z Traceback (most recent call last): 2025-12-04T12:25:16.5973320Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5973563Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5973784Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5973789Z 2025-12-04T12:25:16.5974028Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5974985Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.5975001Z 2025-12-04T12:25:16.5975268Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5975517Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5975639Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5975756Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5976096Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5976330Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5976429Z graph_break [] 2025-12-04T12:25:16.5976644Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5977462Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5977565Z warnings.warn( 2025-12-04T12:25:16.5978140Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.5978263Z Traceback (most recent call last): 2025-12-04T12:25:16.5978735Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5978944Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5979157Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5979162Z 2025-12-04T12:25:16.5979380Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5980340Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.5980345Z 2025-12-04T12:25:16.5980614Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5980840Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5980955Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5981084Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5981428Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5981683Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5981794Z graph_break [] 2025-12-04T12:25:16.5982011Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5982749Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5982865Z warnings.warn( 2025-12-04T12:25:16.5983079Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5983202Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5983315Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5983533Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5983890Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5983985Z graph_break [] 2025-12-04T12:25:16.5984201Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5984966Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5985063Z warnings.warn( 2025-12-04T12:25:16.5985251Z =================================== FAILURES =================================== 2025-12-04T12:25:16.5985816Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.5985937Z Traceback (most recent call last): 2025-12-04T12:25:16.5986420Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.5986619Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.5986856Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5986864Z 2025-12-04T12:25:16.5987091Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5988047Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.5988055Z 2025-12-04T12:25:16.5988331Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5988543Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5988650Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5988779Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5989115Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5989343Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5989443Z graph_break [] 2025-12-04T12:25:16.5989656Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5990396Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5990497Z warnings.warn( 2025-12-04T12:25:16.5990711Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5990829Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5990943Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5991168Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5991503Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5991600Z graph_break [] 2025-12-04T12:25:16.5991823Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5992548Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5992682Z warnings.warn( 2025-12-04T12:25:16.5992909Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.5993017Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.5993134Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.5993350Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.5993686Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.5993790Z graph_break [] 2025-12-04T12:25:16.5994006Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.5994737Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.5994841Z warnings.warn( 2025-12-04T12:25:16.5995661Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-74651aae35dd9844.xml - 2025-12-04T12:25:16.5995869Z =========================== short test summary info ============================ 2025-12-04T12:25:16.5997195Z FAILED [0.1649s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.5997202Z 2025-12-04T12:25:16.5997420Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.5998422Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.5998428Z 2025-12-04T12:25:16.5998697Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.5998888Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.5999088Z ================== 1 failed, 38 deselected, 2 rerun in 5.08s =================== 2025-12-04T12:25:16.5999188Z Got exit code 1 2025-12-04T12:25:16.5999304Z Retrying single test... 2025-12-04T12:25:16.5999951Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-7a5a8b9ba5d60164.xml 2025-12-04T12:25:16.6000121Z ============================= test session starts ============================== 2025-12-04T12:25:16.6000469Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6000577Z cachedir: .pytest_cache 2025-12-04T12:25:16.6001110Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6001236Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6001342Z configfile: pytest.ini 2025-12-04T12:25:16.6001950Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6002179Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.6003227Z stepcurrent: skipping 38 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6003338Z Running 1 items in this shard 2025-12-04T12:25:16.6003343Z 2025-12-04T12:25:16.6004271Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6556s] [100%] 2025-12-04T12:25:16.6005177Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1676s] [100%] 2025-12-04T12:25:16.6006062Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1647s] [100%] 2025-12-04T12:25:16.6006079Z 2025-12-04T12:25:16.6006219Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6006782Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6006913Z Traceback (most recent call last): 2025-12-04T12:25:16.6007384Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6007580Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6007801Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6007849Z 2025-12-04T12:25:16.6008059Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6009061Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6009067Z 2025-12-04T12:25:16.6009334Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6009550Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6009670Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6009777Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6010124Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6010370Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6010471Z graph_break [] 2025-12-04T12:25:16.6010696Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6011437Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6011541Z warnings.warn( 2025-12-04T12:25:16.6012123Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6012244Z Traceback (most recent call last): 2025-12-04T12:25:16.6012721Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6012918Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6013127Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6013132Z 2025-12-04T12:25:16.6013407Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6014364Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6014371Z 2025-12-04T12:25:16.6014652Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6014868Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6014980Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6015104Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6015446Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6015664Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6015771Z graph_break [] 2025-12-04T12:25:16.6015986Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6016774Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6016938Z warnings.warn( 2025-12-04T12:25:16.6017160Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6017281Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6017394Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6017607Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6017955Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6018051Z graph_break [] 2025-12-04T12:25:16.6018274Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6019004Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6019105Z warnings.warn( 2025-12-04T12:25:16.6019295Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6019862Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6020011Z Traceback (most recent call last): 2025-12-04T12:25:16.6020490Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6020685Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6020903Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6020908Z 2025-12-04T12:25:16.6021117Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6022095Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6022116Z 2025-12-04T12:25:16.6022381Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6022595Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6022721Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6022838Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6023178Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6023406Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6023500Z graph_break [] 2025-12-04T12:25:16.6023723Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6024447Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6024548Z warnings.warn( 2025-12-04T12:25:16.6024779Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6024882Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6024993Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6025223Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6025559Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6025658Z graph_break [] 2025-12-04T12:25:16.6025881Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6026605Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6026718Z warnings.warn( 2025-12-04T12:25:16.6026931Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6027078Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6027203Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6027416Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6027751Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6027862Z graph_break [] 2025-12-04T12:25:16.6028073Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6028808Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6028904Z warnings.warn( 2025-12-04T12:25:16.6029717Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-7a5a8b9ba5d60164.xml - 2025-12-04T12:25:16.6029899Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6031000Z FAILED [0.1647s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6031041Z 2025-12-04T12:25:16.6031300Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6032261Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6032267Z 2025-12-04T12:25:16.6032542Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6032722Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6032959Z ================== 1 failed, 204 deselected, 2 rerun in 5.04s ================== 2025-12-04T12:25:16.6033075Z Got exit code 1 2025-12-04T12:25:16.6033186Z Retrying single test... 2025-12-04T12:25:16.6033830Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-26faac4e438f6204.xml 2025-12-04T12:25:16.6034005Z ============================= test session starts ============================== 2025-12-04T12:25:16.6034355Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6034474Z cachedir: .pytest_cache 2025-12-04T12:25:16.6034996Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6035120Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6035243Z configfile: pytest.ini 2025-12-04T12:25:16.6035836Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6036066Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.6037130Z stepcurrent: skipping 38 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6037249Z Running 1 items in this shard 2025-12-04T12:25:16.6037255Z 2025-12-04T12:25:16.6038180Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6540s] [100%] 2025-12-04T12:25:16.6039101Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1664s] [100%] 2025-12-04T12:25:16.6039948Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1615s] [100%] 2025-12-04T12:25:16.6039985Z 2025-12-04T12:25:16.6040126Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6040694Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6040832Z Traceback (most recent call last): 2025-12-04T12:25:16.6041305Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6041517Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6041729Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6041735Z 2025-12-04T12:25:16.6041951Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6042926Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6042966Z 2025-12-04T12:25:16.6043237Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6043494Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6043604Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6043722Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6044121Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6044343Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6044441Z graph_break [] 2025-12-04T12:25:16.6044668Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6045427Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6045549Z warnings.warn( 2025-12-04T12:25:16.6046115Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6046235Z Traceback (most recent call last): 2025-12-04T12:25:16.6046721Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6046915Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6047138Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6047143Z 2025-12-04T12:25:16.6047355Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6048315Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6048322Z 2025-12-04T12:25:16.6048600Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6048812Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6048931Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6049047Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6049385Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6049614Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6049708Z graph_break [] 2025-12-04T12:25:16.6049920Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6050664Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6050790Z warnings.warn( 2025-12-04T12:25:16.6051013Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6051124Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6051235Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6051465Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6051800Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6051893Z graph_break [] 2025-12-04T12:25:16.6052117Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6052843Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6052952Z warnings.warn( 2025-12-04T12:25:16.6053097Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6053660Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6053828Z Traceback (most recent call last): 2025-12-04T12:25:16.6054297Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6054520Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6054742Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6054748Z 2025-12-04T12:25:16.6054958Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6055923Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6055928Z 2025-12-04T12:25:16.6056242Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6056458Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6056582Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6056698Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6057117Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6057343Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6057439Z graph_break [] 2025-12-04T12:25:16.6057662Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6058391Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6058491Z warnings.warn( 2025-12-04T12:25:16.6058716Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6058827Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6058954Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6059172Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6059516Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6059621Z graph_break [] 2025-12-04T12:25:16.6059836Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6060559Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6060670Z warnings.warn( 2025-12-04T12:25:16.6060884Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6061004Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6061118Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6061343Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6061735Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6061834Z graph_break [] 2025-12-04T12:25:16.6062046Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6062788Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6062888Z warnings.warn( 2025-12-04T12:25:16.6063718Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-26faac4e438f6204.xml - 2025-12-04T12:25:16.6063886Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6064978Z FAILED [0.1615s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6065034Z 2025-12-04T12:25:16.6065250Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6066234Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6066240Z 2025-12-04T12:25:16.6066515Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6066695Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6066904Z ================== 1 failed, 204 deselected, 2 rerun in 5.04s ================== 2025-12-04T12:25:16.6067001Z Got exit code 1 2025-12-04T12:25:16.6067901Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6068330Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.6068969Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-aa20002b977e7ab4.xml 2025-12-04T12:25:16.6069133Z ============================= test session starts ============================== 2025-12-04T12:25:16.6069496Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6069602Z cachedir: .pytest_cache 2025-12-04T12:25:16.6070133Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6070255Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6070361Z configfile: pytest.ini 2025-12-04T12:25:16.6070963Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6071192Z collecting ... collected 380 items / 39 deselected / 341 selected 2025-12-04T12:25:16.6071344Z stepcurrent: skipping 39 already run items. 2025-12-04T12:25:16.6071457Z Running 166 items in this shard 2025-12-04T12:25:16.6071462Z 2025-12-04T12:25:16.6072504Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_True_initial_xblock_1_add_1dim_False SKIPPED [0.0039s] (Skip non-critical tests to save resources.) [ 0%] 2025-12-04T12:25:16.6073544Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_True_initial_xblock_1_add_1dim_True SKIPPED [0.0028s] (Skip non-critical tests to save resources.) [ 1%] 2025-12-04T12:25:16.6074579Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_True_initial_xblock_2_add_1dim_False SKIPPED [0.0035s] (Skip non-critical tests to save resources.) [ 1%] 2025-12-04T12:25:16.6075656Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_True_initial_xblock_2_add_1dim_True SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 2%] 2025-12-04T12:25:16.6076682Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_True SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 3%] 2025-12-04T12:25:16.6077712Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_True_initial_xblock_1_add_1dim_True SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 3%] 2025-12-04T12:25:16.6078739Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_True_initial_xblock_2_add_1dim_False SKIPPED [0.0028s] (Skip non-critical tests to save resources.) [ 4%] 2025-12-04T12:25:16.6079828Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_True_initial_xblock_2_add_1dim_True SKIPPED [0.0029s] (Skip non-critical tests to save resources.) [ 4%] 2025-12-04T12:25:16.6080748Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6547s] [ 5%] 2025-12-04T12:25:16.6081664Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1641s] [ 5%] 2025-12-04T12:25:16.6082548Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1620s] [ 5%] 2025-12-04T12:25:16.6082558Z 2025-12-04T12:25:16.6082698Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6083278Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6083398Z Traceback (most recent call last): 2025-12-04T12:25:16.6083865Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6084077Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6084285Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6084291Z 2025-12-04T12:25:16.6084514Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6085473Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6085483Z 2025-12-04T12:25:16.6085756Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6085973Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6086086Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6086210Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6086550Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6086770Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6086881Z graph_break [] 2025-12-04T12:25:16.6087096Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6087833Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6087980Z warnings.warn( 2025-12-04T12:25:16.6088551Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6088686Z Traceback (most recent call last): 2025-12-04T12:25:16.6089158Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6089355Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6089580Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6089585Z 2025-12-04T12:25:16.6089796Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6090768Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6090776Z 2025-12-04T12:25:16.6091041Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6091287Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6091406Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6091520Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6091902Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6092121Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6092221Z graph_break [] 2025-12-04T12:25:16.6092448Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6093211Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6093313Z warnings.warn( 2025-12-04T12:25:16.6093539Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6093650Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6093773Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6093986Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6094323Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6094430Z graph_break [] 2025-12-04T12:25:16.6094642Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6095369Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6095478Z warnings.warn( 2025-12-04T12:25:16.6095621Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6096380Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6096507Z Traceback (most recent call last): 2025-12-04T12:25:16.6097041Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6097253Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6097467Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6097472Z 2025-12-04T12:25:16.6097687Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6098656Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6098661Z 2025-12-04T12:25:16.6098930Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6099161Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6099342Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6099458Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6099814Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6100030Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6100141Z graph_break [] 2025-12-04T12:25:16.6100356Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6101085Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6101199Z warnings.warn( 2025-12-04T12:25:16.6101414Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6101525Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6101653Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6101878Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6102270Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6102367Z graph_break [] 2025-12-04T12:25:16.6102582Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6103380Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6103485Z warnings.warn( 2025-12-04T12:25:16.6103696Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6103818Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6103933Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6104207Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6104549Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6104648Z graph_break [] 2025-12-04T12:25:16.6104875Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6105608Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6105708Z warnings.warn( 2025-12-04T12:25:16.6106544Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-aa20002b977e7ab4.xml - 2025-12-04T12:25:16.6106716Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6107823Z FAILED [0.1620s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6107830Z 2025-12-04T12:25:16.6108044Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6109016Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6109024Z 2025-12-04T12:25:16.6109292Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6109473Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6109704Z ============= 1 failed, 8 skipped, 39 deselected, 2 rerun in 5.07s ============= 2025-12-04T12:25:16.6109804Z Got exit code 1 2025-12-04T12:25:16.6109916Z Retrying single test... 2025-12-04T12:25:16.6110580Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6316dfd7aad65a93.xml 2025-12-04T12:25:16.6110770Z ============================= test session starts ============================== 2025-12-04T12:25:16.6111134Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6111242Z cachedir: .pytest_cache 2025-12-04T12:25:16.6111767Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6111906Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6112013Z configfile: pytest.ini 2025-12-04T12:25:16.6112616Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6112839Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.6113880Z stepcurrent: skipping 47 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6114041Z Running 1 items in this shard 2025-12-04T12:25:16.6114046Z 2025-12-04T12:25:16.6114985Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6484s] [100%] 2025-12-04T12:25:16.6115910Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1634s] [100%] 2025-12-04T12:25:16.6116750Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1616s] [100%] 2025-12-04T12:25:16.6116755Z 2025-12-04T12:25:16.6116933Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6117501Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6117624Z Traceback (most recent call last): 2025-12-04T12:25:16.6118106Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6118303Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6118514Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6118519Z 2025-12-04T12:25:16.6118738Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6119695Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6119700Z 2025-12-04T12:25:16.6119976Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6120194Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6120304Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6120425Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6120769Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6121000Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6121098Z graph_break [] 2025-12-04T12:25:16.6121311Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6122059Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6122158Z warnings.warn( 2025-12-04T12:25:16.6122729Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6122891Z Traceback (most recent call last): 2025-12-04T12:25:16.6123363Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6123566Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6123778Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6123784Z 2025-12-04T12:25:16.6123998Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6124964Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6124970Z 2025-12-04T12:25:16.6125235Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6125462Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6125576Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6125690Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6126076Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6126295Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6126392Z graph_break [] 2025-12-04T12:25:16.6126651Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6127389Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6127501Z warnings.warn( 2025-12-04T12:25:16.6127715Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6127823Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6127976Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6128196Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6128539Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6128647Z graph_break [] 2025-12-04T12:25:16.6128862Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6129608Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6129709Z warnings.warn( 2025-12-04T12:25:16.6129853Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6130431Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6130550Z Traceback (most recent call last): 2025-12-04T12:25:16.6131031Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6131228Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6131439Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6131445Z 2025-12-04T12:25:16.6131666Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6132617Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6132622Z 2025-12-04T12:25:16.6132899Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6133114Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6133224Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6133344Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6133686Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6133941Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6134054Z graph_break [] 2025-12-04T12:25:16.6134267Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6135014Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6135113Z warnings.warn( 2025-12-04T12:25:16.6135326Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6135446Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6135558Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6135777Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6136128Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6136224Z graph_break [] 2025-12-04T12:25:16.6136451Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6137281Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6137384Z warnings.warn( 2025-12-04T12:25:16.6137646Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6137755Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6137867Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6138101Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6138435Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6138543Z graph_break [] 2025-12-04T12:25:16.6138783Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6139513Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6139628Z warnings.warn( 2025-12-04T12:25:16.6140454Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6316dfd7aad65a93.xml - 2025-12-04T12:25:16.6140620Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6141724Z FAILED [0.1616s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6141730Z 2025-12-04T12:25:16.6141944Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6142913Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6142923Z 2025-12-04T12:25:16.6143186Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6143375Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6143575Z ================== 1 failed, 204 deselected, 2 rerun in 5.03s ================== 2025-12-04T12:25:16.6143674Z Got exit code 1 2025-12-04T12:25:16.6143790Z Retrying single test... 2025-12-04T12:25:16.6144442Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d1f50caa5a44c0ee.xml 2025-12-04T12:25:16.6144642Z ============================= test session starts ============================== 2025-12-04T12:25:16.6145001Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6145145Z cachedir: .pytest_cache 2025-12-04T12:25:16.6145681Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6145805Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6145911Z configfile: pytest.ini 2025-12-04T12:25:16.6146515Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6146741Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.6147798Z stepcurrent: skipping 47 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6147910Z Running 1 items in this shard 2025-12-04T12:25:16.6147915Z 2025-12-04T12:25:16.6148833Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6463s] [100%] 2025-12-04T12:25:16.6149839Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1631s] [100%] 2025-12-04T12:25:16.6150675Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1614s] [100%] 2025-12-04T12:25:16.6150681Z 2025-12-04T12:25:16.6150833Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6151398Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6151548Z Traceback (most recent call last): 2025-12-04T12:25:16.6152034Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6152231Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6152454Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6152459Z 2025-12-04T12:25:16.6152675Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6153622Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6153639Z 2025-12-04T12:25:16.6153902Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6154118Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6154242Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6154356Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6154696Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6154926Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6155022Z graph_break [] 2025-12-04T12:25:16.6155233Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6155982Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6156082Z warnings.warn( 2025-12-04T12:25:16.6156660Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6156779Z Traceback (most recent call last): 2025-12-04T12:25:16.6157249Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6157498Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6157708Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6157715Z 2025-12-04T12:25:16.6157940Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6158896Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6158902Z 2025-12-04T12:25:16.6159165Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6159398Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6159507Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6159631Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6159973Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6160192Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6160333Z graph_break [] 2025-12-04T12:25:16.6160549Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6161313Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6161424Z warnings.warn( 2025-12-04T12:25:16.6161639Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6161762Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6161875Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6162093Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6162442Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6162572Z graph_break [] 2025-12-04T12:25:16.6162790Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6163538Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6163637Z warnings.warn( 2025-12-04T12:25:16.6163798Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6164367Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6164489Z Traceback (most recent call last): 2025-12-04T12:25:16.6164976Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6165173Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6165384Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6165402Z 2025-12-04T12:25:16.6165619Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6166571Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6166576Z 2025-12-04T12:25:16.6166859Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6167076Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6167197Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6167316Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6167656Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6167887Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6167990Z graph_break [] 2025-12-04T12:25:16.6168205Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6168979Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6169087Z warnings.warn( 2025-12-04T12:25:16.6169301Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6169421Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6169534Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6169767Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6170211Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6170312Z graph_break [] 2025-12-04T12:25:16.6170618Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6171384Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6171531Z warnings.warn( 2025-12-04T12:25:16.6171756Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6171864Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6171988Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6172236Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6172573Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6172679Z graph_break [] 2025-12-04T12:25:16.6172891Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6173613Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6173749Z warnings.warn( 2025-12-04T12:25:16.6174574Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d1f50caa5a44c0ee.xml - 2025-12-04T12:25:16.6174757Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6175851Z FAILED [0.1614s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6175857Z 2025-12-04T12:25:16.6176081Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6177103Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6177110Z 2025-12-04T12:25:16.6177376Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6177566Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6177767Z ================== 1 failed, 204 deselected, 2 rerun in 5.02s ================== 2025-12-04T12:25:16.6177877Z Got exit code 1 2025-12-04T12:25:16.6178746Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6179152Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.6179805Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-8edf4f7e001e74f9.xml 2025-12-04T12:25:16.6179967Z ============================= test session starts ============================== 2025-12-04T12:25:16.6180326Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6180470Z cachedir: .pytest_cache 2025-12-04T12:25:16.6180989Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6181123Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6181229Z configfile: pytest.ini 2025-12-04T12:25:16.6181822Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6182057Z collecting ... collected 380 items / 48 deselected / 332 selected 2025-12-04T12:25:16.6182202Z stepcurrent: skipping 48 already run items. 2025-12-04T12:25:16.6182325Z Running 157 items in this shard 2025-12-04T12:25:16.6182330Z 2025-12-04T12:25:16.6183367Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_2_add_1dim_True SKIPPED [0.0042s] (Skip non-critical tests to save resources.) [ 0%] 2025-12-04T12:25:16.6184400Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_True_initial_xblock_1_add_1dim_False SKIPPED [0.0035s] (Skip non-critical tests to save resources.) [ 1%] 2025-12-04T12:25:16.6185499Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_True_initial_xblock_2_add_1dim_False SKIPPED [0.0040s] (Skip non-critical tests to save resources.) [ 1%] 2025-12-04T12:25:16.6186405Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.7012s] [ 2%] 2025-12-04T12:25:16.6187350Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1620s] [ 2%] 2025-12-04T12:25:16.6188177Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1588s] [ 2%] 2025-12-04T12:25:16.6188185Z 2025-12-04T12:25:16.6188337Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6188893Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6189014Z Traceback (most recent call last): 2025-12-04T12:25:16.6189497Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6189695Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6189917Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6189922Z 2025-12-04T12:25:16.6190130Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6191078Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6191086Z 2025-12-04T12:25:16.6191365Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6191582Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6191705Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6191816Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6192159Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6192388Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6192486Z graph_break [] 2025-12-04T12:25:16.6192703Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6193506Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6193607Z warnings.warn( 2025-12-04T12:25:16.6194185Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6194305Z Traceback (most recent call last): 2025-12-04T12:25:16.6194774Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6194980Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6195189Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6195194Z 2025-12-04T12:25:16.6195417Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6196561Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6196631Z 2025-12-04T12:25:16.6196900Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6197174Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6197285Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6197412Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6197752Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6197968Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6198081Z graph_break [] 2025-12-04T12:25:16.6198295Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6199067Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6199185Z warnings.warn( 2025-12-04T12:25:16.6199397Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6199516Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6199627Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6199844Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6200195Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6200315Z graph_break [] 2025-12-04T12:25:16.6200620Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6201364Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6201467Z warnings.warn( 2025-12-04T12:25:16.6201623Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6202189Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6202312Z Traceback (most recent call last): 2025-12-04T12:25:16.6202794Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6202988Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6203200Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6203205Z 2025-12-04T12:25:16.6203428Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6204380Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6204386Z 2025-12-04T12:25:16.6204717Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6204934Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6205044Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6205170Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6205510Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6205743Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6205838Z graph_break [] 2025-12-04T12:25:16.6206050Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6206790Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6206888Z warnings.warn( 2025-12-04T12:25:16.6207103Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6207225Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6207369Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6207600Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6207937Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6208064Z graph_break [] 2025-12-04T12:25:16.6208290Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6209017Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6209115Z warnings.warn( 2025-12-04T12:25:16.6209339Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6209481Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6209603Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6209825Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6210163Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6210273Z graph_break [] 2025-12-04T12:25:16.6210484Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6211347Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6211519Z warnings.warn( 2025-12-04T12:25:16.6212348Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-8edf4f7e001e74f9.xml - 2025-12-04T12:25:16.6212530Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6213612Z FAILED [0.1588s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6213623Z 2025-12-04T12:25:16.6213839Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6214795Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6214801Z 2025-12-04T12:25:16.6215069Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6215257Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6215472Z ============= 1 failed, 3 skipped, 48 deselected, 2 rerun in 5.10s ============= 2025-12-04T12:25:16.6215570Z Got exit code 1 2025-12-04T12:25:16.6215687Z Retrying single test... 2025-12-04T12:25:16.6216377Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-defc51812af7e0b0.xml 2025-12-04T12:25:16.6216655Z ============================= test session starts ============================== 2025-12-04T12:25:16.6217117Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6217287Z cachedir: .pytest_cache 2025-12-04T12:25:16.6249685Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6249919Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6250044Z configfile: pytest.ini 2025-12-04T12:25:16.6250652Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6250906Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.6251944Z stepcurrent: skipping 51 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6252178Z Running 1 items in this shard 2025-12-04T12:25:16.6252186Z 2025-12-04T12:25:16.6253156Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.7093s] [100%] 2025-12-04T12:25:16.6254072Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1681s] [100%] 2025-12-04T12:25:16.6254948Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1647s] [100%] 2025-12-04T12:25:16.6254957Z 2025-12-04T12:25:16.6255099Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6255673Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6255792Z Traceback (most recent call last): 2025-12-04T12:25:16.6256265Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6256471Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6256682Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6256688Z 2025-12-04T12:25:16.6256990Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6257953Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6257961Z 2025-12-04T12:25:16.6258229Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6258460Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6258572Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6258691Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6259044Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6259263Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6259370Z graph_break [] 2025-12-04T12:25:16.6259584Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6260331Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6260439Z warnings.warn( 2025-12-04T12:25:16.6261030Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6261154Z Traceback (most recent call last): 2025-12-04T12:25:16.6261628Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6261823Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6262042Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6262047Z 2025-12-04T12:25:16.6262259Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6263210Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6263215Z 2025-12-04T12:25:16.6263489Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6263705Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6263858Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6263969Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6264308Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6264556Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6264649Z graph_break [] 2025-12-04T12:25:16.6264865Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6265611Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6265709Z warnings.warn( 2025-12-04T12:25:16.6265956Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6266067Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6266179Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6266408Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6266743Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6266836Z graph_break [] 2025-12-04T12:25:16.6267054Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6267782Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6267886Z warnings.warn( 2025-12-04T12:25:16.6268023Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6268584Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6268706Z Traceback (most recent call last): 2025-12-04T12:25:16.6269174Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6269372Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6269586Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6269591Z 2025-12-04T12:25:16.6269803Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6270756Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6270761Z 2025-12-04T12:25:16.6271022Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6271242Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6271350Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6271490Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6271836Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6272053Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6272147Z graph_break [] 2025-12-04T12:25:16.6272367Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6273093Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6273198Z warnings.warn( 2025-12-04T12:25:16.6273407Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6273511Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6273629Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6273845Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6274180Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6274307Z graph_break [] 2025-12-04T12:25:16.6274519Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6275269Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6275377Z warnings.warn( 2025-12-04T12:25:16.6275589Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6275702Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6275810Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6276023Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6276390Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6276486Z graph_break [] 2025-12-04T12:25:16.6276696Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6277428Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6277525Z warnings.warn( 2025-12-04T12:25:16.6278353Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-defc51812af7e0b0.xml - 2025-12-04T12:25:16.6278522Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6279597Z FAILED [0.1647s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6279613Z 2025-12-04T12:25:16.6279826Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6280772Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6280780Z 2025-12-04T12:25:16.6281052Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6281227Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6281434Z ================== 1 failed, 204 deselected, 2 rerun in 5.10s ================== 2025-12-04T12:25:16.6281529Z Got exit code 1 2025-12-04T12:25:16.6281635Z Retrying single test... 2025-12-04T12:25:16.6282283Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0bc167e05b941aff.xml 2025-12-04T12:25:16.6282442Z ============================= test session starts ============================== 2025-12-04T12:25:16.6282820Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6282935Z cachedir: .pytest_cache 2025-12-04T12:25:16.6283456Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6283584Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6283692Z configfile: pytest.ini 2025-12-04T12:25:16.6284287Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6284518Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.6285546Z stepcurrent: skipping 51 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6285665Z Running 1 items in this shard 2025-12-04T12:25:16.6285673Z 2025-12-04T12:25:16.6286588Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6683s] [100%] 2025-12-04T12:25:16.6287582Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1669s] [100%] 2025-12-04T12:25:16.6288419Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1606s] [100%] 2025-12-04T12:25:16.6288424Z 2025-12-04T12:25:16.6288560Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6289151Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6289273Z Traceback (most recent call last): 2025-12-04T12:25:16.6289747Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6289947Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6290156Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6290161Z 2025-12-04T12:25:16.6290379Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6291320Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6291326Z 2025-12-04T12:25:16.6291585Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6291809Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6291922Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6292045Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6292385Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6292628Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6292786Z graph_break [] 2025-12-04T12:25:16.6293032Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6293760Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6293872Z warnings.warn( 2025-12-04T12:25:16.6294427Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6294563Z Traceback (most recent call last): 2025-12-04T12:25:16.6295029Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6295265Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6295479Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6295485Z 2025-12-04T12:25:16.6295694Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6296815Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6296821Z 2025-12-04T12:25:16.6297137Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6297350Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6297473Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6297583Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6297935Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6298226Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6298317Z graph_break [] 2025-12-04T12:25:16.6298545Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6299317Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6299413Z warnings.warn( 2025-12-04T12:25:16.6299638Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6299741Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6299862Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6300078Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6300460Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6300568Z graph_break [] 2025-12-04T12:25:16.6300780Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6301503Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6301610Z warnings.warn( 2025-12-04T12:25:16.6301747Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6302316Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6302431Z Traceback (most recent call last): 2025-12-04T12:25:16.6302892Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6303101Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6303355Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6303366Z 2025-12-04T12:25:16.6303671Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6304720Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6304726Z 2025-12-04T12:25:16.6304989Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6305216Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6305323Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6305435Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6305784Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6305997Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6306174Z graph_break [] 2025-12-04T12:25:16.6306387Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6307117Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6307224Z warnings.warn( 2025-12-04T12:25:16.6307433Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6307539Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6307660Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6307872Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6308215Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6308308Z graph_break [] 2025-12-04T12:25:16.6308519Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6309257Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6309386Z warnings.warn( 2025-12-04T12:25:16.6309594Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6309739Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6309849Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6310072Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6310401Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6310496Z graph_break [] 2025-12-04T12:25:16.6310714Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6311462Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6311563Z warnings.warn( 2025-12-04T12:25:16.6312394Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0bc167e05b941aff.xml - 2025-12-04T12:25:16.6312560Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6313668Z FAILED [0.1606s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6313675Z 2025-12-04T12:25:16.6313887Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6314941Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6314949Z 2025-12-04T12:25:16.6315211Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6315387Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6315593Z ================== 1 failed, 204 deselected, 2 rerun in 5.05s ================== 2025-12-04T12:25:16.6315693Z Got exit code 1 2025-12-04T12:25:16.6316554Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6316966Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.6317608Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-98b78c8ed19e0643.xml 2025-12-04T12:25:16.6317781Z ============================= test session starts ============================== 2025-12-04T12:25:16.6318179Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6318285Z cachedir: .pytest_cache 2025-12-04T12:25:16.6318811Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6318933Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6319052Z configfile: pytest.ini 2025-12-04T12:25:16.6319644Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6319864Z collecting ... collected 380 items / 52 deselected / 328 selected 2025-12-04T12:25:16.6320017Z stepcurrent: skipping 52 already run items. 2025-12-04T12:25:16.6320129Z Running 153 items in this shard 2025-12-04T12:25:16.6320135Z 2025-12-04T12:25:16.6321055Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6673s] [ 0%] 2025-12-04T12:25:16.6321998Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1645s] [ 0%] 2025-12-04T12:25:16.6322844Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1600s] [ 0%] 2025-12-04T12:25:16.6322850Z 2025-12-04T12:25:16.6323000Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6323557Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6323715Z Traceback (most recent call last): 2025-12-04T12:25:16.6324184Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6324378Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6324589Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6324594Z 2025-12-04T12:25:16.6324805Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6325898Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6325905Z 2025-12-04T12:25:16.6326167Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6326383Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6326495Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6326607Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6326954Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6327169Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6327265Z graph_break [] 2025-12-04T12:25:16.6327480Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6328215Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6328313Z warnings.warn( 2025-12-04T12:25:16.6328874Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6328987Z Traceback (most recent call last): 2025-12-04T12:25:16.6329468Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6329663Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6329926Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6329934Z 2025-12-04T12:25:16.6330154Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6331095Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6331101Z 2025-12-04T12:25:16.6331370Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6331584Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6331687Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6331812Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6332151Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6332363Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6332472Z graph_break [] 2025-12-04T12:25:16.6332711Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6333476Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6333576Z warnings.warn( 2025-12-04T12:25:16.6333786Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6333900Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6334005Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6334214Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6334560Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6334679Z graph_break [] 2025-12-04T12:25:16.6334901Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6335629Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6335780Z warnings.warn( 2025-12-04T12:25:16.6336017Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6336621Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6336741Z Traceback (most recent call last): 2025-12-04T12:25:16.6337282Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6337475Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6337698Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6337703Z 2025-12-04T12:25:16.6337913Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6338856Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6338879Z 2025-12-04T12:25:16.6339144Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6339360Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6339480Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6339593Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6339933Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6340166Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6340258Z graph_break [] 2025-12-04T12:25:16.6340470Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6341260Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6341360Z warnings.warn( 2025-12-04T12:25:16.6341583Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6341692Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6341805Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6342028Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6342362Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6342456Z graph_break [] 2025-12-04T12:25:16.6342677Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6343405Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6343518Z warnings.warn( 2025-12-04T12:25:16.6343755Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6343863Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6343984Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6344225Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6344557Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6344662Z graph_break [] 2025-12-04T12:25:16.6344871Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6345603Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6345727Z warnings.warn( 2025-12-04T12:25:16.6346539Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-98b78c8ed19e0643.xml - 2025-12-04T12:25:16.6346720Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6347805Z FAILED [0.1600s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6347811Z 2025-12-04T12:25:16.6348037Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6348973Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6348979Z 2025-12-04T12:25:16.6349250Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6349431Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6349624Z ================== 1 failed, 52 deselected, 2 rerun in 5.05s =================== 2025-12-04T12:25:16.6349731Z Got exit code 1 2025-12-04T12:25:16.6349834Z Retrying single test... 2025-12-04T12:25:16.6350473Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-cc69c969909d9707.xml 2025-12-04T12:25:16.6350645Z ============================= test session starts ============================== 2025-12-04T12:25:16.6350989Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6351104Z cachedir: .pytest_cache 2025-12-04T12:25:16.6351620Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6351739Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6351878Z configfile: pytest.ini 2025-12-04T12:25:16.6352469Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6352692Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.6353734Z stepcurrent: skipping 52 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6353844Z Running 1 items in this shard 2025-12-04T12:25:16.6353849Z 2025-12-04T12:25:16.6354771Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6847s] [100%] 2025-12-04T12:25:16.6355684Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1656s] [100%] 2025-12-04T12:25:16.6356548Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1601s] [100%] 2025-12-04T12:25:16.6356554Z 2025-12-04T12:25:16.6356719Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6357273Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6357403Z Traceback (most recent call last): 2025-12-04T12:25:16.6357878Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6358079Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6358317Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6358325Z 2025-12-04T12:25:16.6358534Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6359489Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6359496Z 2025-12-04T12:25:16.6359759Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6359980Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6360089Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6360201Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6360545Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6360763Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6360856Z graph_break [] 2025-12-04T12:25:16.6361080Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6361816Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6361919Z warnings.warn( 2025-12-04T12:25:16.6362481Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6362597Z Traceback (most recent call last): 2025-12-04T12:25:16.6363072Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6363261Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6363464Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6363481Z 2025-12-04T12:25:16.6363697Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6364667Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6364674Z 2025-12-04T12:25:16.6364947Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6365161Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6365280Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6365391Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6365727Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6365944Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6366036Z graph_break [] 2025-12-04T12:25:16.6366245Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6366984Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6367114Z warnings.warn( 2025-12-04T12:25:16.6367334Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6367441Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6367577Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6367802Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6368135Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6368227Z graph_break [] 2025-12-04T12:25:16.6368446Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6369212Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6369321Z warnings.warn( 2025-12-04T12:25:16.6369471Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6370032Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6370160Z Traceback (most recent call last): 2025-12-04T12:25:16.6370628Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6370827Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6371048Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6371053Z 2025-12-04T12:25:16.6371265Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6372229Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6372237Z 2025-12-04T12:25:16.6372501Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6372721Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6372842Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6372959Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6373304Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6373519Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6373615Z graph_break [] 2025-12-04T12:25:16.6373842Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6374572Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6374672Z warnings.warn( 2025-12-04T12:25:16.6374927Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6375037Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6375164Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6375381Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6375719Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6375827Z graph_break [] 2025-12-04T12:25:16.6376039Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6376763Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6376939Z warnings.warn( 2025-12-04T12:25:16.6377157Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6377274Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6377388Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6377605Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6377985Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6378084Z graph_break [] 2025-12-04T12:25:16.6378329Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6379063Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6379162Z warnings.warn( 2025-12-04T12:25:16.6379990Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-cc69c969909d9707.xml - 2025-12-04T12:25:16.6380187Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6381271Z FAILED [0.1601s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6381280Z 2025-12-04T12:25:16.6381506Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6382454Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6382459Z 2025-12-04T12:25:16.6382736Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6382913Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6383113Z ================== 1 failed, 204 deselected, 2 rerun in 5.06s ================== 2025-12-04T12:25:16.6383222Z Got exit code 1 2025-12-04T12:25:16.6383325Z Retrying single test... 2025-12-04T12:25:16.6383976Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9977f3a62a4ac5d4.xml 2025-12-04T12:25:16.6384136Z ============================= test session starts ============================== 2025-12-04T12:25:16.6384483Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6384603Z cachedir: .pytest_cache 2025-12-04T12:25:16.6385121Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6385241Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6385356Z configfile: pytest.ini 2025-12-04T12:25:16.6385951Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6386189Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.6387248Z stepcurrent: skipping 52 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6387362Z Running 1 items in this shard 2025-12-04T12:25:16.6387368Z 2025-12-04T12:25:16.6388294Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6646s] [100%] 2025-12-04T12:25:16.6389206Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1710s] [100%] 2025-12-04T12:25:16.6390048Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1620s] [100%] 2025-12-04T12:25:16.6390085Z 2025-12-04T12:25:16.6390219Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6390790Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6390936Z Traceback (most recent call last): 2025-12-04T12:25:16.6391407Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6391613Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6391820Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6391826Z 2025-12-04T12:25:16.6392048Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6393029Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6393038Z 2025-12-04T12:25:16.6393299Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6393526Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6393637Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6393761Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6394099Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6394315Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6394425Z graph_break [] 2025-12-04T12:25:16.6394636Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6395373Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6395486Z warnings.warn( 2025-12-04T12:25:16.6396237Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6396370Z Traceback (most recent call last): 2025-12-04T12:25:16.6396838Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6397031Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6397250Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6397255Z 2025-12-04T12:25:16.6397464Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6398422Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6398490Z 2025-12-04T12:25:16.6398752Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6398970Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6399090Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6399201Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6399544Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6399774Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6399870Z graph_break [] 2025-12-04T12:25:16.6400098Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6400830Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6400933Z warnings.warn( 2025-12-04T12:25:16.6401155Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6401264Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6401419Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6401649Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6401983Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6402127Z graph_break [] 2025-12-04T12:25:16.6402342Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6403067Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6403172Z warnings.warn( 2025-12-04T12:25:16.6403313Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6403909Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6404042Z Traceback (most recent call last): 2025-12-04T12:25:16.6404513Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6404718Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6404928Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6404933Z 2025-12-04T12:25:16.6405144Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6406106Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6406111Z 2025-12-04T12:25:16.6406374Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6406601Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6406712Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6406825Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6407172Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6407386Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6407485Z graph_break [] 2025-12-04T12:25:16.6407713Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6408438Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6408548Z warnings.warn( 2025-12-04T12:25:16.6408758Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6408864Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6408986Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6409238Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6409573Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6409685Z graph_break [] 2025-12-04T12:25:16.6409898Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6410633Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6410733Z warnings.warn( 2025-12-04T12:25:16.6410944Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6411063Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6411174Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6411392Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6411737Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6411835Z graph_break [] 2025-12-04T12:25:16.6412057Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6412817Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6412961Z warnings.warn( 2025-12-04T12:25:16.6413794Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9977f3a62a4ac5d4.xml - 2025-12-04T12:25:16.6413960Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6415098Z FAILED [0.1620s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6415108Z 2025-12-04T12:25:16.6415324Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6416279Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6416292Z 2025-12-04T12:25:16.6416559Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6416743Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6417013Z ================== 1 failed, 204 deselected, 2 rerun in 5.05s ================== 2025-12-04T12:25:16.6417110Z Got exit code 1 2025-12-04T12:25:16.6417971Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6418388Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.6419030Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ec39337b11cd8a4d.xml 2025-12-04T12:25:16.6419202Z ============================= test session starts ============================== 2025-12-04T12:25:16.6419555Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6419663Z cachedir: .pytest_cache 2025-12-04T12:25:16.6420190Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6420315Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6420429Z configfile: pytest.ini 2025-12-04T12:25:16.6421023Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6421283Z collecting ... collected 380 items / 53 deselected / 327 selected 2025-12-04T12:25:16.6421440Z stepcurrent: skipping 53 already run items. 2025-12-04T12:25:16.6421552Z Running 152 items in this shard 2025-12-04T12:25:16.6421557Z 2025-12-04T12:25:16.6422471Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [4.6584s] [ 0%] 2025-12-04T12:25:16.6423382Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [0.1703s] [ 0%] 2025-12-04T12:25:16.6424206Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True FAILED [0.1669s] [ 0%] 2025-12-04T12:25:16.6424213Z 2025-12-04T12:25:16.6424363Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6424917Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.6425091Z Traceback (most recent call last): 2025-12-04T12:25:16.6425583Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6425784Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6426002Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6426007Z 2025-12-04T12:25:16.6426218Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6427208Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.6427213Z 2025-12-04T12:25:16.6427479Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6427700Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6427819Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6427932Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6428272Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6428500Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6428599Z graph_break [] 2025-12-04T12:25:16.6428823Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6431533Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6431653Z return x.grad, w.grad 2025-12-04T12:25:16.6432381Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6432480Z warnings.warn( 2025-12-04T12:25:16.6435198Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6435335Z return x.grad, w.grad 2025-12-04T12:25:16.6435905Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.6436022Z Traceback (most recent call last): 2025-12-04T12:25:16.6436489Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6436695Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6436902Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6436907Z 2025-12-04T12:25:16.6437126Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6438071Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.6438108Z 2025-12-04T12:25:16.6438385Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6438596Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6438735Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6438868Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6439206Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6439423Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6439530Z graph_break [] 2025-12-04T12:25:16.6439748Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6442498Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6442608Z return x.grad, w.grad 2025-12-04T12:25:16.6443339Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6443448Z warnings.warn( 2025-12-04T12:25:16.6446149Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6446274Z return x.grad, w.grad 2025-12-04T12:25:16.6446489Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6446608Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6446723Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6446943Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6447295Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6447393Z graph_break [] 2025-12-04T12:25:16.6447607Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6450330Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6450439Z return x.grad, w.grad 2025-12-04T12:25:16.6451178Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6451280Z warnings.warn( 2025-12-04T12:25:16.6454604Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6454749Z return x.grad, w.grad 2025-12-04T12:25:16.6454908Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6455468Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.6455617Z Traceback (most recent call last): 2025-12-04T12:25:16.6456103Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6456304Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6456520Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6456535Z 2025-12-04T12:25:16.6456907Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6457875Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.6457882Z 2025-12-04T12:25:16.6458149Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6458364Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6458486Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6458605Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6458948Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6459179Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6459273Z graph_break [] 2025-12-04T12:25:16.6459489Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6462205Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6467883Z return x.grad, w.grad 2025-12-04T12:25:16.6468738Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6468837Z warnings.warn( 2025-12-04T12:25:16.6471553Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6471672Z return x.grad, w.grad 2025-12-04T12:25:16.6471903Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6472052Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6472176Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6472401Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6472742Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6472892Z graph_break [] 2025-12-04T12:25:16.6473110Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6475863Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6475972Z return x.grad, w.grad 2025-12-04T12:25:16.6476706Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6476817Z warnings.warn( 2025-12-04T12:25:16.6479515Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6479636Z return x.grad, w.grad 2025-12-04T12:25:16.6479856Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6479978Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6480089Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6480315Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6480665Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6480761Z graph_break [] 2025-12-04T12:25:16.6480975Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6481717Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6481817Z warnings.warn( 2025-12-04T12:25:16.6484524Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6484736Z return x.grad, w.grad 2025-12-04T12:25:16.6485568Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ec39337b11cd8a4d.xml - 2025-12-04T12:25:16.6485739Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6486817Z FAILED [0.1669s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6486830Z 2025-12-04T12:25:16.6487055Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6488036Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.6488043Z 2025-12-04T12:25:16.6488398Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6488598Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6488794Z ================== 1 failed, 53 deselected, 2 rerun in 5.05s =================== 2025-12-04T12:25:16.6488906Z Got exit code 1 2025-12-04T12:25:16.6489047Z Retrying single test... 2025-12-04T12:25:16.6489701Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f9d493fa40deaa3c.xml 2025-12-04T12:25:16.6489866Z ============================= test session starts ============================== 2025-12-04T12:25:16.6490213Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6490332Z cachedir: .pytest_cache 2025-12-04T12:25:16.6490854Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6490975Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6491091Z configfile: pytest.ini 2025-12-04T12:25:16.6491679Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6491919Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.6492947Z stepcurrent: skipping 53 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.6493065Z Running 1 items in this shard 2025-12-04T12:25:16.6493071Z 2025-12-04T12:25:16.6493994Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [4.6712s] [100%] 2025-12-04T12:25:16.6494896Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [0.1688s] [100%] 2025-12-04T12:25:16.6495730Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True FAILED [0.1639s] [100%] 2025-12-04T12:25:16.6495813Z 2025-12-04T12:25:16.6495955Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6496699Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.6496819Z Traceback (most recent call last): 2025-12-04T12:25:16.6497375Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6497587Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6497798Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6497804Z 2025-12-04T12:25:16.6498028Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6499045Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.6499061Z 2025-12-04T12:25:16.6499414Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6499647Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6499756Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6499868Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6500295Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6500514Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6500622Z graph_break [] 2025-12-04T12:25:16.6500840Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6503605Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6503735Z return x.grad, w.grad 2025-12-04T12:25:16.6504469Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6504579Z warnings.warn( 2025-12-04T12:25:16.6507287Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6507406Z return x.grad, w.grad 2025-12-04T12:25:16.6507969Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.6508100Z Traceback (most recent call last): 2025-12-04T12:25:16.6508567Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6508765Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6508986Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6508992Z 2025-12-04T12:25:16.6509206Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6510213Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.6510271Z 2025-12-04T12:25:16.6510535Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6510752Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6510872Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6510985Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6511327Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6511556Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6511656Z graph_break [] 2025-12-04T12:25:16.6511884Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6514615Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6514733Z return x.grad, w.grad 2025-12-04T12:25:16.6515463Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6515562Z warnings.warn( 2025-12-04T12:25:16.6518324Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6518435Z return x.grad, w.grad 2025-12-04T12:25:16.6518663Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6518771Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6518884Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6519112Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6519451Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6519551Z graph_break [] 2025-12-04T12:25:16.6519779Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6522473Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6522590Z return x.grad, w.grad 2025-12-04T12:25:16.6523318Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6523429Z warnings.warn( 2025-12-04T12:25:16.6526173Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6526290Z return x.grad, w.grad 2025-12-04T12:25:16.6526434Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6526991Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.6527120Z Traceback (most recent call last): 2025-12-04T12:25:16.6527588Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6527798Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6528005Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6528011Z 2025-12-04T12:25:16.6528252Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6529210Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.6529216Z 2025-12-04T12:25:16.6529480Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6529705Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6529844Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6529962Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6530307Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6530525Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6530622Z graph_break [] 2025-12-04T12:25:16.6530848Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6533563Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6533675Z return x.grad, w.grad 2025-12-04T12:25:16.6534405Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6534519Z warnings.warn( 2025-12-04T12:25:16.6537308Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6537433Z return x.grad, w.grad 2025-12-04T12:25:16.6537724Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6537850Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6537966Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6538189Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6538554Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6538654Z graph_break [] 2025-12-04T12:25:16.6538869Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6541582Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6541693Z return x.grad, w.grad 2025-12-04T12:25:16.6542473Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6542575Z warnings.warn( 2025-12-04T12:25:16.6545307Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6545419Z return x.grad, w.grad 2025-12-04T12:25:16.6545647Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6545757Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6545891Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6546132Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6546474Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6546572Z graph_break [] 2025-12-04T12:25:16.6546801Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6547534Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6547646Z warnings.warn( 2025-12-04T12:25:16.6550353Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6550471Z return x.grad, w.grad 2025-12-04T12:25:16.6551298Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f9d493fa40deaa3c.xml - 2025-12-04T12:25:16.6551471Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6552551Z FAILED [0.1639s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6552616Z 2025-12-04T12:25:16.6552833Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6553789Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.6553795Z 2025-12-04T12:25:16.6554062Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6554242Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6554452Z ================== 1 failed, 204 deselected, 2 rerun in 5.06s ================== 2025-12-04T12:25:16.6554556Z Got exit code 1 2025-12-04T12:25:16.6554674Z Retrying single test... 2025-12-04T12:25:16.6555322Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-492f33e438c88cbb.xml 2025-12-04T12:25:16.6555484Z ============================= test session starts ============================== 2025-12-04T12:25:16.6555872Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6555981Z cachedir: .pytest_cache 2025-12-04T12:25:16.6556504Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6556641Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6556747Z configfile: pytest.ini 2025-12-04T12:25:16.6557344Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6557599Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.6558639Z stepcurrent: skipping 53 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.6558762Z Running 1 items in this shard 2025-12-04T12:25:16.6558768Z 2025-12-04T12:25:16.6559676Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [4.6675s] [100%] 2025-12-04T12:25:16.6560594Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [0.1663s] [100%] 2025-12-04T12:25:16.6561421Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True FAILED [0.1638s] [100%] 2025-12-04T12:25:16.6561431Z 2025-12-04T12:25:16.6561585Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6562142Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.6562269Z Traceback (most recent call last): 2025-12-04T12:25:16.6562752Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6562948Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6563159Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6563173Z 2025-12-04T12:25:16.6563384Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6564330Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.6564397Z 2025-12-04T12:25:16.6564678Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6564900Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6565027Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6565145Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6565488Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6565715Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6565813Z graph_break [] 2025-12-04T12:25:16.6566027Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6568797Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6568911Z return x.grad, w.grad 2025-12-04T12:25:16.6569661Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6569761Z warnings.warn( 2025-12-04T12:25:16.6572499Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6572610Z return x.grad, w.grad 2025-12-04T12:25:16.6573169Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.6573302Z Traceback (most recent call last): 2025-12-04T12:25:16.6573772Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6573977Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6574187Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6574195Z 2025-12-04T12:25:16.6574406Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6575366Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.6575372Z 2025-12-04T12:25:16.6575641Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6575869Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6575983Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6576098Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6576448Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6576665Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6576774Z graph_break [] 2025-12-04T12:25:16.6577066Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6579857Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6579980Z return x.grad, w.grad 2025-12-04T12:25:16.6580713Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6580826Z warnings.warn( 2025-12-04T12:25:16.6583546Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6583673Z return x.grad, w.grad 2025-12-04T12:25:16.6583890Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6584003Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6584132Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6584354Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6584720Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6584834Z graph_break [] 2025-12-04T12:25:16.6585050Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6587747Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6587857Z return x.grad, w.grad 2025-12-04T12:25:16.6588606Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6588711Z warnings.warn( 2025-12-04T12:25:16.6591394Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6591513Z return x.grad, w.grad 2025-12-04T12:25:16.6591658Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6592230Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.6592411Z Traceback (most recent call last): 2025-12-04T12:25:16.6592884Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6593094Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6593306Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6593313Z 2025-12-04T12:25:16.6593539Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6594485Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.6594490Z 2025-12-04T12:25:16.6594769Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6594989Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6595103Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6595231Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6595571Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6595789Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6595899Z graph_break [] 2025-12-04T12:25:16.6596337Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6599109Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6599223Z return x.grad, w.grad 2025-12-04T12:25:16.6599953Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6600067Z warnings.warn( 2025-12-04T12:25:16.6602760Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6602881Z return x.grad, w.grad 2025-12-04T12:25:16.6603097Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6603218Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6603330Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6603553Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6603919Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6604056Z graph_break [] 2025-12-04T12:25:16.6604320Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6607025Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6607233Z return x.grad, w.grad 2025-12-04T12:25:16.6607983Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6608085Z warnings.warn( 2025-12-04T12:25:16.6610794Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6610903Z return x.grad, w.grad 2025-12-04T12:25:16.6611129Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6611239Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6611384Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6611620Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6611961Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6612058Z graph_break [] 2025-12-04T12:25:16.6612290Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6613057Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6613174Z warnings.warn( 2025-12-04T12:25:16.6615871Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6615991Z return x.grad, w.grad 2025-12-04T12:25:16.6616817Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-492f33e438c88cbb.xml - 2025-12-04T12:25:16.6617051Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6618153Z FAILED [0.1638s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6618159Z 2025-12-04T12:25:16.6618381Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6619344Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.6619349Z 2025-12-04T12:25:16.6619614Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6619791Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6620006Z ================== 1 failed, 204 deselected, 2 rerun in 5.05s ================== 2025-12-04T12:25:16.6620194Z Got exit code 1 2025-12-04T12:25:16.6621071Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.6621485Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.6622131Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-cfe26d7f4e2f142c.xml 2025-12-04T12:25:16.6622305Z ============================= test session starts ============================== 2025-12-04T12:25:16.6622655Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6622774Z cachedir: .pytest_cache 2025-12-04T12:25:16.6623294Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6623424Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6623543Z configfile: pytest.ini 2025-12-04T12:25:16.6624136Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6624364Z collecting ... collected 380 items / 54 deselected / 326 selected 2025-12-04T12:25:16.6624551Z stepcurrent: skipping 54 already run items. 2025-12-04T12:25:16.6624667Z Running 151 items in this shard 2025-12-04T12:25:16.6624673Z 2025-12-04T12:25:16.6625712Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_True_initial_xblock_2_add_1dim_True SKIPPED [0.0039s] (Skip non-critical tests to save resources.) [ 0%] 2025-12-04T12:25:16.6626651Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6274s] [ 1%] 2025-12-04T12:25:16.6627580Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1651s] [ 1%] 2025-12-04T12:25:16.6628416Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1615s] [ 1%] 2025-12-04T12:25:16.6628422Z 2025-12-04T12:25:16.6628563Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6629134Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6629260Z Traceback (most recent call last): 2025-12-04T12:25:16.6629748Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6629945Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6630164Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6630170Z 2025-12-04T12:25:16.6630399Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6631354Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6631359Z 2025-12-04T12:25:16.6631639Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6631859Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6631971Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6632103Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6632449Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6632733Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6632850Z graph_break [] 2025-12-04T12:25:16.6633067Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6633819Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6633921Z warnings.warn( 2025-12-04T12:25:16.6634481Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6634621Z Traceback (most recent call last): 2025-12-04T12:25:16.6635088Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6635302Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6635514Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6635523Z 2025-12-04T12:25:16.6635737Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6636707Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6636740Z 2025-12-04T12:25:16.6637011Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6637245Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6637360Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6637474Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6637829Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6638078Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6638179Z graph_break [] 2025-12-04T12:25:16.6638411Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6639142Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6639264Z warnings.warn( 2025-12-04T12:25:16.6639484Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6639594Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6639727Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6639947Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6640284Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6640392Z graph_break [] 2025-12-04T12:25:16.6640607Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6641349Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6641450Z warnings.warn( 2025-12-04T12:25:16.6641595Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6642168Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6642292Z Traceback (most recent call last): 2025-12-04T12:25:16.6642762Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6642967Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6643175Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6643181Z 2025-12-04T12:25:16.6643406Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6644359Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6644427Z 2025-12-04T12:25:16.6644693Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6644918Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6645030Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6645154Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6645492Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6645710Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6645816Z graph_break [] 2025-12-04T12:25:16.6646030Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6646760Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6646876Z warnings.warn( 2025-12-04T12:25:16.6647090Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6647211Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6647323Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6647568Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6647914Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6648010Z graph_break [] 2025-12-04T12:25:16.6648223Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6648963Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6649090Z warnings.warn( 2025-12-04T12:25:16.6649319Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6649426Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6649541Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6649771Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6650115Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6650214Z graph_break [] 2025-12-04T12:25:16.6650440Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6651167Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6651277Z warnings.warn( 2025-12-04T12:25:16.6652103Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-cfe26d7f4e2f142c.xml - 2025-12-04T12:25:16.6652275Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6653376Z FAILED [0.1615s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6653384Z 2025-12-04T12:25:16.6653599Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6654558Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6654564Z 2025-12-04T12:25:16.6654826Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6655005Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6655296Z ============= 1 failed, 1 skipped, 54 deselected, 2 rerun in 5.01s ============= 2025-12-04T12:25:16.6655394Z Got exit code 1 2025-12-04T12:25:16.6655512Z Retrying single test... 2025-12-04T12:25:16.6656154Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-85dd835c0392b89d.xml 2025-12-04T12:25:16.6656322Z ============================= test session starts ============================== 2025-12-04T12:25:16.6656685Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6656794Z cachedir: .pytest_cache 2025-12-04T12:25:16.6657404Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6657533Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6657642Z configfile: pytest.ini 2025-12-04T12:25:16.6658255Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6658489Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.6659556Z stepcurrent: skipping 55 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6659687Z Running 1 items in this shard 2025-12-04T12:25:16.6659693Z 2025-12-04T12:25:16.6660609Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6447s] [100%] 2025-12-04T12:25:16.6661561Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1639s] [100%] 2025-12-04T12:25:16.6662400Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1613s] [100%] 2025-12-04T12:25:16.6662406Z 2025-12-04T12:25:16.6662558Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6663121Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6663241Z Traceback (most recent call last): 2025-12-04T12:25:16.6663722Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6663919Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6664142Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6664147Z 2025-12-04T12:25:16.6664361Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6665312Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6665318Z 2025-12-04T12:25:16.6665592Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6665810Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6665935Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6666050Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6666389Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6666620Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6666719Z graph_break [] 2025-12-04T12:25:16.6666937Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6667761Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6667864Z warnings.warn( 2025-12-04T12:25:16.6668439Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6668562Z Traceback (most recent call last): 2025-12-04T12:25:16.6669031Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6669239Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6669448Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6669453Z 2025-12-04T12:25:16.6669666Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6670629Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6670639Z 2025-12-04T12:25:16.6670904Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6671135Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6671274Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6671389Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6671740Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6671958Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6672066Z graph_break [] 2025-12-04T12:25:16.6672281Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6673039Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6673156Z warnings.warn( 2025-12-04T12:25:16.6673368Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6673478Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6673603Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6673823Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6674172Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6674270Z graph_break [] 2025-12-04T12:25:16.6674483Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6675223Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6675322Z warnings.warn( 2025-12-04T12:25:16.6675467Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6676046Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6676170Z Traceback (most recent call last): 2025-12-04T12:25:16.6676652Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6676850Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6677060Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6677065Z 2025-12-04T12:25:16.6677287Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6678233Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6678241Z 2025-12-04T12:25:16.6678521Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6678801Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6678912Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6679040Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6679383Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6679612Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6679709Z graph_break [] 2025-12-04T12:25:16.6679924Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6680674Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6680773Z warnings.warn( 2025-12-04T12:25:16.6680990Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6681117Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6681250Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6681468Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6681816Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6681914Z graph_break [] 2025-12-04T12:25:16.6682164Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6682890Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6682990Z warnings.warn( 2025-12-04T12:25:16.6683212Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6683321Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6683464Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6683696Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6684040Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6684148Z graph_break [] 2025-12-04T12:25:16.6684364Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6685088Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6685198Z warnings.warn( 2025-12-04T12:25:16.6686009Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-85dd835c0392b89d.xml - 2025-12-04T12:25:16.6686182Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6687266Z FAILED [0.1613s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6687276Z 2025-12-04T12:25:16.6687494Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6688458Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6688464Z 2025-12-04T12:25:16.6688730Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6688923Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6689124Z ================== 1 failed, 204 deselected, 2 rerun in 5.02s ================== 2025-12-04T12:25:16.6689224Z Got exit code 1 2025-12-04T12:25:16.6689350Z Retrying single test... 2025-12-04T12:25:16.6690005Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d745daffa2211db3.xml 2025-12-04T12:25:16.6690260Z ============================= test session starts ============================== 2025-12-04T12:25:16.6690610Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6690721Z cachedir: .pytest_cache 2025-12-04T12:25:16.6691264Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6691390Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6691499Z configfile: pytest.ini 2025-12-04T12:25:16.6692104Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6692332Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.6693387Z stepcurrent: skipping 55 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6693509Z Running 1 items in this shard 2025-12-04T12:25:16.6693514Z 2025-12-04T12:25:16.6694463Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6350s] [100%] 2025-12-04T12:25:16.6695395Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1622s] [100%] 2025-12-04T12:25:16.6696490Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1611s] [100%] 2025-12-04T12:25:16.6696496Z 2025-12-04T12:25:16.6696657Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6697277Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6697419Z Traceback (most recent call last): 2025-12-04T12:25:16.6697892Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6698092Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6698320Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6698325Z 2025-12-04T12:25:16.6698539Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6699505Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6699515Z 2025-12-04T12:25:16.6699783Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6700003Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6700133Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6700250Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6700595Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6700827Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6700926Z graph_break [] 2025-12-04T12:25:16.6701155Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6701892Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6701994Z warnings.warn( 2025-12-04T12:25:16.6702572Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6702784Z Traceback (most recent call last): 2025-12-04T12:25:16.6703273Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6703473Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6703684Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6703689Z 2025-12-04T12:25:16.6703913Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6704866Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6704871Z 2025-12-04T12:25:16.6705149Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6705370Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6705480Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6705607Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6705950Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6706211Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6706321Z graph_break [] 2025-12-04T12:25:16.6706536Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6707281Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6707382Z warnings.warn( 2025-12-04T12:25:16.6707595Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6707766Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6707885Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6708104Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6708453Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6708551Z graph_break [] 2025-12-04T12:25:16.6708768Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6709514Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6709617Z warnings.warn( 2025-12-04T12:25:16.6709771Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6710329Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6710454Z Traceback (most recent call last): 2025-12-04T12:25:16.6710941Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6711137Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6711362Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6711368Z 2025-12-04T12:25:16.6711585Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6712527Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6712532Z 2025-12-04T12:25:16.6712811Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6713027Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6713153Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6713267Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6713664Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6713893Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6713991Z graph_break [] 2025-12-04T12:25:16.6714206Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6714941Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6715040Z warnings.warn( 2025-12-04T12:25:16.6715264Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6715373Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6715485Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6715716Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6716051Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6716152Z graph_break [] 2025-12-04T12:25:16.6716380Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6717133Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6717246Z warnings.warn( 2025-12-04T12:25:16.6717459Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6717568Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6717694Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6717913Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6718247Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6718385Z graph_break [] 2025-12-04T12:25:16.6718600Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6719340Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6719440Z warnings.warn( 2025-12-04T12:25:16.6720265Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d745daffa2211db3.xml - 2025-12-04T12:25:16.6720446Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6721535Z FAILED [0.1611s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6721540Z 2025-12-04T12:25:16.6721769Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6722721Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6722726Z 2025-12-04T12:25:16.6722989Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6723181Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6723380Z ================== 1 failed, 204 deselected, 2 rerun in 5.01s ================== 2025-12-04T12:25:16.6723491Z Got exit code 1 2025-12-04T12:25:16.6724357Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6724771Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.6725420Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-29d661e244421a0d.xml 2025-12-04T12:25:16.6725643Z ============================= test session starts ============================== 2025-12-04T12:25:16.6726005Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6726116Z cachedir: .pytest_cache 2025-12-04T12:25:16.6726636Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6726774Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6726884Z configfile: pytest.ini 2025-12-04T12:25:16.6727476Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6727715Z collecting ... collected 380 items / 56 deselected / 324 selected 2025-12-04T12:25:16.6727860Z stepcurrent: skipping 56 already run items. 2025-12-04T12:25:16.6727991Z Running 149 items in this shard 2025-12-04T12:25:16.6727996Z 2025-12-04T12:25:16.6729033Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_True SKIPPED [0.0041s] (Skip non-critical tests to save resources.) [ 0%] 2025-12-04T12:25:16.6730078Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_True_initial_xblock_1_add_1dim_True SKIPPED [0.0031s] (Skip non-critical tests to save resources.) [ 1%] 2025-12-04T12:25:16.6731108Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_True_initial_xblock_2_add_1dim_False SKIPPED [0.0037s] (Skip non-critical tests to save resources.) [ 2%] 2025-12-04T12:25:16.6732045Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6357s] [ 2%] 2025-12-04T12:25:16.6732969Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1613s] [ 2%] 2025-12-04T12:25:16.6733800Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1589s] [ 2%] 2025-12-04T12:25:16.6733806Z 2025-12-04T12:25:16.6733957Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6734516Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6734650Z Traceback (most recent call last): 2025-12-04T12:25:16.6735122Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6735322Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6735544Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6735549Z 2025-12-04T12:25:16.6735763Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6736715Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6736733Z 2025-12-04T12:25:16.6737086Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6737305Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6737429Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6737546Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6737953Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6738186Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6738283Z graph_break [] 2025-12-04T12:25:16.6738514Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6739256Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6739357Z warnings.warn( 2025-12-04T12:25:16.6739928Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6740049Z Traceback (most recent call last): 2025-12-04T12:25:16.6740520Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6740728Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6740943Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6740948Z 2025-12-04T12:25:16.6741171Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6742152Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6742158Z 2025-12-04T12:25:16.6742424Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6742652Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6742764Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6742891Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6743257Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6743478Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6743592Z graph_break [] 2025-12-04T12:25:16.6743806Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6744542Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6744657Z warnings.warn( 2025-12-04T12:25:16.6744872Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6744991Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6745103Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6745321Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6745674Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6745774Z graph_break [] 2025-12-04T12:25:16.6745986Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6746736Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6746835Z warnings.warn( 2025-12-04T12:25:16.6746990Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6747554Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6747673Z Traceback (most recent call last): 2025-12-04T12:25:16.6748153Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6748347Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6748556Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6748576Z 2025-12-04T12:25:16.6748787Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6749791Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6749796Z 2025-12-04T12:25:16.6750075Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6750290Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6750416Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6750532Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6750870Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6751103Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6751203Z graph_break [] 2025-12-04T12:25:16.6751420Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6752165Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6752267Z warnings.warn( 2025-12-04T12:25:16.6752495Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6752654Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6752770Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6753003Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6753343Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6753441Z graph_break [] 2025-12-04T12:25:16.6753671Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6754431Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6754551Z warnings.warn( 2025-12-04T12:25:16.6754767Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6754881Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6755011Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6755237Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6755574Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6755690Z graph_break [] 2025-12-04T12:25:16.6755903Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6756632Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6756746Z warnings.warn( 2025-12-04T12:25:16.6757565Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-29d661e244421a0d.xml - 2025-12-04T12:25:16.6757754Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6758836Z FAILED [0.1589s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6758842Z 2025-12-04T12:25:16.6759073Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6760016Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6760021Z 2025-12-04T12:25:16.6760289Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6760543Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6760762Z ============= 1 failed, 3 skipped, 56 deselected, 2 rerun in 5.03s ============= 2025-12-04T12:25:16.6760873Z Got exit code 1 2025-12-04T12:25:16.6760981Z Retrying single test... 2025-12-04T12:25:16.6761625Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6a080fc2c141366f.xml 2025-12-04T12:25:16.6761802Z ============================= test session starts ============================== 2025-12-04T12:25:16.6762149Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6762257Z cachedir: .pytest_cache 2025-12-04T12:25:16.6762789Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6762916Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6763042Z configfile: pytest.ini 2025-12-04T12:25:16.6763634Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6763861Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.6764931Z stepcurrent: skipping 59 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6765046Z Running 1 items in this shard 2025-12-04T12:25:16.6765051Z 2025-12-04T12:25:16.6765967Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6342s] [100%] 2025-12-04T12:25:16.6766911Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1652s] [100%] 2025-12-04T12:25:16.6767746Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1635s] [100%] 2025-12-04T12:25:16.6767764Z 2025-12-04T12:25:16.6767904Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6768462Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6768594Z Traceback (most recent call last): 2025-12-04T12:25:16.6769062Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6769258Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6769482Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6769491Z 2025-12-04T12:25:16.6769702Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6770663Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6770668Z 2025-12-04T12:25:16.6770934Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6771151Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6771276Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6771389Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6771739Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6771960Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6772056Z graph_break [] 2025-12-04T12:25:16.6772341Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6773085Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6773185Z warnings.warn( 2025-12-04T12:25:16.6773761Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6773883Z Traceback (most recent call last): 2025-12-04T12:25:16.6774365Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6774565Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6774776Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6774782Z 2025-12-04T12:25:16.6775009Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6775954Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6775959Z 2025-12-04T12:25:16.6776236Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6776480Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6776592Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6776717Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6777136Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6777371Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6777469Z graph_break [] 2025-12-04T12:25:16.6777720Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6778466Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6778571Z warnings.warn( 2025-12-04T12:25:16.6778785Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6778907Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6779022Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6779245Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6779594Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6779692Z graph_break [] 2025-12-04T12:25:16.6779919Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6780653Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6780757Z warnings.warn( 2025-12-04T12:25:16.6780914Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6781476Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6781608Z Traceback (most recent call last): 2025-12-04T12:25:16.6782078Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6782277Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6782502Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6782507Z 2025-12-04T12:25:16.6782722Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6783677Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6783751Z 2025-12-04T12:25:16.6784019Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6784233Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6784357Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6784470Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6784811Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6785039Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6785136Z graph_break [] 2025-12-04T12:25:16.6785361Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6786088Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6786188Z warnings.warn( 2025-12-04T12:25:16.6786417Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6786527Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6786640Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6786870Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6787238Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6787349Z graph_break [] 2025-12-04T12:25:16.6787562Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6788287Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6788400Z warnings.warn( 2025-12-04T12:25:16.6788640Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6788751Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6788881Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6789097Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6789444Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6789541Z graph_break [] 2025-12-04T12:25:16.6789757Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6790496Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6790595Z warnings.warn( 2025-12-04T12:25:16.6791413Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6a080fc2c141366f.xml - 2025-12-04T12:25:16.6791596Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6792686Z FAILED [0.1635s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6792694Z 2025-12-04T12:25:16.6792923Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6793871Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6793876Z 2025-12-04T12:25:16.6794153Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6794332Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6794531Z ================== 1 failed, 204 deselected, 2 rerun in 5.02s ================== 2025-12-04T12:25:16.6794642Z Got exit code 1 2025-12-04T12:25:16.6794832Z Retrying single test... 2025-12-04T12:25:16.6795476Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b727a3d5cb43a100.xml 2025-12-04T12:25:16.6795653Z ============================= test session starts ============================== 2025-12-04T12:25:16.6796185Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6796308Z cachedir: .pytest_cache 2025-12-04T12:25:16.6796830Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6796954Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6797077Z configfile: pytest.ini 2025-12-04T12:25:16.6797668Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6797895Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.6798943Z stepcurrent: skipping 59 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6799057Z Running 1 items in this shard 2025-12-04T12:25:16.6799063Z 2025-12-04T12:25:16.6800043Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6284s] [100%] 2025-12-04T12:25:16.6800956Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1634s] [100%] 2025-12-04T12:25:16.6801836Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1600s] [100%] 2025-12-04T12:25:16.6801846Z 2025-12-04T12:25:16.6801988Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6802545Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6802682Z Traceback (most recent call last): 2025-12-04T12:25:16.6803156Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6803364Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6803577Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6803582Z 2025-12-04T12:25:16.6803796Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6804758Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6804767Z 2025-12-04T12:25:16.6805033Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6805262Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6805377Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6805492Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6805844Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6806063Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6806174Z graph_break [] 2025-12-04T12:25:16.6806389Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6807127Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6807326Z warnings.warn( 2025-12-04T12:25:16.6807886Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6808005Z Traceback (most recent call last): 2025-12-04T12:25:16.6808489Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6808684Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6808909Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6808914Z 2025-12-04T12:25:16.6809125Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6810075Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6810080Z 2025-12-04T12:25:16.6810366Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6810586Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6810710Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6810825Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6811189Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6811422Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6811519Z graph_break [] 2025-12-04T12:25:16.6811734Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6812478Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6812577Z warnings.warn( 2025-12-04T12:25:16.6812832Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6812946Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6813059Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6813293Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6813630Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6813728Z graph_break [] 2025-12-04T12:25:16.6813957Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6814697Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6814811Z warnings.warn( 2025-12-04T12:25:16.6814956Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6815520Z _ MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.6815660Z Traceback (most recent call last): 2025-12-04T12:25:16.6816138Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6816338Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6816564Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6816570Z 2025-12-04T12:25:16.6816782Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6817824Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6817832Z 2025-12-04T12:25:16.6818099Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6818334Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6818514Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6818628Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6818981Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6819198Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6819297Z graph_break [] 2025-12-04T12:25:16.6819531Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6820266Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6820379Z warnings.warn( 2025-12-04T12:25:16.6820593Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6820700Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6820835Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6821057Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6821399Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6821508Z graph_break [] 2025-12-04T12:25:16.6821722Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6822481Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6822595Z warnings.warn( 2025-12-04T12:25:16.6822805Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6822925Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6823038Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6823253Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6823629Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6823730Z graph_break [] 2025-12-04T12:25:16.6823939Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6824680Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6824776Z warnings.warn( 2025-12-04T12:25:16.6825610Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b727a3d5cb43a100.xml - 2025-12-04T12:25:16.6825778Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6826861Z FAILED [0.1600s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6826881Z 2025-12-04T12:25:16.6827098Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6828044Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6828049Z 2025-12-04T12:25:16.6828324Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6828503Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6828709Z ================== 1 failed, 204 deselected, 2 rerun in 5.01s ================== 2025-12-04T12:25:16.6828808Z Got exit code 1 2025-12-04T12:25:16.6829664Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.6830087Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.6830785Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9820049c2fc9b56d.xml 2025-12-04T12:25:16.6830944Z ============================= test session starts ============================== 2025-12-04T12:25:16.6831303Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6831409Z cachedir: .pytest_cache 2025-12-04T12:25:16.6831937Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6832062Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6832167Z configfile: pytest.ini 2025-12-04T12:25:16.6832773Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6833001Z collecting ... collected 380 items / 60 deselected / 320 selected 2025-12-04T12:25:16.6833155Z stepcurrent: skipping 60 already run items. 2025-12-04T12:25:16.6833266Z Running 145 items in this shard 2025-12-04T12:25:16.6833271Z 2025-12-04T12:25:16.6834326Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_True_initial_xblock_1_add_1dim_False SKIPPED [0.0040s] (Skip non-critical tests to save resources.) [ 0%] 2025-12-04T12:25:16.6835363Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_True_initial_xblock_1_add_1dim_True SKIPPED [0.0029s] (Skip non-critical tests to save resources.) [ 1%] 2025-12-04T12:25:16.6836410Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_True_initial_xblock_2_add_1dim_True SKIPPED [0.0035s] (Skip non-critical tests to save resources.) [ 2%] 2025-12-04T12:25:16.6837444Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_True_initial_xblock_1_add_1dim_False SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 2%] 2025-12-04T12:25:16.6838455Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_True_initial_xblock_1_add_1dim_True SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 3%] 2025-12-04T12:25:16.6839480Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_True_initial_xblock_2_add_1dim_True SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 4%] 2025-12-04T12:25:16.6840392Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6549s] [ 4%] 2025-12-04T12:25:16.6841304Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1610s] [ 4%] 2025-12-04T12:25:16.6842136Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1570s] [ 4%] 2025-12-04T12:25:16.6842142Z 2025-12-04T12:25:16.6842280Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6842846Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6842967Z Traceback (most recent call last): 2025-12-04T12:25:16.6843447Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6843692Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6843937Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6843942Z 2025-12-04T12:25:16.6844167Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6845119Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6845125Z 2025-12-04T12:25:16.6845400Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6845618Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6845728Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6845852Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6846192Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6846413Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6846521Z graph_break [] 2025-12-04T12:25:16.6846736Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6847507Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6847608Z warnings.warn( 2025-12-04T12:25:16.6848166Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6848298Z Traceback (most recent call last): 2025-12-04T12:25:16.6848765Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6848958Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6849208Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6849218Z 2025-12-04T12:25:16.6849430Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6850395Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6850403Z 2025-12-04T12:25:16.6850669Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6850895Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6851005Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6851120Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6851474Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6851692Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6851787Z graph_break [] 2025-12-04T12:25:16.6852014Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6852748Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6852863Z warnings.warn( 2025-12-04T12:25:16.6853077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6853185Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6853306Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6853524Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6853859Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6853966Z graph_break [] 2025-12-04T12:25:16.6854179Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6854907Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6855076Z warnings.warn( 2025-12-04T12:25:16.6855219Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6855793Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6855915Z Traceback (most recent call last): 2025-12-04T12:25:16.6856385Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6856587Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6856798Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6856803Z 2025-12-04T12:25:16.6857104Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6858054Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6858063Z 2025-12-04T12:25:16.6858327Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6858592Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6858703Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6858829Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6859166Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6859382Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6859489Z graph_break [] 2025-12-04T12:25:16.6859702Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6860463Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6860577Z warnings.warn( 2025-12-04T12:25:16.6860791Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6860906Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6861020Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6861240Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6861584Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6861681Z graph_break [] 2025-12-04T12:25:16.6861891Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6862634Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6862735Z warnings.warn( 2025-12-04T12:25:16.6862960Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6863067Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6863182Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6863411Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6863746Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6863840Z graph_break [] 2025-12-04T12:25:16.6864063Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6864785Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6864896Z warnings.warn( 2025-12-04T12:25:16.6865721Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9820049c2fc9b56d.xml - 2025-12-04T12:25:16.6865945Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6867037Z FAILED [0.1570s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6867045Z 2025-12-04T12:25:16.6867257Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6868217Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6868223Z 2025-12-04T12:25:16.6868484Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6868662Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6874406Z ============= 1 failed, 6 skipped, 60 deselected, 2 rerun in 5.05s ============= 2025-12-04T12:25:16.6874552Z Got exit code 1 2025-12-04T12:25:16.6874657Z Retrying single test... 2025-12-04T12:25:16.6875322Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-bcfbf334a5fb1b19.xml 2025-12-04T12:25:16.6875573Z ============================= test session starts ============================== 2025-12-04T12:25:16.6875937Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6876043Z cachedir: .pytest_cache 2025-12-04T12:25:16.6876561Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6876691Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6876796Z configfile: pytest.ini 2025-12-04T12:25:16.6877425Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6877667Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.6878703Z stepcurrent: skipping 66 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6878823Z Running 1 items in this shard 2025-12-04T12:25:16.6878829Z 2025-12-04T12:25:16.6879744Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6648s] [100%] 2025-12-04T12:25:16.6880658Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1623s] [100%] 2025-12-04T12:25:16.6881487Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1566s] [100%] 2025-12-04T12:25:16.6881493Z 2025-12-04T12:25:16.6881630Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6882199Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6882319Z Traceback (most recent call last): 2025-12-04T12:25:16.6882792Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6882984Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6883190Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6883196Z 2025-12-04T12:25:16.6883415Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6884451Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6884456Z 2025-12-04T12:25:16.6884725Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6884942Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6885048Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6885164Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6885499Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6885722Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6885818Z graph_break [] 2025-12-04T12:25:16.6886033Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6886779Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6886876Z warnings.warn( 2025-12-04T12:25:16.6887437Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6887593Z Traceback (most recent call last): 2025-12-04T12:25:16.6888061Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6888263Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6888469Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6888475Z 2025-12-04T12:25:16.6888682Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6889688Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6889699Z 2025-12-04T12:25:16.6889961Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6890242Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6890348Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6890464Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6890805Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6891019Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6891111Z graph_break [] 2025-12-04T12:25:16.6891329Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6892056Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6892162Z warnings.warn( 2025-12-04T12:25:16.6892369Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6892473Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6892587Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6892799Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6893130Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6893228Z graph_break [] 2025-12-04T12:25:16.6893440Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6894173Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6894266Z warnings.warn( 2025-12-04T12:25:16.6894405Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6895022Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6895138Z Traceback (most recent call last): 2025-12-04T12:25:16.6895603Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6895806Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6896200Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6896205Z 2025-12-04T12:25:16.6896423Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6897436Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6897446Z 2025-12-04T12:25:16.6897716Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6897934Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6898037Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6898154Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6898557Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6898770Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6898869Z graph_break [] 2025-12-04T12:25:16.6899081Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6899810Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6899914Z warnings.warn( 2025-12-04T12:25:16.6900166Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6900286Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6900411Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6900739Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6901084Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6901178Z graph_break [] 2025-12-04T12:25:16.6901389Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6902119Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6902215Z warnings.warn( 2025-12-04T12:25:16.6902431Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6902536Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6902650Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6902873Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6903210Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6903302Z graph_break [] 2025-12-04T12:25:16.6903521Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6904244Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6904348Z warnings.warn( 2025-12-04T12:25:16.6905167Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-bcfbf334a5fb1b19.xml - 2025-12-04T12:25:16.6905333Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6906425Z FAILED [0.1566s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6906529Z 2025-12-04T12:25:16.6906745Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6907706Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6907712Z 2025-12-04T12:25:16.6907974Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6908161Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6908354Z ================== 1 failed, 204 deselected, 2 rerun in 5.04s ================== 2025-12-04T12:25:16.6908450Z Got exit code 1 2025-12-04T12:25:16.6908566Z Retrying single test... 2025-12-04T12:25:16.6909215Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-55cd5c19b34efcf0.xml 2025-12-04T12:25:16.6909378Z ============================= test session starts ============================== 2025-12-04T12:25:16.6909736Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6909845Z cachedir: .pytest_cache 2025-12-04T12:25:16.6910399Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6910522Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6910628Z configfile: pytest.ini 2025-12-04T12:25:16.6911225Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6911449Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.6912507Z stepcurrent: skipping 66 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6912634Z Running 1 items in this shard 2025-12-04T12:25:16.6912639Z 2025-12-04T12:25:16.6913552Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6613s] [100%] 2025-12-04T12:25:16.6914465Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1627s] [100%] 2025-12-04T12:25:16.6915299Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1613s] [100%] 2025-12-04T12:25:16.6915304Z 2025-12-04T12:25:16.6915452Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6916004Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6916125Z Traceback (most recent call last): 2025-12-04T12:25:16.6916597Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6916789Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6917004Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6917010Z 2025-12-04T12:25:16.6917216Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6918163Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6918229Z 2025-12-04T12:25:16.6918506Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6918724Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6918837Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6918950Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6919291Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6919515Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6919607Z graph_break [] 2025-12-04T12:25:16.6919819Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6920565Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6920666Z warnings.warn( 2025-12-04T12:25:16.6921231Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6921354Z Traceback (most recent call last): 2025-12-04T12:25:16.6921816Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6922043Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6922254Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6922259Z 2025-12-04T12:25:16.6922473Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6923412Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6923417Z 2025-12-04T12:25:16.6923704Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6923932Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6924035Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6924161Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6924502Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6924722Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6924825Z graph_break [] 2025-12-04T12:25:16.6925036Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6925759Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6925863Z warnings.warn( 2025-12-04T12:25:16.6926074Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6926188Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6926303Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6926518Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6926859Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6926951Z graph_break [] 2025-12-04T12:25:16.6927159Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6927892Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6927991Z warnings.warn( 2025-12-04T12:25:16.6928137Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6928695Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.6928813Z Traceback (most recent call last): 2025-12-04T12:25:16.6929348Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6929540Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6929745Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6929750Z 2025-12-04T12:25:16.6929969Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6930909Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6930914Z 2025-12-04T12:25:16.6931176Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6931388Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6931496Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6931617Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6931956Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6932178Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6932270Z graph_break [] 2025-12-04T12:25:16.6932485Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6933265Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6933364Z warnings.warn( 2025-12-04T12:25:16.6933575Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6933687Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6933797Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6934047Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6934385Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6934483Z graph_break [] 2025-12-04T12:25:16.6934704Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6935432Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6935528Z warnings.warn( 2025-12-04T12:25:16.6935747Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6935857Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6935977Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6936195Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6936525Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6936630Z graph_break [] 2025-12-04T12:25:16.6936916Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6937649Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6937758Z warnings.warn( 2025-12-04T12:25:16.6938581Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-55cd5c19b34efcf0.xml - 2025-12-04T12:25:16.6938755Z =========================== short test summary info ============================ 2025-12-04T12:25:16.6939837Z FAILED [0.1613s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6939843Z 2025-12-04T12:25:16.6940058Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6941083Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6941088Z 2025-12-04T12:25:16.6941349Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6941535Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.6941733Z ================== 1 failed, 204 deselected, 2 rerun in 5.04s ================== 2025-12-04T12:25:16.6941830Z Got exit code 1 2025-12-04T12:25:16.6942705Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.6943122Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.6943770Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3f17acc66e50cb98.xml 2025-12-04T12:25:16.6943934Z ============================= test session starts ============================== 2025-12-04T12:25:16.6944281Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.6944426Z cachedir: .pytest_cache 2025-12-04T12:25:16.6944949Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.6945074Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.6945178Z configfile: pytest.ini 2025-12-04T12:25:16.6945767Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.6946028Z collecting ... collected 380 items / 67 deselected / 313 selected 2025-12-04T12:25:16.6946171Z stepcurrent: skipping 67 already run items. 2025-12-04T12:25:16.6946284Z Running 138 items in this shard 2025-12-04T12:25:16.6946289Z 2025-12-04T12:25:16.6947210Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [4.6268s] [ 0%] 2025-12-04T12:25:16.6948113Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [0.1662s] [ 0%] 2025-12-04T12:25:16.6948930Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True FAILED [0.1630s] [ 0%] 2025-12-04T12:25:16.6948935Z 2025-12-04T12:25:16.6949075Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.6949636Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.6949757Z Traceback (most recent call last): 2025-12-04T12:25:16.6950225Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6950425Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6950630Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6950635Z 2025-12-04T12:25:16.6950850Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6951792Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.6951797Z 2025-12-04T12:25:16.6952067Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6952347Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6952455Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6952570Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6952907Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6953126Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6953227Z graph_break [] 2025-12-04T12:25:16.6953440Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6956169Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6956277Z return x.grad, w.grad 2025-12-04T12:25:16.6957034Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6957136Z warnings.warn( 2025-12-04T12:25:16.6959861Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6959975Z return x.grad, w.grad 2025-12-04T12:25:16.6960534Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.6960657Z Traceback (most recent call last): 2025-12-04T12:25:16.6961123Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6961319Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6961535Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6961541Z 2025-12-04T12:25:16.6961749Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6962697Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.6962713Z 2025-12-04T12:25:16.6962974Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6963188Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6963304Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6963413Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6963751Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6963973Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6964068Z graph_break [] 2025-12-04T12:25:16.6964289Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6966999Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6967166Z return x.grad, w.grad 2025-12-04T12:25:16.6967889Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6967986Z warnings.warn( 2025-12-04T12:25:16.6970681Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6970785Z return x.grad, w.grad 2025-12-04T12:25:16.6971029Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6971134Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6971241Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6971471Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6971805Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6971899Z graph_break [] 2025-12-04T12:25:16.6972116Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6974873Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6974992Z return x.grad, w.grad 2025-12-04T12:25:16.6975715Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6975820Z warnings.warn( 2025-12-04T12:25:16.6978589Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6978709Z return x.grad, w.grad 2025-12-04T12:25:16.6978853Z =================================== FAILURES =================================== 2025-12-04T12:25:16.6979405Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.6979536Z Traceback (most recent call last): 2025-12-04T12:25:16.6980007Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.6980213Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.6980498Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.6980504Z 2025-12-04T12:25:16.6980709Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.6981663Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.6981668Z 2025-12-04T12:25:16.6981932Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.6982152Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6982261Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6982369Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6982715Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6982935Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6983028Z graph_break [] 2025-12-04T12:25:16.6983250Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6986000Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6986110Z return x.grad, w.grad 2025-12-04T12:25:16.6986867Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6986977Z warnings.warn( 2025-12-04T12:25:16.6989680Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6989792Z return x.grad, w.grad 2025-12-04T12:25:16.6990004Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6990114Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6990235Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6990455Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6990793Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6990896Z graph_break [] 2025-12-04T12:25:16.6991108Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6993820Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6993979Z return x.grad, w.grad 2025-12-04T12:25:16.6994704Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.6994808Z warnings.warn( 2025-12-04T12:25:16.6997693Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.6997808Z return x.grad, w.grad 2025-12-04T12:25:16.6998023Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.6998146Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.6998257Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.6998476Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.6998878Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.6998972Z graph_break [] 2025-12-04T12:25:16.6999184Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.6999921Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7000020Z warnings.warn( 2025-12-04T12:25:16.7002762Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7002870Z return x.grad, w.grad 2025-12-04T12:25:16.7003692Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3f17acc66e50cb98.xml - 2025-12-04T12:25:16.7003857Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7004931Z FAILED [0.1630s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7004948Z 2025-12-04T12:25:16.7005157Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7006094Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.7006100Z 2025-12-04T12:25:16.7006369Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7006541Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7006743Z ================== 1 failed, 67 deselected, 2 rerun in 5.01s =================== 2025-12-04T12:25:16.7006839Z Got exit code 1 2025-12-04T12:25:16.7006940Z Retrying single test... 2025-12-04T12:25:16.7007590Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a7cd72d0cd16fd7c.xml 2025-12-04T12:25:16.7007817Z ============================= test session starts ============================== 2025-12-04T12:25:16.7008160Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7008275Z cachedir: .pytest_cache 2025-12-04T12:25:16.7008788Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7008918Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7009021Z configfile: pytest.ini 2025-12-04T12:25:16.7009606Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7009839Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.7010873Z stepcurrent: skipping 67 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.7010987Z Running 1 items in this shard 2025-12-04T12:25:16.7011003Z 2025-12-04T12:25:16.7011944Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [4.6524s] [100%] 2025-12-04T12:25:16.7012841Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [0.1679s] [100%] 2025-12-04T12:25:16.7013667Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True FAILED [0.1643s] [100%] 2025-12-04T12:25:16.7013672Z 2025-12-04T12:25:16.7013834Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7014397Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.7014514Z Traceback (most recent call last): 2025-12-04T12:25:16.7014980Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7015188Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7015393Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7015398Z 2025-12-04T12:25:16.7015615Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7016551Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.7016556Z 2025-12-04T12:25:16.7016818Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7017102Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7017211Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7017332Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7017668Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7017885Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7018001Z graph_break [] 2025-12-04T12:25:16.7018215Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7020932Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7021096Z return x.grad, w.grad 2025-12-04T12:25:16.7021824Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7021930Z warnings.warn( 2025-12-04T12:25:16.7024625Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7024746Z return x.grad, w.grad 2025-12-04T12:25:16.7025303Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.7025465Z Traceback (most recent call last): 2025-12-04T12:25:16.7025930Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7026128Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7026344Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7026350Z 2025-12-04T12:25:16.7026561Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7027546Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.7027556Z 2025-12-04T12:25:16.7027820Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7028035Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7028155Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7028269Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7028614Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7028845Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7028943Z graph_break [] 2025-12-04T12:25:16.7029168Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7031897Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7032020Z return x.grad, w.grad 2025-12-04T12:25:16.7032746Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7032846Z warnings.warn( 2025-12-04T12:25:16.7035567Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7035742Z return x.grad, w.grad 2025-12-04T12:25:16.7035974Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7036084Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7036197Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7036422Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7036761Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7036864Z graph_break [] 2025-12-04T12:25:16.7037082Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7039809Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7039927Z return x.grad, w.grad 2025-12-04T12:25:16.7040655Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7040765Z warnings.warn( 2025-12-04T12:25:16.7043493Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7043618Z return x.grad, w.grad 2025-12-04T12:25:16.7043755Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7044306Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.7044434Z Traceback (most recent call last): 2025-12-04T12:25:16.7044899Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7045111Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7045317Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7045322Z 2025-12-04T12:25:16.7045529Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7046490Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.7046495Z 2025-12-04T12:25:16.7046761Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7046988Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7047096Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7047209Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7047557Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7047833Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7047931Z graph_break [] 2025-12-04T12:25:16.7048150Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7050853Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7050970Z return x.grad, w.grad 2025-12-04T12:25:16.7051697Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7051815Z warnings.warn( 2025-12-04T12:25:16.7054532Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7054651Z return x.grad, w.grad 2025-12-04T12:25:16.7054896Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7055005Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7055135Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7055354Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7055691Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7055797Z graph_break [] 2025-12-04T12:25:16.7056008Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7058790Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7058898Z return x.grad, w.grad 2025-12-04T12:25:16.7059639Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7059734Z warnings.warn( 2025-12-04T12:25:16.7062435Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7062589Z return x.grad, w.grad 2025-12-04T12:25:16.7062834Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7062955Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7063063Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7063281Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7063631Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7063725Z graph_break [] 2025-12-04T12:25:16.7063937Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7064677Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7064776Z warnings.warn( 2025-12-04T12:25:16.7067508Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7067616Z return x.grad, w.grad 2025-12-04T12:25:16.7068451Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a7cd72d0cd16fd7c.xml - 2025-12-04T12:25:16.7068616Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7069725Z FAILED [0.1643s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7069747Z 2025-12-04T12:25:16.7069960Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7070898Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.7070904Z 2025-12-04T12:25:16.7071177Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7071352Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7071561Z ================== 1 failed, 204 deselected, 2 rerun in 5.04s ================== 2025-12-04T12:25:16.7071655Z Got exit code 1 2025-12-04T12:25:16.7071760Z Retrying single test... 2025-12-04T12:25:16.7072417Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-45e1674e77992a96.xml 2025-12-04T12:25:16.7072579Z ============================= test session starts ============================== 2025-12-04T12:25:16.7072924Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7073039Z cachedir: .pytest_cache 2025-12-04T12:25:16.7073559Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7073691Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7073797Z configfile: pytest.ini 2025-12-04T12:25:16.7074391Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7074625Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.7075647Z stepcurrent: skipping 67 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.7075822Z Running 1 items in this shard 2025-12-04T12:25:16.7075828Z 2025-12-04T12:25:16.7076744Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [4.6676s] [100%] 2025-12-04T12:25:16.7077652Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [0.1681s] [100%] 2025-12-04T12:25:16.7078479Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True FAILED [0.1648s] [100%] 2025-12-04T12:25:16.7078487Z 2025-12-04T12:25:16.7078624Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7079192Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.7079312Z Traceback (most recent call last): 2025-12-04T12:25:16.7079812Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7080019Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7080226Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7080231Z 2025-12-04T12:25:16.7080451Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7081435Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.7081440Z 2025-12-04T12:25:16.7081712Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7081937Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7082046Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7082168Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7082508Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7082720Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7082824Z graph_break [] 2025-12-04T12:25:16.7083037Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7085755Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7085866Z return x.grad, w.grad 2025-12-04T12:25:16.7086603Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7086710Z warnings.warn( 2025-12-04T12:25:16.7089403Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7089589Z return x.grad, w.grad 2025-12-04T12:25:16.7090147Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.7090280Z Traceback (most recent call last): 2025-12-04T12:25:16.7090742Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7090938Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7091160Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7091165Z 2025-12-04T12:25:16.7091373Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7092335Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.7092345Z 2025-12-04T12:25:16.7092608Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7092821Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7092975Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7093083Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7093423Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7093646Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7093740Z graph_break [] 2025-12-04T12:25:16.7093967Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7096968Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7097107Z return x.grad, w.grad 2025-12-04T12:25:16.7097840Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7097937Z warnings.warn( 2025-12-04T12:25:16.7100656Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7100762Z return x.grad, w.grad 2025-12-04T12:25:16.7100988Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7101093Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7101205Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7101434Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7101768Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7101876Z graph_break [] 2025-12-04T12:25:16.7102088Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7104870Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7104984Z return x.grad, w.grad 2025-12-04T12:25:16.7105708Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7105820Z warnings.warn( 2025-12-04T12:25:16.7108538Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7108660Z return x.grad, w.grad 2025-12-04T12:25:16.7108799Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7109349Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.7109510Z Traceback (most recent call last): 2025-12-04T12:25:16.7109978Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7110189Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7110398Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7110403Z 2025-12-04T12:25:16.7110610Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7111568Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.7111574Z 2025-12-04T12:25:16.7111836Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7112062Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7112171Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7112281Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7112636Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7112854Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7112950Z graph_break [] 2025-12-04T12:25:16.7113173Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7115869Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7115987Z return x.grad, w.grad 2025-12-04T12:25:16.7116798Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7116909Z warnings.warn( 2025-12-04T12:25:16.7119598Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7119713Z return x.grad, w.grad 2025-12-04T12:25:16.7119930Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7120039Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7120167Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7120391Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7120728Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7120869Z graph_break [] 2025-12-04T12:25:16.7121084Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7123824Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7123937Z return x.grad, w.grad 2025-12-04T12:25:16.7124667Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7124770Z warnings.warn( 2025-12-04T12:25:16.7127467Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7127586Z return x.grad, w.grad 2025-12-04T12:25:16.7127802Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7127917Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7128030Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7128253Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7128597Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7128694Z graph_break [] 2025-12-04T12:25:16.7128916Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7129645Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7129744Z warnings.warn( 2025-12-04T12:25:16.7132441Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7132604Z return x.grad, w.grad 2025-12-04T12:25:16.7133434Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-45e1674e77992a96.xml - 2025-12-04T12:25:16.7133607Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7134702Z FAILED [0.1648s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7134712Z 2025-12-04T12:25:16.7134923Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7135912Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.7135917Z 2025-12-04T12:25:16.7136189Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7136364Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7136574Z ================== 1 failed, 204 deselected, 2 rerun in 5.05s ================== 2025-12-04T12:25:16.7136671Z Got exit code 1 2025-12-04T12:25:16.7137627Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.7138054Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.7138691Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-384ea3a6b9159dd5.xml 2025-12-04T12:25:16.7138864Z ============================= test session starts ============================== 2025-12-04T12:25:16.7139210Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7139319Z cachedir: .pytest_cache 2025-12-04T12:25:16.7139850Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7139969Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7140080Z configfile: pytest.ini 2025-12-04T12:25:16.7140684Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7140910Z collecting ... collected 380 items / 68 deselected / 312 selected 2025-12-04T12:25:16.7141060Z stepcurrent: skipping 68 already run items. 2025-12-04T12:25:16.7141170Z Running 137 items in this shard 2025-12-04T12:25:16.7141177Z 2025-12-04T12:25:16.7142203Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_True_initial_xblock_1_add_1dim_True SKIPPED [0.0041s] (Skip non-critical tests to save resources.) [ 0%] 2025-12-04T12:25:16.7143239Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_True_initial_xblock_2_add_1dim_False SKIPPED [0.0032s] (Skip non-critical tests to save resources.) [ 1%] 2025-12-04T12:25:16.7144251Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_True_initial_xblock_2_add_1dim_True SKIPPED [0.0039s] (Skip non-critical tests to save resources.) [ 2%] 2025-12-04T12:25:16.7145346Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_True SKIPPED [0.0030s] (Skip non-critical tests to save resources.) [ 2%] 2025-12-04T12:25:16.7146360Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_True SKIPPED [0.0028s] (Skip non-critical tests to save resources.) [ 3%] 2025-12-04T12:25:16.7147387Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape1_max_autotune_True_initial_xblock_1_add_1dim_False SKIPPED [0.0028s] (Skip non-critical tests to save resources.) [ 4%] 2025-12-04T12:25:16.7148396Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape1_max_autotune_True_initial_xblock_1_add_1dim_True SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 5%] 2025-12-04T12:25:16.7149453Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape1_max_autotune_True_initial_xblock_2_add_1dim_False SKIPPED [0.0029s] (Skip non-critical tests to save resources.) [ 5%] 2025-12-04T12:25:16.7150473Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_True SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 6%] 2025-12-04T12:25:16.7151526Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_True SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 7%] 2025-12-04T12:25:16.7152559Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_True_initial_xblock_1_add_1dim_False SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 8%] 2025-12-04T12:25:16.7153578Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_True_initial_xblock_1_add_1dim_True SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 8%] 2025-12-04T12:25:16.7154615Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_True_initial_xblock_2_add_1dim_False SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 9%] 2025-12-04T12:25:16.7155525Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6466s] [ 10%] 2025-12-04T12:25:16.7156447Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1622s] [ 10%] 2025-12-04T12:25:16.7157273Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1588s] [ 10%] 2025-12-04T12:25:16.7157278Z 2025-12-04T12:25:16.7157423Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7157976Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.7158097Z Traceback (most recent call last): 2025-12-04T12:25:16.7158575Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7158774Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7159052Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7159058Z 2025-12-04T12:25:16.7159268Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7160212Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7160217Z 2025-12-04T12:25:16.7160487Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7160702Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7160819Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7160926Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7161272Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7161498Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7161593Z graph_break [] 2025-12-04T12:25:16.7161806Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7162570Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7162672Z warnings.warn( 2025-12-04T12:25:16.7163245Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.7163364Z Traceback (most recent call last): 2025-12-04T12:25:16.7163832Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7164033Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7164271Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7164281Z 2025-12-04T12:25:16.7164497Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7165433Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7165439Z 2025-12-04T12:25:16.7165702Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7165928Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7166035Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7166154Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7166489Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7166707Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7166810Z graph_break [] 2025-12-04T12:25:16.7167026Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7167759Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7167868Z warnings.warn( 2025-12-04T12:25:16.7168082Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7168196Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7168309Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7168527Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7168868Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7168962Z graph_break [] 2025-12-04T12:25:16.7169171Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7169905Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7170061Z warnings.warn( 2025-12-04T12:25:16.7170212Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7170770Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.7170893Z Traceback (most recent call last): 2025-12-04T12:25:16.7171371Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7171567Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7171772Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7171777Z 2025-12-04T12:25:16.7171994Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7172939Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7172949Z 2025-12-04T12:25:16.7173220Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7173466Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7173576Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7173700Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7174040Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7174260Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7174355Z graph_break [] 2025-12-04T12:25:16.7174565Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7175325Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7175433Z warnings.warn( 2025-12-04T12:25:16.7175643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7175764Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7175876Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7176101Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7176437Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7176531Z graph_break [] 2025-12-04T12:25:16.7176752Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7177547Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7177646Z warnings.warn( 2025-12-04T12:25:16.7177873Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7177983Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7178105Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7178325Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7178665Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7178772Z graph_break [] 2025-12-04T12:25:16.7178984Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7179702Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7179810Z warnings.warn( 2025-12-04T12:25:16.7180635Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-384ea3a6b9159dd5.xml - 2025-12-04T12:25:16.7180891Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7181973Z FAILED [0.1588s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7181981Z 2025-12-04T12:25:16.7182195Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7183153Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7183158Z 2025-12-04T12:25:16.7183423Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7183608Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7183827Z ============ 1 failed, 13 skipped, 68 deselected, 2 rerun in 5.08s ============= 2025-12-04T12:25:16.7183922Z Got exit code 1 2025-12-04T12:25:16.7184039Z Retrying single test... 2025-12-04T12:25:16.7184679Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-40502088a31692e3.xml 2025-12-04T12:25:16.7184878Z ============================= test session starts ============================== 2025-12-04T12:25:16.7185227Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7185334Z cachedir: .pytest_cache 2025-12-04T12:25:16.7185862Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7185984Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7186089Z configfile: pytest.ini 2025-12-04T12:25:16.7186714Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7186944Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.7187972Z stepcurrent: skipping 81 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7188089Z Running 1 items in this shard 2025-12-04T12:25:16.7188094Z 2025-12-04T12:25:16.7189016Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6592s] [100%] 2025-12-04T12:25:16.7189923Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1605s] [100%] 2025-12-04T12:25:16.7190755Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1584s] [100%] 2025-12-04T12:25:16.7190760Z 2025-12-04T12:25:16.7190909Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7191471Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.7191599Z Traceback (most recent call last): 2025-12-04T12:25:16.7192064Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7192262Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7192478Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7192483Z 2025-12-04T12:25:16.7192694Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7193715Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7193721Z 2025-12-04T12:25:16.7193985Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7194205Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7194326Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7194440Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7194790Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7195008Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7195108Z graph_break [] 2025-12-04T12:25:16.7195338Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7196266Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7196372Z warnings.warn( 2025-12-04T12:25:16.7196948Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.7197134Z Traceback (most recent call last): 2025-12-04T12:25:16.7197615Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7197813Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7198027Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7198032Z 2025-12-04T12:25:16.7198260Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7199246Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7199256Z 2025-12-04T12:25:16.7199534Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7199759Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7199875Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7200002Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7200341Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7200560Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7200668Z graph_break [] 2025-12-04T12:25:16.7200883Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7201634Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7201740Z warnings.warn( 2025-12-04T12:25:16.7201955Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7202077Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7202191Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7202411Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7202759Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7202856Z graph_break [] 2025-12-04T12:25:16.7203080Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7203810Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7203913Z warnings.warn( 2025-12-04T12:25:16.7204072Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7204707Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.7204825Z Traceback (most recent call last): 2025-12-04T12:25:16.7205307Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7205506Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7205729Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7205734Z 2025-12-04T12:25:16.7205948Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7206893Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7206912Z 2025-12-04T12:25:16.7207175Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7207391Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7207513Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7207626Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7207963Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7208224Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7208320Z graph_break [] 2025-12-04T12:25:16.7208530Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7209265Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7209365Z warnings.warn( 2025-12-04T12:25:16.7209615Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7209727Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7209835Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7210066Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7210399Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7210492Z graph_break [] 2025-12-04T12:25:16.7210715Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7211441Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7211548Z warnings.warn( 2025-12-04T12:25:16.7211758Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7211865Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7211990Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7212205Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7212543Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7212652Z graph_break [] 2025-12-04T12:25:16.7212865Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7213604Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7213703Z warnings.warn( 2025-12-04T12:25:16.7214516Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-40502088a31692e3.xml - 2025-12-04T12:25:16.7214698Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7215774Z FAILED [0.1584s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7215840Z 2025-12-04T12:25:16.7216070Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7217079Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7217086Z 2025-12-04T12:25:16.7217362Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7217540Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7217741Z ================== 1 failed, 204 deselected, 2 rerun in 5.03s ================== 2025-12-04T12:25:16.7217851Z Got exit code 1 2025-12-04T12:25:16.7217958Z Retrying single test... 2025-12-04T12:25:16.7218600Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-1a1d3c182ee91d42.xml 2025-12-04T12:25:16.7218778Z ============================= test session starts ============================== 2025-12-04T12:25:16.7219126Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7219246Z cachedir: .pytest_cache 2025-12-04T12:25:16.7219801Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7219921Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7220041Z configfile: pytest.ini 2025-12-04T12:25:16.7220636Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7220862Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.7221939Z stepcurrent: skipping 81 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7222057Z Running 1 items in this shard 2025-12-04T12:25:16.7222063Z 2025-12-04T12:25:16.7222991Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6803s] [100%] 2025-12-04T12:25:16.7223901Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1649s] [100%] 2025-12-04T12:25:16.7224743Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1614s] [100%] 2025-12-04T12:25:16.7224748Z 2025-12-04T12:25:16.7224892Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7225450Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.7225586Z Traceback (most recent call last): 2025-12-04T12:25:16.7226060Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7226268Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7226479Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7226484Z 2025-12-04T12:25:16.7226697Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7227653Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7227730Z 2025-12-04T12:25:16.7227997Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7228226Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7228370Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7228540Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7228983Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7229204Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7229301Z graph_break [] 2025-12-04T12:25:16.7229527Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7230267Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7230380Z warnings.warn( 2025-12-04T12:25:16.7230941Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.7231070Z Traceback (most recent call last): 2025-12-04T12:25:16.7231549Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7231784Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7231996Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7232012Z 2025-12-04T12:25:16.7232225Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7233173Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7233178Z 2025-12-04T12:25:16.7233494Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7233717Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7233838Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7233951Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7234289Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7234524Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7234621Z graph_break [] 2025-12-04T12:25:16.7234835Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7235582Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7235683Z warnings.warn( 2025-12-04T12:25:16.7235909Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7236024Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7236143Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7236373Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7236706Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7236802Z graph_break [] 2025-12-04T12:25:16.7237025Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7237750Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7237864Z warnings.warn( 2025-12-04T12:25:16.7238003Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7238563Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.7238695Z Traceback (most recent call last): 2025-12-04T12:25:16.7239195Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7239424Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7239643Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7239648Z 2025-12-04T12:25:16.7239858Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7240811Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7240816Z 2025-12-04T12:25:16.7241078Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7241293Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7241415Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7241527Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7241883Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7242102Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7242201Z graph_break [] 2025-12-04T12:25:16.7242423Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7243187Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7243288Z warnings.warn( 2025-12-04T12:25:16.7243512Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7243620Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7243752Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7243995Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7244331Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7244440Z graph_break [] 2025-12-04T12:25:16.7244654Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7245385Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7245495Z warnings.warn( 2025-12-04T12:25:16.7245708Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7245823Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7245935Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7246151Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7246497Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7246596Z graph_break [] 2025-12-04T12:25:16.7246802Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7247547Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7247644Z warnings.warn( 2025-12-04T12:25:16.7248477Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-1a1d3c182ee91d42.xml - 2025-12-04T12:25:16.7248648Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7249728Z FAILED [0.1614s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7249734Z 2025-12-04T12:25:16.7250021Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7251030Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7251035Z 2025-12-04T12:25:16.7251306Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7251490Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7251688Z ================== 1 failed, 204 deselected, 2 rerun in 5.06s ================== 2025-12-04T12:25:16.7251797Z Got exit code 1 2025-12-04T12:25:16.7252663Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7253084Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.7253726Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9bd3fb708dd83987.xml 2025-12-04T12:25:16.7253893Z ============================= test session starts ============================== 2025-12-04T12:25:16.7254256Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7254395Z cachedir: .pytest_cache 2025-12-04T12:25:16.7254932Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7255054Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7255167Z configfile: pytest.ini 2025-12-04T12:25:16.7255768Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7256019Z collecting ... collected 380 items / 82 deselected / 298 selected 2025-12-04T12:25:16.7256162Z stepcurrent: skipping 82 already run items. 2025-12-04T12:25:16.7256293Z Running 123 items in this shard 2025-12-04T12:25:16.7256299Z 2025-12-04T12:25:16.7257401Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_True SKIPPED [0.0040s] (Skip non-critical tests to save resources.) [ 0%] 2025-12-04T12:25:16.7258440Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_2_add_1dim_True SKIPPED [0.0029s] (Skip non-critical tests to save resources.) [ 1%] 2025-12-04T12:25:16.7259457Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_True_initial_xblock_2_add_1dim_False SKIPPED [0.0036s] (Skip non-critical tests to save resources.) [ 2%] 2025-12-04T12:25:16.7260481Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_True_initial_xblock_2_add_1dim_True SKIPPED [0.0028s] (Skip non-critical tests to save resources.) [ 3%] 2025-12-04T12:25:16.7261388Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [4.7458s] [ 4%] 2025-12-04T12:25:16.7262308Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [0.1822s] [ 4%] 2025-12-04T12:25:16.7263122Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True FAILED [0.1775s] [ 4%] 2025-12-04T12:25:16.7263128Z 2025-12-04T12:25:16.7263270Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7263868Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.7264019Z Traceback (most recent call last): 2025-12-04T12:25:16.7264501Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7264701Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7264912Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7264917Z 2025-12-04T12:25:16.7265145Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7266091Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.7266096Z 2025-12-04T12:25:16.7266376Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7266599Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7266716Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7266845Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7267187Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7267440Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7267553Z graph_break [] 2025-12-04T12:25:16.7267770Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7270540Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7270653Z return x.grad, w.grad 2025-12-04T12:25:16.7271404Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7271502Z warnings.warn( 2025-12-04T12:25:16.7274202Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7274330Z return x.grad, w.grad 2025-12-04T12:25:16.7274883Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.7275014Z Traceback (most recent call last): 2025-12-04T12:25:16.7275483Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7275678Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7275902Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7275907Z 2025-12-04T12:25:16.7276119Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7277078Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.7277140Z 2025-12-04T12:25:16.7277409Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7277635Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7277746Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7277859Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7278212Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7278427Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7278523Z graph_break [] 2025-12-04T12:25:16.7278749Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7281476Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7281600Z return x.grad, w.grad 2025-12-04T12:25:16.7282328Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7282426Z warnings.warn( 2025-12-04T12:25:16.7285168Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7285279Z return x.grad, w.grad 2025-12-04T12:25:16.7285508Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7285617Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7285744Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7285965Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7286302Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7286406Z graph_break [] 2025-12-04T12:25:16.7286622Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7289336Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7289446Z return x.grad, w.grad 2025-12-04T12:25:16.7290172Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7290281Z warnings.warn( 2025-12-04T12:25:16.7292972Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7293178Z return x.grad, w.grad 2025-12-04T12:25:16.7293328Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7293887Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.7294006Z Traceback (most recent call last): 2025-12-04T12:25:16.7294477Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7294687Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7294895Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7294901Z 2025-12-04T12:25:16.7295126Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7296299Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.7296306Z 2025-12-04T12:25:16.7296574Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7296802Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7296977Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7297095Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7297498Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7297725Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7297835Z graph_break [] 2025-12-04T12:25:16.7298051Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7300745Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7300864Z return x.grad, w.grad 2025-12-04T12:25:16.7301591Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7301708Z warnings.warn( 2025-12-04T12:25:16.7304580Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7304698Z return x.grad, w.grad 2025-12-04T12:25:16.7304916Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7305028Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7305242Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7305464Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7305814Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7305909Z graph_break [] 2025-12-04T12:25:16.7306126Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7308845Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7308955Z return x.grad, w.grad 2025-12-04T12:25:16.7309695Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7309793Z warnings.warn( 2025-12-04T12:25:16.7312555Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7312662Z return x.grad, w.grad 2025-12-04T12:25:16.7312884Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7313001Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7313115Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7313350Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7313691Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7313788Z graph_break [] 2025-12-04T12:25:16.7314013Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7314747Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7314845Z warnings.warn( 2025-12-04T12:25:16.7317563Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7317670Z return x.grad, w.grad 2025-12-04T12:25:16.7318500Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9bd3fb708dd83987.xml - 2025-12-04T12:25:16.7318672Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7319751Z FAILED [0.1775s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7319814Z 2025-12-04T12:25:16.7320029Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7320975Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.7320991Z 2025-12-04T12:25:16.7321254Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7321428Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7321654Z ============= 1 failed, 4 skipped, 82 deselected, 2 rerun in 5.18s ============= 2025-12-04T12:25:16.7321751Z Got exit code 1 2025-12-04T12:25:16.7321857Z Retrying single test... 2025-12-04T12:25:16.7322507Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-991a5e87e1946535.xml 2025-12-04T12:25:16.7322672Z ============================= test session starts ============================== 2025-12-04T12:25:16.7323033Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7323138Z cachedir: .pytest_cache 2025-12-04T12:25:16.7323701Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7323832Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7323937Z configfile: pytest.ini 2025-12-04T12:25:16.7324532Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7324770Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.7325820Z stepcurrent: skipping 86 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.7325951Z Running 1 items in this shard 2025-12-04T12:25:16.7325955Z 2025-12-04T12:25:16.7326863Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [4.6722s] [100%] 2025-12-04T12:25:16.7327771Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [0.1689s] [100%] 2025-12-04T12:25:16.7328585Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True FAILED [0.1651s] [100%] 2025-12-04T12:25:16.7328591Z 2025-12-04T12:25:16.7328731Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7329291Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.7329413Z Traceback (most recent call last): 2025-12-04T12:25:16.7329891Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7330087Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7330299Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7330304Z 2025-12-04T12:25:16.7330526Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7331458Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.7331466Z 2025-12-04T12:25:16.7331739Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7332013Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7332123Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7332248Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7332590Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7332823Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7332919Z graph_break [] 2025-12-04T12:25:16.7333136Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7335876Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7335989Z return x.grad, w.grad 2025-12-04T12:25:16.7336758Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7336923Z warnings.warn( 2025-12-04T12:25:16.7339655Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7339776Z return x.grad, w.grad 2025-12-04T12:25:16.7340327Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.7340458Z Traceback (most recent call last): 2025-12-04T12:25:16.7340924Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7341129Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7341343Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7341349Z 2025-12-04T12:25:16.7341560Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7342511Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.7342520Z 2025-12-04T12:25:16.7342782Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7343006Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7343115Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7343228Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7343578Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7343800Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7343901Z graph_break [] 2025-12-04T12:25:16.7344129Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7346829Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7347007Z return x.grad, w.grad 2025-12-04T12:25:16.7347749Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7347867Z warnings.warn( 2025-12-04T12:25:16.7350577Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7350732Z return x.grad, w.grad 2025-12-04T12:25:16.7350951Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7351062Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7351191Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7351413Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7351753Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7351868Z graph_break [] 2025-12-04T12:25:16.7352109Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7354837Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7354943Z return x.grad, w.grad 2025-12-04T12:25:16.7355669Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7355784Z warnings.warn( 2025-12-04T12:25:16.7358482Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7358608Z return x.grad, w.grad 2025-12-04T12:25:16.7358753Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7359320Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.7359443Z Traceback (most recent call last): 2025-12-04T12:25:16.7359915Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7360180Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7360393Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7360398Z 2025-12-04T12:25:16.7360627Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7361565Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.7361571Z 2025-12-04T12:25:16.7361834Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7362064Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7362175Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7362305Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7362649Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7362871Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7362981Z graph_break [] 2025-12-04T12:25:16.7363197Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7365950Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7366089Z return x.grad, w.grad 2025-12-04T12:25:16.7366825Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7366947Z warnings.warn( 2025-12-04T12:25:16.7369656Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7369779Z return x.grad, w.grad 2025-12-04T12:25:16.7369998Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7370112Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7370239Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7370457Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7370806Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7370902Z graph_break [] 2025-12-04T12:25:16.7371120Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7373833Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7374006Z return x.grad, w.grad 2025-12-04T12:25:16.7374750Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7374848Z warnings.warn( 2025-12-04T12:25:16.7377624Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7377732Z return x.grad, w.grad 2025-12-04T12:25:16.7377946Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7378067Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7378179Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7378410Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7378780Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7378877Z graph_break [] 2025-12-04T12:25:16.7379105Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7379838Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7379937Z warnings.warn( 2025-12-04T12:25:16.7382675Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7382789Z return x.grad, w.grad 2025-12-04T12:25:16.7383618Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-991a5e87e1946535.xml - 2025-12-04T12:25:16.7383788Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7384866Z FAILED [0.1651s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7384876Z 2025-12-04T12:25:16.7385092Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7386048Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.7386054Z 2025-12-04T12:25:16.7386320Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7386497Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7386708Z ================== 1 failed, 204 deselected, 2 rerun in 5.06s ================== 2025-12-04T12:25:16.7386805Z Got exit code 1 2025-12-04T12:25:16.7386909Z Retrying single test... 2025-12-04T12:25:16.7387563Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-61bd448d788d4193.xml 2025-12-04T12:25:16.7387781Z ============================= test session starts ============================== 2025-12-04T12:25:16.7388140Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7388247Z cachedir: .pytest_cache 2025-12-04T12:25:16.7388775Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7388910Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7389017Z configfile: pytest.ini 2025-12-04T12:25:16.7389611Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7389848Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.7390869Z stepcurrent: skipping 86 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.7391004Z Running 1 items in this shard 2025-12-04T12:25:16.7391008Z 2025-12-04T12:25:16.7391936Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [4.6477s] [100%] 2025-12-04T12:25:16.7392842Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [0.1648s] [100%] 2025-12-04T12:25:16.7393663Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True FAILED [0.1633s] [100%] 2025-12-04T12:25:16.7393696Z 2025-12-04T12:25:16.7393834Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7394404Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.7394525Z Traceback (most recent call last): 2025-12-04T12:25:16.7395012Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7395211Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7395425Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7395430Z 2025-12-04T12:25:16.7395659Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7396774Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.7396779Z 2025-12-04T12:25:16.7397062Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7397280Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7397393Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7397523Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7397871Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7398106Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7398206Z graph_break [] 2025-12-04T12:25:16.7398423Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7401139Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7401344Z return x.grad, w.grad 2025-12-04T12:25:16.7402097Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7402202Z warnings.warn( 2025-12-04T12:25:16.7404938Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7405049Z return x.grad, w.grad 2025-12-04T12:25:16.7405638Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.7405779Z Traceback (most recent call last): 2025-12-04T12:25:16.7406251Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7406459Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7406669Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7406674Z 2025-12-04T12:25:16.7406888Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7407874Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.7407883Z 2025-12-04T12:25:16.7408151Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7408382Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7408496Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7408610Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7408964Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7409182Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7409281Z graph_break [] 2025-12-04T12:25:16.7409508Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7412235Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7412357Z return x.grad, w.grad 2025-12-04T12:25:16.7413086Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7413199Z warnings.warn( 2025-12-04T12:25:16.7415904Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7416079Z return x.grad, w.grad 2025-12-04T12:25:16.7416292Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7416401Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7416526Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7416747Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7417147Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7417258Z graph_break [] 2025-12-04T12:25:16.7417475Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7420229Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7420338Z return x.grad, w.grad 2025-12-04T12:25:16.7421077Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7421206Z warnings.warn( 2025-12-04T12:25:16.7423907Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7424033Z return x.grad, w.grad 2025-12-04T12:25:16.7424179Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7424743Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.7424864Z Traceback (most recent call last): 2025-12-04T12:25:16.7425339Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7425552Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7425765Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7425770Z 2025-12-04T12:25:16.7425993Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7426937Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.7426943Z 2025-12-04T12:25:16.7427207Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7427438Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7427548Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7427678Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7428018Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7428309Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7428418Z graph_break [] 2025-12-04T12:25:16.7428634Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7431365Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7431472Z return x.grad, w.grad 2025-12-04T12:25:16.7432205Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7432317Z warnings.warn( 2025-12-04T12:25:16.7435043Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7435159Z return x.grad, w.grad 2025-12-04T12:25:16.7435402Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7435525Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7435637Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7435860Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7436210Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7436306Z graph_break [] 2025-12-04T12:25:16.7436522Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7439232Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7439341Z return x.grad, w.grad 2025-12-04T12:25:16.7440078Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7440180Z warnings.warn( 2025-12-04T12:25:16.7442898Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7443058Z return x.grad, w.grad 2025-12-04T12:25:16.7443272Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7443392Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7443506Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7443741Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7444080Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7444177Z graph_break [] 2025-12-04T12:25:16.7444408Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7445138Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7445241Z warnings.warn( 2025-12-04T12:25:16.7447974Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.7448088Z return x.grad, w.grad 2025-12-04T12:25:16.7448919Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-61bd448d788d4193.xml - 2025-12-04T12:25:16.7449090Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7450544Z FAILED [0.1633s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7450556Z 2025-12-04T12:25:16.7450774Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7451730Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.7451735Z 2025-12-04T12:25:16.7452002Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7452181Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7452399Z ================== 1 failed, 204 deselected, 2 rerun in 5.03s ================== 2025-12-04T12:25:16.7452496Z Got exit code 1 2025-12-04T12:25:16.7453360Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.7453779Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.7454423Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0c87f6ca176b74e4.xml 2025-12-04T12:25:16.7454604Z ============================= test session starts ============================== 2025-12-04T12:25:16.7454952Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7455073Z cachedir: .pytest_cache 2025-12-04T12:25:16.7455596Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7455722Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7455846Z configfile: pytest.ini 2025-12-04T12:25:16.7456437Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7456724Z collecting ... collected 380 items / 87 deselected / 293 selected 2025-12-04T12:25:16.7456946Z stepcurrent: skipping 87 already run items. 2025-12-04T12:25:16.7457062Z Running 118 items in this shard 2025-12-04T12:25:16.7457067Z 2025-12-04T12:25:16.7458113Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_True_initial_xblock_1_add_1dim_False SKIPPED [0.0042s] (Skip non-critical tests to save resources.) [ 0%] 2025-12-04T12:25:16.7459135Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_True SKIPPED [0.0029s] (Skip non-critical tests to save resources.) [ 1%] 2025-12-04T12:25:16.7460157Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_True SKIPPED [0.0037s] (Skip non-critical tests to save resources.) [ 2%] 2025-12-04T12:25:16.7461221Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_True_initial_xblock_1_add_1dim_False SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 3%] 2025-12-04T12:25:16.7462229Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_True_initial_xblock_2_add_1dim_True SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 4%] 2025-12-04T12:25:16.7463287Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape2_max_autotune_False_initial_xblock_1_add_1dim_True SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 5%] 2025-12-04T12:25:16.7464301Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape2_max_autotune_True_initial_xblock_1_add_1dim_False SKIPPED [0.0026s] (Skip non-critical tests to save resources.) [ 5%] 2025-12-04T12:25:16.7465228Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6332s] [ 6%] 2025-12-04T12:25:16.7466127Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1594s] [ 6%] 2025-12-04T12:25:16.7466971Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1581s] [ 6%] 2025-12-04T12:25:16.7466977Z 2025-12-04T12:25:16.7467120Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7467676Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.7467810Z Traceback (most recent call last): 2025-12-04T12:25:16.7468277Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7468488Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7468697Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7468702Z 2025-12-04T12:25:16.7468920Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7469874Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7469882Z 2025-12-04T12:25:16.7470147Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7470434Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7470543Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7470656Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7471008Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7471225Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7471337Z graph_break [] 2025-12-04T12:25:16.7471553Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7472291Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7472402Z warnings.warn( 2025-12-04T12:25:16.7472959Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.7473081Z Traceback (most recent call last): 2025-12-04T12:25:16.7473559Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7473755Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7474006Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7474011Z 2025-12-04T12:25:16.7474227Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7475165Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7475170Z 2025-12-04T12:25:16.7475449Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7475712Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7475839Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7475953Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7476290Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7476525Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7476626Z graph_break [] 2025-12-04T12:25:16.7476839Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7477579Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7477677Z warnings.warn( 2025-12-04T12:25:16.7477902Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7478011Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7478127Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7478361Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7478697Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7478794Z graph_break [] 2025-12-04T12:25:16.7479021Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7479744Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7479856Z warnings.warn( 2025-12-04T12:25:16.7479998Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7480553Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.7480687Z Traceback (most recent call last): 2025-12-04T12:25:16.7481157Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7481408Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7481632Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7481637Z 2025-12-04T12:25:16.7481846Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7482802Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7482808Z 2025-12-04T12:25:16.7483074Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7483302Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7483412Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7483527Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7483880Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7484101Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7484196Z graph_break [] 2025-12-04T12:25:16.7484421Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7485175Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7485289Z warnings.warn( 2025-12-04T12:25:16.7485502Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7485611Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7485736Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7485955Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7486317Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7486430Z graph_break [] 2025-12-04T12:25:16.7486645Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7487371Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7487480Z warnings.warn( 2025-12-04T12:25:16.7487692Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7487810Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7487926Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7488143Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7488492Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7488587Z graph_break [] 2025-12-04T12:25:16.7488803Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7489543Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7489643Z warnings.warn( 2025-12-04T12:25:16.7490479Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0c87f6ca176b74e4.xml - 2025-12-04T12:25:16.7490647Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7491718Z FAILED [0.1581s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7491736Z 2025-12-04T12:25:16.7491951Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7492889Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7492952Z 2025-12-04T12:25:16.7493230Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7493409Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7493638Z ============= 1 failed, 7 skipped, 87 deselected, 2 rerun in 5.04s ============= 2025-12-04T12:25:16.7493736Z Got exit code 1 2025-12-04T12:25:16.7493841Z Retrying single test... 2025-12-04T12:25:16.7494493Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-00981048468ad3fe.xml 2025-12-04T12:25:16.7494655Z ============================= test session starts ============================== 2025-12-04T12:25:16.7495004Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7495127Z cachedir: .pytest_cache 2025-12-04T12:25:16.7495643Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7495780Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7495890Z configfile: pytest.ini 2025-12-04T12:25:16.7496697Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7496995Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.7498018Z stepcurrent: skipping 94 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7498146Z Running 1 items in this shard 2025-12-04T12:25:16.7498201Z 2025-12-04T12:25:16.7499102Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6537s] [100%] 2025-12-04T12:25:16.7500006Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1609s] [100%] 2025-12-04T12:25:16.7500844Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1598s] [100%] 2025-12-04T12:25:16.7500850Z 2025-12-04T12:25:16.7500990Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7501557Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.7501683Z Traceback (most recent call last): 2025-12-04T12:25:16.7502156Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7502367Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7502580Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7502585Z 2025-12-04T12:25:16.7502812Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7503754Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7503759Z 2025-12-04T12:25:16.7504023Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7504256Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7504370Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7504501Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7504930Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7505150Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7505260Z graph_break [] 2025-12-04T12:25:16.7505476Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7506217Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7506328Z warnings.warn( 2025-12-04T12:25:16.7506878Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.7507013Z Traceback (most recent call last): 2025-12-04T12:25:16.7507484Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7507684Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7507908Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7507913Z 2025-12-04T12:25:16.7508126Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7509116Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7509122Z 2025-12-04T12:25:16.7509392Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7509608Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7509730Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7509843Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7510221Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7510445Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7510542Z graph_break [] 2025-12-04T12:25:16.7510766Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7511504Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7511604Z warnings.warn( 2025-12-04T12:25:16.7511837Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7511944Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7512056Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7512285Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7512620Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7512729Z graph_break [] 2025-12-04T12:25:16.7512944Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7513664Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7513774Z warnings.warn( 2025-12-04T12:25:16.7513914Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7514480Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.7514599Z Traceback (most recent call last): 2025-12-04T12:25:16.7515066Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7515269Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7515482Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7515487Z 2025-12-04T12:25:16.7515754Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7516706Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7516712Z 2025-12-04T12:25:16.7516975Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7517203Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7517316Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7517430Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7517783Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7518059Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7518169Z graph_break [] 2025-12-04T12:25:16.7518382Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7519119Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7519233Z warnings.warn( 2025-12-04T12:25:16.7519441Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7519664Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7519791Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7520009Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7520357Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7520454Z graph_break [] 2025-12-04T12:25:16.7520665Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7521435Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7521537Z warnings.warn( 2025-12-04T12:25:16.7521813Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7521937Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7522047Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7522280Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7522617Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7522712Z graph_break [] 2025-12-04T12:25:16.7522936Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7523661Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7523760Z warnings.warn( 2025-12-04T12:25:16.7524585Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-00981048468ad3fe.xml - 2025-12-04T12:25:16.7524755Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7525846Z FAILED [0.1598s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7525852Z 2025-12-04T12:25:16.7526062Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7527011Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7527017Z 2025-12-04T12:25:16.7527282Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7527516Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7527722Z ================== 1 failed, 204 deselected, 2 rerun in 5.03s ================== 2025-12-04T12:25:16.7527818Z Got exit code 1 2025-12-04T12:25:16.7527920Z Retrying single test... 2025-12-04T12:25:16.7528572Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-7f9686778c56ad66.xml 2025-12-04T12:25:16.7528728Z ============================= test session starts ============================== 2025-12-04T12:25:16.7529088Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7529195Z cachedir: .pytest_cache 2025-12-04T12:25:16.7529717Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7529846Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7529958Z configfile: pytest.ini 2025-12-04T12:25:16.7530547Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7530785Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.7531829Z stepcurrent: skipping 94 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7531958Z Running 1 items in this shard 2025-12-04T12:25:16.7531963Z 2025-12-04T12:25:16.7532870Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6501s] [100%] 2025-12-04T12:25:16.7533810Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1609s] [100%] 2025-12-04T12:25:16.7534639Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1591s] [100%] 2025-12-04T12:25:16.7534644Z 2025-12-04T12:25:16.7534779Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7535339Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.7535457Z Traceback (most recent call last): 2025-12-04T12:25:16.7535937Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7536136Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7536346Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7536355Z 2025-12-04T12:25:16.7536578Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7537583Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7537592Z 2025-12-04T12:25:16.7537866Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7538085Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7538196Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7538318Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7538659Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7538891Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7538986Z graph_break [] 2025-12-04T12:25:16.7539273Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7540023Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7540123Z warnings.warn( 2025-12-04T12:25:16.7540681Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.7540808Z Traceback (most recent call last): 2025-12-04T12:25:16.7541277Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7541486Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7541692Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7541697Z 2025-12-04T12:25:16.7541911Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7542865Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7542870Z 2025-12-04T12:25:16.7543132Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7543381Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7543489Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7543600Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7543949Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7544164Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7544259Z graph_break [] 2025-12-04T12:25:16.7544507Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7545246Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7545357Z warnings.warn( 2025-12-04T12:25:16.7545571Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7545681Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7545802Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7546018Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7546350Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7546454Z graph_break [] 2025-12-04T12:25:16.7546663Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7547403Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7547505Z warnings.warn( 2025-12-04T12:25:16.7547646Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7548209Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.7548327Z Traceback (most recent call last): 2025-12-04T12:25:16.7548798Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7549000Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7549211Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7549216Z 2025-12-04T12:25:16.7549434Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7550375Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7550439Z 2025-12-04T12:25:16.7550717Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7550928Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7551038Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7551164Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7551498Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7551716Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7551820Z graph_break [] 2025-12-04T12:25:16.7552032Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7552762Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7552874Z warnings.warn( 2025-12-04T12:25:16.7553089Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7553204Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7553313Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7553532Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7553903Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7554000Z graph_break [] 2025-12-04T12:25:16.7554210Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7554942Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7555041Z warnings.warn( 2025-12-04T12:25:16.7555286Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7555393Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7555515Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7555742Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7556078Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7556173Z graph_break [] 2025-12-04T12:25:16.7556396Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7557124Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7557230Z warnings.warn( 2025-12-04T12:25:16.7558053Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-7f9686778c56ad66.xml - 2025-12-04T12:25:16.7558222Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7559300Z FAILED [0.1591s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7559307Z 2025-12-04T12:25:16.7559520Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7560474Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7560480Z 2025-12-04T12:25:16.7560742Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7560933Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7561132Z ================== 1 failed, 204 deselected, 2 rerun in 5.02s ================== 2025-12-04T12:25:16.7561229Z Got exit code 1 2025-12-04T12:25:16.7562176Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.7562582Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.7563225Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dc76d780af365e97.xml 2025-12-04T12:25:16.7563393Z ============================= test session starts ============================== 2025-12-04T12:25:16.7563737Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7563856Z cachedir: .pytest_cache 2025-12-04T12:25:16.7564375Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7564497Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7564621Z configfile: pytest.ini 2025-12-04T12:25:16.7565209Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7565445Z collecting ... collected 380 items / 95 deselected / 285 selected 2025-12-04T12:25:16.7565614Z stepcurrent: skipping 95 already run items. 2025-12-04T12:25:16.7565727Z Running 110 items in this shard 2025-12-04T12:25:16.7565732Z 2025-12-04T12:25:16.7566778Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_True SKIPPED [0.0041s] (Skip non-critical tests to save resources.) [ 0%] 2025-12-04T12:25:16.7573502Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6422s] [ 1%] 2025-12-04T12:25:16.7574504Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1610s] [ 1%] 2025-12-04T12:25:16.7575349Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1584s] [ 1%] 2025-12-04T12:25:16.7575357Z 2025-12-04T12:25:16.7575497Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7576055Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.7576173Z Traceback (most recent call last): 2025-12-04T12:25:16.7576654Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7576939Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7577160Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7577166Z 2025-12-04T12:25:16.7577386Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7578334Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.7578340Z 2025-12-04T12:25:16.7578613Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7578830Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7578941Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7579059Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7579402Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7579618Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7579811Z graph_break [] 2025-12-04T12:25:16.7580026Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7580772Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7580872Z warnings.warn( 2025-12-04T12:25:16.7581425Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.7581550Z Traceback (most recent call last): 2025-12-04T12:25:16.7582155Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7582439Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7582665Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7582676Z 2025-12-04T12:25:16.7582887Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7583837Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.7583843Z 2025-12-04T12:25:16.7584156Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7584383Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7584491Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7584602Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7584954Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7585166Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7585296Z graph_break [] 2025-12-04T12:25:16.7585517Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7586252Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7586356Z warnings.warn( 2025-12-04T12:25:16.7586569Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7586672Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7586790Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7587002Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7587341Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7587442Z graph_break [] 2025-12-04T12:25:16.7587651Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7588388Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7588492Z warnings.warn( 2025-12-04T12:25:16.7588632Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7589193Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.7589307Z Traceback (most recent call last): 2025-12-04T12:25:16.7589771Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7589969Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7590181Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7590187Z 2025-12-04T12:25:16.7590402Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7591345Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.7591416Z 2025-12-04T12:25:16.7591676Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7591889Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7591996Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7592109Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7592441Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7592655Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7592753Z graph_break [] 2025-12-04T12:25:16.7592959Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7593688Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7593791Z warnings.warn( 2025-12-04T12:25:16.7593999Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7594110Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7594218Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7594461Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7594797Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7594891Z graph_break [] 2025-12-04T12:25:16.7595099Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7595830Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7595953Z warnings.warn( 2025-12-04T12:25:16.7596325Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7596437Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7596545Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7596768Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7597101Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7597192Z graph_break [] 2025-12-04T12:25:16.7597414Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7598138Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7598241Z warnings.warn( 2025-12-04T12:25:16.7599063Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dc76d780af365e97.xml - 2025-12-04T12:25:16.7599232Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7600311Z FAILED [0.1584s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7600317Z 2025-12-04T12:25:16.7600528Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7601474Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.7601479Z 2025-12-04T12:25:16.7601740Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7601915Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7602136Z ============= 1 failed, 1 skipped, 95 deselected, 2 rerun in 5.02s ============= 2025-12-04T12:25:16.7602345Z Got exit code 1 2025-12-04T12:25:16.7602462Z Retrying single test... 2025-12-04T12:25:16.7603101Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3d712c264776fdc2.xml 2025-12-04T12:25:16.7603262Z ============================= test session starts ============================== 2025-12-04T12:25:16.7603622Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7603727Z cachedir: .pytest_cache 2025-12-04T12:25:16.7604247Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7604379Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7604483Z configfile: pytest.ini 2025-12-04T12:25:16.7605077Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7605304Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.7606368Z stepcurrent: skipping 96 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.7606489Z Running 1 items in this shard 2025-12-04T12:25:16.7606494Z 2025-12-04T12:25:16.7607393Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6453s] [100%] 2025-12-04T12:25:16.7608336Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1644s] [100%] 2025-12-04T12:25:16.7609156Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1615s] [100%] 2025-12-04T12:25:16.7609166Z 2025-12-04T12:25:16.7609308Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7609861Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.7609979Z Traceback (most recent call last): 2025-12-04T12:25:16.7610450Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7610641Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7610848Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7610864Z 2025-12-04T12:25:16.7611073Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7612009Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.7612016Z 2025-12-04T12:25:16.7612282Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7612497Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7612618Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7612726Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7613063Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7613287Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7613377Z graph_break [] 2025-12-04T12:25:16.7613591Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7614329Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7614501Z warnings.warn( 2025-12-04T12:25:16.7615062Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.7615178Z Traceback (most recent call last): 2025-12-04T12:25:16.7615642Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7615845Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7616048Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7616053Z 2025-12-04T12:25:16.7616262Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7617279Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.7617291Z 2025-12-04T12:25:16.7617555Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7617778Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7617886Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7618031Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7618378Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7618591Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7618695Z graph_break [] 2025-12-04T12:25:16.7618905Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7619668Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7619778Z warnings.warn( 2025-12-04T12:25:16.7619985Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7620089Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7620206Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7620417Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7620759Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7620850Z graph_break [] 2025-12-04T12:25:16.7621058Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7621802Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7621896Z warnings.warn( 2025-12-04T12:25:16.7622038Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7622597Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.7622712Z Traceback (most recent call last): 2025-12-04T12:25:16.7623181Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7623375Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7623581Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7623586Z 2025-12-04T12:25:16.7623799Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7624731Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.7624736Z 2025-12-04T12:25:16.7625009Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7625281Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7625385Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7625504Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7625839Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7626053Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7626154Z graph_break [] 2025-12-04T12:25:16.7626362Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7627092Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7627189Z warnings.warn( 2025-12-04T12:25:16.7627400Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7627517Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7627628Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7627843Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7628182Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7628274Z graph_break [] 2025-12-04T12:25:16.7628519Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7629242Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7629335Z warnings.warn( 2025-12-04T12:25:16.7629551Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7629656Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7629769Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7630018Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7630353Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7630453Z graph_break [] 2025-12-04T12:25:16.7630660Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7631389Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7631489Z warnings.warn( 2025-12-04T12:25:16.7632309Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3d712c264776fdc2.xml - 2025-12-04T12:25:16.7632482Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7633555Z FAILED [0.1615s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7633566Z 2025-12-04T12:25:16.7633778Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7634729Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.7634735Z 2025-12-04T12:25:16.7634995Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7635178Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7635370Z ================== 1 failed, 204 deselected, 2 rerun in 5.03s ================== 2025-12-04T12:25:16.7635465Z Got exit code 1 2025-12-04T12:25:16.7635577Z Retrying single test... 2025-12-04T12:25:16.7636228Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4da37e7aafa09519.xml 2025-12-04T12:25:16.7636459Z ============================= test session starts ============================== 2025-12-04T12:25:16.7636803Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7636906Z cachedir: .pytest_cache 2025-12-04T12:25:16.7637433Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7637552Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7637653Z configfile: pytest.ini 2025-12-04T12:25:16.7638246Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7638469Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.7639502Z stepcurrent: skipping 96 already run items. Running only test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.7639616Z Running 1 items in this shard 2025-12-04T12:25:16.7639622Z 2025-12-04T12:25:16.7640559Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6420s] [100%] 2025-12-04T12:25:16.7641466Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1604s] [100%] 2025-12-04T12:25:16.7642290Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1581s] [100%] 2025-12-04T12:25:16.7642327Z 2025-12-04T12:25:16.7642472Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7643030Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.7643155Z Traceback (most recent call last): 2025-12-04T12:25:16.7643623Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7643818Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7644033Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7644038Z 2025-12-04T12:25:16.7644246Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7645180Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.7645191Z 2025-12-04T12:25:16.7645457Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7645671Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7645784Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7645894Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7646232Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7646450Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7646544Z graph_break [] 2025-12-04T12:25:16.7646761Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7647491Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7647585Z warnings.warn( 2025-12-04T12:25:16.7648147Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.7648322Z Traceback (most recent call last): 2025-12-04T12:25:16.7648783Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7648986Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7649193Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7649198Z 2025-12-04T12:25:16.7649415Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7650351Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.7650357Z 2025-12-04T12:25:16.7650615Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7650842Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7650952Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7651072Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7651407Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7651618Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7651751Z graph_break [] 2025-12-04T12:25:16.7651964Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7652688Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7652796Z warnings.warn( 2025-12-04T12:25:16.7653006Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7653164Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7653273Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7653492Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7653838Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7653930Z graph_break [] 2025-12-04T12:25:16.7654141Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7654874Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7654972Z warnings.warn( 2025-12-04T12:25:16.7655120Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7655666Z _ MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.7655781Z Traceback (most recent call last): 2025-12-04T12:25:16.7656254Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.7656451Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7656667Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7656672Z 2025-12-04T12:25:16.7656952Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7657897Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.7657902Z 2025-12-04T12:25:16.7658173Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7658386Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7658499Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7658611Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7658944Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7659235Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7659328Z graph_break [] 2025-12-04T12:25:16.7659537Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7660277Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7660372Z warnings.warn( 2025-12-04T12:25:16.7660585Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7660693Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7660800Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7661029Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7661364Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7661463Z graph_break [] 2025-12-04T12:25:16.7661688Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7662413Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7662547Z warnings.warn( 2025-12-04T12:25:16.7662757Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7662860Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7662977Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.7663192Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7663523Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7663624Z graph_break [] 2025-12-04T12:25:16.7663863Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7664598Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7664693Z warnings.warn( 2025-12-04T12:25:16.7665508Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4da37e7aafa09519.xml - 2025-12-04T12:25:16.7665681Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7666742Z FAILED [0.1581s] inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7666748Z 2025-12-04T12:25:16.7666972Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7667913Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py MixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.7667923Z 2025-12-04T12:25:16.7668181Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7668370Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7668565Z ================== 1 failed, 204 deselected, 2 rerun in 5.01s ================== 2025-12-04T12:25:16.7668670Z Got exit code 1 2025-12-04T12:25:16.7669522Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.7669932Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.7670579Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4f9e658df3b9ce6d.xml 2025-12-04T12:25:16.7670801Z ============================= test session starts ============================== 2025-12-04T12:25:16.7671156Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7671265Z cachedir: .pytest_cache 2025-12-04T12:25:16.7671783Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7671912Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7672016Z configfile: pytest.ini 2025-12-04T12:25:16.7672607Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7672841Z collecting ... collected 380 items / 97 deselected / 283 selected 2025-12-04T12:25:16.7672978Z stepcurrent: skipping 97 already run items. 2025-12-04T12:25:16.7673100Z Running 108 items in this shard 2025-12-04T12:25:16.7673106Z 2025-12-04T12:25:16.7674115Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_True_initial_xblock_2_add_1dim_True SKIPPED [0.0041s] (Skip non-critical tests to save resources.) [ 0%] 2025-12-04T12:25:16.7674711Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_with_dynamic_shape_dynamic_dims1 PASSED [6.6441s] [ 1%] 2025-12-04T12:25:16.7675102Z inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_xmask PASSED [0.5306s] [ 2%] 2025-12-04T12:25:16.7675753Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_avoid_non_coalesced_access SKIPPED [0.0033s] (Mix order reduction not enabled) [ 3%] 2025-12-04T12:25:16.7676439Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_independent_split_size SKIPPED [0.0033s] (Mix order reduction not enabled) [ 4%] 2025-12-04T12:25:16.7677045Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_no_bias_split_reductions_False_shape0 PASSED [0.8208s] [ 5%] 2025-12-04T12:25:16.7677646Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_no_bias_split_reductions_False_shape1 PASSED [1.0974s] [ 6%] 2025-12-04T12:25:16.7678413Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 ('RERUN', {'yellow': True}) [0.1782s] [ 7%] 2025-12-04T12:25:16.7679161Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 ('RERUN', {'yellow': True}) [0.1416s] [ 7%] 2025-12-04T12:25:16.7679832Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 FAILED [0.1402s] [ 7%] 2025-12-04T12:25:16.7679837Z 2025-12-04T12:25:16.7679975Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7680377Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 _ 2025-12-04T12:25:16.7680493Z Traceback (most recent call last): 2025-12-04T12:25:16.7681016Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7681218Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7681426Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7681431Z 2025-12-04T12:25:16.7681639Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7682428Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.7682434Z 2025-12-04T12:25:16.7682696Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7682984Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7683088Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7683199Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7683421Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7683538Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7683640Z graph_break [] 2025-12-04T12:25:16.7683851Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7684579Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7684685Z warnings.warn( 2025-12-04T12:25:16.7685077Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 _ 2025-12-04T12:25:16.7685196Z Traceback (most recent call last): 2025-12-04T12:25:16.7685737Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7685932Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7686146Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7686153Z 2025-12-04T12:25:16.7686392Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7687171Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.7687177Z 2025-12-04T12:25:16.7687450Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7687663Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7687775Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7687922Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7688136Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7688262Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7688355Z graph_break [] 2025-12-04T12:25:16.7688563Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7689302Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7689394Z warnings.warn( 2025-12-04T12:25:16.7689616Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7689721Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7689828Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7690053Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7690166Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7690258Z graph_break [] 2025-12-04T12:25:16.7690474Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7691199Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7691301Z warnings.warn( 2025-12-04T12:25:16.7691448Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7691841Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 _ 2025-12-04T12:25:16.7691967Z Traceback (most recent call last): 2025-12-04T12:25:16.7692495Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7692689Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7692906Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7692913Z 2025-12-04T12:25:16.7693125Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7693987Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.7693993Z 2025-12-04T12:25:16.7694253Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7694470Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7694591Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7694702Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7694929Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7695045Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7695190Z graph_break [] 2025-12-04T12:25:16.7695413Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7696360Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7696463Z warnings.warn( 2025-12-04T12:25:16.7696685Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7696790Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7696976Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7697291Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7697411Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7697516Z graph_break [] 2025-12-04T12:25:16.7697726Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7698448Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7698553Z warnings.warn( 2025-12-04T12:25:16.7698804Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7698918Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7699040Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7699257Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7699384Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7699476Z graph_break [] 2025-12-04T12:25:16.7699690Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7700416Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7700513Z warnings.warn( 2025-12-04T12:25:16.7701327Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4f9e658df3b9ce6d.xml - 2025-12-04T12:25:16.7701503Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7702435Z FAILED [0.1402s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7702441Z 2025-12-04T12:25:16.7702658Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7703436Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.7703442Z 2025-12-04T12:25:16.7703715Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7703892Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7704127Z ======== 1 failed, 4 passed, 3 skipped, 97 deselected, 2 rerun in 9.63s ======== 2025-12-04T12:25:16.7704235Z Got exit code 1 2025-12-04T12:25:16.7704339Z Retrying single test... 2025-12-04T12:25:16.7705131Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3d80a9d76d3be932.xml 2025-12-04T12:25:16.7705305Z ============================= test session starts ============================== 2025-12-04T12:25:16.7705654Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7705767Z cachedir: .pytest_cache 2025-12-04T12:25:16.7706283Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7706402Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7706512Z configfile: pytest.ini 2025-12-04T12:25:16.7707102Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7707329Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.7708200Z stepcurrent: skipping 104 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.7708311Z Running 1 items in this shard 2025-12-04T12:25:16.7708317Z 2025-12-04T12:25:16.7709111Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 ('RERUN', {'yellow': True}) [4.6155s] [100%] 2025-12-04T12:25:16.7709858Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 ('RERUN', {'yellow': True}) [0.1416s] [100%] 2025-12-04T12:25:16.7710529Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 FAILED [0.1382s] [100%] 2025-12-04T12:25:16.7710535Z 2025-12-04T12:25:16.7710701Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7711093Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 _ 2025-12-04T12:25:16.7711225Z Traceback (most recent call last): 2025-12-04T12:25:16.7711751Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7711954Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7712162Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7712167Z 2025-12-04T12:25:16.7712375Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7713169Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.7713174Z 2025-12-04T12:25:16.7713437Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7713666Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7713774Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7713884Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7714015Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7714232Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7714328Z graph_break [] 2025-12-04T12:25:16.7714550Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7715276Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7715373Z warnings.warn( 2025-12-04T12:25:16.7715770Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 _ 2025-12-04T12:25:16.7715891Z Traceback (most recent call last): 2025-12-04T12:25:16.7716434Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7716683Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7716889Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7716895Z 2025-12-04T12:25:16.7717118Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7717894Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.7717899Z 2025-12-04T12:25:16.7718169Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7718379Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7718483Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7718607Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7718724Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7718946Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7719047Z graph_break [] 2025-12-04T12:25:16.7719258Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7720027Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7720125Z warnings.warn( 2025-12-04T12:25:16.7720335Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7720449Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7720557Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7720777Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7720906Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7721033Z graph_break [] 2025-12-04T12:25:16.7721256Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7721984Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7722082Z warnings.warn( 2025-12-04T12:25:16.7722229Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7722626Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 _ 2025-12-04T12:25:16.7722745Z Traceback (most recent call last): 2025-12-04T12:25:16.7723285Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7723479Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7723693Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7723701Z 2025-12-04T12:25:16.7723916Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7724697Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.7724703Z 2025-12-04T12:25:16.7724970Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7725185Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7725303Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7725414Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7725532Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7725756Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7725851Z graph_break [] 2025-12-04T12:25:16.7726061Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7726795Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7726950Z warnings.warn( 2025-12-04T12:25:16.7727166Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7727276Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7727388Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7727612Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7727731Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7727824Z graph_break [] 2025-12-04T12:25:16.7728043Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7728769Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7728872Z warnings.warn( 2025-12-04T12:25:16.7729087Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7729201Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7729324Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7729547Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7729666Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7729768Z graph_break [] 2025-12-04T12:25:16.7730008Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7730735Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7730839Z warnings.warn( 2025-12-04T12:25:16.7731660Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3d80a9d76d3be932.xml - 2025-12-04T12:25:16.7731863Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7732792Z FAILED [0.1382s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7732798Z 2025-12-04T12:25:16.7733019Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7733807Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.7733812Z 2025-12-04T12:25:16.7734076Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7734255Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7734454Z ================== 1 failed, 204 deselected, 2 rerun in 4.95s ================== 2025-12-04T12:25:16.7734552Z Got exit code 1 2025-12-04T12:25:16.7734661Z Retrying single test... 2025-12-04T12:25:16.7735308Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ede94f48d428153a.xml 2025-12-04T12:25:16.7735470Z ============================= test session starts ============================== 2025-12-04T12:25:16.7735817Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7735927Z cachedir: .pytest_cache 2025-12-04T12:25:16.7736452Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7736572Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7736679Z configfile: pytest.ini 2025-12-04T12:25:16.7737370Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7737603Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.7738565Z stepcurrent: skipping 104 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.7738680Z Running 1 items in this shard 2025-12-04T12:25:16.7738685Z 2025-12-04T12:25:16.7739449Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 ('RERUN', {'yellow': True}) [4.6432s] [100%] 2025-12-04T12:25:16.7740200Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 ('RERUN', {'yellow': True}) [0.1459s] [100%] 2025-12-04T12:25:16.7740859Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 FAILED [0.1399s] [100%] 2025-12-04T12:25:16.7740865Z 2025-12-04T12:25:16.7741013Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7741407Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 _ 2025-12-04T12:25:16.7741539Z Traceback (most recent call last): 2025-12-04T12:25:16.7742071Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7742299Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7742517Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7742522Z 2025-12-04T12:25:16.7742731Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7743526Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.7743531Z 2025-12-04T12:25:16.7743825Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7744045Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7744162Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7744272Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7744391Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7744620Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7744717Z graph_break [] 2025-12-04T12:25:16.7744937Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7745667Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7745764Z warnings.warn( 2025-12-04T12:25:16.7746169Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 _ 2025-12-04T12:25:16.7746288Z Traceback (most recent call last): 2025-12-04T12:25:16.7746816Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7747020Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7747225Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7747230Z 2025-12-04T12:25:16.7747447Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7748230Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.7748235Z 2025-12-04T12:25:16.7748498Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7748718Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7748828Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7748949Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7749065Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7749342Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7749443Z graph_break [] 2025-12-04T12:25:16.7749653Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7750380Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7750489Z warnings.warn( 2025-12-04T12:25:16.7750704Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7750825Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7750937Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7751153Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7751281Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7751376Z graph_break [] 2025-12-04T12:25:16.7751582Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7752322Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7752417Z warnings.warn( 2025-12-04T12:25:16.7752554Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7752988Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 _ 2025-12-04T12:25:16.7753105Z Traceback (most recent call last): 2025-12-04T12:25:16.7753644Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7753839Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7754043Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7754049Z 2025-12-04T12:25:16.7754299Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7755081Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.7755087Z 2025-12-04T12:25:16.7755363Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7755577Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7755687Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7755812Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7755930Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7756148Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7756255Z graph_break [] 2025-12-04T12:25:16.7756466Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7757198Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7757298Z warnings.warn( 2025-12-04T12:25:16.7757506Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7757620Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7757731Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7757945Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7758074Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7758166Z graph_break [] 2025-12-04T12:25:16.7758380Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7759101Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7759197Z warnings.warn( 2025-12-04T12:25:16.7759424Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7759590Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7759702Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7759925Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7760042Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7760147Z graph_break [] 2025-12-04T12:25:16.7760360Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7761081Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7761188Z warnings.warn( 2025-12-04T12:25:16.7762004Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ede94f48d428153a.xml - 2025-12-04T12:25:16.7762174Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7763109Z FAILED [0.1399s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7763115Z 2025-12-04T12:25:16.7763327Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7764145Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.7764150Z 2025-12-04T12:25:16.7764416Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7764601Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7764800Z ================== 1 failed, 204 deselected, 2 rerun in 4.98s ================== 2025-12-04T12:25:16.7764930Z Got exit code 1 2025-12-04T12:25:16.7765636Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0 2025-12-04T12:25:16.7766047Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.7766683Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6ae027693c5dbbf4.xml 2025-12-04T12:25:16.7766849Z ============================= test session starts ============================== 2025-12-04T12:25:16.7767194Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7767308Z cachedir: .pytest_cache 2025-12-04T12:25:16.7767828Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7767947Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7768066Z configfile: pytest.ini 2025-12-04T12:25:16.7768658Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7768892Z collecting ... collected 380 items / 105 deselected / 275 selected 2025-12-04T12:25:16.7769038Z stepcurrent: skipping 105 already run items. 2025-12-04T12:25:16.7769149Z Running 100 items in this shard 2025-12-04T12:25:16.7769156Z 2025-12-04T12:25:16.7769924Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 ('RERUN', {'yellow': True}) [4.6265s] [ 1%] 2025-12-04T12:25:16.7770664Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 ('RERUN', {'yellow': True}) [0.1405s] [ 1%] 2025-12-04T12:25:16.7771329Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 FAILED [0.1385s] [ 1%] 2025-12-04T12:25:16.7771394Z 2025-12-04T12:25:16.7771536Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7771917Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 _ 2025-12-04T12:25:16.7772045Z Traceback (most recent call last): 2025-12-04T12:25:16.7772574Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7772777Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7772987Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7772992Z 2025-12-04T12:25:16.7773203Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7773994Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.7773999Z 2025-12-04T12:25:16.7774266Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7774487Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7774595Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7774706Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7774833Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7775083Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7775177Z graph_break [] 2025-12-04T12:25:16.7775398Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7776124Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7776235Z warnings.warn( 2025-12-04T12:25:16.7776651Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 _ 2025-12-04T12:25:16.7776775Z Traceback (most recent call last): 2025-12-04T12:25:16.7777397Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7777591Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7777799Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7777806Z 2025-12-04T12:25:16.7778032Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7778806Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.7778811Z 2025-12-04T12:25:16.7779088Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7779303Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7779412Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7779538Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7779657Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7779875Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7779982Z graph_break [] 2025-12-04T12:25:16.7780196Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7780936Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7781033Z warnings.warn( 2025-12-04T12:25:16.7781245Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7781361Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7781470Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7781687Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7781810Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7781983Z graph_break [] 2025-12-04T12:25:16.7782199Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7782918Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7783015Z warnings.warn( 2025-12-04T12:25:16.7783166Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7783551Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 _ 2025-12-04T12:25:16.7783667Z Traceback (most recent call last): 2025-12-04T12:25:16.7784206Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7784399Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7784621Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7784630Z 2025-12-04T12:25:16.7784844Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7785623Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.7785629Z 2025-12-04T12:25:16.7785929Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7786145Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7786262Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7786370Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7786485Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7786711Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7786805Z graph_break [] 2025-12-04T12:25:16.7787045Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7787783Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7787881Z warnings.warn( 2025-12-04T12:25:16.7788101Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7788213Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7788319Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7788552Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7788667Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7788759Z graph_break [] 2025-12-04T12:25:16.7788979Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7789702Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7789809Z warnings.warn( 2025-12-04T12:25:16.7790022Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7790129Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7790252Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7790462Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7790580Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7790685Z graph_break [] 2025-12-04T12:25:16.7790893Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7791618Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7791723Z warnings.warn( 2025-12-04T12:25:16.7792540Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6ae027693c5dbbf4.xml - 2025-12-04T12:25:16.7792776Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7793698Z FAILED [0.1385s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7793704Z 2025-12-04T12:25:16.7793931Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7794704Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.7794709Z 2025-12-04T12:25:16.7794967Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7795150Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7795350Z ================== 1 failed, 105 deselected, 2 rerun in 4.96s ================== 2025-12-04T12:25:16.7795451Z Got exit code 1 2025-12-04T12:25:16.7795560Z Retrying single test... 2025-12-04T12:25:16.7796429Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-48db12bdca4a53d6.xml 2025-12-04T12:25:16.7796598Z ============================= test session starts ============================== 2025-12-04T12:25:16.7797006Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7797114Z cachedir: .pytest_cache 2025-12-04T12:25:16.7797647Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7797768Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7797875Z configfile: pytest.ini 2025-12-04T12:25:16.7798511Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7798745Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.7799617Z stepcurrent: skipping 105 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.7799736Z Running 1 items in this shard 2025-12-04T12:25:16.7799741Z 2025-12-04T12:25:16.7800503Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 ('RERUN', {'yellow': True}) [4.6390s] [100%] 2025-12-04T12:25:16.7801248Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 ('RERUN', {'yellow': True}) [0.1430s] [100%] 2025-12-04T12:25:16.7801912Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 FAILED [0.1407s] [100%] 2025-12-04T12:25:16.7801921Z 2025-12-04T12:25:16.7802072Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7802458Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 _ 2025-12-04T12:25:16.7802590Z Traceback (most recent call last): 2025-12-04T12:25:16.7803123Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7803320Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7803546Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7803551Z 2025-12-04T12:25:16.7803764Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7804562Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.7804567Z 2025-12-04T12:25:16.7804910Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7805123Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7805248Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7805360Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7805480Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7805719Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7805816Z graph_break [] 2025-12-04T12:25:16.7806042Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7806772Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7806872Z warnings.warn( 2025-12-04T12:25:16.7807278Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 _ 2025-12-04T12:25:16.7807401Z Traceback (most recent call last): 2025-12-04T12:25:16.7807929Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7808132Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7808343Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7808348Z 2025-12-04T12:25:16.7808605Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7809384Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.7809389Z 2025-12-04T12:25:16.7809652Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7809877Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7810020Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7810147Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7810264Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7810484Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7810590Z graph_break [] 2025-12-04T12:25:16.7810800Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7811531Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7811638Z warnings.warn( 2025-12-04T12:25:16.7811849Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7811969Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7812080Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7812299Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7812432Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7812533Z graph_break [] 2025-12-04T12:25:16.7812743Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7813477Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7813574Z warnings.warn( 2025-12-04T12:25:16.7813725Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7814114Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 _ 2025-12-04T12:25:16.7814232Z Traceback (most recent call last): 2025-12-04T12:25:16.7814772Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7814968Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7815176Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7815237Z 2025-12-04T12:25:16.7815459Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7816237Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.7816242Z 2025-12-04T12:25:16.7816520Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7816737Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7816903Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7817033Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7817152Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7817367Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7817474Z graph_break [] 2025-12-04T12:25:16.7817687Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7818433Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7818529Z warnings.warn( 2025-12-04T12:25:16.7818738Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7818863Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7819007Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7819223Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7819355Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7819447Z graph_break [] 2025-12-04T12:25:16.7819670Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7820437Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7820537Z warnings.warn( 2025-12-04T12:25:16.7820762Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7820869Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7820982Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7821205Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7821323Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7821431Z graph_break [] 2025-12-04T12:25:16.7821640Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7822366Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7822471Z warnings.warn( 2025-12-04T12:25:16.7823298Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-48db12bdca4a53d6.xml - 2025-12-04T12:25:16.7823468Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7824399Z FAILED [0.1407s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7824404Z 2025-12-04T12:25:16.7824621Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7825406Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.7825411Z 2025-12-04T12:25:16.7825673Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7825861Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7826059Z ================== 1 failed, 204 deselected, 2 rerun in 4.98s ================== 2025-12-04T12:25:16.7826216Z Got exit code 1 2025-12-04T12:25:16.7826330Z Retrying single test... 2025-12-04T12:25:16.7826978Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0c861fbac71335e0.xml 2025-12-04T12:25:16.7827139Z ============================= test session starts ============================== 2025-12-04T12:25:16.7827498Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7827605Z cachedir: .pytest_cache 2025-12-04T12:25:16.7828133Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7828255Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7828362Z configfile: pytest.ini 2025-12-04T12:25:16.7828963Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7829190Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.7830052Z stepcurrent: skipping 105 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.7830179Z Running 1 items in this shard 2025-12-04T12:25:16.7830185Z 2025-12-04T12:25:16.7830966Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 ('RERUN', {'yellow': True}) [4.6135s] [100%] 2025-12-04T12:25:16.7831723Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 ('RERUN', {'yellow': True}) [0.1425s] [100%] 2025-12-04T12:25:16.7832410Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 FAILED [0.1387s] [100%] 2025-12-04T12:25:16.7832417Z 2025-12-04T12:25:16.7832570Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7832958Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 _ 2025-12-04T12:25:16.7833074Z Traceback (most recent call last): 2025-12-04T12:25:16.7833618Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7833813Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7834028Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7834033Z 2025-12-04T12:25:16.7834246Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7835027Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.7835034Z 2025-12-04T12:25:16.7835307Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7835527Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7835645Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7835757Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7835878Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7836104Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7836200Z graph_break [] 2025-12-04T12:25:16.7836413Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7837159Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7837260Z warnings.warn( 2025-12-04T12:25:16.7837649Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 _ 2025-12-04T12:25:16.7837778Z Traceback (most recent call last): 2025-12-04T12:25:16.7838370Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7838576Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7838784Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7838790Z 2025-12-04T12:25:16.7838999Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7839793Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.7839798Z 2025-12-04T12:25:16.7840060Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7840288Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7840397Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7840513Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7840644Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7840858Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7840954Z graph_break [] 2025-12-04T12:25:16.7841174Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7841936Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7842047Z warnings.warn( 2025-12-04T12:25:16.7842258Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7842371Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7842491Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7842704Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7842852Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7842959Z graph_break [] 2025-12-04T12:25:16.7843171Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7843901Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7843999Z warnings.warn( 2025-12-04T12:25:16.7844139Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7844538Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 _ 2025-12-04T12:25:16.7844660Z Traceback (most recent call last): 2025-12-04T12:25:16.7845188Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7845393Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7845605Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7845615Z 2025-12-04T12:25:16.7845836Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7846610Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.7846616Z 2025-12-04T12:25:16.7846881Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7847101Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7847213Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7847334Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7847452Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7847669Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7847776Z graph_break [] 2025-12-04T12:25:16.7847992Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7848755Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7848896Z warnings.warn( 2025-12-04T12:25:16.7849108Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7849226Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7849341Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7849556Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7849682Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7849776Z graph_break [] 2025-12-04T12:25:16.7849984Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7850726Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7850823Z warnings.warn( 2025-12-04T12:25:16.7851049Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7851157Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7851270Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7851496Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7851614Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7851738Z graph_break [] 2025-12-04T12:25:16.7851964Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7852688Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7852785Z warnings.warn( 2025-12-04T12:25:16.7853635Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0c861fbac71335e0.xml - 2025-12-04T12:25:16.7853805Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7854741Z FAILED [0.1387s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7854747Z 2025-12-04T12:25:16.7854964Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7855752Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.7855758Z 2025-12-04T12:25:16.7856020Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7856197Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7856410Z ================== 1 failed, 204 deselected, 2 rerun in 4.95s ================== 2025-12-04T12:25:16.7856510Z Got exit code 1 2025-12-04T12:25:16.7857282Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0 2025-12-04T12:25:16.7857707Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.7858351Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ffbb324af5af0672.xml 2025-12-04T12:25:16.7858524Z ============================= test session starts ============================== 2025-12-04T12:25:16.7858877Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7858986Z cachedir: .pytest_cache 2025-12-04T12:25:16.7859516Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7859641Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7859830Z configfile: pytest.ini 2025-12-04T12:25:16.7860420Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7860648Z collecting ... collected 380 items / 106 deselected / 274 selected 2025-12-04T12:25:16.7860809Z stepcurrent: skipping 106 already run items. 2025-12-04T12:25:16.7860926Z Running 99 items in this shard 2025-12-04T12:25:16.7860932Z 2025-12-04T12:25:16.7861693Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 ('RERUN', {'yellow': True}) [4.6021s] [ 1%] 2025-12-04T12:25:16.7862447Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 ('RERUN', {'yellow': True}) [0.1417s] [ 1%] 2025-12-04T12:25:16.7863105Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 FAILED [0.1393s] [ 1%] 2025-12-04T12:25:16.7863115Z 2025-12-04T12:25:16.7863270Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7863659Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 _ 2025-12-04T12:25:16.7863797Z Traceback (most recent call last): 2025-12-04T12:25:16.7864379Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7864575Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7864797Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7864802Z 2025-12-04T12:25:16.7865015Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7865839Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.7865849Z 2025-12-04T12:25:16.7866116Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7866333Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7866455Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7866567Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7866692Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7866922Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7867018Z graph_break [] 2025-12-04T12:25:16.7867241Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7867970Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7868072Z warnings.warn( 2025-12-04T12:25:16.7868471Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 _ 2025-12-04T12:25:16.7868596Z Traceback (most recent call last): 2025-12-04T12:25:16.7869130Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7869337Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7869552Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7869557Z 2025-12-04T12:25:16.7869780Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7870558Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.7870563Z 2025-12-04T12:25:16.7870828Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7871050Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7871235Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7871358Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7871475Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7871692Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7871798Z graph_break [] 2025-12-04T12:25:16.7872016Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7872747Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7872858Z warnings.warn( 2025-12-04T12:25:16.7873066Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7873184Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7873296Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7873515Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7873651Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7873746Z graph_break [] 2025-12-04T12:25:16.7873958Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7874726Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7874826Z warnings.warn( 2025-12-04T12:25:16.7874977Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7875364Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 _ 2025-12-04T12:25:16.7875484Z Traceback (most recent call last): 2025-12-04T12:25:16.7876024Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7876249Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7876461Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7876467Z 2025-12-04T12:25:16.7876686Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7877464Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.7877469Z 2025-12-04T12:25:16.7877739Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7877952Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7878062Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7878181Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7878300Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7878517Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7878627Z graph_break [] 2025-12-04T12:25:16.7878838Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7879576Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7879676Z warnings.warn( 2025-12-04T12:25:16.7879890Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7880010Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7880121Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7880336Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7880464Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7880560Z graph_break [] 2025-12-04T12:25:16.7880779Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7881506Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7881735Z warnings.warn( 2025-12-04T12:25:16.7881957Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7882063Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7882172Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7882404Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7882527Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7882635Z graph_break [] 2025-12-04T12:25:16.7882850Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7883574Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7883684Z warnings.warn( 2025-12-04T12:25:16.7884508Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ffbb324af5af0672.xml - 2025-12-04T12:25:16.7884679Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7885641Z FAILED [0.1393s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7885647Z 2025-12-04T12:25:16.7885861Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7886646Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.7886651Z 2025-12-04T12:25:16.7886915Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7887128Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7887325Z ================== 1 failed, 106 deselected, 2 rerun in 4.94s ================== 2025-12-04T12:25:16.7887425Z Got exit code 1 2025-12-04T12:25:16.7887539Z Retrying single test... 2025-12-04T12:25:16.7888183Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-042dc549bdd77a45.xml 2025-12-04T12:25:16.7888346Z ============================= test session starts ============================== 2025-12-04T12:25:16.7888706Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7888815Z cachedir: .pytest_cache 2025-12-04T12:25:16.7889347Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7889468Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7889573Z configfile: pytest.ini 2025-12-04T12:25:16.7890179Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7890409Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.7891271Z stepcurrent: skipping 106 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.7891401Z Running 1 items in this shard 2025-12-04T12:25:16.7891406Z 2025-12-04T12:25:16.7892156Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 ('RERUN', {'yellow': True}) [4.6460s] [100%] 2025-12-04T12:25:16.7892905Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 ('RERUN', {'yellow': True}) [0.1425s] [100%] 2025-12-04T12:25:16.7893563Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 FAILED [0.1399s] [100%] 2025-12-04T12:25:16.7893641Z 2025-12-04T12:25:16.7893789Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7894177Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 _ 2025-12-04T12:25:16.7894295Z Traceback (most recent call last): 2025-12-04T12:25:16.7894830Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7895027Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7895245Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7895250Z 2025-12-04T12:25:16.7895461Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7896384Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.7896394Z 2025-12-04T12:25:16.7896669Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7896938Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7897062Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7897174Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7897356Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7897590Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7897687Z graph_break [] 2025-12-04T12:25:16.7897902Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7898648Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7898747Z warnings.warn( 2025-12-04T12:25:16.7899177Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 _ 2025-12-04T12:25:16.7899314Z Traceback (most recent call last): 2025-12-04T12:25:16.7899848Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7900053Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7900268Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7900274Z 2025-12-04T12:25:16.7900490Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7901274Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.7901280Z 2025-12-04T12:25:16.7901543Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7901771Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7901883Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7901998Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7902129Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7902349Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7902444Z graph_break [] 2025-12-04T12:25:16.7902674Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7903399Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7903506Z warnings.warn( 2025-12-04T12:25:16.7903718Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7903826Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7903948Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7904170Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7904389Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7904496Z graph_break [] 2025-12-04T12:25:16.7904751Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7905632Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7905732Z warnings.warn( 2025-12-04T12:25:16.7905873Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7906277Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 _ 2025-12-04T12:25:16.7906397Z Traceback (most recent call last): 2025-12-04T12:25:16.7906924Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7907136Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7907350Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7907355Z 2025-12-04T12:25:16.7907578Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7908358Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.7908406Z 2025-12-04T12:25:16.7908682Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7908912Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7909023Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7909152Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7909274Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7909493Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7909636Z graph_break [] 2025-12-04T12:25:16.7909852Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7910595Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7911607Z warnings.warn( 2025-12-04T12:25:16.7912004Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7912486Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7912812Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7913250Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7913740Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7914072Z graph_break [] 2025-12-04T12:25:16.7914452Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7915546Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7916521Z warnings.warn( 2025-12-04T12:25:16.7916893Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7917379Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7917709Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7918170Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7918692Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7919032Z graph_break [] 2025-12-04T12:25:16.7919403Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7920487Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7921453Z warnings.warn( 2025-12-04T12:25:16.7922442Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-042dc549bdd77a45.xml - 2025-12-04T12:25:16.7923651Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7924890Z FAILED [0.1399s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7925961Z 2025-12-04T12:25:16.7926179Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7927313Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.7928227Z 2025-12-04T12:25:16.7928508Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7929096Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.7929611Z ================== 1 failed, 204 deselected, 2 rerun in 4.98s ================== 2025-12-04T12:25:16.7930060Z Got exit code 1 2025-12-04T12:25:16.7930331Z Retrying single test... 2025-12-04T12:25:16.7931149Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0942d088a0709320.xml 2025-12-04T12:25:16.7932124Z ============================= test session starts ============================== 2025-12-04T12:25:16.7932787Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.7933393Z cachedir: .pytest_cache 2025-12-04T12:25:16.7934093Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.7934880Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.7935231Z configfile: pytest.ini 2025-12-04T12:25:16.7936027Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.7937071Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.7938307Z stepcurrent: skipping 106 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.7939440Z Running 1 items in this shard 2025-12-04T12:25:16.7939653Z 2025-12-04T12:25:16.7940413Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 ('RERUN', {'yellow': True}) [4.6298s] [100%] 2025-12-04T12:25:16.7942058Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 ('RERUN', {'yellow': True}) [0.1434s] [100%] 2025-12-04T12:25:16.7943613Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 FAILED [0.1410s] [100%] 2025-12-04T12:25:16.7944412Z 2025-12-04T12:25:16.7944567Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.7945251Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 _ 2025-12-04T12:25:16.7945899Z Traceback (most recent call last): 2025-12-04T12:25:16.7946672Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7947541Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7948080Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7948432Z 2025-12-04T12:25:16.7948649Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7949785Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.7950772Z 2025-12-04T12:25:16.7951056Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7951677Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7952153Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7952489Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7952829Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7953280Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7953741Z graph_break [] 2025-12-04T12:25:16.7954120Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7955208Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7956175Z warnings.warn( 2025-12-04T12:25:16.7956735Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 _ 2025-12-04T12:25:16.7957393Z Traceback (most recent call last): 2025-12-04T12:25:16.7958154Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7959025Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7959608Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7959949Z 2025-12-04T12:25:16.7960164Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7961302Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.7962230Z 2025-12-04T12:25:16.7962500Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7963160Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7963627Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7963964Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7964309Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7964762Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7965226Z graph_break [] 2025-12-04T12:25:16.7965609Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7966716Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7967677Z warnings.warn( 2025-12-04T12:25:16.7968061Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7968536Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7968860Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7969300Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7969791Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7970129Z graph_break [] 2025-12-04T12:25:16.7970489Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7971580Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7972556Z warnings.warn( 2025-12-04T12:25:16.7972854Z =================================== FAILURES =================================== 2025-12-04T12:25:16.7973536Z _ NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 _ 2025-12-04T12:25:16.7974187Z Traceback (most recent call last): 2025-12-04T12:25:16.7974971Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 392, in test_layer_norm_bwd_with_bias 2025-12-04T12:25:16.7975826Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.7976376Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7976779Z 2025-12-04T12:25:16.7977073Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.7978192Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.7979124Z 2025-12-04T12:25:16.7979392Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.7980017Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7980495Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7980822Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7981157Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7981621Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7982081Z graph_break [] 2025-12-04T12:25:16.7982448Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7983550Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7984516Z warnings.warn( 2025-12-04T12:25:16.7984885Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7985395Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7985732Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7986157Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7986644Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7986989Z graph_break [] 2025-12-04T12:25:16.7987359Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7988475Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7989452Z warnings.warn( 2025-12-04T12:25:16.7989836Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.7990292Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.7990625Z stats [('calls_captured', 3)] 2025-12-04T12:25:16.7991062Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.7991546Z inductor [('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.7991871Z graph_break [] 2025-12-04T12:25:16.7992241Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.7993337Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.7994288Z warnings.warn( 2025-12-04T12:25:16.7995271Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0942d088a0709320.xml - 2025-12-04T12:25:16.7996604Z =========================== short test summary info ============================ 2025-12-04T12:25:16.7997857Z FAILED [0.1410s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.7998908Z 2025-12-04T12:25:16.7999130Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8000272Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.8001200Z 2025-12-04T12:25:16.8001466Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8002055Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.8002572Z ================== 1 failed, 204 deselected, 2 rerun in 4.97s ================== 2025-12-04T12:25:16.8003154Z Got exit code 1 2025-12-04T12:25:16.8004019Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1 2025-12-04T12:25:16.8005269Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.8006458Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c3fbd4a060fe6236.xml 2025-12-04T12:25:16.8007407Z ============================= test session starts ============================== 2025-12-04T12:25:16.8008071Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.8008672Z cachedir: .pytest_cache 2025-12-04T12:25:16.8009374Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.8010165Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.8010522Z configfile: pytest.ini 2025-12-04T12:25:16.8011285Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.8012303Z collecting ... collected 380 items / 107 deselected / 273 selected 2025-12-04T12:25:16.8012901Z stepcurrent: skipping 107 already run items. 2025-12-04T12:25:16.8013296Z Running 98 items in this shard 2025-12-04T12:25:16.8013509Z 2025-12-04T12:25:16.8014237Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_float32_split_reductions_False_shape1 PASSED [5.6890s] [ 1%] 2025-12-04T12:25:16.8015707Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_float32_split_reductions_True_shape1 PASSED [1.1304s] [ 2%] 2025-12-04T12:25:16.8017381Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_dynamic_shape_dynamic_dims1 SKIPPED [0.0033s] (Mix order reduction not enabled) [ 3%] 2025-12-04T12:25:16.8018995Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_mean_swap_False_split_reductions_False_shape1 PASSED [0.3765s] [ 4%] 2025-12-04T12:25:16.8020542Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_mean_swap_False_split_reductions_False_shape2 PASSED [0.4053s] [ 5%] 2025-12-04T12:25:16.8022075Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_mean_swap_False_split_reductions_True_shape1 PASSED [0.4789s] [ 6%] 2025-12-04T12:25:16.8023602Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_mean_swap_False_split_reductions_True_shape2 PASSED [0.7217s] [ 7%] 2025-12-04T12:25:16.8025135Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_mean_swap_True_split_reductions_False_shape0 PASSED [0.5985s] [ 8%] 2025-12-04T12:25:16.8026673Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_mean_swap_True_split_reductions_False_shape1 PASSED [0.5914s] [ 9%] 2025-12-04T12:25:16.8028208Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_mean_swap_True_split_reductions_False_shape2 PASSED [0.6084s] [ 10%] 2025-12-04T12:25:16.8029725Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_mean_swap_True_split_reductions_True_shape0 PASSED [0.6159s] [ 11%] 2025-12-04T12:25:16.8031233Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_mean_swap_True_split_reductions_True_shape2 PASSED [0.6311s] [ 12%] 2025-12-04T12:25:16.8032763Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_prod_swap_False_split_reductions_False_shape0 PASSED [0.5077s] [ 13%] 2025-12-04T12:25:16.8034482Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_prod_swap_False_split_reductions_False_shape2 SKIPPED [0.0032s] (Invalid combination) [ 14%] 2025-12-04T12:25:16.8036130Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_prod_swap_False_split_reductions_True_shape0 PASSED [0.2632s] [ 15%] 2025-12-04T12:25:16.8037779Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_prod_swap_False_split_reductions_True_shape2 SKIPPED [0.0032s] (Invalid combination) [ 16%] 2025-12-04T12:25:16.8039416Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_prod_swap_True_split_reductions_False_shape1 PASSED [0.2260s] [ 17%] 2025-12-04T12:25:16.8040942Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_prod_swap_True_split_reductions_True_shape1 PASSED [0.2551s] [ 18%] 2025-12-04T12:25:16.8042573Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_prod_swap_True_split_reductions_True_shape2 SKIPPED [0.0033s] (Invalid combination) [ 19%] 2025-12-04T12:25:16.8044213Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_sum_swap_False_split_reductions_False_shape1 PASSED [0.2785s] [ 20%] 2025-12-04T12:25:16.8045762Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_sum_swap_False_split_reductions_False_shape2 PASSED [0.2925s] [ 21%] 2025-12-04T12:25:16.8047281Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_sum_swap_False_split_reductions_True_shape0 PASSED [0.3049s] [ 22%] 2025-12-04T12:25:16.8048803Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_sum_swap_False_split_reductions_True_shape1 PASSED [0.3048s] [ 23%] 2025-12-04T12:25:16.8050348Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_sum_swap_True_split_reductions_False_shape0 PASSED [0.2744s] [ 24%] 2025-12-04T12:25:16.8051857Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_mix_order_reduction_name_sum_swap_True_split_reductions_True_shape2 PASSED [0.5307s] [ 25%] 2025-12-04T12:25:16.8053615Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [0.2389s] [ 26%] 2025-12-04T12:25:16.8055622Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [0.1684s] [ 26%] 2025-12-04T12:25:16.8057610Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True FAILED [0.1662s] [ 26%] 2025-12-04T12:25:16.8058594Z 2025-12-04T12:25:16.8058754Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.8059613Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.8060620Z Traceback (most recent call last): 2025-12-04T12:25:16.8061347Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8062170Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8062710Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8063067Z 2025-12-04T12:25:16.8063286Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8064620Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.8065808Z 2025-12-04T12:25:16.8066091Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8066710Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8067188Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8067526Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8067953Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8068652Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8069236Z graph_break [] 2025-12-04T12:25:16.8069609Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8072691Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8075708Z return x.grad, w.grad 2025-12-04T12:25:16.8076622Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8077596Z warnings.warn( 2025-12-04T12:25:16.8080509Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8083452Z return x.grad, w.grad 2025-12-04T12:25:16.8084216Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.8085048Z Traceback (most recent call last): 2025-12-04T12:25:16.8085763Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8086568Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8087104Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8087457Z 2025-12-04T12:25:16.8087675Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8088993Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.8090095Z 2025-12-04T12:25:16.8090376Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8091006Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8091469Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8091805Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8092244Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8092937Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8093524Z graph_break [] 2025-12-04T12:25:16.8093904Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8097216Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8100268Z return x.grad, w.grad 2025-12-04T12:25:16.8101181Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8102149Z warnings.warn( 2025-12-04T12:25:16.8105072Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8107999Z return x.grad, w.grad 2025-12-04T12:25:16.8108393Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8108868Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8109203Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8109642Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8110397Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8110975Z graph_break [] 2025-12-04T12:25:16.8111355Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8114431Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8117342Z return x.grad, w.grad 2025-12-04T12:25:16.8118259Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8119234Z warnings.warn( 2025-12-04T12:25:16.8122122Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8125090Z return x.grad, w.grad 2025-12-04T12:25:16.8125489Z =================================== FAILURES =================================== 2025-12-04T12:25:16.8126402Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.8127246Z Traceback (most recent call last): 2025-12-04T12:25:16.8128044Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8128835Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8129382Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8129723Z 2025-12-04T12:25:16.8129957Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8131273Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.8132369Z 2025-12-04T12:25:16.8132634Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8133260Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8133740Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8134069Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8134507Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8135212Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8135798Z graph_break [] 2025-12-04T12:25:16.8136194Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8139383Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8149257Z return x.grad, w.grad 2025-12-04T12:25:16.8150268Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8151245Z warnings.warn( 2025-12-04T12:25:16.8154150Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8157097Z return x.grad, w.grad 2025-12-04T12:25:16.8157501Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8157968Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8158293Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8158736Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8159449Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8160024Z graph_break [] 2025-12-04T12:25:16.8160395Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8163456Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8166522Z return x.grad, w.grad 2025-12-04T12:25:16.8167444Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8168398Z warnings.warn( 2025-12-04T12:25:16.8171264Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8174182Z return x.grad, w.grad 2025-12-04T12:25:16.8174590Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8175073Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8175397Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8175862Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8176565Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8177243Z graph_break [] 2025-12-04T12:25:16.8177618Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8178883Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8179853Z warnings.warn( 2025-12-04T12:25:16.8182714Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8185631Z return x.grad, w.grad 2025-12-04T12:25:16.8186638Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c3fbd4a060fe6236.xml - 2025-12-04T12:25:16.8187779Z =========================== short test summary info ============================ 2025-12-04T12:25:16.8189195Z FAILED [0.1662s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8190420Z 2025-12-04T12:25:16.8190642Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8191947Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.8193053Z 2025-12-04T12:25:16.8193319Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8193901Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.8194464Z ====== 1 failed, 21 passed, 4 skipped, 107 deselected, 2 rerun in 15.77s ======= 2025-12-04T12:25:16.8194938Z Got exit code 1 2025-12-04T12:25:16.8195241Z Retrying single test... 2025-12-04T12:25:16.8196289Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f308a9dbba7be53a.xml 2025-12-04T12:25:16.8197225Z ============================= test session starts ============================== 2025-12-04T12:25:16.8197888Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.8198488Z cachedir: .pytest_cache 2025-12-04T12:25:16.8199202Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.8199977Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.8200321Z configfile: pytest.ini 2025-12-04T12:25:16.8201090Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.8202051Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.8203464Z stepcurrent: skipping 132 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.8204747Z Running 1 items in this shard 2025-12-04T12:25:16.8204955Z 2025-12-04T12:25:16.8205962Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [4.6606s] [100%] 2025-12-04T12:25:16.8207968Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [0.1705s] [100%] 2025-12-04T12:25:16.8209910Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True FAILED [0.1666s] [100%] 2025-12-04T12:25:16.8210911Z 2025-12-04T12:25:16.8211050Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.8211913Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.8212756Z Traceback (most recent call last): 2025-12-04T12:25:16.8213455Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8214263Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8214808Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8215144Z 2025-12-04T12:25:16.8215373Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8216687Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.8217870Z 2025-12-04T12:25:16.8218135Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8218764Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8219243Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8219564Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8220124Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8220823Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8221265Z graph_break [] 2025-12-04T12:25:16.8221631Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8224708Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8227737Z return x.grad, w.grad 2025-12-04T12:25:16.8228654Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8229619Z warnings.warn( 2025-12-04T12:25:16.8232486Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8235422Z return x.grad, w.grad 2025-12-04T12:25:16.8236247Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.8237084Z Traceback (most recent call last): 2025-12-04T12:25:16.8237778Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8238581Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8239156Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8239499Z 2025-12-04T12:25:16.8239718Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8241019Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.8242115Z 2025-12-04T12:25:16.8242382Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8243006Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8243477Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8243801Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8244357Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8245060Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8245509Z graph_break [] 2025-12-04T12:25:16.8245885Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8248953Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8251871Z return x.grad, w.grad 2025-12-04T12:25:16.8252794Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8253762Z warnings.warn( 2025-12-04T12:25:16.8256633Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8259692Z return x.grad, w.grad 2025-12-04T12:25:16.8260098Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8260569Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8260892Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8261329Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8262029Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8262602Z graph_break [] 2025-12-04T12:25:16.8262966Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8266068Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8268988Z return x.grad, w.grad 2025-12-04T12:25:16.8269940Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8270903Z warnings.warn( 2025-12-04T12:25:16.8273762Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8276685Z return x.grad, w.grad 2025-12-04T12:25:16.8277015Z =================================== FAILURES =================================== 2025-12-04T12:25:16.8277886Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.8278714Z Traceback (most recent call last): 2025-12-04T12:25:16.8279410Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8280216Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8280762Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8281100Z 2025-12-04T12:25:16.8281314Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8282627Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.8283711Z 2025-12-04T12:25:16.8283985Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8284613Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8285163Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8285493Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8286055Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8286746Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8287204Z graph_break [] 2025-12-04T12:25:16.8287578Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8290644Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8293567Z return x.grad, w.grad 2025-12-04T12:25:16.8294477Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8295472Z warnings.warn( 2025-12-04T12:25:16.8299220Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8302169Z return x.grad, w.grad 2025-12-04T12:25:16.8302569Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8303032Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8303372Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8303810Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8304505Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8305083Z graph_break [] 2025-12-04T12:25:16.8305451Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8308514Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8311550Z return x.grad, w.grad 2025-12-04T12:25:16.8312463Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8313430Z warnings.warn( 2025-12-04T12:25:16.8316314Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8319333Z return x.grad, w.grad 2025-12-04T12:25:16.8319733Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8320194Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8320523Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8320955Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8321640Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8322217Z graph_break [] 2025-12-04T12:25:16.8322583Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8323670Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8324639Z warnings.warn( 2025-12-04T12:25:16.8327573Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8330510Z return x.grad, w.grad 2025-12-04T12:25:16.8331524Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f308a9dbba7be53a.xml - 2025-12-04T12:25:16.8332683Z =========================== short test summary info ============================ 2025-12-04T12:25:16.8334097Z FAILED [0.1666s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8335334Z 2025-12-04T12:25:16.8335551Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8336948Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.8338052Z 2025-12-04T12:25:16.8338337Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8338935Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.8339459Z ================== 1 failed, 204 deselected, 2 rerun in 5.05s ================== 2025-12-04T12:25:16.8339919Z Got exit code 1 2025-12-04T12:25:16.8340193Z Retrying single test... 2025-12-04T12:25:16.8341020Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-bd2ede9c3a5cd1d1.xml 2025-12-04T12:25:16.8341981Z ============================= test session starts ============================== 2025-12-04T12:25:16.8342647Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.8343258Z cachedir: .pytest_cache 2025-12-04T12:25:16.8343960Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.8344749Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.8345108Z configfile: pytest.ini 2025-12-04T12:25:16.8345876Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.8346928Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.8348357Z stepcurrent: skipping 132 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.8349659Z Running 1 items in this shard 2025-12-04T12:25:16.8349872Z 2025-12-04T12:25:16.8350802Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [4.6968s] [100%] 2025-12-04T12:25:16.8352799Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [0.1754s] [100%] 2025-12-04T12:25:16.8354712Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True FAILED [0.1745s] [100%] 2025-12-04T12:25:16.8355706Z 2025-12-04T12:25:16.8355845Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.8356743Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.8357582Z Traceback (most recent call last): 2025-12-04T12:25:16.8358286Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8359104Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8359659Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8359998Z 2025-12-04T12:25:16.8360244Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8361560Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.8362666Z 2025-12-04T12:25:16.8362932Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8363569Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8364033Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8364370Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8364934Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8365623Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8366085Z graph_break [] 2025-12-04T12:25:16.8366460Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8369532Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8372461Z return x.grad, w.grad 2025-12-04T12:25:16.8373371Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8374350Z warnings.warn( 2025-12-04T12:25:16.8377285Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8380292Z return x.grad, w.grad 2025-12-04T12:25:16.8381053Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.8381894Z Traceback (most recent call last): 2025-12-04T12:25:16.8382601Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8383416Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8383954Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8384308Z 2025-12-04T12:25:16.8384523Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8385844Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.8386962Z 2025-12-04T12:25:16.8387243Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8387877Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8388345Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8388681Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8389247Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8389977Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8390440Z graph_break [] 2025-12-04T12:25:16.8390818Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8393912Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8397012Z return x.grad, w.grad 2025-12-04T12:25:16.8397925Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8398900Z warnings.warn( 2025-12-04T12:25:16.8401767Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8404697Z return x.grad, w.grad 2025-12-04T12:25:16.8405104Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8405569Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8405903Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8406347Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8407146Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8407732Z graph_break [] 2025-12-04T12:25:16.8408116Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8411199Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8414140Z return x.grad, w.grad 2025-12-04T12:25:16.8415049Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8416023Z warnings.warn( 2025-12-04T12:25:16.8419002Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8421937Z return x.grad, w.grad 2025-12-04T12:25:16.8422320Z =================================== FAILURES =================================== 2025-12-04T12:25:16.8423186Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.8424032Z Traceback (most recent call last): 2025-12-04T12:25:16.8424745Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8425542Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8426088Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8426446Z 2025-12-04T12:25:16.8426661Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8427972Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.8429066Z 2025-12-04T12:25:16.8429334Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8429969Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8430445Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8430780Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8431327Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8432032Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8432488Z graph_break [] 2025-12-04T12:25:16.8432850Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8435923Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8438911Z return x.grad, w.grad 2025-12-04T12:25:16.8439833Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8440802Z warnings.warn( 2025-12-04T12:25:16.8443492Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8443617Z return x.grad, w.grad 2025-12-04T12:25:16.8443833Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8443946Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8444071Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8444326Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8444670Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8444778Z graph_break [] 2025-12-04T12:25:16.8444991Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8447731Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8447846Z return x.grad, w.grad 2025-12-04T12:25:16.8448586Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8448687Z warnings.warn( 2025-12-04T12:25:16.8451373Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8451498Z return x.grad, w.grad 2025-12-04T12:25:16.8451716Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8451836Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8451950Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8452170Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8452520Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8452616Z graph_break [] 2025-12-04T12:25:16.8452841Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8453569Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8453748Z warnings.warn( 2025-12-04T12:25:16.8456468Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8456573Z return x.grad, w.grad 2025-12-04T12:25:16.8457471Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-bd2ede9c3a5cd1d1.xml - 2025-12-04T12:25:16.8457649Z =========================== short test summary info ============================ 2025-12-04T12:25:16.8458793Z FAILED [0.1745s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8458799Z 2025-12-04T12:25:16.8459017Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8459982Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.8460000Z 2025-12-04T12:25:16.8460271Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8460480Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.8460696Z ================== 1 failed, 204 deselected, 2 rerun in 5.10s ================== 2025-12-04T12:25:16.8460793Z Got exit code 1 2025-12-04T12:25:16.8461678Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.8462104Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.8462751Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d27c0d840d6fe564.xml 2025-12-04T12:25:16.8462927Z ============================= test session starts ============================== 2025-12-04T12:25:16.8463278Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.8463390Z cachedir: .pytest_cache 2025-12-04T12:25:16.8463922Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.8464051Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.8464160Z configfile: pytest.ini 2025-12-04T12:25:16.8464768Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.8464997Z collecting ... collected 380 items / 133 deselected / 247 selected 2025-12-04T12:25:16.8465158Z stepcurrent: skipping 133 already run items. 2025-12-04T12:25:16.8465272Z Running 72 items in this shard 2025-12-04T12:25:16.8465277Z 2025-12-04T12:25:16.8466215Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6825s] [ 1%] 2025-12-04T12:25:16.8467173Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1663s] [ 1%] 2025-12-04T12:25:16.8468091Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1647s] [ 1%] 2025-12-04T12:25:16.8468097Z 2025-12-04T12:25:16.8468251Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.8468829Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.8468964Z Traceback (most recent call last): 2025-12-04T12:25:16.8469437Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8469635Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8469853Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8469862Z 2025-12-04T12:25:16.8470076Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8471088Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8471094Z 2025-12-04T12:25:16.8471358Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8471575Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8471703Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8471816Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8472152Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8472412Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8472517Z graph_break [] 2025-12-04T12:25:16.8472742Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8473479Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8473579Z warnings.warn( 2025-12-04T12:25:16.8474173Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.8474301Z Traceback (most recent call last): 2025-12-04T12:25:16.8474770Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8474979Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8475187Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8475195Z 2025-12-04T12:25:16.8475419Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8476398Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8476404Z 2025-12-04T12:25:16.8476680Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8476899Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8477012Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8477136Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8477474Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8477693Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8477799Z graph_break [] 2025-12-04T12:25:16.8478014Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8478824Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8478924Z warnings.warn( 2025-12-04T12:25:16.8479135Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8479256Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8479369Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8479586Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8479931Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8480040Z graph_break [] 2025-12-04T12:25:16.8480341Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8481090Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8481194Z warnings.warn( 2025-12-04T12:25:16.8481349Z =================================== FAILURES =================================== 2025-12-04T12:25:16.8481927Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.8482086Z Traceback (most recent call last): 2025-12-04T12:25:16.8482569Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8482769Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8482994Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8482999Z 2025-12-04T12:25:16.8483216Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8484219Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8484229Z 2025-12-04T12:25:16.8484510Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8484728Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8484849Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8484965Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8485302Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8485530Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8485627Z graph_break [] 2025-12-04T12:25:16.8485841Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8486584Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8486689Z warnings.warn( 2025-12-04T12:25:16.8486913Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8487022Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8487136Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8487365Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8487706Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8487804Z graph_break [] 2025-12-04T12:25:16.8488029Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8488755Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8488867Z warnings.warn( 2025-12-04T12:25:16.8489081Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8489251Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8489377Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8489598Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8489933Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8490042Z graph_break [] 2025-12-04T12:25:16.8490257Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8490993Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8491092Z warnings.warn( 2025-12-04T12:25:16.8491914Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d27c0d840d6fe564.xml - 2025-12-04T12:25:16.8492098Z =========================== short test summary info ============================ 2025-12-04T12:25:16.8493198Z FAILED [0.1647s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8493204Z 2025-12-04T12:25:16.8493464Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8494430Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8494435Z 2025-12-04T12:25:16.8494701Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8494889Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.8495143Z ================== 1 failed, 133 deselected, 2 rerun in 5.07s ================== 2025-12-04T12:25:16.8495260Z Got exit code 1 2025-12-04T12:25:16.8495365Z Retrying single test... 2025-12-04T12:25:16.8496172Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2f600ca197676d4b.xml 2025-12-04T12:25:16.8496336Z ============================= test session starts ============================== 2025-12-04T12:25:16.8496687Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.8496814Z cachedir: .pytest_cache 2025-12-04T12:25:16.8497395Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.8497525Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.8497648Z configfile: pytest.ini 2025-12-04T12:25:16.8498244Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.8498490Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.8499547Z stepcurrent: skipping 133 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8499665Z Running 1 items in this shard 2025-12-04T12:25:16.8499671Z 2025-12-04T12:25:16.8500619Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6493s] [100%] 2025-12-04T12:25:16.8501549Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1704s] [100%] 2025-12-04T12:25:16.8502413Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1662s] [100%] 2025-12-04T12:25:16.8502522Z 2025-12-04T12:25:16.8502665Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.8503264Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.8503388Z Traceback (most recent call last): 2025-12-04T12:25:16.8503857Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8504072Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8504285Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8504290Z 2025-12-04T12:25:16.8504517Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8505500Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8505510Z 2025-12-04T12:25:16.8505775Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8506046Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8506162Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8506296Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8506640Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8506862Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8506976Z graph_break [] 2025-12-04T12:25:16.8507194Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8507969Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8508089Z warnings.warn( 2025-12-04T12:25:16.8508666Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.8508801Z Traceback (most recent call last): 2025-12-04T12:25:16.8509274Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8509472Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8509698Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8509703Z 2025-12-04T12:25:16.8509916Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8510906Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8510916Z 2025-12-04T12:25:16.8511183Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8511402Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8511533Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8511652Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8511992Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8512223Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8512321Z graph_break [] 2025-12-04T12:25:16.8512545Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8513286Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8513386Z warnings.warn( 2025-12-04T12:25:16.8513670Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8513783Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8513896Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8514124Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8514464Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8514571Z graph_break [] 2025-12-04T12:25:16.8514784Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8515509Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8515617Z warnings.warn( 2025-12-04T12:25:16.8515762Z =================================== FAILURES =================================== 2025-12-04T12:25:16.8516344Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.8516481Z Traceback (most recent call last): 2025-12-04T12:25:16.8516953Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8517192Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8517406Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8517411Z 2025-12-04T12:25:16.8517623Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8518613Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8518618Z 2025-12-04T12:25:16.8518914Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8519147Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8519260Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8519374Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8519727Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8519946Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8520043Z graph_break [] 2025-12-04T12:25:16.8520271Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8521000Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8521111Z warnings.warn( 2025-12-04T12:25:16.8521324Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8521435Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8521565Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8521782Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8522116Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8522223Z graph_break [] 2025-12-04T12:25:16.8522436Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8523176Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8523275Z warnings.warn( 2025-12-04T12:25:16.8523487Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8523609Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8523724Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8523942Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8524348Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8524450Z graph_break [] 2025-12-04T12:25:16.8524677Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8525401Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8525503Z warnings.warn( 2025-12-04T12:25:16.8526333Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2f600ca197676d4b.xml - 2025-12-04T12:25:16.8526504Z =========================== short test summary info ============================ 2025-12-04T12:25:16.8527619Z FAILED [0.1662s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8527629Z 2025-12-04T12:25:16.8527846Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8528856Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8528874Z 2025-12-04T12:25:16.8529142Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8529321Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.8529534Z ================== 1 failed, 204 deselected, 2 rerun in 5.04s ================== 2025-12-04T12:25:16.8529633Z Got exit code 1 2025-12-04T12:25:16.8529742Z Retrying single test... 2025-12-04T12:25:16.8530436Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ecad5133d6245a6a.xml 2025-12-04T12:25:16.8530607Z ============================= test session starts ============================== 2025-12-04T12:25:16.8530966Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.8531074Z cachedir: .pytest_cache 2025-12-04T12:25:16.8531598Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.8531734Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.8531844Z configfile: pytest.ini 2025-12-04T12:25:16.8532433Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.8532671Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.8533734Z stepcurrent: skipping 133 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8533864Z Running 1 items in this shard 2025-12-04T12:25:16.8533870Z 2025-12-04T12:25:16.8534802Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6678s] [100%] 2025-12-04T12:25:16.8535750Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1653s] [100%] 2025-12-04T12:25:16.8536592Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1621s] [100%] 2025-12-04T12:25:16.8536600Z 2025-12-04T12:25:16.8536741Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.8537462Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.8537586Z Traceback (most recent call last): 2025-12-04T12:25:16.8538075Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8538274Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8538485Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8538491Z 2025-12-04T12:25:16.8538719Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8539695Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8539700Z 2025-12-04T12:25:16.8539984Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8540202Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8540315Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8540444Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8540842Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8541062Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8541176Z graph_break [] 2025-12-04T12:25:16.8541391Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8542132Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8542231Z warnings.warn( 2025-12-04T12:25:16.8542854Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.8542994Z Traceback (most recent call last): 2025-12-04T12:25:16.8543464Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8543673Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8543886Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8543892Z 2025-12-04T12:25:16.8544108Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8545090Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8545095Z 2025-12-04T12:25:16.8545365Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8545591Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8545709Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8545824Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8546171Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8546390Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8546489Z graph_break [] 2025-12-04T12:25:16.8546714Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8547449Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8547566Z warnings.warn( 2025-12-04T12:25:16.8547779Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8547892Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8548022Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8548310Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8548646Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8548759Z graph_break [] 2025-12-04T12:25:16.8548975Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8549714Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8549815Z warnings.warn( 2025-12-04T12:25:16.8549961Z =================================== FAILURES =================================== 2025-12-04T12:25:16.8550557Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.8550680Z Traceback (most recent call last): 2025-12-04T12:25:16.8551153Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8551364Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8551574Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8551579Z 2025-12-04T12:25:16.8551801Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8552803Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8552809Z 2025-12-04T12:25:16.8553077Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8553315Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8553426Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8553583Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8553926Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8554144Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8554253Z graph_break [] 2025-12-04T12:25:16.8554467Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8555198Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8555309Z warnings.warn( 2025-12-04T12:25:16.8555524Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8555646Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8555759Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8555974Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8556327Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8556429Z graph_break [] 2025-12-04T12:25:16.8556643Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8557379Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8557484Z warnings.warn( 2025-12-04T12:25:16.8557709Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8557821Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8557936Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8558169Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8558507Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8558605Z graph_break [] 2025-12-04T12:25:16.8558841Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8559630Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8559745Z warnings.warn( 2025-12-04T12:25:16.8560575Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ecad5133d6245a6a.xml - 2025-12-04T12:25:16.8560746Z =========================== short test summary info ============================ 2025-12-04T12:25:16.8561868Z FAILED [0.1621s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8561874Z 2025-12-04T12:25:16.8562095Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8563084Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8563093Z 2025-12-04T12:25:16.8563359Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8563573Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.8563788Z ================== 1 failed, 204 deselected, 2 rerun in 5.05s ================== 2025-12-04T12:25:16.8563889Z Got exit code 1 2025-12-04T12:25:16.8564794Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8565242Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.8565889Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d38b619605c8f606.xml 2025-12-04T12:25:16.8566070Z ============================= test session starts ============================== 2025-12-04T12:25:16.8566420Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.8566551Z cachedir: .pytest_cache 2025-12-04T12:25:16.8567072Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.8567199Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.8567322Z configfile: pytest.ini 2025-12-04T12:25:16.8567914Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.8568162Z collecting ... collected 380 items / 134 deselected / 246 selected 2025-12-04T12:25:16.8568313Z stepcurrent: skipping 134 already run items. 2025-12-04T12:25:16.8568433Z Running 71 items in this shard 2025-12-04T12:25:16.8568439Z 2025-12-04T12:25:16.8569390Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [4.6680s] [ 1%] 2025-12-04T12:25:16.8570314Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [0.1691s] [ 1%] 2025-12-04T12:25:16.8571177Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True FAILED [0.1666s] [ 1%] 2025-12-04T12:25:16.8571182Z 2025-12-04T12:25:16.8571323Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.8571902Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.8572100Z Traceback (most recent call last): 2025-12-04T12:25:16.8572568Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8572780Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8572992Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8572997Z 2025-12-04T12:25:16.8573210Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8574192Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.8574197Z 2025-12-04T12:25:16.8574465Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8574696Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8574811Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8574928Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8575281Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8575604Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8575702Z graph_break [] 2025-12-04T12:25:16.8575928Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8578742Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8578867Z return x.grad, w.grad 2025-12-04T12:25:16.8579600Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8579717Z warnings.warn( 2025-12-04T12:25:16.8582427Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8582552Z return x.grad, w.grad 2025-12-04T12:25:16.8583127Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.8583248Z Traceback (most recent call last): 2025-12-04T12:25:16.8583738Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8583940Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8584150Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8584171Z 2025-12-04T12:25:16.8584381Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8585343Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.8585426Z 2025-12-04T12:25:16.8585709Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8585926Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8586039Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8586166Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8586508Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8586739Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8586836Z graph_break [] 2025-12-04T12:25:16.8587049Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8589772Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8589957Z return x.grad, w.grad 2025-12-04T12:25:16.8590714Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8590822Z warnings.warn( 2025-12-04T12:25:16.8593571Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8593684Z return x.grad, w.grad 2025-12-04T12:25:16.8593917Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8594030Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8594143Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8594379Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8594722Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8594819Z graph_break [] 2025-12-04T12:25:16.8595049Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8597958Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8598087Z return x.grad, w.grad 2025-12-04T12:25:16.8598818Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8598938Z warnings.warn( 2025-12-04T12:25:16.8601641Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8601871Z return x.grad, w.grad 2025-12-04T12:25:16.8602020Z =================================== FAILURES =================================== 2025-12-04T12:25:16.8602629Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.8602766Z Traceback (most recent call last): 2025-12-04T12:25:16.8603242Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8603478Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8603696Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8603701Z 2025-12-04T12:25:16.8603916Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8604935Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.8604941Z 2025-12-04T12:25:16.8605208Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8605442Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8605552Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8605667Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8606060Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8606288Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8606388Z graph_break [] 2025-12-04T12:25:16.8606617Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8609325Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8609444Z return x.grad, w.grad 2025-12-04T12:25:16.8610177Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8610300Z warnings.warn( 2025-12-04T12:25:16.8612991Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8613109Z return x.grad, w.grad 2025-12-04T12:25:16.8613325Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8613437Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8613628Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8613848Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8614187Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8614292Z graph_break [] 2025-12-04T12:25:16.8614509Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8617308Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8617419Z return x.grad, w.grad 2025-12-04T12:25:16.8618182Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8618282Z warnings.warn( 2025-12-04T12:25:16.8621050Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8621202Z return x.grad, w.grad 2025-12-04T12:25:16.8621425Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8621550Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8621664Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8621887Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8622241Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8622338Z graph_break [] 2025-12-04T12:25:16.8622552Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8623300Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8623402Z warnings.warn( 2025-12-04T12:25:16.8626115Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8626228Z return x.grad, w.grad 2025-12-04T12:25:16.8627060Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d38b619605c8f606.xml - 2025-12-04T12:25:16.8627233Z =========================== short test summary info ============================ 2025-12-04T12:25:16.8628350Z FAILED [0.1666s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8628429Z 2025-12-04T12:25:16.8628645Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8629609Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.8629614Z 2025-12-04T12:25:16.8629895Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8630074Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.8630293Z ================== 1 failed, 134 deselected, 2 rerun in 5.06s ================== 2025-12-04T12:25:16.8630391Z Got exit code 1 2025-12-04T12:25:16.8630498Z Retrying single test... 2025-12-04T12:25:16.8631161Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ca04900b7cf26510.xml 2025-12-04T12:25:16.8631328Z ============================= test session starts ============================== 2025-12-04T12:25:16.8631678Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.8631798Z cachedir: .pytest_cache 2025-12-04T12:25:16.8632351Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.8632485Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.8632594Z configfile: pytest.ini 2025-12-04T12:25:16.8633186Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.8633446Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.8634534Z stepcurrent: skipping 134 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.8634664Z Running 1 items in this shard 2025-12-04T12:25:16.8634670Z 2025-12-04T12:25:16.8635598Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [4.6823s] [100%] 2025-12-04T12:25:16.8636526Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [0.1701s] [100%] 2025-12-04T12:25:16.8637385Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True FAILED [0.1653s] [100%] 2025-12-04T12:25:16.8637391Z 2025-12-04T12:25:16.8637541Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.8638136Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.8638258Z Traceback (most recent call last): 2025-12-04T12:25:16.8638731Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8638948Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8639163Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8639169Z 2025-12-04T12:25:16.8639396Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8640372Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.8640380Z 2025-12-04T12:25:16.8640647Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8640955Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8641068Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8641199Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8641542Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8641763Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8641877Z graph_break [] 2025-12-04T12:25:16.8642097Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8644834Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8644949Z return x.grad, w.grad 2025-12-04T12:25:16.8645716Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8645833Z warnings.warn( 2025-12-04T12:25:16.8648565Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8648697Z return x.grad, w.grad 2025-12-04T12:25:16.8649274Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.8649415Z Traceback (most recent call last): 2025-12-04T12:25:16.8649973Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8650258Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8650612Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8650623Z 2025-12-04T12:25:16.8650849Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8652000Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.8652011Z 2025-12-04T12:25:16.8652285Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8652509Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8652642Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8652765Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8653108Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8653345Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8653443Z graph_break [] 2025-12-04T12:25:16.8653672Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8656387Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8656590Z return x.grad, w.grad 2025-12-04T12:25:16.8657384Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8657487Z warnings.warn( 2025-12-04T12:25:16.8660200Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8660347Z return x.grad, w.grad 2025-12-04T12:25:16.8660583Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8660692Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8660809Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8661046Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8661387Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8661498Z graph_break [] 2025-12-04T12:25:16.8661744Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8664470Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8664588Z return x.grad, w.grad 2025-12-04T12:25:16.8665321Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8665433Z warnings.warn( 2025-12-04T12:25:16.8668140Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8668262Z return x.grad, w.grad 2025-12-04T12:25:16.8668405Z =================================== FAILURES =================================== 2025-12-04T12:25:16.8668984Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.8669117Z Traceback (most recent call last): 2025-12-04T12:25:16.8669590Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8669858Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8670069Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8670075Z 2025-12-04T12:25:16.8670291Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8671266Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.8671272Z 2025-12-04T12:25:16.8671537Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8671766Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8671878Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8671995Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8672347Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8672571Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8672669Z graph_break [] 2025-12-04T12:25:16.8672898Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8675685Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8675809Z return x.grad, w.grad 2025-12-04T12:25:16.8676548Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8676662Z warnings.warn( 2025-12-04T12:25:16.8679378Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8679497Z return x.grad, w.grad 2025-12-04T12:25:16.8679715Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8679830Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8679955Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8680178Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8680516Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8680625Z graph_break [] 2025-12-04T12:25:16.8680843Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8683567Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8683733Z return x.grad, w.grad 2025-12-04T12:25:16.8684472Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8684575Z warnings.warn( 2025-12-04T12:25:16.8687274Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8687397Z return x.grad, w.grad 2025-12-04T12:25:16.8687614Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8687738Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8687853Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8688073Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8688454Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8688555Z graph_break [] 2025-12-04T12:25:16.8688782Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8689508Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8689631Z warnings.warn( 2025-12-04T12:25:16.8692415Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8692529Z return x.grad, w.grad 2025-12-04T12:25:16.8693368Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ca04900b7cf26510.xml - 2025-12-04T12:25:16.8693542Z =========================== short test summary info ============================ 2025-12-04T12:25:16.8694661Z FAILED [0.1653s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8694671Z 2025-12-04T12:25:16.8694889Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8695858Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.8695868Z 2025-12-04T12:25:16.8696320Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8696503Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.8696715Z ================== 1 failed, 204 deselected, 2 rerun in 5.07s ================== 2025-12-04T12:25:16.8696818Z Got exit code 1 2025-12-04T12:25:16.8696987Z Retrying single test... 2025-12-04T12:25:16.8697645Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-8a1dc4f258524121.xml 2025-12-04T12:25:16.8698552Z ============================= test session starts ============================== 2025-12-04T12:25:16.8698906Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.8699034Z cachedir: .pytest_cache 2025-12-04T12:25:16.8699559Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.8699700Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.8699809Z configfile: pytest.ini 2025-12-04T12:25:16.8700402Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.8700646Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.8701702Z stepcurrent: skipping 134 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.8701834Z Running 1 items in this shard 2025-12-04T12:25:16.8701841Z 2025-12-04T12:25:16.8702813Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [4.6620s] [100%] 2025-12-04T12:25:16.8703742Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [0.1671s] [100%] 2025-12-04T12:25:16.8704642Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True FAILED [0.1646s] [100%] 2025-12-04T12:25:16.8704652Z 2025-12-04T12:25:16.8704793Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.8705383Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.8705509Z Traceback (most recent call last): 2025-12-04T12:25:16.8706005Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8706203Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8706417Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8706422Z 2025-12-04T12:25:16.8706648Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8707617Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.8707627Z 2025-12-04T12:25:16.8707908Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8708127Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8708239Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8708370Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8708715Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8708934Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8709044Z graph_break [] 2025-12-04T12:25:16.8709258Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8711978Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8712154Z return x.grad, w.grad 2025-12-04T12:25:16.8712900Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8713002Z warnings.warn( 2025-12-04T12:25:16.8715728Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8715853Z return x.grad, w.grad 2025-12-04T12:25:16.8716464Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.8716601Z Traceback (most recent call last): 2025-12-04T12:25:16.8717070Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8717271Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8717499Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8717505Z 2025-12-04T12:25:16.8717750Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8718740Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.8718746Z 2025-12-04T12:25:16.8719013Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8719235Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8719359Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8719473Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8719829Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8720047Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8720144Z graph_break [] 2025-12-04T12:25:16.8720375Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8723208Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8723338Z return x.grad, w.grad 2025-12-04T12:25:16.8724073Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8724175Z warnings.warn( 2025-12-04T12:25:16.8726880Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8727056Z return x.grad, w.grad 2025-12-04T12:25:16.8727288Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8727399Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8727526Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8727749Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8728092Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8728202Z graph_break [] 2025-12-04T12:25:16.8728421Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8731162Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8731271Z return x.grad, w.grad 2025-12-04T12:25:16.8732032Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8732151Z warnings.warn( 2025-12-04T12:25:16.8734843Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8734964Z return x.grad, w.grad 2025-12-04T12:25:16.8735111Z =================================== FAILURES =================================== 2025-12-04T12:25:16.8735704Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:16.8735829Z Traceback (most recent call last): 2025-12-04T12:25:16.8736307Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8736523Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8736740Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8736745Z 2025-12-04T12:25:16.8737036Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8738026Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.8738031Z 2025-12-04T12:25:16.8738299Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8738531Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8738648Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8738835Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8739195Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8739441Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8739555Z graph_break [] 2025-12-04T12:25:16.8739774Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8742501Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8742630Z return x.grad, w.grad 2025-12-04T12:25:16.8743362Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8743478Z warnings.warn( 2025-12-04T12:25:16.8746280Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8746406Z return x.grad, w.grad 2025-12-04T12:25:16.8746630Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8746742Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8746872Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8747093Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8747451Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8747550Z graph_break [] 2025-12-04T12:25:16.8747769Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8750495Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8750611Z return x.grad, w.grad 2025-12-04T12:25:16.8751362Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8751464Z warnings.warn( 2025-12-04T12:25:16.8754180Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8754348Z return x.grad, w.grad 2025-12-04T12:25:16.8754566Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8754690Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8754806Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8755034Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8755388Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8755487Z graph_break [] 2025-12-04T12:25:16.8755719Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8756449Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8756551Z warnings.warn( 2025-12-04T12:25:16.8759302Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.8759412Z return x.grad, w.grad 2025-12-04T12:25:16.8760248Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-8a1dc4f258524121.xml - 2025-12-04T12:25:16.8760452Z =========================== short test summary info ============================ 2025-12-04T12:25:16.8761565Z FAILED [0.1646s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8761575Z 2025-12-04T12:25:16.8761791Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8762756Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.8762773Z 2025-12-04T12:25:16.8763039Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8763218Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.8763430Z ================== 1 failed, 204 deselected, 2 rerun in 5.05s ================== 2025-12-04T12:25:16.8763528Z Got exit code 1 2025-12-04T12:25:16.8764407Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:16.8764829Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.8765493Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a0112c740677bec4.xml 2025-12-04T12:25:16.8765667Z ============================= test session starts ============================== 2025-12-04T12:25:16.8766020Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.8766129Z cachedir: .pytest_cache 2025-12-04T12:25:16.8766686Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.8766813Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.8766997Z configfile: pytest.ini 2025-12-04T12:25:16.8767589Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.8767817Z collecting ... collected 380 items / 135 deselected / 245 selected 2025-12-04T12:25:16.8767980Z stepcurrent: skipping 135 already run items. 2025-12-04T12:25:16.8768098Z Running 70 items in this shard 2025-12-04T12:25:16.8768103Z 2025-12-04T12:25:16.8769038Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6549s] [ 1%] 2025-12-04T12:25:16.8769983Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1639s] [ 1%] 2025-12-04T12:25:16.8770836Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1612s] [ 1%] 2025-12-04T12:25:16.8770844Z 2025-12-04T12:25:16.8770999Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.8771610Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.8771750Z Traceback (most recent call last): 2025-12-04T12:25:16.8772221Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8772419Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8772644Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8772649Z 2025-12-04T12:25:16.8772892Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8773882Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.8773888Z 2025-12-04T12:25:16.8774155Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8774375Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8774500Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8774618Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8774963Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8775195Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8775295Z graph_break [] 2025-12-04T12:25:16.8775524Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8776260Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8776368Z warnings.warn( 2025-12-04T12:25:16.8777033Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.8777162Z Traceback (most recent call last): 2025-12-04T12:25:16.8777651Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8777850Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8778065Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8778070Z 2025-12-04T12:25:16.8778302Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8779270Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.8779341Z 2025-12-04T12:25:16.8779626Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8779844Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8779959Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8780090Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8780435Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8780654Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8780763Z graph_break [] 2025-12-04T12:25:16.8780981Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8781727Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8781834Z warnings.warn( 2025-12-04T12:25:16.8782048Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8782174Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8782290Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8782514Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8782895Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8782992Z graph_break [] 2025-12-04T12:25:16.8783206Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8783948Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8784049Z warnings.warn( 2025-12-04T12:25:16.8784241Z =================================== FAILURES =================================== 2025-12-04T12:25:16.8784823Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.8784944Z Traceback (most recent call last): 2025-12-04T12:25:16.8785426Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8785627Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8785849Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8785855Z 2025-12-04T12:25:16.8786068Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8787043Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.8787051Z 2025-12-04T12:25:16.8787330Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8787550Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8787679Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8787794Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8788135Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8788370Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8788468Z graph_break [] 2025-12-04T12:25:16.8788681Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8789429Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8789530Z warnings.warn( 2025-12-04T12:25:16.8789759Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8790017Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8790131Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8790365Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8790701Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8790796Z graph_break [] 2025-12-04T12:25:16.8791024Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8791748Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8791861Z warnings.warn( 2025-12-04T12:25:16.8792073Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8792182Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8792311Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8792529Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8792870Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8792980Z graph_break [] 2025-12-04T12:25:16.8793193Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8793955Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8794056Z warnings.warn( 2025-12-04T12:25:16.8794878Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a0112c740677bec4.xml - 2025-12-04T12:25:16.8795061Z =========================== short test summary info ============================ 2025-12-04T12:25:16.8796392Z FAILED [0.1612s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8796403Z 2025-12-04T12:25:16.8796643Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8797615Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.8797621Z 2025-12-04T12:25:16.8797887Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8798082Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.8798284Z ================== 1 failed, 135 deselected, 2 rerun in 5.03s ================== 2025-12-04T12:25:16.8798397Z Got exit code 1 2025-12-04T12:25:16.8798504Z Retrying single test... 2025-12-04T12:25:16.8799148Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-67764f47d32c222e.xml 2025-12-04T12:25:16.8799328Z ============================= test session starts ============================== 2025-12-04T12:25:16.8799680Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.8799789Z cachedir: .pytest_cache 2025-12-04T12:25:16.8800327Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.8800452Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.8800573Z configfile: pytest.ini 2025-12-04T12:25:16.8801165Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.8801394Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.8802464Z stepcurrent: skipping 135 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.8802659Z Running 1 items in this shard 2025-12-04T12:25:16.8802664Z 2025-12-04T12:25:16.8803614Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6806s] [100%] 2025-12-04T12:25:16.8804550Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1649s] [100%] 2025-12-04T12:25:16.8805415Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1608s] [100%] 2025-12-04T12:25:16.8805421Z 2025-12-04T12:25:16.8805564Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.8806151Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.8806290Z Traceback (most recent call last): 2025-12-04T12:25:16.8806800Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8807013Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8807226Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8807231Z 2025-12-04T12:25:16.8807446Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8808463Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.8808474Z 2025-12-04T12:25:16.8808742Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8808974Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8809086Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8809204Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8809562Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8809782Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8809882Z graph_break [] 2025-12-04T12:25:16.8810109Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8810841Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8810960Z warnings.warn( 2025-12-04T12:25:16.8811542Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.8811665Z Traceback (most recent call last): 2025-12-04T12:25:16.8812152Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8812352Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8812566Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8812585Z 2025-12-04T12:25:16.8812800Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8813768Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.8813773Z 2025-12-04T12:25:16.8814055Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8814336Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8814462Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8814579Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8814919Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8815156Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8815254Z graph_break [] 2025-12-04T12:25:16.8815469Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8816222Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8816324Z warnings.warn( 2025-12-04T12:25:16.8816540Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8816668Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8816788Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8817085Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8817426Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8817524Z graph_break [] 2025-12-04T12:25:16.8817789Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8818519Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8818621Z warnings.warn( 2025-12-04T12:25:16.8818780Z =================================== FAILURES =================================== 2025-12-04T12:25:16.8819398Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.8819537Z Traceback (most recent call last): 2025-12-04T12:25:16.8820015Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8820214Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8820440Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8820445Z 2025-12-04T12:25:16.8820660Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8821650Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.8821656Z 2025-12-04T12:25:16.8821922Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8822135Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8822266Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8822385Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8822764Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8822982Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8823078Z graph_break [] 2025-12-04T12:25:16.8823303Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8824031Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8824131Z warnings.warn( 2025-12-04T12:25:16.8824355Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8824462Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8824585Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8824808Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8825207Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8825315Z graph_break [] 2025-12-04T12:25:16.8825527Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8826263Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8826373Z warnings.warn( 2025-12-04T12:25:16.8826587Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8826707Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8826818Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8827034Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8827382Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8827479Z graph_break [] 2025-12-04T12:25:16.8827700Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8828443Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8828541Z warnings.warn( 2025-12-04T12:25:16.8829407Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-67764f47d32c222e.xml - 2025-12-04T12:25:16.8829580Z =========================== short test summary info ============================ 2025-12-04T12:25:16.8830682Z FAILED [0.1608s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8830688Z 2025-12-04T12:25:16.8830948Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8831914Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.8831919Z 2025-12-04T12:25:16.8832196Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8832376Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.8832576Z ================== 1 failed, 204 deselected, 2 rerun in 5.06s ================== 2025-12-04T12:25:16.8832685Z Got exit code 1 2025-12-04T12:25:16.8832792Z Retrying single test... 2025-12-04T12:25:16.8833448Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c39fdfd87cf64578.xml 2025-12-04T12:25:16.8833613Z ============================= test session starts ============================== 2025-12-04T12:25:16.8833967Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.8834087Z cachedir: .pytest_cache 2025-12-04T12:25:16.8834611Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.8834733Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.8834853Z configfile: pytest.ini 2025-12-04T12:25:16.8835445Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.8835687Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.8836751Z stepcurrent: skipping 135 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.8836867Z Running 1 items in this shard 2025-12-04T12:25:16.8836950Z 2025-12-04T12:25:16.8837899Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6567s] [100%] 2025-12-04T12:25:16.8838827Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1676s] [100%] 2025-12-04T12:25:16.8839714Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1646s] [100%] 2025-12-04T12:25:16.8839719Z 2025-12-04T12:25:16.8839863Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.8840455Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.8840582Z Traceback (most recent call last): 2025-12-04T12:25:16.8841054Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8841265Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8841506Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8841513Z 2025-12-04T12:25:16.8841738Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8842703Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.8842709Z 2025-12-04T12:25:16.8842975Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8843233Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8843351Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8843479Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8843818Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8844039Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8844154Z graph_break [] 2025-12-04T12:25:16.8844370Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8845111Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8845224Z warnings.warn( 2025-12-04T12:25:16.8845805Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.8845941Z Traceback (most recent call last): 2025-12-04T12:25:16.8846417Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8846615Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8846839Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8846844Z 2025-12-04T12:25:16.8847059Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8848051Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.8848056Z 2025-12-04T12:25:16.8848321Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8848540Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8848666Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8848783Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8849190Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8849418Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8849515Z graph_break [] 2025-12-04T12:25:16.8849742Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8850482Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8850584Z warnings.warn( 2025-12-04T12:25:16.8850811Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8850919Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8851035Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8851271Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8851605Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8851717Z graph_break [] 2025-12-04T12:25:16.8851931Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8852691Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8852804Z warnings.warn( 2025-12-04T12:25:16.8852946Z =================================== FAILURES =================================== 2025-12-04T12:25:16.8853524Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.8853657Z Traceback (most recent call last): 2025-12-04T12:25:16.8854128Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8854365Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8854584Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8854589Z 2025-12-04T12:25:16.8854797Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8855779Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.8855784Z 2025-12-04T12:25:16.8856050Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8856277Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8856388Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8856502Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8856929Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8857150Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8857252Z graph_break [] 2025-12-04T12:25:16.8857484Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8858214Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8858332Z warnings.warn( 2025-12-04T12:25:16.8858551Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8858660Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8858788Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8859004Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8859340Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8867568Z graph_break [] 2025-12-04T12:25:16.8867912Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8868815Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8868930Z warnings.warn( 2025-12-04T12:25:16.8869151Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8869281Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8869396Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8869620Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8869979Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8870071Z graph_break [] 2025-12-04T12:25:16.8870287Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8871032Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8871132Z warnings.warn( 2025-12-04T12:25:16.8871961Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c39fdfd87cf64578.xml - 2025-12-04T12:25:16.8872168Z =========================== short test summary info ============================ 2025-12-04T12:25:16.8873280Z FAILED [0.1646s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8873299Z 2025-12-04T12:25:16.8873518Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8874553Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.8874565Z 2025-12-04T12:25:16.8874838Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8875016Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.8875223Z ================== 1 failed, 204 deselected, 2 rerun in 5.04s ================== 2025-12-04T12:25:16.8875321Z Got exit code 1 2025-12-04T12:25:16.8876203Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.8876625Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.8877262Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-48a46382e3216972.xml 2025-12-04T12:25:16.8877429Z ============================= test session starts ============================== 2025-12-04T12:25:16.8877782Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.8877890Z cachedir: .pytest_cache 2025-12-04T12:25:16.8878416Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.8878537Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.8878644Z configfile: pytest.ini 2025-12-04T12:25:16.8879245Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.8879470Z collecting ... collected 380 items / 136 deselected / 244 selected 2025-12-04T12:25:16.8879628Z stepcurrent: skipping 136 already run items. 2025-12-04T12:25:16.8879737Z Running 69 items in this shard 2025-12-04T12:25:16.8879746Z 2025-12-04T12:25:16.8880679Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6440s] [ 1%] 2025-12-04T12:25:16.8881686Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1625s] [ 1%] 2025-12-04T12:25:16.8882529Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1596s] [ 1%] 2025-12-04T12:25:16.8882535Z 2025-12-04T12:25:16.8882685Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.8883264Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.8883383Z Traceback (most recent call last): 2025-12-04T12:25:16.8883868Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8884062Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8884282Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8884288Z 2025-12-04T12:25:16.8884532Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8885500Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8885517Z 2025-12-04T12:25:16.8885776Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8885989Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8886160Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8886275Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8886610Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8886832Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8886926Z graph_break [] 2025-12-04T12:25:16.8887165Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8887913Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8888009Z warnings.warn( 2025-12-04T12:25:16.8888599Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.8888717Z Traceback (most recent call last): 2025-12-04T12:25:16.8889193Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8889396Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8889606Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8889611Z 2025-12-04T12:25:16.8889828Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8890790Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8890795Z 2025-12-04T12:25:16.8891059Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8891279Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8891388Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8891506Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8891843Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8892119Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8892222Z graph_break [] 2025-12-04T12:25:16.8892435Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8893165Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8893267Z warnings.warn( 2025-12-04T12:25:16.8893476Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8893589Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8893699Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8893913Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8894262Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8894358Z graph_break [] 2025-12-04T12:25:16.8894575Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8895303Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8895401Z warnings.warn( 2025-12-04T12:25:16.8895579Z =================================== FAILURES =================================== 2025-12-04T12:25:16.8896339Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.8896462Z Traceback (most recent call last): 2025-12-04T12:25:16.8897031Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8897232Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8897516Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8897536Z 2025-12-04T12:25:16.8897748Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8898720Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8898725Z 2025-12-04T12:25:16.8898999Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8899215Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8899324Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8899450Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8899787Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8900020Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8900120Z graph_break [] 2025-12-04T12:25:16.8900336Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8901075Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8901178Z warnings.warn( 2025-12-04T12:25:16.8901387Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8901507Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8901621Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8901846Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8902177Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8902270Z graph_break [] 2025-12-04T12:25:16.8902492Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8903219Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8903398Z warnings.warn( 2025-12-04T12:25:16.8903615Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8903723Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8903845Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8904063Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8904396Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8904505Z graph_break [] 2025-12-04T12:25:16.8904714Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8905440Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8905545Z warnings.warn( 2025-12-04T12:25:16.8906362Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-48a46382e3216972.xml - 2025-12-04T12:25:16.8906541Z =========================== short test summary info ============================ 2025-12-04T12:25:16.8907691Z FAILED [0.1596s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8907698Z 2025-12-04T12:25:16.8907914Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8908894Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8908929Z 2025-12-04T12:25:16.8909186Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8909377Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.8909571Z ================== 1 failed, 136 deselected, 2 rerun in 5.02s ================== 2025-12-04T12:25:16.8909665Z Got exit code 1 2025-12-04T12:25:16.8909779Z Retrying single test... 2025-12-04T12:25:16.8910421Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d40d195c41f12813.xml 2025-12-04T12:25:16.8910596Z ============================= test session starts ============================== 2025-12-04T12:25:16.8910940Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.8911047Z cachedir: .pytest_cache 2025-12-04T12:25:16.8911576Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.8911702Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.8911811Z configfile: pytest.ini 2025-12-04T12:25:16.8912581Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.8912811Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.8913887Z stepcurrent: skipping 136 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8913997Z Running 1 items in this shard 2025-12-04T12:25:16.8914003Z 2025-12-04T12:25:16.8914951Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6391s] [100%] 2025-12-04T12:25:16.8915876Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1643s] [100%] 2025-12-04T12:25:16.8916802Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1612s] [100%] 2025-12-04T12:25:16.8916818Z 2025-12-04T12:25:16.8916956Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.8917529Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.8917662Z Traceback (most recent call last): 2025-12-04T12:25:16.8918130Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8918325Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8918549Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8918554Z 2025-12-04T12:25:16.8918767Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8919771Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8919777Z 2025-12-04T12:25:16.8920042Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8920258Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8920377Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8920489Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8920837Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8921083Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8921184Z graph_break [] 2025-12-04T12:25:16.8921405Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8922140Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8922241Z warnings.warn( 2025-12-04T12:25:16.8922914Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.8923089Z Traceback (most recent call last): 2025-12-04T12:25:16.8923670Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8923868Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8924082Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8924088Z 2025-12-04T12:25:16.8924312Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8925273Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8925278Z 2025-12-04T12:25:16.8925555Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8925771Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8925881Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8926001Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8926343Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8926555Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8926662Z graph_break [] 2025-12-04T12:25:16.8926875Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8927688Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8927787Z warnings.warn( 2025-12-04T12:25:16.8928000Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8928116Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8928225Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8928437Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8928784Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8928879Z graph_break [] 2025-12-04T12:25:16.8929097Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8929826Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8929928Z warnings.warn( 2025-12-04T12:25:16.8930082Z =================================== FAILURES =================================== 2025-12-04T12:25:16.8930663Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.8930841Z Traceback (most recent call last): 2025-12-04T12:25:16.8931310Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8931501Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8931716Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8931722Z 2025-12-04T12:25:16.8931934Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8932936Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8932954Z 2025-12-04T12:25:16.8933217Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8933428Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8933548Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8933659Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8933993Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8934224Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8934319Z graph_break [] 2025-12-04T12:25:16.8934542Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8935270Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8935371Z warnings.warn( 2025-12-04T12:25:16.8935594Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8935700Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8935816Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8936047Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8936386Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8936493Z graph_break [] 2025-12-04T12:25:16.8936702Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8937510Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8937624Z warnings.warn( 2025-12-04T12:25:16.8937834Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8938012Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8938135Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8938352Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8938697Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8938793Z graph_break [] 2025-12-04T12:25:16.8939006Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8939737Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8939836Z warnings.warn( 2025-12-04T12:25:16.8940652Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d40d195c41f12813.xml - 2025-12-04T12:25:16.8940829Z =========================== short test summary info ============================ 2025-12-04T12:25:16.8941928Z FAILED [0.1612s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8941935Z 2025-12-04T12:25:16.8942184Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8943152Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8943157Z 2025-12-04T12:25:16.8943430Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8943640Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.8943843Z ================== 1 failed, 204 deselected, 2 rerun in 5.02s ================== 2025-12-04T12:25:16.8943954Z Got exit code 1 2025-12-04T12:25:16.8944058Z Retrying single test... 2025-12-04T12:25:16.8944696Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-35b201d0b4c8de0a.xml 2025-12-04T12:25:16.8944866Z ============================= test session starts ============================== 2025-12-04T12:25:16.8945210Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.8945324Z cachedir: .pytest_cache 2025-12-04T12:25:16.8945869Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.8945989Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.8946104Z configfile: pytest.ini 2025-12-04T12:25:16.8946696Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.8946923Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.8947996Z stepcurrent: skipping 136 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8948102Z Running 1 items in this shard 2025-12-04T12:25:16.8948107Z 2025-12-04T12:25:16.8949046Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6486s] [100%] 2025-12-04T12:25:16.8949977Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1619s] [100%] 2025-12-04T12:25:16.8950861Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1581s] [100%] 2025-12-04T12:25:16.8950899Z 2025-12-04T12:25:16.8951041Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.8951618Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.8951744Z Traceback (most recent call last): 2025-12-04T12:25:16.8952211Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8952412Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8952623Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8952628Z 2025-12-04T12:25:16.8952847Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8953827Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8953832Z 2025-12-04T12:25:16.8954095Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8954346Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8954455Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8954564Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8954911Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8955132Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8955233Z graph_break [] 2025-12-04T12:25:16.8955477Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8956212Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8956325Z warnings.warn( 2025-12-04T12:25:16.8956901Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.8957020Z Traceback (most recent call last): 2025-12-04T12:25:16.8957500Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8957697Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8957913Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8957919Z 2025-12-04T12:25:16.8958133Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8959099Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8959109Z 2025-12-04T12:25:16.8959383Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8959598Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8959721Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8959835Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8960173Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8960398Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8960493Z graph_break [] 2025-12-04T12:25:16.8960706Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8961448Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8961611Z warnings.warn( 2025-12-04T12:25:16.8961834Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8961939Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8962053Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8962278Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8962612Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8962709Z graph_break [] 2025-12-04T12:25:16.8962926Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8963654Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8963761Z warnings.warn( 2025-12-04T12:25:16.8963905Z =================================== FAILURES =================================== 2025-12-04T12:25:16.8964494Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.8964621Z Traceback (most recent call last): 2025-12-04T12:25:16.8965091Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8965315Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8965532Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8965537Z 2025-12-04T12:25:16.8965747Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8966726Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8966767Z 2025-12-04T12:25:16.8967036Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8967262Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8967369Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8967480Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8967828Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8968045Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8968140Z graph_break [] 2025-12-04T12:25:16.8968362Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8969095Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8969201Z warnings.warn( 2025-12-04T12:25:16.8969415Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8969527Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8969646Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8969862Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8970200Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8970303Z graph_break [] 2025-12-04T12:25:16.8970517Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8971241Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8971348Z warnings.warn( 2025-12-04T12:25:16.8971557Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8971674Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8971788Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8972005Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8972421Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8972515Z graph_break [] 2025-12-04T12:25:16.8972727Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8973464Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8973561Z warnings.warn( 2025-12-04T12:25:16.8974517Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-35b201d0b4c8de0a.xml - 2025-12-04T12:25:16.8974694Z =========================== short test summary info ============================ 2025-12-04T12:25:16.8975798Z FAILED [0.1581s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8975818Z 2025-12-04T12:25:16.8976030Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8977112Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8977118Z 2025-12-04T12:25:16.8977392Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8977569Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.8977778Z ================== 1 failed, 204 deselected, 2 rerun in 5.02s ================== 2025-12-04T12:25:16.8977875Z Got exit code 1 2025-12-04T12:25:16.8978790Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.8979245Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.8979890Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e10bae22111bfdec.xml 2025-12-04T12:25:16.8980063Z ============================= test session starts ============================== 2025-12-04T12:25:16.8980410Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.8980523Z cachedir: .pytest_cache 2025-12-04T12:25:16.8981046Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.8981168Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.8981278Z configfile: pytest.ini 2025-12-04T12:25:16.8981875Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.8982108Z collecting ... collected 380 items / 137 deselected / 243 selected 2025-12-04T12:25:16.8982265Z stepcurrent: skipping 137 already run items. 2025-12-04T12:25:16.8982376Z Running 68 items in this shard 2025-12-04T12:25:16.8982382Z 2025-12-04T12:25:16.8983439Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_True SKIPPED [0.0041s] (Skip non-critical tests to save resources.) [ 1%] 2025-12-04T12:25:16.8984546Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_True_initial_xblock_2_add_1dim_True SKIPPED [0.0030s] (Skip non-critical tests to save resources.) [ 2%] 2025-12-04T12:25:16.8985554Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6305s] [ 4%] 2025-12-04T12:25:16.8986638Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1630s] [ 4%] 2025-12-04T12:25:16.8987495Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1594s] [ 4%] 2025-12-04T12:25:16.8987501Z 2025-12-04T12:25:16.8987649Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.8988225Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.8988348Z Traceback (most recent call last): 2025-12-04T12:25:16.8988831Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8989032Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8989243Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8989269Z 2025-12-04T12:25:16.8989510Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8990480Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.8990485Z 2025-12-04T12:25:16.8990765Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8990983Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8991138Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8991258Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8991604Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8991837Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8991936Z graph_break [] 2025-12-04T12:25:16.8992153Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8992907Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8993008Z warnings.warn( 2025-12-04T12:25:16.8993625Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.8993746Z Traceback (most recent call last): 2025-12-04T12:25:16.8994220Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.8994436Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.8994648Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.8994654Z 2025-12-04T12:25:16.8994872Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.8996106Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.8996114Z 2025-12-04T12:25:16.8996385Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.8996620Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8996732Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8996849Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8997207Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8997540Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8997656Z graph_break [] 2025-12-04T12:25:16.8997871Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.8998608Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.8998719Z warnings.warn( 2025-12-04T12:25:16.8998937Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.8999049Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.8999181Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.8999404Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.8999756Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.8999857Z graph_break [] 2025-12-04T12:25:16.9000074Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9000822Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9000922Z warnings.warn( 2025-12-04T12:25:16.9001109Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9001706Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9001832Z Traceback (most recent call last): 2025-12-04T12:25:16.9002320Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9002519Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9002770Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9002775Z 2025-12-04T12:25:16.9003013Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9003983Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9003988Z 2025-12-04T12:25:16.9004266Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9004480Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9004589Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9004717Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9005053Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9005291Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9005399Z graph_break [] 2025-12-04T12:25:16.9005612Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9006566Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9006668Z warnings.warn( 2025-12-04T12:25:16.9006886Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9007008Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9007120Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9007336Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9007683Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9007783Z graph_break [] 2025-12-04T12:25:16.9008007Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9008738Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9008929Z warnings.warn( 2025-12-04T12:25:16.9009151Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9009262Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9009379Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9009611Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9009947Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9010059Z graph_break [] 2025-12-04T12:25:16.9010273Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9010999Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9011114Z warnings.warn( 2025-12-04T12:25:16.9011937Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e10bae22111bfdec.xml - 2025-12-04T12:25:16.9012123Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9013263Z FAILED [0.1594s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9013270Z 2025-12-04T12:25:16.9013488Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9014473Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9014478Z 2025-12-04T12:25:16.9014771Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9014971Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9015213Z ============ 1 failed, 2 skipped, 137 deselected, 2 rerun in 5.02s ============= 2025-12-04T12:25:16.9015310Z Got exit code 1 2025-12-04T12:25:16.9015428Z Retrying single test... 2025-12-04T12:25:16.9016075Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-5822c2fc7d137525.xml 2025-12-04T12:25:16.9016252Z ============================= test session starts ============================== 2025-12-04T12:25:16.9016599Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9016707Z cachedir: .pytest_cache 2025-12-04T12:25:16.9017319Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9017443Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9017557Z configfile: pytest.ini 2025-12-04T12:25:16.9018167Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9018397Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.9019478Z stepcurrent: skipping 139 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9019591Z Running 1 items in this shard 2025-12-04T12:25:16.9019596Z 2025-12-04T12:25:16.9020531Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6775s] [100%] 2025-12-04T12:25:16.9021480Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1656s] [100%] 2025-12-04T12:25:16.9022415Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1633s] [100%] 2025-12-04T12:25:16.9022424Z 2025-12-04T12:25:16.9022582Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9023164Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9023305Z Traceback (most recent call last): 2025-12-04T12:25:16.9023776Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9023980Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9024209Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9024216Z 2025-12-04T12:25:16.9024432Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9025445Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9025450Z 2025-12-04T12:25:16.9025717Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9025935Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9026060Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9026174Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9026514Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9026774Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9026879Z graph_break [] 2025-12-04T12:25:16.9027107Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9027843Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9027943Z warnings.warn( 2025-12-04T12:25:16.9028536Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9028656Z Traceback (most recent call last): 2025-12-04T12:25:16.9029124Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9029329Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9029540Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9029551Z 2025-12-04T12:25:16.9029775Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9030756Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9030761Z 2025-12-04T12:25:16.9031041Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9031255Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9031366Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9031493Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9031852Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9032069Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9032178Z graph_break [] 2025-12-04T12:25:16.9032395Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9033204Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9033303Z warnings.warn( 2025-12-04T12:25:16.9033520Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9033644Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9033756Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9033973Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9034320Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9034415Z graph_break [] 2025-12-04T12:25:16.9034630Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9035370Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9035474Z warnings.warn( 2025-12-04T12:25:16.9035630Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9036214Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9036364Z Traceback (most recent call last): 2025-12-04T12:25:16.9036846Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9037040Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9037256Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9037261Z 2025-12-04T12:25:16.9037474Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9038477Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9038487Z 2025-12-04T12:25:16.9038763Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9038977Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9039099Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9039212Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9039552Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9039783Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9039880Z graph_break [] 2025-12-04T12:25:16.9040098Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9040846Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9040949Z warnings.warn( 2025-12-04T12:25:16.9041178Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9041289Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9041406Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9041638Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9041975Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9042072Z graph_break [] 2025-12-04T12:25:16.9042298Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9043026Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9043136Z warnings.warn( 2025-12-04T12:25:16.9043346Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9043517Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9043641Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9043862Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9044198Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9044306Z graph_break [] 2025-12-04T12:25:16.9044522Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9045265Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9045367Z warnings.warn( 2025-12-04T12:25:16.9046189Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-5822c2fc7d137525.xml - 2025-12-04T12:25:16.9046374Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9047484Z FAILED [0.1633s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9047490Z 2025-12-04T12:25:16.9047747Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9048713Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9048719Z 2025-12-04T12:25:16.9048987Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9049177Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9049404Z ================== 1 failed, 204 deselected, 2 rerun in 5.06s ================== 2025-12-04T12:25:16.9049522Z Got exit code 1 2025-12-04T12:25:16.9049628Z Retrying single test... 2025-12-04T12:25:16.9050280Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a6ba20f02ccefade.xml 2025-12-04T12:25:16.9050458Z ============================= test session starts ============================== 2025-12-04T12:25:16.9050806Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9050915Z cachedir: .pytest_cache 2025-12-04T12:25:16.9051450Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9051575Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9051698Z configfile: pytest.ini 2025-12-04T12:25:16.9052295Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9052528Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.9053623Z stepcurrent: skipping 139 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9053739Z Running 1 items in this shard 2025-12-04T12:25:16.9053745Z 2025-12-04T12:25:16.9054694Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6833s] [100%] 2025-12-04T12:25:16.9055622Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1677s] [100%] 2025-12-04T12:25:16.9056485Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1641s] [100%] 2025-12-04T12:25:16.9056552Z 2025-12-04T12:25:16.9056695Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9057365Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9057505Z Traceback (most recent call last): 2025-12-04T12:25:16.9057979Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9058191Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9058403Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9058408Z 2025-12-04T12:25:16.9058629Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9059609Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9059618Z 2025-12-04T12:25:16.9059884Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9060152Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9060267Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9060382Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9060737Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9060955Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9061055Z graph_break [] 2025-12-04T12:25:16.9061284Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9062064Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9062186Z warnings.warn( 2025-12-04T12:25:16.9062789Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9062911Z Traceback (most recent call last): 2025-12-04T12:25:16.9063399Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9063594Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9063808Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9063827Z 2025-12-04T12:25:16.9064045Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9065020Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9065030Z 2025-12-04T12:25:16.9065312Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9065528Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9065641Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9065815Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9066225Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9066490Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9066636Z graph_break [] 2025-12-04T12:25:16.9066855Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9067688Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9067866Z warnings.warn( 2025-12-04T12:25:16.9068079Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9068202Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9068314Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9068545Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9068884Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9068978Z graph_break [] 2025-12-04T12:25:16.9069205Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9069935Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9070035Z warnings.warn( 2025-12-04T12:25:16.9070196Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9070775Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9070913Z Traceback (most recent call last): 2025-12-04T12:25:16.9071385Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9071610Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9071834Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9071840Z 2025-12-04T12:25:16.9072050Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9073035Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9073040Z 2025-12-04T12:25:16.9073335Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9073557Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9073681Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9073796Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9074145Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9074364Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9074463Z graph_break [] 2025-12-04T12:25:16.9074689Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9075417Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9075517Z warnings.warn( 2025-12-04T12:25:16.9075745Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9075853Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9075971Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9076201Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9076533Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9076636Z graph_break [] 2025-12-04T12:25:16.9076851Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9077573Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9077685Z warnings.warn( 2025-12-04T12:25:16.9077898Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9078008Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9078134Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9078354Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9078761Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9078860Z graph_break [] 2025-12-04T12:25:16.9079072Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9079810Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9079908Z warnings.warn( 2025-12-04T12:25:16.9080742Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a6ba20f02ccefade.xml - 2025-12-04T12:25:16.9080911Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9082017Z FAILED [0.1641s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9082027Z 2025-12-04T12:25:16.9082255Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9083253Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9083259Z 2025-12-04T12:25:16.9083534Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9083712Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9083909Z ================== 1 failed, 204 deselected, 2 rerun in 5.07s ================== 2025-12-04T12:25:16.9084019Z Got exit code 1 2025-12-04T12:25:16.9084936Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9085362Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.9086004Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dc63b1e627261305.xml 2025-12-04T12:25:16.9086170Z ============================= test session starts ============================== 2025-12-04T12:25:16.9086531Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9086641Z cachedir: .pytest_cache 2025-12-04T12:25:16.9087178Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9087301Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9087409Z configfile: pytest.ini 2025-12-04T12:25:16.9088018Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9088249Z collecting ... collected 380 items / 140 deselected / 240 selected 2025-12-04T12:25:16.9088401Z stepcurrent: skipping 140 already run items. 2025-12-04T12:25:16.9088527Z Running 65 items in this shard 2025-12-04T12:25:16.9088532Z 2025-12-04T12:25:16.9089478Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6657s] [ 1%] 2025-12-04T12:25:16.9090420Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1665s] [ 1%] 2025-12-04T12:25:16.9091271Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1635s] [ 1%] 2025-12-04T12:25:16.9091340Z 2025-12-04T12:25:16.9091492Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9092071Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9092194Z Traceback (most recent call last): 2025-12-04T12:25:16.9092677Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9092876Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9093086Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9093103Z 2025-12-04T12:25:16.9093313Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9094277Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9094286Z 2025-12-04T12:25:16.9094561Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9094782Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9094903Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9095049Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9095391Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9095624Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9095721Z graph_break [] 2025-12-04T12:25:16.9095933Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9097158Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9097270Z warnings.warn( 2025-12-04T12:25:16.9097863Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9097984Z Traceback (most recent call last): 2025-12-04T12:25:16.9098456Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9098665Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9098875Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9098881Z 2025-12-04T12:25:16.9099093Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9100079Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9100084Z 2025-12-04T12:25:16.9100357Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9100583Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9100694Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9100808Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9101164Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9101386Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9101496Z graph_break [] 2025-12-04T12:25:16.9101709Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9102450Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9102567Z warnings.warn( 2025-12-04T12:25:16.9102784Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9102975Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9103102Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9103321Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9103668Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9103766Z graph_break [] 2025-12-04T12:25:16.9103979Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9104720Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9104818Z warnings.warn( 2025-12-04T12:25:16.9104959Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9105552Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9105676Z Traceback (most recent call last): 2025-12-04T12:25:16.9106158Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9106354Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9106620Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9106626Z 2025-12-04T12:25:16.9106849Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9107817Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9107822Z 2025-12-04T12:25:16.9108099Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9108345Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9108462Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9108590Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9108926Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9109144Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9109253Z graph_break [] 2025-12-04T12:25:16.9109469Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9110228Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9110331Z warnings.warn( 2025-12-04T12:25:16.9110542Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9110667Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9110785Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9111003Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9111355Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9111455Z graph_break [] 2025-12-04T12:25:16.9111684Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9112418Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9112520Z warnings.warn( 2025-12-04T12:25:16.9112752Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9112862Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9112979Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9113215Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9113556Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9113726Z graph_break [] 2025-12-04T12:25:16.9113940Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9114670Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9114789Z warnings.warn( 2025-12-04T12:25:16.9115613Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dc63b1e627261305.xml - 2025-12-04T12:25:16.9115799Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9116915Z FAILED [0.1635s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9116925Z 2025-12-04T12:25:16.9117142Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9118133Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9118139Z 2025-12-04T12:25:16.9118435Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9118636Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9118838Z ================== 1 failed, 140 deselected, 2 rerun in 5.05s ================== 2025-12-04T12:25:16.9118939Z Got exit code 1 2025-12-04T12:25:16.9119063Z Retrying single test... 2025-12-04T12:25:16.9119749Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9dda8854f56bbc5e.xml 2025-12-04T12:25:16.9119925Z ============================= test session starts ============================== 2025-12-04T12:25:16.9120283Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9120394Z cachedir: .pytest_cache 2025-12-04T12:25:16.9120934Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9121058Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9121166Z configfile: pytest.ini 2025-12-04T12:25:16.9121774Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9122007Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.9123081Z stepcurrent: skipping 140 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9123202Z Running 1 items in this shard 2025-12-04T12:25:16.9123208Z 2025-12-04T12:25:16.9124139Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6375s] [100%] 2025-12-04T12:25:16.9125087Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1645s] [100%] 2025-12-04T12:25:16.9125935Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1609s] [100%] 2025-12-04T12:25:16.9125942Z 2025-12-04T12:25:16.9126102Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9126683Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9126879Z Traceback (most recent call last): 2025-12-04T12:25:16.9127348Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9127546Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9127771Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9127776Z 2025-12-04T12:25:16.9127992Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9128970Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9128975Z 2025-12-04T12:25:16.9129242Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9129465Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9129592Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9129707Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9130051Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9130315Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9130416Z graph_break [] 2025-12-04T12:25:16.9130648Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9131382Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9131483Z warnings.warn( 2025-12-04T12:25:16.9132109Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9132237Z Traceback (most recent call last): 2025-12-04T12:25:16.9132705Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9132919Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9133128Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9133136Z 2025-12-04T12:25:16.9133365Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9134330Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9134335Z 2025-12-04T12:25:16.9134611Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9134831Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9134945Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9135074Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9135415Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9135634Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9135745Z graph_break [] 2025-12-04T12:25:16.9135964Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9136703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9136804Z warnings.warn( 2025-12-04T12:25:16.9137091Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9137215Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9137330Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9137552Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9137975Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9138074Z graph_break [] 2025-12-04T12:25:16.9138285Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9139029Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9139128Z warnings.warn( 2025-12-04T12:25:16.9139281Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9139862Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9139984Z Traceback (most recent call last): 2025-12-04T12:25:16.9140472Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9140676Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9140902Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9140907Z 2025-12-04T12:25:16.9141120Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9142125Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9142131Z 2025-12-04T12:25:16.9142409Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9142624Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9142749Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9142862Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9143231Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9143465Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9143561Z graph_break [] 2025-12-04T12:25:16.9143776Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9144522Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9144622Z warnings.warn( 2025-12-04T12:25:16.9144846Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9144956Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9145070Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9145300Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9145639Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9145736Z graph_break [] 2025-12-04T12:25:16.9145965Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9146694Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9146806Z warnings.warn( 2025-12-04T12:25:16.9147022Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9147132Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9147256Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9147473Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9147811Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9147918Z graph_break [] 2025-12-04T12:25:16.9148139Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9148876Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9149052Z warnings.warn( 2025-12-04T12:25:16.9149874Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9dda8854f56bbc5e.xml - 2025-12-04T12:25:16.9150063Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9151169Z FAILED [0.1609s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9151175Z 2025-12-04T12:25:16.9151402Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9152377Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9152386Z 2025-12-04T12:25:16.9152651Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9152847Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9153089Z ================== 1 failed, 204 deselected, 2 rerun in 5.02s ================== 2025-12-04T12:25:16.9153205Z Got exit code 1 2025-12-04T12:25:16.9153312Z Retrying single test... 2025-12-04T12:25:16.9153959Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d35f802df30bd39d.xml 2025-12-04T12:25:16.9154132Z ============================= test session starts ============================== 2025-12-04T12:25:16.9154508Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9154616Z cachedir: .pytest_cache 2025-12-04T12:25:16.9155152Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9155274Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9155392Z configfile: pytest.ini 2025-12-04T12:25:16.9155989Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9156216Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.9157292Z stepcurrent: skipping 140 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9157410Z Running 1 items in this shard 2025-12-04T12:25:16.9157415Z 2025-12-04T12:25:16.9158362Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.7093s] [100%] 2025-12-04T12:25:16.9159297Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1705s] [100%] 2025-12-04T12:25:16.9160159Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1684s] [100%] 2025-12-04T12:25:16.9160165Z 2025-12-04T12:25:16.9160304Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9160886Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9161024Z Traceback (most recent call last): 2025-12-04T12:25:16.9161498Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9161766Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9161976Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9161981Z 2025-12-04T12:25:16.9162200Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9163179Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9163184Z 2025-12-04T12:25:16.9163450Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9163678Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9163792Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9163910Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9164271Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9164490Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9164588Z graph_break [] 2025-12-04T12:25:16.9164817Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9165589Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9165704Z warnings.warn( 2025-12-04T12:25:16.9166279Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9166404Z Traceback (most recent call last): 2025-12-04T12:25:16.9166989Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9167187Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9167403Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9167420Z 2025-12-04T12:25:16.9167632Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9168608Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9168614Z 2025-12-04T12:25:16.9168892Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9169107Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9169220Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9169349Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9169692Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9169927Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9170028Z graph_break [] 2025-12-04T12:25:16.9170243Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9171000Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9171103Z warnings.warn( 2025-12-04T12:25:16.9171317Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9171442Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9171559Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9171790Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9172128Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9172229Z graph_break [] 2025-12-04T12:25:16.9172458Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9173248Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9173350Z warnings.warn( 2025-12-04T12:25:16.9173511Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9174095Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9174232Z Traceback (most recent call last): 2025-12-04T12:25:16.9174705Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9174902Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9175131Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9175136Z 2025-12-04T12:25:16.9175356Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9176343Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9176348Z 2025-12-04T12:25:16.9176644Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9176935Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9177065Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9177185Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9177540Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9177762Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9177898Z graph_break [] 2025-12-04T12:25:16.9178130Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9178873Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9178976Z warnings.warn( 2025-12-04T12:25:16.9179206Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9179319Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9179434Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9179673Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9180010Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9180123Z graph_break [] 2025-12-04T12:25:16.9180338Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9181068Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9181188Z warnings.warn( 2025-12-04T12:25:16.9181402Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9181513Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9181640Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9181860Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9182216Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9182311Z graph_break [] 2025-12-04T12:25:16.9182523Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9183260Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9183362Z warnings.warn( 2025-12-04T12:25:16.9184198Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d35f802df30bd39d.xml - 2025-12-04T12:25:16.9184429Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9185536Z FAILED [0.1684s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9185543Z 2025-12-04T12:25:16.9185770Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9186735Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9186740Z 2025-12-04T12:25:16.9187020Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9187202Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9187400Z ================== 1 failed, 204 deselected, 2 rerun in 5.10s ================== 2025-12-04T12:25:16.9187511Z Got exit code 1 2025-12-04T12:25:16.9188431Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9188854Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.9189496Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d1473b4b18944240.xml 2025-12-04T12:25:16.9189658Z ============================= test session starts ============================== 2025-12-04T12:25:16.9190055Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9190169Z cachedir: .pytest_cache 2025-12-04T12:25:16.9190707Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9190829Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9190941Z configfile: pytest.ini 2025-12-04T12:25:16.9191547Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9191776Z collecting ... collected 380 items / 141 deselected / 239 selected 2025-12-04T12:25:16.9191929Z stepcurrent: skipping 141 already run items. 2025-12-04T12:25:16.9192057Z Running 64 items in this shard 2025-12-04T12:25:16.9192063Z 2025-12-04T12:25:16.9193120Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_True SKIPPED [0.0041s] (Skip non-critical tests to save resources.) [ 1%] 2025-12-04T12:25:16.9194180Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_True_initial_xblock_1_add_1dim_False SKIPPED [0.0031s] (Skip non-critical tests to save resources.) [ 3%] 2025-12-04T12:25:16.9195114Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6657s] [ 4%] 2025-12-04T12:25:16.9196244Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1640s] [ 4%] 2025-12-04T12:25:16.9197098Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1611s] [ 4%] 2025-12-04T12:25:16.9197222Z 2025-12-04T12:25:16.9197364Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9197960Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9198081Z Traceback (most recent call last): 2025-12-04T12:25:16.9198567Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9198768Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9198979Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9198984Z 2025-12-04T12:25:16.9199211Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9200189Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9200199Z 2025-12-04T12:25:16.9200479Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9200697Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9200810Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9200988Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9201328Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9201558Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9201655Z graph_break [] 2025-12-04T12:25:16.9201870Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9202654Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9202762Z warnings.warn( 2025-12-04T12:25:16.9203344Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9203482Z Traceback (most recent call last): 2025-12-04T12:25:16.9203955Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9204170Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9204381Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9204386Z 2025-12-04T12:25:16.9204600Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9205583Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9205591Z 2025-12-04T12:25:16.9205856Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9206094Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9206208Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9206325Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9206683Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9206902Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9207000Z graph_break [] 2025-12-04T12:25:16.9207227Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9207961Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9208075Z warnings.warn( 2025-12-04T12:25:16.9208292Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9208459Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9208585Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9208806Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9209144Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9209251Z graph_break [] 2025-12-04T12:25:16.9209467Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9210207Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9210308Z warnings.warn( 2025-12-04T12:25:16.9210452Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9211047Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9211173Z Traceback (most recent call last): 2025-12-04T12:25:16.9211643Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9211853Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9212062Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9212097Z 2025-12-04T12:25:16.9212324Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9213300Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9213306Z 2025-12-04T12:25:16.9213581Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9213824Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9213941Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9214073Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9214412Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9214629Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9214738Z graph_break [] 2025-12-04T12:25:16.9214957Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9215702Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9215803Z warnings.warn( 2025-12-04T12:25:16.9216018Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9216141Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9216256Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9216478Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9216833Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9216998Z graph_break [] 2025-12-04T12:25:16.9217214Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9217959Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9218059Z warnings.warn( 2025-12-04T12:25:16.9218286Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9218396Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9218510Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9218743Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9219086Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9219252Z graph_break [] 2025-12-04T12:25:16.9219483Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9220209Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9220320Z warnings.warn( 2025-12-04T12:25:16.9221147Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d1473b4b18944240.xml - 2025-12-04T12:25:16.9221319Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9222440Z FAILED [0.1611s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9222446Z 2025-12-04T12:25:16.9222667Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9223652Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9223657Z 2025-12-04T12:25:16.9223955Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9224151Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9224375Z ============ 1 failed, 2 skipped, 141 deselected, 2 rerun in 5.06s ============= 2025-12-04T12:25:16.9224477Z Got exit code 1 2025-12-04T12:25:16.9224595Z Retrying single test... 2025-12-04T12:25:16.9225240Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c84649193ffdba8a.xml 2025-12-04T12:25:16.9225433Z ============================= test session starts ============================== 2025-12-04T12:25:16.9225801Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9225910Z cachedir: .pytest_cache 2025-12-04T12:25:16.9226449Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9226573Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9226681Z configfile: pytest.ini 2025-12-04T12:25:16.9227282Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9227511Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.9228583Z stepcurrent: skipping 143 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9228701Z Running 1 items in this shard 2025-12-04T12:25:16.9228706Z 2025-12-04T12:25:16.9229641Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6772s] [100%] 2025-12-04T12:25:16.9230595Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1669s] [100%] 2025-12-04T12:25:16.9231439Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1648s] [100%] 2025-12-04T12:25:16.9231444Z 2025-12-04T12:25:16.9231597Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9232179Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9232364Z Traceback (most recent call last): 2025-12-04T12:25:16.9232846Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9233047Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9233284Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9233289Z 2025-12-04T12:25:16.9233504Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9234483Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9234502Z 2025-12-04T12:25:16.9234774Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9234996Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9235130Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9235248Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9235590Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9235860Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9235963Z graph_break [] 2025-12-04T12:25:16.9236182Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9236932Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9237035Z warnings.warn( 2025-12-04T12:25:16.9237677Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9237802Z Traceback (most recent call last): 2025-12-04T12:25:16.9238279Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9238496Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9238710Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9238715Z 2025-12-04T12:25:16.9238948Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9239921Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9239926Z 2025-12-04T12:25:16.9240193Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9240433Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9240546Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9240682Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9241026Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9241246Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9241360Z graph_break [] 2025-12-04T12:25:16.9241584Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9242323Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9242438Z warnings.warn( 2025-12-04T12:25:16.9242656Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9242781Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9242897Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9243120Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9243534Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9243634Z graph_break [] 2025-12-04T12:25:16.9243849Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9244591Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9244691Z warnings.warn( 2025-12-04T12:25:16.9244845Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9245427Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9245547Z Traceback (most recent call last): 2025-12-04T12:25:16.9246033Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9246232Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9246446Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9246462Z 2025-12-04T12:25:16.9246676Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9247682Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9247688Z 2025-12-04T12:25:16.9247966Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9248182Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9248293Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9248419Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9248786Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9249022Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9249120Z graph_break [] 2025-12-04T12:25:16.9249335Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9250079Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9250182Z warnings.warn( 2025-12-04T12:25:16.9250393Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9250514Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9250628Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9250858Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9251194Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9251292Z graph_break [] 2025-12-04T12:25:16.9251519Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9252252Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9252355Z warnings.warn( 2025-12-04T12:25:16.9252583Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9252694Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9252822Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9253041Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9253375Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9253485Z graph_break [] 2025-12-04T12:25:16.9253698Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9254428Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9254597Z warnings.warn( 2025-12-04T12:25:16.9255420Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c84649193ffdba8a.xml - 2025-12-04T12:25:16.9255608Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9256709Z FAILED [0.1648s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9256715Z 2025-12-04T12:25:16.9257016Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9257993Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9258003Z 2025-12-04T12:25:16.9258268Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9258462Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9258696Z ================== 1 failed, 204 deselected, 2 rerun in 5.06s ================== 2025-12-04T12:25:16.9258808Z Got exit code 1 2025-12-04T12:25:16.9258916Z Retrying single test... 2025-12-04T12:25:16.9259557Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b127809f84c30c37.xml 2025-12-04T12:25:16.9259732Z ============================= test session starts ============================== 2025-12-04T12:25:16.9260112Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9260224Z cachedir: .pytest_cache 2025-12-04T12:25:16.9260765Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9260889Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9261009Z configfile: pytest.ini 2025-12-04T12:25:16.9261600Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9261829Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.9262894Z stepcurrent: skipping 143 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9263009Z Running 1 items in this shard 2025-12-04T12:25:16.9263013Z 2025-12-04T12:25:16.9263965Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6454s] [100%] 2025-12-04T12:25:16.9264906Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1623s] [100%] 2025-12-04T12:25:16.9265762Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1600s] [100%] 2025-12-04T12:25:16.9265781Z 2025-12-04T12:25:16.9265922Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9266504Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9266638Z Traceback (most recent call last): 2025-12-04T12:25:16.9267116Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9267372Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9267596Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9267601Z 2025-12-04T12:25:16.9267816Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9268795Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9268800Z 2025-12-04T12:25:16.9269066Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9269282Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9269406Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9269523Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9269880Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9270099Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9270197Z graph_break [] 2025-12-04T12:25:16.9270422Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9271187Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9271289Z warnings.warn( 2025-12-04T12:25:16.9271882Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9272003Z Traceback (most recent call last): 2025-12-04T12:25:16.9272513Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9272711Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9272929Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9272934Z 2025-12-04T12:25:16.9273157Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9274136Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9274141Z 2025-12-04T12:25:16.9274419Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9274636Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9274752Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9274882Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9275228Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9275461Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9275563Z graph_break [] 2025-12-04T12:25:16.9275778Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9276526Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9276628Z warnings.warn( 2025-12-04T12:25:16.9276844Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9276969Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9277083Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9277304Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9277652Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9277753Z graph_break [] 2025-12-04T12:25:16.9277979Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9278762Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9278862Z warnings.warn( 2025-12-04T12:25:16.9279019Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9279605Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9279738Z Traceback (most recent call last): 2025-12-04T12:25:16.9280209Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9280406Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9280630Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9280637Z 2025-12-04T12:25:16.9280851Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9281837Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9281854Z 2025-12-04T12:25:16.9282167Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9282382Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9282506Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9282620Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9282959Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9283191Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9283292Z graph_break [] 2025-12-04T12:25:16.9283556Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9284291Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9284392Z warnings.warn( 2025-12-04T12:25:16.9284618Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9284732Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9284850Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9285078Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9285417Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9285528Z graph_break [] 2025-12-04T12:25:16.9285741Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9286466Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9286584Z warnings.warn( 2025-12-04T12:25:16.9286794Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9286904Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9287029Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9287249Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9287593Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9287690Z graph_break [] 2025-12-04T12:25:16.9287903Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9288641Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9288741Z warnings.warn( 2025-12-04T12:25:16.9289556Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b127809f84c30c37.xml - 2025-12-04T12:25:16.9289806Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9290908Z FAILED [0.1600s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9290914Z 2025-12-04T12:25:16.9291142Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9292113Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9292117Z 2025-12-04T12:25:16.9292398Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9292583Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9292782Z ================== 1 failed, 204 deselected, 2 rerun in 5.02s ================== 2025-12-04T12:25:16.9292894Z Got exit code 1 2025-12-04T12:25:16.9293808Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9294233Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.9294878Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-1a7c4cad159549d9.xml 2025-12-04T12:25:16.9295044Z ============================= test session starts ============================== 2025-12-04T12:25:16.9295438Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9295555Z cachedir: .pytest_cache 2025-12-04T12:25:16.9296245Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9296391Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9296501Z configfile: pytest.ini 2025-12-04T12:25:16.9297180Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9297427Z collecting ... collected 380 items / 144 deselected / 236 selected 2025-12-04T12:25:16.9297650Z stepcurrent: skipping 144 already run items. 2025-12-04T12:25:16.9297839Z Running 61 items in this shard 2025-12-04T12:25:16.9297847Z 2025-12-04T12:25:16.9299002Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_True SKIPPED [0.0041s] (Skip non-critical tests to save resources.) [ 1%] 2025-12-04T12:25:16.9300072Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_True_initial_xblock_2_add_1dim_False SKIPPED [0.0032s] (Skip non-critical tests to save resources.) [ 3%] 2025-12-04T12:25:16.9301006Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6579s] [ 4%] 2025-12-04T12:25:16.9301939Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1710s] [ 4%] 2025-12-04T12:25:16.9302788Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1624s] [ 4%] 2025-12-04T12:25:16.9302913Z 2025-12-04T12:25:16.9303056Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9303646Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9303770Z Traceback (most recent call last): 2025-12-04T12:25:16.9304264Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9304465Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9304680Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9304685Z 2025-12-04T12:25:16.9304916Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9305887Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9305897Z 2025-12-04T12:25:16.9306183Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9306404Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9306519Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9306649Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9307030Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9307251Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9307366Z graph_break [] 2025-12-04T12:25:16.9307587Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9308374Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9308480Z warnings.warn( 2025-12-04T12:25:16.9309138Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9309279Z Traceback (most recent call last): 2025-12-04T12:25:16.9309752Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9309953Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9310178Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9310184Z 2025-12-04T12:25:16.9310398Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9311372Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9311378Z 2025-12-04T12:25:16.9311645Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9311876Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9311987Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9312101Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9312459Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9312683Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9312782Z graph_break [] 2025-12-04T12:25:16.9313011Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9313746Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9313859Z warnings.warn( 2025-12-04T12:25:16.9314075Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9314184Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9314376Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9314596Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9314938Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9315047Z graph_break [] 2025-12-04T12:25:16.9315262Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9316001Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9316104Z warnings.warn( 2025-12-04T12:25:16.9316247Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9316839Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9316961Z Traceback (most recent call last): 2025-12-04T12:25:16.9317437Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9317646Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9317858Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9317863Z 2025-12-04T12:25:16.9318118Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9319083Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9319088Z 2025-12-04T12:25:16.9319354Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9319613Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9319728Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9319864Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9320203Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9320421Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9320531Z graph_break [] 2025-12-04T12:25:16.9320748Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9321478Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9321589Z warnings.warn( 2025-12-04T12:25:16.9321803Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9321924Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9322038Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9322260Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9322612Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9322708Z graph_break [] 2025-12-04T12:25:16.9322921Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9323661Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9323763Z warnings.warn( 2025-12-04T12:25:16.9323989Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9324100Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9324215Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9324447Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9324785Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9324883Z graph_break [] 2025-12-04T12:25:16.9325192Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9325920Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9326034Z warnings.warn( 2025-12-04T12:25:16.9326858Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-1a7c4cad159549d9.xml - 2025-12-04T12:25:16.9327030Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9328134Z FAILED [0.1624s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9328142Z 2025-12-04T12:25:16.9328360Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9329337Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9329342Z 2025-12-04T12:25:16.9329640Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9329823Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9330058Z ============ 1 failed, 2 skipped, 144 deselected, 2 rerun in 5.06s ============= 2025-12-04T12:25:16.9330158Z Got exit code 1 2025-12-04T12:25:16.9330279Z Retrying single test... 2025-12-04T12:25:16.9330919Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2f388a73f046ed87.xml 2025-12-04T12:25:16.9331112Z ============================= test session starts ============================== 2025-12-04T12:25:16.9331478Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9331586Z cachedir: .pytest_cache 2025-12-04T12:25:16.9332113Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9332250Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9332361Z configfile: pytest.ini 2025-12-04T12:25:16.9332965Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9333191Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.9334246Z stepcurrent: skipping 146 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9334375Z Running 1 items in this shard 2025-12-04T12:25:16.9334382Z 2025-12-04T12:25:16.9335303Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6544s] [100%] 2025-12-04T12:25:16.9336237Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1674s] [100%] 2025-12-04T12:25:16.9337160Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1624s] [100%] 2025-12-04T12:25:16.9337167Z 2025-12-04T12:25:16.9337322Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9337902Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9338093Z Traceback (most recent call last): 2025-12-04T12:25:16.9338577Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9338777Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9339004Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9339009Z 2025-12-04T12:25:16.9339222Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9340192Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9340198Z 2025-12-04T12:25:16.9340474Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9340696Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9340831Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9340946Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9341284Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9341514Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9341641Z graph_break [] 2025-12-04T12:25:16.9341857Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9342605Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9342706Z warnings.warn( 2025-12-04T12:25:16.9343295Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9343446Z Traceback (most recent call last): 2025-12-04T12:25:16.9343920Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9344130Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9344341Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9344347Z 2025-12-04T12:25:16.9344560Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9345540Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9345545Z 2025-12-04T12:25:16.9345809Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9346038Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9346150Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9346265Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9346620Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9346838Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9346952Z graph_break [] 2025-12-04T12:25:16.9347166Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9347905Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9348021Z warnings.warn( 2025-12-04T12:25:16.9348235Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9348348Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9348478Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9348699Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9349053Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9349209Z graph_break [] 2025-12-04T12:25:16.9349422Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9350168Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9350270Z warnings.warn( 2025-12-04T12:25:16.9350414Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9350999Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9351120Z Traceback (most recent call last): 2025-12-04T12:25:16.9351604Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9351804Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9352018Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9352023Z 2025-12-04T12:25:16.9352250Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9353304Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9353309Z 2025-12-04T12:25:16.9353589Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9353805Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9353917Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9354046Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9354419Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9354653Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9354763Z graph_break [] 2025-12-04T12:25:16.9354977Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9355723Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9355828Z warnings.warn( 2025-12-04T12:25:16.9356044Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9356169Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9356286Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9356506Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9356856Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9356952Z graph_break [] 2025-12-04T12:25:16.9357186Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9357919Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9358021Z warnings.warn( 2025-12-04T12:25:16.9358249Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9358363Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9358479Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9358713Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9359051Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9359165Z graph_break [] 2025-12-04T12:25:16.9359381Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9360111Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9360347Z warnings.warn( 2025-12-04T12:25:16.9361169Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2f388a73f046ed87.xml - 2025-12-04T12:25:16.9361361Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9362459Z FAILED [0.1624s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9362464Z 2025-12-04T12:25:16.9362684Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9363669Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9363679Z 2025-12-04T12:25:16.9363946Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9364141Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9364340Z ================== 1 failed, 204 deselected, 2 rerun in 5.04s ================== 2025-12-04T12:25:16.9364470Z Got exit code 1 2025-12-04T12:25:16.9364593Z Retrying single test... 2025-12-04T12:25:16.9365241Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d68d80f1ba61c89f.xml 2025-12-04T12:25:16.9365419Z ============================= test session starts ============================== 2025-12-04T12:25:16.9365773Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9365912Z cachedir: .pytest_cache 2025-12-04T12:25:16.9366452Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9366581Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9366691Z configfile: pytest.ini 2025-12-04T12:25:16.9367300Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9367528Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.9368599Z stepcurrent: skipping 146 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9368717Z Running 1 items in this shard 2025-12-04T12:25:16.9368722Z 2025-12-04T12:25:16.9369657Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6635s] [100%] 2025-12-04T12:25:16.9370601Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1676s] [100%] 2025-12-04T12:25:16.9371445Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1640s] [100%] 2025-12-04T12:25:16.9371451Z 2025-12-04T12:25:16.9371605Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9372177Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9372308Z Traceback (most recent call last): 2025-12-04T12:25:16.9372777Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9373053Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9373275Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9373281Z 2025-12-04T12:25:16.9373497Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9374474Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9374479Z 2025-12-04T12:25:16.9374742Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9374957Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9375079Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9375194Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9375535Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9375771Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9375868Z graph_break [] 2025-12-04T12:25:16.9376093Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9376928Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9377039Z warnings.warn( 2025-12-04T12:25:16.9377626Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9377746Z Traceback (most recent call last): 2025-12-04T12:25:16.9378229Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9378458Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9378673Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9378681Z 2025-12-04T12:25:16.9378910Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9379878Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9379883Z 2025-12-04T12:25:16.9380164Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9380380Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9380492Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9380623Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9380962Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9381181Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9381292Z graph_break [] 2025-12-04T12:25:16.9381506Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9382250Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9382350Z warnings.warn( 2025-12-04T12:25:16.9382570Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9382691Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9382805Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9383023Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9383370Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9383466Z graph_break [] 2025-12-04T12:25:16.9383697Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9384492Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9384592Z warnings.warn( 2025-12-04T12:25:16.9384747Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9385324Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9385445Z Traceback (most recent call last): 2025-12-04T12:25:16.9385927Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9386126Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9386350Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9386356Z 2025-12-04T12:25:16.9386569Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9387531Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9387537Z 2025-12-04T12:25:16.9387818Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9388062Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9388187Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9388300Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9388640Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9388870Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9388966Z graph_break [] 2025-12-04T12:25:16.9389221Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9390158Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9390268Z warnings.warn( 2025-12-04T12:25:16.9390496Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9390605Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9390738Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9391040Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9391380Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9391476Z graph_break [] 2025-12-04T12:25:16.9391702Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9392432Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9392550Z warnings.warn( 2025-12-04T12:25:16.9392763Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9392871Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9392998Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9393216Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9393554Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9393663Z graph_break [] 2025-12-04T12:25:16.9393874Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9394612Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9394713Z warnings.warn( 2025-12-04T12:25:16.9395537Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d68d80f1ba61c89f.xml - 2025-12-04T12:25:16.9395795Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9397032Z FAILED [0.1640s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9397038Z 2025-12-04T12:25:16.9397269Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9398232Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9398238Z 2025-12-04T12:25:16.9398508Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9398702Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9398907Z ================== 1 failed, 204 deselected, 2 rerun in 5.05s ================== 2025-12-04T12:25:16.9399021Z Got exit code 1 2025-12-04T12:25:16.9399960Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9400370Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.9401028Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b8a30a3f316b7f0b.xml 2025-12-04T12:25:16.9401190Z ============================= test session starts ============================== 2025-12-04T12:25:16.9401595Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9401706Z cachedir: .pytest_cache 2025-12-04T12:25:16.9402227Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9402364Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9402472Z configfile: pytest.ini 2025-12-04T12:25:16.9403064Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9403302Z collecting ... collected 380 items / 147 deselected / 233 selected 2025-12-04T12:25:16.9403450Z stepcurrent: skipping 147 already run items. 2025-12-04T12:25:16.9403578Z Running 58 items in this shard 2025-12-04T12:25:16.9403583Z 2025-12-04T12:25:16.9404515Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [4.7006s] [ 1%] 2025-12-04T12:25:16.9405443Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [0.1741s] [ 1%] 2025-12-04T12:25:16.9406295Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True FAILED [0.1693s] [ 1%] 2025-12-04T12:25:16.9406303Z 2025-12-04T12:25:16.9406442Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9407025Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.9407146Z Traceback (most recent call last): 2025-12-04T12:25:16.9407630Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9407832Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9408120Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9408125Z 2025-12-04T12:25:16.9408353Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9409313Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.9409318Z 2025-12-04T12:25:16.9409602Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9409821Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9409934Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9410063Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9410405Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9410629Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9410746Z graph_break [] 2025-12-04T12:25:16.9410966Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9413722Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9413833Z return x.grad, w.grad 2025-12-04T12:25:16.9414624Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9414730Z warnings.warn( 2025-12-04T12:25:16.9417500Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9423687Z return x.grad, w.grad 2025-12-04T12:25:16.9424373Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.9424517Z Traceback (most recent call last): 2025-12-04T12:25:16.9424993Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9425201Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9425424Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9425432Z 2025-12-04T12:25:16.9425645Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9426629Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.9426636Z 2025-12-04T12:25:16.9426900Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9427122Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9427241Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9427356Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9427838Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9428061Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9428157Z graph_break [] 2025-12-04T12:25:16.9428384Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9431107Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9431229Z return x.grad, w.grad 2025-12-04T12:25:16.9431966Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9432063Z warnings.warn( 2025-12-04T12:25:16.9434805Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9434911Z return x.grad, w.grad 2025-12-04T12:25:16.9435175Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9435286Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9435396Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9435625Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9435963Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9436073Z graph_break [] 2025-12-04T12:25:16.9436288Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9438992Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9439112Z return x.grad, w.grad 2025-12-04T12:25:16.9439844Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9439959Z warnings.warn( 2025-12-04T12:25:16.9442654Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9442826Z return x.grad, w.grad 2025-12-04T12:25:16.9442972Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9443538Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.9443668Z Traceback (most recent call last): 2025-12-04T12:25:16.9444135Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9444341Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9444553Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9444560Z 2025-12-04T12:25:16.9444772Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9445738Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.9445748Z 2025-12-04T12:25:16.9446013Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9446233Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9446343Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9446486Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9446833Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9447049Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9447141Z graph_break [] 2025-12-04T12:25:16.9447363Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9450097Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9450213Z return x.grad, w.grad 2025-12-04T12:25:16.9450948Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9451053Z warnings.warn( 2025-12-04T12:25:16.9453760Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9453878Z return x.grad, w.grad 2025-12-04T12:25:16.9454095Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9454204Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9454319Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9454539Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9454880Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9454985Z graph_break [] 2025-12-04T12:25:16.9455200Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9458078Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9458183Z return x.grad, w.grad 2025-12-04T12:25:16.9458923Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9459017Z warnings.warn( 2025-12-04T12:25:16.9461734Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9461857Z return x.grad, w.grad 2025-12-04T12:25:16.9462066Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9462184Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9462291Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9462506Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9462884Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9462982Z graph_break [] 2025-12-04T12:25:16.9463203Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9463927Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9464026Z warnings.warn( 2025-12-04T12:25:16.9466730Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9466835Z return x.grad, w.grad 2025-12-04T12:25:16.9467663Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b8a30a3f316b7f0b.xml - 2025-12-04T12:25:16.9467829Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9468928Z FAILED [0.1693s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9468935Z 2025-12-04T12:25:16.9469150Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9470104Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.9470169Z 2025-12-04T12:25:16.9470443Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9470619Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9470821Z ================== 1 failed, 147 deselected, 2 rerun in 5.10s ================== 2025-12-04T12:25:16.9470917Z Got exit code 1 2025-12-04T12:25:16.9471024Z Retrying single test... 2025-12-04T12:25:16.9471673Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d3917b026aff3e64.xml 2025-12-04T12:25:16.9471835Z ============================= test session starts ============================== 2025-12-04T12:25:16.9472184Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9472299Z cachedir: .pytest_cache 2025-12-04T12:25:16.9472821Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9472961Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9473063Z configfile: pytest.ini 2025-12-04T12:25:16.9473653Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9473939Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.9474980Z stepcurrent: skipping 147 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.9475105Z Running 1 items in this shard 2025-12-04T12:25:16.9475110Z 2025-12-04T12:25:16.9476061Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [4.6808s] [100%] 2025-12-04T12:25:16.9476983Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [0.1700s] [100%] 2025-12-04T12:25:16.9477834Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True FAILED [0.1654s] [100%] 2025-12-04T12:25:16.9477840Z 2025-12-04T12:25:16.9477977Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9478556Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.9478674Z Traceback (most recent call last): 2025-12-04T12:25:16.9479146Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9479357Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9479570Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9479576Z 2025-12-04T12:25:16.9479799Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9480768Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.9480774Z 2025-12-04T12:25:16.9481046Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9481264Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9481374Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9481500Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9481840Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9482108Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9482219Z graph_break [] 2025-12-04T12:25:16.9482431Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9485151Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9485257Z return x.grad, w.grad 2025-12-04T12:25:16.9486001Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9486105Z warnings.warn( 2025-12-04T12:25:16.9488830Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9488948Z return x.grad, w.grad 2025-12-04T12:25:16.9489547Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.9489676Z Traceback (most recent call last): 2025-12-04T12:25:16.9490259Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9490459Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9490678Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9490683Z 2025-12-04T12:25:16.9490897Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9491864Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.9491870Z 2025-12-04T12:25:16.9492135Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9492358Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9492477Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9492597Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9492940Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9493158Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9493255Z graph_break [] 2025-12-04T12:25:16.9493478Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9496364Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9496609Z return x.grad, w.grad 2025-12-04T12:25:16.9497424Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9497527Z warnings.warn( 2025-12-04T12:25:16.9500239Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9500343Z return x.grad, w.grad 2025-12-04T12:25:16.9500571Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9500676Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9500802Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9501021Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9501405Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9501511Z graph_break [] 2025-12-04T12:25:16.9501721Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9504444Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9504558Z return x.grad, w.grad 2025-12-04T12:25:16.9505288Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9505399Z warnings.warn( 2025-12-04T12:25:16.9508081Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9508201Z return x.grad, w.grad 2025-12-04T12:25:16.9508342Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9508924Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.9509046Z Traceback (most recent call last): 2025-12-04T12:25:16.9509507Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9509713Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9509920Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9509927Z 2025-12-04T12:25:16.9510137Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9511100Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.9511168Z 2025-12-04T12:25:16.9511434Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9511657Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9511765Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9511877Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9512226Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9512440Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9512542Z graph_break [] 2025-12-04T12:25:16.9512758Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9515488Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9515608Z return x.grad, w.grad 2025-12-04T12:25:16.9516339Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9516444Z warnings.warn( 2025-12-04T12:25:16.9519183Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9519298Z return x.grad, w.grad 2025-12-04T12:25:16.9519515Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9519623Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9519742Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9519964Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9520313Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9520413Z graph_break [] 2025-12-04T12:25:16.9520629Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9523338Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9523444Z return x.grad, w.grad 2025-12-04T12:25:16.9524179Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9524281Z warnings.warn( 2025-12-04T12:25:16.9527058Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9527160Z return x.grad, w.grad 2025-12-04T12:25:16.9527373Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9527491Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9527608Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9527829Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9528173Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9528273Z graph_break [] 2025-12-04T12:25:16.9528496Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9529260Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9529358Z warnings.warn( 2025-12-04T12:25:16.9532094Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9532206Z return x.grad, w.grad 2025-12-04T12:25:16.9533032Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d3917b026aff3e64.xml - 2025-12-04T12:25:16.9533207Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9534309Z FAILED [0.1654s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9534315Z 2025-12-04T12:25:16.9534531Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9535488Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.9535507Z 2025-12-04T12:25:16.9535771Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9535954Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9536169Z ================== 1 failed, 204 deselected, 2 rerun in 5.07s ================== 2025-12-04T12:25:16.9536265Z Got exit code 1 2025-12-04T12:25:16.9536373Z Retrying single test... 2025-12-04T12:25:16.9537096Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c0c657ac1df627f9.xml 2025-12-04T12:25:16.9537260Z ============================= test session starts ============================== 2025-12-04T12:25:16.9537618Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9537733Z cachedir: .pytest_cache 2025-12-04T12:25:16.9538320Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9538454Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9538561Z configfile: pytest.ini 2025-12-04T12:25:16.9539155Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9539386Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.9540432Z stepcurrent: skipping 147 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.9540559Z Running 1 items in this shard 2025-12-04T12:25:16.9540564Z 2025-12-04T12:25:16.9541484Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [4.6671s] [100%] 2025-12-04T12:25:16.9542412Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [0.1722s] [100%] 2025-12-04T12:25:16.9543280Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True FAILED [0.1673s] [100%] 2025-12-04T12:25:16.9543286Z 2025-12-04T12:25:16.9543426Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9544003Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.9544122Z Traceback (most recent call last): 2025-12-04T12:25:16.9544636Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9544834Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9545044Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9545049Z 2025-12-04T12:25:16.9545267Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9546225Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.9546231Z 2025-12-04T12:25:16.9546509Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9546721Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9546831Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9546959Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9547301Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9547518Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9547622Z graph_break [] 2025-12-04T12:25:16.9547835Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9550562Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9550665Z return x.grad, w.grad 2025-12-04T12:25:16.9551467Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9551563Z warnings.warn( 2025-12-04T12:25:16.9554269Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9554383Z return x.grad, w.grad 2025-12-04T12:25:16.9554953Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.9555088Z Traceback (most recent call last): 2025-12-04T12:25:16.9555551Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9555743Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9555993Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9555999Z 2025-12-04T12:25:16.9556207Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9557171Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.9557176Z 2025-12-04T12:25:16.9557440Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9557692Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9557807Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9557918Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9558263Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9558480Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9558574Z graph_break [] 2025-12-04T12:25:16.9558796Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9561518Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9561635Z return x.grad, w.grad 2025-12-04T12:25:16.9562367Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9562466Z warnings.warn( 2025-12-04T12:25:16.9565177Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9565419Z return x.grad, w.grad 2025-12-04T12:25:16.9565646Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9565756Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9565881Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9566102Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9566443Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9566550Z graph_break [] 2025-12-04T12:25:16.9566768Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9569476Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9569587Z return x.grad, w.grad 2025-12-04T12:25:16.9570350Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9570462Z warnings.warn( 2025-12-04T12:25:16.9573186Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9573305Z return x.grad, w.grad 2025-12-04T12:25:16.9573448Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9574026Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:16.9574144Z Traceback (most recent call last): 2025-12-04T12:25:16.9574612Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9574815Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9575027Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9575032Z 2025-12-04T12:25:16.9575259Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9576218Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.9576224Z 2025-12-04T12:25:16.9576488Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9576713Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9576823Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9577002Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9577354Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9577571Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9577679Z graph_break [] 2025-12-04T12:25:16.9577894Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9580697Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9580812Z return x.grad, w.grad 2025-12-04T12:25:16.9581537Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9581646Z warnings.warn( 2025-12-04T12:25:16.9584356Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9584475Z return x.grad, w.grad 2025-12-04T12:25:16.9584690Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9584802Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9584922Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9585143Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9585520Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9585618Z graph_break [] 2025-12-04T12:25:16.9585834Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9588537Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9588645Z return x.grad, w.grad 2025-12-04T12:25:16.9589389Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9589489Z warnings.warn( 2025-12-04T12:25:16.9592194Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9592297Z return x.grad, w.grad 2025-12-04T12:25:16.9592509Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9592630Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9592744Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9592973Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9593371Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9593464Z graph_break [] 2025-12-04T12:25:16.9593688Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9594414Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9594514Z warnings.warn( 2025-12-04T12:25:16.9597414Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:16.9597524Z return x.grad, w.grad 2025-12-04T12:25:16.9598415Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c0c657ac1df627f9.xml - 2025-12-04T12:25:16.9598582Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9599691Z FAILED [0.1673s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9599697Z 2025-12-04T12:25:16.9599912Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9600906Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.9600927Z 2025-12-04T12:25:16.9601196Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9601378Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9601590Z ================== 1 failed, 204 deselected, 2 rerun in 5.06s ================== 2025-12-04T12:25:16.9601688Z Got exit code 1 2025-12-04T12:25:16.9602562Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:16.9602981Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.9603624Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-175aa275ea059250.xml 2025-12-04T12:25:16.9603797Z ============================= test session starts ============================== 2025-12-04T12:25:16.9604148Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9604256Z cachedir: .pytest_cache 2025-12-04T12:25:16.9604785Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9604908Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9605023Z configfile: pytest.ini 2025-12-04T12:25:16.9605609Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9605834Z collecting ... collected 380 items / 148 deselected / 232 selected 2025-12-04T12:25:16.9605992Z stepcurrent: skipping 148 already run items. 2025-12-04T12:25:16.9606104Z Running 57 items in this shard 2025-12-04T12:25:16.9606191Z 2025-12-04T12:25:16.9607249Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_True_initial_xblock_1_add_1dim_False SKIPPED [0.0041s] (Skip non-critical tests to save resources.) [ 1%] 2025-12-04T12:25:16.9608294Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_True_initial_xblock_2_add_1dim_False SKIPPED [0.0030s] (Skip non-critical tests to save resources.) [ 3%] 2025-12-04T12:25:16.9609315Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_True_initial_xblock_2_add_1dim_True SKIPPED [0.0037s] (Skip non-critical tests to save resources.) [ 5%] 2025-12-04T12:25:16.9610258Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6708s] [ 7%] 2025-12-04T12:25:16.9611183Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1676s] [ 7%] 2025-12-04T12:25:16.9612071Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1646s] [ 7%] 2025-12-04T12:25:16.9612077Z 2025-12-04T12:25:16.9612218Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9612801Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9612930Z Traceback (most recent call last): 2025-12-04T12:25:16.9613430Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9613645Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9613857Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9613864Z 2025-12-04T12:25:16.9614093Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9615056Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9615061Z 2025-12-04T12:25:16.9615327Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9615559Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9615675Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9615802Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9616146Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9616369Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9616480Z graph_break [] 2025-12-04T12:25:16.9616696Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9617492Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9617612Z warnings.warn( 2025-12-04T12:25:16.9618188Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9618322Z Traceback (most recent call last): 2025-12-04T12:25:16.9618792Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9618994Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9619287Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9619293Z 2025-12-04T12:25:16.9619508Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9620490Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9620496Z 2025-12-04T12:25:16.9620762Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9620978Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9621104Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9621222Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9621560Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9621797Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9621898Z graph_break [] 2025-12-04T12:25:16.9622126Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9622865Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9622996Z warnings.warn( 2025-12-04T12:25:16.9623225Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9623337Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9623448Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9623675Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9624012Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9624122Z graph_break [] 2025-12-04T12:25:16.9624386Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9625120Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9625234Z warnings.warn( 2025-12-04T12:25:16.9625377Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9625954Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9626087Z Traceback (most recent call last): 2025-12-04T12:25:16.9626557Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9626765Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9626973Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9626978Z 2025-12-04T12:25:16.9627192Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9628171Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9628176Z 2025-12-04T12:25:16.9628439Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9628668Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9628777Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9628890Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9629243Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9629458Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9629554Z graph_break [] 2025-12-04T12:25:16.9629785Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9630575Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9630689Z warnings.warn( 2025-12-04T12:25:16.9630904Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9631012Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9631140Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9631357Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9631691Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9631798Z graph_break [] 2025-12-04T12:25:16.9632013Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9632757Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9632857Z warnings.warn( 2025-12-04T12:25:16.9633071Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9633193Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9633304Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9633521Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9633897Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9633997Z graph_break [] 2025-12-04T12:25:16.9634227Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9634958Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9635056Z warnings.warn( 2025-12-04T12:25:16.9635921Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-175aa275ea059250.xml - 2025-12-04T12:25:16.9636097Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9637210Z FAILED [0.1646s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9637216Z 2025-12-04T12:25:16.9637431Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9638386Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9638402Z 2025-12-04T12:25:16.9638670Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9638849Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9639083Z ============ 1 failed, 3 skipped, 148 deselected, 2 rerun in 5.07s ============= 2025-12-04T12:25:16.9639181Z Got exit code 1 2025-12-04T12:25:16.9639286Z Retrying single test... 2025-12-04T12:25:16.9639946Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-5ce17ea59d8fd893.xml 2025-12-04T12:25:16.9640107Z ============================= test session starts ============================== 2025-12-04T12:25:16.9640466Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9640573Z cachedir: .pytest_cache 2025-12-04T12:25:16.9641095Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9641234Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9641343Z configfile: pytest.ini 2025-12-04T12:25:16.9641993Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9642231Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.9643283Z stepcurrent: skipping 151 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9643411Z Running 1 items in this shard 2025-12-04T12:25:16.9643416Z 2025-12-04T12:25:16.9644354Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.9424s] [100%] 2025-12-04T12:25:16.9645299Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1858s] [100%] 2025-12-04T12:25:16.9646152Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1827s] [100%] 2025-12-04T12:25:16.9646157Z 2025-12-04T12:25:16.9646328Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9646919Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9647041Z Traceback (most recent call last): 2025-12-04T12:25:16.9647520Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9647716Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9647957Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9647965Z 2025-12-04T12:25:16.9648194Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9649158Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9649165Z 2025-12-04T12:25:16.9649443Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9649660Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9649773Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9649898Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9650242Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9650462Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9650573Z graph_break [] 2025-12-04T12:25:16.9650789Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9651534Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9651634Z warnings.warn( 2025-12-04T12:25:16.9652210Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9652342Z Traceback (most recent call last): 2025-12-04T12:25:16.9652810Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9653017Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9653227Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9653232Z 2025-12-04T12:25:16.9653447Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9654418Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9654487Z 2025-12-04T12:25:16.9654752Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9654981Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9655092Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9655206Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9655553Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9655771Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9655867Z graph_break [] 2025-12-04T12:25:16.9656093Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9656828Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9657039Z warnings.warn( 2025-12-04T12:25:16.9657254Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9657364Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9657489Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9657748Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9658086Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9658196Z graph_break [] 2025-12-04T12:25:16.9658409Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9659746Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9659848Z warnings.warn( 2025-12-04T12:25:16.9659997Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9660584Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9660704Z Traceback (most recent call last): 2025-12-04T12:25:16.9661176Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9661385Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9661598Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9661603Z 2025-12-04T12:25:16.9661827Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9662798Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9662808Z 2025-12-04T12:25:16.9663074Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9663303Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9663415Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9663544Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9663889Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9664105Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9664214Z graph_break [] 2025-12-04T12:25:16.9664430Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9665157Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9665271Z warnings.warn( 2025-12-04T12:25:16.9665555Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9665677Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9665791Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9666004Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9666355Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9666452Z graph_break [] 2025-12-04T12:25:16.9666662Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9667397Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9667493Z warnings.warn( 2025-12-04T12:25:16.9667714Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9667824Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9667940Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9668172Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9668507Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9668603Z graph_break [] 2025-12-04T12:25:16.9668880Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9669608Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9669717Z warnings.warn( 2025-12-04T12:25:16.9670539Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-5ce17ea59d8fd893.xml - 2025-12-04T12:25:16.9670737Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9671857Z FAILED [0.1827s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9671868Z 2025-12-04T12:25:16.9672085Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9673068Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9673074Z 2025-12-04T12:25:16.9673341Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9673523Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9673737Z ================== 1 failed, 204 deselected, 2 rerun in 5.37s ================== 2025-12-04T12:25:16.9673840Z Got exit code 1 2025-12-04T12:25:16.9673972Z Retrying single test... 2025-12-04T12:25:16.9674624Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4874d0c887fabf8b.xml 2025-12-04T12:25:16.9674786Z ============================= test session starts ============================== 2025-12-04T12:25:16.9675149Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9675259Z cachedir: .pytest_cache 2025-12-04T12:25:16.9675794Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9675919Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9676028Z configfile: pytest.ini 2025-12-04T12:25:16.9676632Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9676863Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.9677979Z stepcurrent: skipping 151 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9678107Z Running 1 items in this shard 2025-12-04T12:25:16.9678112Z 2025-12-04T12:25:16.9679040Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6538s] [100%] 2025-12-04T12:25:16.9679978Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1657s] [100%] 2025-12-04T12:25:16.9680820Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1622s] [100%] 2025-12-04T12:25:16.9680830Z 2025-12-04T12:25:16.9680979Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9681555Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9681716Z Traceback (most recent call last): 2025-12-04T12:25:16.9682203Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9682402Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9682627Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9682632Z 2025-12-04T12:25:16.9682848Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9683845Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9683856Z 2025-12-04T12:25:16.9684139Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9684357Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9684485Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9684601Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9684943Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9685174Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9685271Z graph_break [] 2025-12-04T12:25:16.9685485Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9686234Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9686338Z warnings.warn( 2025-12-04T12:25:16.9686922Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9687042Z Traceback (most recent call last): 2025-12-04T12:25:16.9687516Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9687724Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9687933Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9687938Z 2025-12-04T12:25:16.9688162Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9689125Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9689130Z 2025-12-04T12:25:16.9689455Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9689679Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9689792Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9689908Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9690258Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9690477Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9690585Z graph_break [] 2025-12-04T12:25:16.9690799Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9691540Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9691648Z warnings.warn( 2025-12-04T12:25:16.9691864Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9691976Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9692090Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9692310Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9692658Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9692783Z graph_break [] 2025-12-04T12:25:16.9692997Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9693735Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9693836Z warnings.warn( 2025-12-04T12:25:16.9693977Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9694590Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9694718Z Traceback (most recent call last): 2025-12-04T12:25:16.9695199Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9695396Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9695609Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9695614Z 2025-12-04T12:25:16.9695836Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9697036Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9697043Z 2025-12-04T12:25:16.9697324Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9697539Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9697655Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9697783Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9698121Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9698352Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9698449Z graph_break [] 2025-12-04T12:25:16.9698661Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9699403Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9699503Z warnings.warn( 2025-12-04T12:25:16.9699716Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9699837Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9699949Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9700178Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9700620Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9700715Z graph_break [] 2025-12-04T12:25:16.9700941Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9701673Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9701774Z warnings.warn( 2025-12-04T12:25:16.9702001Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9702109Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9702223Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9702452Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9702788Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9702905Z graph_break [] 2025-12-04T12:25:16.9703121Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9703853Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9704000Z warnings.warn( 2025-12-04T12:25:16.9704822Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4874d0c887fabf8b.xml - 2025-12-04T12:25:16.9705005Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9706147Z FAILED [0.1622s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9706158Z 2025-12-04T12:25:16.9706376Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9707344Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9707350Z 2025-12-04T12:25:16.9707617Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9707808Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9708005Z ================== 1 failed, 204 deselected, 2 rerun in 5.04s ================== 2025-12-04T12:25:16.9708101Z Got exit code 1 2025-12-04T12:25:16.9708990Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9709399Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.9710066Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a28d30f3a161bea7.xml 2025-12-04T12:25:16.9710229Z ============================= test session starts ============================== 2025-12-04T12:25:16.9710576Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9710698Z cachedir: .pytest_cache 2025-12-04T12:25:16.9711215Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9711349Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9711455Z configfile: pytest.ini 2025-12-04T12:25:16.9712050Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9712369Z collecting ... collected 380 items / 152 deselected / 228 selected 2025-12-04T12:25:16.9712517Z stepcurrent: skipping 152 already run items. 2025-12-04T12:25:16.9712635Z Running 53 items in this shard 2025-12-04T12:25:16.9712640Z 2025-12-04T12:25:16.9713695Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_True SKIPPED [0.0041s] (Skip non-critical tests to save resources.) [ 1%] 2025-12-04T12:25:16.9714617Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6388s] [ 3%] 2025-12-04T12:25:16.9715559Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1628s] [ 3%] 2025-12-04T12:25:16.9716406Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1593s] [ 3%] 2025-12-04T12:25:16.9716412Z 2025-12-04T12:25:16.9716562Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9717164Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9717288Z Traceback (most recent call last): 2025-12-04T12:25:16.9717775Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9717972Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9718192Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9718198Z 2025-12-04T12:25:16.9718441Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9719411Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9719416Z 2025-12-04T12:25:16.9719696Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9719916Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9720036Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9720150Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9720489Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9720714Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9720810Z graph_break [] 2025-12-04T12:25:16.9721026Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9721773Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9721875Z warnings.warn( 2025-12-04T12:25:16.9722459Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9722582Z Traceback (most recent call last): 2025-12-04T12:25:16.9723052Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9723260Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9723473Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9723478Z 2025-12-04T12:25:16.9723699Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9724663Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9724728Z 2025-12-04T12:25:16.9724995Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9725221Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9725334Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9725458Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9725796Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9726015Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9726123Z graph_break [] 2025-12-04T12:25:16.9726339Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9727073Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9727188Z warnings.warn( 2025-12-04T12:25:16.9727401Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9727523Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9727638Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9727886Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9728234Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9728332Z graph_break [] 2025-12-04T12:25:16.9728545Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9729286Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9729386Z warnings.warn( 2025-12-04T12:25:16.9729569Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9730148Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9730269Z Traceback (most recent call last): 2025-12-04T12:25:16.9730748Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9730947Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9731156Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9731161Z 2025-12-04T12:25:16.9731387Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9732350Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9732357Z 2025-12-04T12:25:16.9732635Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9732851Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9732967Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9733160Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9733569Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9733891Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9733991Z graph_break [] 2025-12-04T12:25:16.9734209Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9735049Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9735153Z warnings.warn( 2025-12-04T12:25:16.9735371Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9735571Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9735688Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9735922Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9736261Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9736359Z graph_break [] 2025-12-04T12:25:16.9736590Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9737427Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9737580Z warnings.warn( 2025-12-04T12:25:16.9737825Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9737941Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9738073Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9738296Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9738655Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9738756Z graph_break [] 2025-12-04T12:25:16.9738972Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9739766Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9739870Z warnings.warn( 2025-12-04T12:25:16.9740692Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a28d30f3a161bea7.xml - 2025-12-04T12:25:16.9740880Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9742008Z FAILED [0.1593s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9742019Z 2025-12-04T12:25:16.9742257Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9743222Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9743227Z 2025-12-04T12:25:16.9743512Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9743691Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9743911Z ============ 1 failed, 1 skipped, 152 deselected, 2 rerun in 5.02s ============= 2025-12-04T12:25:16.9744025Z Got exit code 1 2025-12-04T12:25:16.9744133Z Retrying single test... 2025-12-04T12:25:16.9744777Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-23cf59059171b3d3.xml 2025-12-04T12:25:16.9744959Z ============================= test session starts ============================== 2025-12-04T12:25:16.9745310Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9745436Z cachedir: .pytest_cache 2025-12-04T12:25:16.9745966Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9746095Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9746216Z configfile: pytest.ini 2025-12-04T12:25:16.9746810Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9747050Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.9748107Z stepcurrent: skipping 153 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9748281Z Running 1 items in this shard 2025-12-04T12:25:16.9748287Z 2025-12-04T12:25:16.9749231Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6551s] [100%] 2025-12-04T12:25:16.9750153Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1639s] [100%] 2025-12-04T12:25:16.9751011Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1612s] [100%] 2025-12-04T12:25:16.9751016Z 2025-12-04T12:25:16.9751159Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9751747Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9751870Z Traceback (most recent call last): 2025-12-04T12:25:16.9752372Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9752587Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9752797Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9752802Z 2025-12-04T12:25:16.9753013Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9754020Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9754031Z 2025-12-04T12:25:16.9754296Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9754525Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9754635Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9754749Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9755101Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9755319Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9755427Z graph_break [] 2025-12-04T12:25:16.9755641Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9756374Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9756491Z warnings.warn( 2025-12-04T12:25:16.9757069Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9757192Z Traceback (most recent call last): 2025-12-04T12:25:16.9757675Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9757873Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9758100Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9758105Z 2025-12-04T12:25:16.9758318Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9759278Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9759296Z 2025-12-04T12:25:16.9759560Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9759924Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9760050Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9760163Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9760502Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9760737Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9760834Z graph_break [] 2025-12-04T12:25:16.9761049Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9761795Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9761895Z warnings.warn( 2025-12-04T12:25:16.9762124Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9762234Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9762354Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9762587Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9762931Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9763030Z graph_break [] 2025-12-04T12:25:16.9763291Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9764023Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9764136Z warnings.warn( 2025-12-04T12:25:16.9764280Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9764884Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9765023Z Traceback (most recent call last): 2025-12-04T12:25:16.9765499Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9765707Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9765918Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9765923Z 2025-12-04T12:25:16.9766137Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9767118Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9767123Z 2025-12-04T12:25:16.9767411Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9767641Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9767754Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9767876Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9768229Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9768449Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9768546Z graph_break [] 2025-12-04T12:25:16.9768775Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9769509Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9769621Z warnings.warn( 2025-12-04T12:25:16.9769836Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9769945Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9770073Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9770299Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9770666Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9770804Z graph_break [] 2025-12-04T12:25:16.9771017Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9771758Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9771858Z warnings.warn( 2025-12-04T12:25:16.9772071Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9772195Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9772309Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9772526Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9772875Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9772975Z graph_break [] 2025-12-04T12:25:16.9773204Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9773929Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9774029Z warnings.warn( 2025-12-04T12:25:16.9774891Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-23cf59059171b3d3.xml - 2025-12-04T12:25:16.9775063Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9776172Z FAILED [0.1612s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9776179Z 2025-12-04T12:25:16.9776423Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9777547Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9777555Z 2025-12-04T12:25:16.9777834Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9778018Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9778232Z ================== 1 failed, 204 deselected, 2 rerun in 5.03s ================== 2025-12-04T12:25:16.9778330Z Got exit code 1 2025-12-04T12:25:16.9778439Z Retrying single test... 2025-12-04T12:25:16.9779095Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-14e4b4194810001e.xml 2025-12-04T12:25:16.9779261Z ============================= test session starts ============================== 2025-12-04T12:25:16.9779613Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9779736Z cachedir: .pytest_cache 2025-12-04T12:25:16.9780257Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9780393Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9780502Z configfile: pytest.ini 2025-12-04T12:25:16.9781093Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9781334Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.9782393Z stepcurrent: skipping 153 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9782519Z Running 1 items in this shard 2025-12-04T12:25:16.9782590Z 2025-12-04T12:25:16.9783516Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6760s] [100%] 2025-12-04T12:25:16.9784441Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1631s] [100%] 2025-12-04T12:25:16.9785295Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1622s] [100%] 2025-12-04T12:25:16.9785301Z 2025-12-04T12:25:16.9785441Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9786030Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9786158Z Traceback (most recent call last): 2025-12-04T12:25:16.9786628Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9786843Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9787084Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9787090Z 2025-12-04T12:25:16.9787316Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9788279Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9788284Z 2025-12-04T12:25:16.9788562Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9788809Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9788926Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9789052Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9789417Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9789638Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9789748Z graph_break [] 2025-12-04T12:25:16.9789966Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9790717Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9790817Z warnings.warn( 2025-12-04T12:25:16.9791392Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9791528Z Traceback (most recent call last): 2025-12-04T12:25:16.9791994Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9792195Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9792418Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9792423Z 2025-12-04T12:25:16.9792634Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9793612Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9793617Z 2025-12-04T12:25:16.9793880Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9794098Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9794222Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9794337Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9794780Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9795000Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9795101Z graph_break [] 2025-12-04T12:25:16.9795328Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9796245Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9796355Z warnings.warn( 2025-12-04T12:25:16.9796587Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9796697Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9796826Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9797044Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9797386Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9797503Z graph_break [] 2025-12-04T12:25:16.9797721Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9798454Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9798631Z warnings.warn( 2025-12-04T12:25:16.9798778Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9799364Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9799486Z Traceback (most recent call last): 2025-12-04T12:25:16.9799957Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9800223Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9800445Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9800450Z 2025-12-04T12:25:16.9800679Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9801655Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9801661Z 2025-12-04T12:25:16.9801927Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9802158Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9802272Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9802387Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9802742Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9802967Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9803080Z graph_break [] 2025-12-04T12:25:16.9803295Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9804028Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9804149Z warnings.warn( 2025-12-04T12:25:16.9804368Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9804482Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9804612Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9804832Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9805189Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9805287Z graph_break [] 2025-12-04T12:25:16.9805504Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9806329Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9806431Z warnings.warn( 2025-12-04T12:25:16.9806648Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9806776Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9806893Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9807126Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9807460Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9807557Z graph_break [] 2025-12-04T12:25:16.9807785Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9808516Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9808620Z warnings.warn( 2025-12-04T12:25:16.9809447Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-14e4b4194810001e.xml - 2025-12-04T12:25:16.9809617Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9810765Z FAILED [0.1622s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9810772Z 2025-12-04T12:25:16.9810990Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9811997Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9812007Z 2025-12-04T12:25:16.9812274Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9812451Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9812659Z ================== 1 failed, 204 deselected, 2 rerun in 5.06s ================== 2025-12-04T12:25:16.9812758Z Got exit code 1 2025-12-04T12:25:16.9813649Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9814058Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.9814700Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4b58d6ae0af928ee.xml 2025-12-04T12:25:16.9814878Z ============================= test session starts ============================== 2025-12-04T12:25:16.9815228Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9815350Z cachedir: .pytest_cache 2025-12-04T12:25:16.9815869Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9815997Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9816118Z configfile: pytest.ini 2025-12-04T12:25:16.9816712Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9817005Z collecting ... collected 380 items / 154 deselected / 226 selected 2025-12-04T12:25:16.9817170Z stepcurrent: skipping 154 already run items. 2025-12-04T12:25:16.9817284Z Running 51 items in this shard 2025-12-04T12:25:16.9817290Z 2025-12-04T12:25:16.9818367Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_True SKIPPED [0.0040s] (Skip non-critical tests to save resources.) [ 1%] 2025-12-04T12:25:16.9819480Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_True_initial_xblock_1_add_1dim_False SKIPPED [0.0029s] (Skip non-critical tests to save resources.) [ 3%] 2025-12-04T12:25:16.9820526Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_True_initial_xblock_1_add_1dim_True SKIPPED [0.0035s] (Skip non-critical tests to save resources.) [ 5%] 2025-12-04T12:25:16.9821565Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_1_add_1dim_True SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 7%] 2025-12-04T12:25:16.9822491Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6333s] [ 9%] 2025-12-04T12:25:16.9823465Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1633s] [ 9%] 2025-12-04T12:25:16.9824311Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1610s] [ 9%] 2025-12-04T12:25:16.9824316Z 2025-12-04T12:25:16.9824465Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9825063Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9825196Z Traceback (most recent call last): 2025-12-04T12:25:16.9825674Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9825874Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9826098Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9826103Z 2025-12-04T12:25:16.9826317Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9827275Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9827292Z 2025-12-04T12:25:16.9827557Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9827779Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9827906Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9828025Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9828365Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9828595Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9828694Z graph_break [] 2025-12-04T12:25:16.9828924Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9829660Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9829761Z warnings.warn( 2025-12-04T12:25:16.9830349Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9830470Z Traceback (most recent call last): 2025-12-04T12:25:16.9830944Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9831211Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9831425Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9831430Z 2025-12-04T12:25:16.9831656Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9832620Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9832625Z 2025-12-04T12:25:16.9832903Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9833118Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9833232Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9833361Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9833701Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9833921Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9834031Z graph_break [] 2025-12-04T12:25:16.9834247Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9835007Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9835121Z warnings.warn( 2025-12-04T12:25:16.9835334Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9835460Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9835577Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9835795Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9836175Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9836278Z graph_break [] 2025-12-04T12:25:16.9836492Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9837236Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9837338Z warnings.warn( 2025-12-04T12:25:16.9837497Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9838071Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9838194Z Traceback (most recent call last): 2025-12-04T12:25:16.9838678Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9838876Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9839103Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9839112Z 2025-12-04T12:25:16.9839326Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9840293Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9840300Z 2025-12-04T12:25:16.9840579Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9840800Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9840926Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9841039Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9841380Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9841613Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9841712Z graph_break [] 2025-12-04T12:25:16.9841995Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9842735Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9842833Z warnings.warn( 2025-12-04T12:25:16.9843062Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9843171Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9843284Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9843515Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9843853Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9843949Z graph_break [] 2025-12-04T12:25:16.9844176Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9844903Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9845018Z warnings.warn( 2025-12-04T12:25:16.9845230Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9845340Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9845515Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9845736Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9846070Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9846179Z graph_break [] 2025-12-04T12:25:16.9846392Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9847161Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9847261Z warnings.warn( 2025-12-04T12:25:16.9848114Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4b58d6ae0af928ee.xml - 2025-12-04T12:25:16.9848299Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9849396Z FAILED [0.1610s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9849402Z 2025-12-04T12:25:16.9849629Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9850596Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9850601Z 2025-12-04T12:25:16.9850875Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9851071Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9851290Z ============ 1 failed, 4 skipped, 154 deselected, 2 rerun in 5.03s ============= 2025-12-04T12:25:16.9851403Z Got exit code 1 2025-12-04T12:25:16.9851510Z Retrying single test... 2025-12-04T12:25:16.9852151Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9605f50d5f418e61.xml 2025-12-04T12:25:16.9852326Z ============================= test session starts ============================== 2025-12-04T12:25:16.9852669Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9852779Z cachedir: .pytest_cache 2025-12-04T12:25:16.9853314Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9853499Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9853618Z configfile: pytest.ini 2025-12-04T12:25:16.9854209Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9854436Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.9855502Z stepcurrent: skipping 158 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9855617Z Running 1 items in this shard 2025-12-04T12:25:16.9855621Z 2025-12-04T12:25:16.9856562Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6595s] [100%] 2025-12-04T12:25:16.9857562Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1636s] [100%] 2025-12-04T12:25:16.9858464Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1605s] [100%] 2025-12-04T12:25:16.9858471Z 2025-12-04T12:25:16.9858621Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9859200Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9859338Z Traceback (most recent call last): 2025-12-04T12:25:16.9859811Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9860055Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9860274Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9860279Z 2025-12-04T12:25:16.9860493Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9861470Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9861475Z 2025-12-04T12:25:16.9861742Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9861975Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9862089Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9862206Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9862567Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9862812Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9862914Z graph_break [] 2025-12-04T12:25:16.9863150Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9863887Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9864005Z warnings.warn( 2025-12-04T12:25:16.9864581Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9864706Z Traceback (most recent call last): 2025-12-04T12:25:16.9865193Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9865394Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9865607Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9865614Z 2025-12-04T12:25:16.9865902Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9866864Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9866869Z 2025-12-04T12:25:16.9867157Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9867376Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9867490Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9867622Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9867965Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9868198Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9868298Z graph_break [] 2025-12-04T12:25:16.9868514Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9869270Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9869374Z warnings.warn( 2025-12-04T12:25:16.9869620Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9869746Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9869862Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9870098Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9870437Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9870536Z graph_break [] 2025-12-04T12:25:16.9870764Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9871522Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9871631Z warnings.warn( 2025-12-04T12:25:16.9871793Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9872368Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9872505Z Traceback (most recent call last): 2025-12-04T12:25:16.9872977Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9873171Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9873392Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9873398Z 2025-12-04T12:25:16.9873609Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9874579Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9874589Z 2025-12-04T12:25:16.9874854Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9875067Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9875191Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9875302Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9875640Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9875866Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9875964Z graph_break [] 2025-12-04T12:25:16.9876192Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9876925Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9877083Z warnings.warn( 2025-12-04T12:25:16.9877309Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9877417Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9877530Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9877762Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9878101Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9878212Z graph_break [] 2025-12-04T12:25:16.9878424Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9879145Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9879263Z warnings.warn( 2025-12-04T12:25:16.9879476Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9879592Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9879718Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9879940Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9880290Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9880417Z graph_break [] 2025-12-04T12:25:16.9880630Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9881362Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9881461Z warnings.warn( 2025-12-04T12:25:16.9882313Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9605f50d5f418e61.xml - 2025-12-04T12:25:16.9882500Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9883594Z FAILED [0.1605s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9883599Z 2025-12-04T12:25:16.9883830Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9884790Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9884795Z 2025-12-04T12:25:16.9885073Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9885254Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9885452Z ================== 1 failed, 204 deselected, 2 rerun in 5.04s ================== 2025-12-04T12:25:16.9885566Z Got exit code 1 2025-12-04T12:25:16.9885675Z Retrying single test... 2025-12-04T12:25:16.9886333Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2668b63c8efdb7a5.xml 2025-12-04T12:25:16.9886499Z ============================= test session starts ============================== 2025-12-04T12:25:16.9886848Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9886970Z cachedir: .pytest_cache 2025-12-04T12:25:16.9887496Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9887621Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9887742Z configfile: pytest.ini 2025-12-04T12:25:16.9888338Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9888659Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.9889707Z stepcurrent: skipping 158 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9889824Z Running 1 items in this shard 2025-12-04T12:25:16.9889829Z 2025-12-04T12:25:16.9890776Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6822s] [100%] 2025-12-04T12:25:16.9891704Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1626s] [100%] 2025-12-04T12:25:16.9892565Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1607s] [100%] 2025-12-04T12:25:16.9892573Z 2025-12-04T12:25:16.9892714Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9893336Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9893460Z Traceback (most recent call last): 2025-12-04T12:25:16.9893933Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9894141Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9894357Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9894362Z 2025-12-04T12:25:16.9894607Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9895595Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9895601Z 2025-12-04T12:25:16.9895865Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9896257Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9896372Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9896488Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9896905Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9897131Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9897249Z graph_break [] 2025-12-04T12:25:16.9897469Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9898204Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9898325Z warnings.warn( 2025-12-04T12:25:16.9898902Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9899027Z Traceback (most recent call last): 2025-12-04T12:25:16.9899517Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9899714Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9899941Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9899946Z 2025-12-04T12:25:16.9900162Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9901158Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9901285Z 2025-12-04T12:25:16.9901555Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9901772Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9901901Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9902019Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9902360Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9902592Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9902690Z graph_break [] 2025-12-04T12:25:16.9902905Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9903660Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9903765Z warnings.warn( 2025-12-04T12:25:16.9903992Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9904102Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9904218Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9904451Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9904855Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9904955Z graph_break [] 2025-12-04T12:25:16.9905179Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9905917Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9906030Z warnings.warn( 2025-12-04T12:25:16.9906227Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9906810Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:16.9906945Z Traceback (most recent call last): 2025-12-04T12:25:16.9907416Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9907628Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9907839Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9907844Z 2025-12-04T12:25:16.9908056Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9909043Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9909048Z 2025-12-04T12:25:16.9909315Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9909548Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9909666Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9909783Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9910137Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9910360Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9910460Z graph_break [] 2025-12-04T12:25:16.9910689Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9911425Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9911537Z warnings.warn( 2025-12-04T12:25:16.9911755Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9911930Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9912056Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9912274Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9912614Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9912726Z graph_break [] 2025-12-04T12:25:16.9912943Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9913688Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9913789Z warnings.warn( 2025-12-04T12:25:16.9914003Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9914126Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9914239Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9914460Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9914819Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9914919Z graph_break [] 2025-12-04T12:25:16.9915146Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9915901Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9916001Z warnings.warn( 2025-12-04T12:25:16.9916839Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2668b63c8efdb7a5.xml - 2025-12-04T12:25:16.9917012Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9918162Z FAILED [0.1607s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9918172Z 2025-12-04T12:25:16.9918390Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9919358Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9919363Z 2025-12-04T12:25:16.9919645Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9919826Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9920040Z ================== 1 failed, 204 deselected, 2 rerun in 5.06s ================== 2025-12-04T12:25:16.9920140Z Got exit code 1 2025-12-04T12:25:16.9921023Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:16.9921455Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:16.9922093Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-8d527629568e40ed.xml 2025-12-04T12:25:16.9922275Z ============================= test session starts ============================== 2025-12-04T12:25:16.9922625Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9922737Z cachedir: .pytest_cache 2025-12-04T12:25:16.9923278Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9923405Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9923517Z configfile: pytest.ini 2025-12-04T12:25:16.9924124Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9924414Z collecting ... collected 380 items / 159 deselected / 221 selected 2025-12-04T12:25:16.9924582Z stepcurrent: skipping 159 already run items. 2025-12-04T12:25:16.9924700Z Running 46 items in this shard 2025-12-04T12:25:16.9924705Z 2025-12-04T12:25:16.9925763Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_True SKIPPED [0.0041s] (Skip non-critical tests to save resources.) [ 2%] 2025-12-04T12:25:16.9926817Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_True_initial_xblock_2_add_1dim_False SKIPPED [0.0032s] (Skip non-critical tests to save resources.) [ 4%] 2025-12-04T12:25:16.9927855Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_True_initial_xblock_2_add_1dim_True SKIPPED [0.0038s] (Skip non-critical tests to save resources.) [ 6%] 2025-12-04T12:25:16.9928799Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.7005s] [ 8%] 2025-12-04T12:25:16.9929762Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1648s] [ 8%] 2025-12-04T12:25:16.9930627Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1642s] [ 8%] 2025-12-04T12:25:16.9930632Z 2025-12-04T12:25:16.9930806Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9931400Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9931525Z Traceback (most recent call last): 2025-12-04T12:25:16.9931999Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9932216Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9932433Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9932438Z 2025-12-04T12:25:16.9932652Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9933635Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9933641Z 2025-12-04T12:25:16.9933910Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9934148Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9934262Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9934380Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9934737Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9934959Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9935073Z graph_break [] 2025-12-04T12:25:16.9935289Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9936024Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9936145Z warnings.warn( 2025-12-04T12:25:16.9936720Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9936994Z Traceback (most recent call last): 2025-12-04T12:25:16.9937488Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9937687Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9937910Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9937918Z 2025-12-04T12:25:16.9938130Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9939098Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9939116Z 2025-12-04T12:25:16.9939381Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9939601Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9939731Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9939847Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9940187Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9940419Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9940515Z graph_break [] 2025-12-04T12:25:16.9940785Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9941537Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9941638Z warnings.warn( 2025-12-04T12:25:16.9941865Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9941975Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9942195Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9942431Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9942775Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9942873Z graph_break [] 2025-12-04T12:25:16.9943102Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9943839Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9943956Z warnings.warn( 2025-12-04T12:25:16.9944102Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9944684Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9944821Z Traceback (most recent call last): 2025-12-04T12:25:16.9945296Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9945512Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9945727Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9945733Z 2025-12-04T12:25:16.9945947Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9946924Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9946930Z 2025-12-04T12:25:16.9947197Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9947425Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9947537Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9947651Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9948000Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9948276Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9948372Z graph_break [] 2025-12-04T12:25:16.9948597Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9949335Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9949449Z warnings.warn( 2025-12-04T12:25:16.9949663Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9949771Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9949896Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9950117Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9950457Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9950576Z graph_break [] 2025-12-04T12:25:16.9950789Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9951528Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9951627Z warnings.warn( 2025-12-04T12:25:16.9951875Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9951998Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9952115Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9952333Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9952684Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9952781Z graph_break [] 2025-12-04T12:25:16.9953030Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9953798Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9953904Z warnings.warn( 2025-12-04T12:25:16.9954740Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-8d527629568e40ed.xml - 2025-12-04T12:25:16.9954910Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9956005Z FAILED [0.1642s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9956026Z 2025-12-04T12:25:16.9956244Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9957216Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9957226Z 2025-12-04T12:25:16.9957507Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9957690Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9957927Z ============ 1 failed, 3 skipped, 159 deselected, 2 rerun in 5.10s ============= 2025-12-04T12:25:16.9958025Z Got exit code 1 2025-12-04T12:25:16.9958135Z Retrying single test... 2025-12-04T12:25:16.9958788Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a273900f3029c317.xml 2025-12-04T12:25:16.9958953Z ============================= test session starts ============================== 2025-12-04T12:25:16.9959305Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9959496Z cachedir: .pytest_cache 2025-12-04T12:25:16.9960016Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9960156Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9960263Z configfile: pytest.ini 2025-12-04T12:25:16.9960855Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9961095Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.9962148Z stepcurrent: skipping 162 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9962278Z Running 1 items in this shard 2025-12-04T12:25:16.9962283Z 2025-12-04T12:25:16.9963212Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6432s] [100%] 2025-12-04T12:25:16.9964178Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1626s] [100%] 2025-12-04T12:25:16.9965040Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1600s] [100%] 2025-12-04T12:25:16.9965045Z 2025-12-04T12:25:16.9965186Z ==================================== RERUNS ==================================== 2025-12-04T12:25:16.9965774Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9965925Z Traceback (most recent call last): 2025-12-04T12:25:16.9966399Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9966611Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9966823Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9966828Z 2025-12-04T12:25:16.9967056Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9968023Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9968028Z 2025-12-04T12:25:16.9968304Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9968523Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9968637Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9968764Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9969109Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9969327Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9969435Z graph_break [] 2025-12-04T12:25:16.9969648Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9970393Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9970493Z warnings.warn( 2025-12-04T12:25:16.9971068Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9971202Z Traceback (most recent call last): 2025-12-04T12:25:16.9971680Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9971939Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9972161Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9972167Z 2025-12-04T12:25:16.9972379Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9973358Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9973363Z 2025-12-04T12:25:16.9973628Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9973843Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9973973Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9974091Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9974450Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9974673Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9974774Z graph_break [] 2025-12-04T12:25:16.9975001Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9975770Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9975873Z warnings.warn( 2025-12-04T12:25:16.9976100Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9976211Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9976339Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9976558Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9977031Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9977149Z graph_break [] 2025-12-04T12:25:16.9977369Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9978096Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9978211Z warnings.warn( 2025-12-04T12:25:16.9978359Z =================================== FAILURES =================================== 2025-12-04T12:25:16.9978949Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:16.9979071Z Traceback (most recent call last): 2025-12-04T12:25:16.9979543Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:16.9979752Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:16.9979965Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9979971Z 2025-12-04T12:25:16.9980189Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9981164Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9981169Z 2025-12-04T12:25:16.9981444Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9981672Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9981782Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9981900Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9982258Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9982477Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9982595Z graph_break [] 2025-12-04T12:25:16.9982809Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9986329Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9986473Z warnings.warn( 2025-12-04T12:25:16.9986704Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9986817Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9986947Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9987170Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9987511Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9987625Z graph_break [] 2025-12-04T12:25:16.9987844Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9988608Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9988717Z warnings.warn( 2025-12-04T12:25:16.9988933Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:16.9989094Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:16.9989210Z stats [('calls_captured', 10)] 2025-12-04T12:25:16.9989432Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:16.9989785Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:16.9989885Z graph_break [] 2025-12-04T12:25:16.9990101Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:16.9990843Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:16.9991003Z warnings.warn( 2025-12-04T12:25:16.9991971Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a273900f3029c317.xml - 2025-12-04T12:25:16.9992152Z =========================== short test summary info ============================ 2025-12-04T12:25:16.9993262Z FAILED [0.1600s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:16.9993269Z 2025-12-04T12:25:16.9993506Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:16.9994468Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9994474Z 2025-12-04T12:25:16.9994760Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:16.9994944Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:16.9995146Z ================== 1 failed, 204 deselected, 2 rerun in 5.02s ================== 2025-12-04T12:25:16.9995263Z Got exit code 1 2025-12-04T12:25:16.9995372Z Retrying single test... 2025-12-04T12:25:16.9996285Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dcefced86a4c35d8.xml 2025-12-04T12:25:16.9996450Z ============================= test session starts ============================== 2025-12-04T12:25:16.9996805Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:16.9996932Z cachedir: .pytest_cache 2025-12-04T12:25:16.9997462Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:16.9997586Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:16.9997824Z configfile: pytest.ini 2025-12-04T12:25:16.9998421Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:16.9998749Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:16.9999806Z stepcurrent: skipping 162 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:16.9999922Z Running 1 items in this shard 2025-12-04T12:25:16.9999928Z 2025-12-04T12:25:17.0000877Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6480s] [100%] 2025-12-04T12:25:17.0001809Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1615s] [100%] 2025-12-04T12:25:17.0002672Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1603s] [100%] 2025-12-04T12:25:17.0002679Z 2025-12-04T12:25:17.0002817Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0003401Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0003521Z Traceback (most recent call last): 2025-12-04T12:25:17.0003994Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0004242Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0004457Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0004465Z 2025-12-04T12:25:17.0004692Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0005658Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0005664Z 2025-12-04T12:25:17.0005927Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0006160Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0006271Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0006386Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0006740Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0006962Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0007076Z graph_break [] 2025-12-04T12:25:17.0007290Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0008032Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0008146Z warnings.warn( 2025-12-04T12:25:17.0008724Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0008859Z Traceback (most recent call last): 2025-12-04T12:25:17.0009332Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0009527Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0009748Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0009756Z 2025-12-04T12:25:17.0009967Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0011070Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0011091Z 2025-12-04T12:25:17.0011356Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0011572Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0011695Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0011811Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0012151Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0012383Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0012481Z graph_break [] 2025-12-04T12:25:17.0012709Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0013452Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0013554Z warnings.warn( 2025-12-04T12:25:17.0013785Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0013897Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0014013Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0014251Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0014592Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0014703Z graph_break [] 2025-12-04T12:25:17.0014913Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0015680Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0015799Z warnings.warn( 2025-12-04T12:25:17.0015942Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0016516Z _ NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0016649Z Traceback (most recent call last): 2025-12-04T12:25:17.0017192Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0017409Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0017621Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0017627Z 2025-12-04T12:25:17.0017842Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0018825Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0018834Z 2025-12-04T12:25:17.0019102Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0019337Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0019452Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0019569Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0019924Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0020144Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0020241Z graph_break [] 2025-12-04T12:25:17.0020469Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0021207Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0021393Z warnings.warn( 2025-12-04T12:25:17.0021608Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0021718Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0021920Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0022143Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0022478Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0022586Z graph_break [] 2025-12-04T12:25:17.0022799Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0023540Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0023640Z warnings.warn( 2025-12-04T12:25:17.0023856Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0023981Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0024097Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0024317Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0024667Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0024765Z graph_break [] 2025-12-04T12:25:17.0024996Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0025725Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0025826Z warnings.warn( 2025-12-04T12:25:17.0026666Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dcefced86a4c35d8.xml - 2025-12-04T12:25:17.0026868Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0028008Z FAILED [0.1603s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0028014Z 2025-12-04T12:25:17.0028229Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0029189Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0029205Z 2025-12-04T12:25:17.0029469Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0029648Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0029862Z ================== 1 failed, 204 deselected, 2 rerun in 5.02s ================== 2025-12-04T12:25:17.0029964Z Got exit code 1 2025-12-04T12:25:17.0030838Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0031262Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:17.0031898Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e15b12282035b26d.xml 2025-12-04T12:25:17.0032070Z ============================= test session starts ============================== 2025-12-04T12:25:17.0032420Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0032527Z cachedir: .pytest_cache 2025-12-04T12:25:17.0033064Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0033252Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0033359Z configfile: pytest.ini 2025-12-04T12:25:17.0033990Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0034226Z collecting ... collected 380 items / 163 deselected / 217 selected 2025-12-04T12:25:17.0034387Z stepcurrent: skipping 163 already run items. 2025-12-04T12:25:17.0034505Z Running 42 items in this shard 2025-12-04T12:25:17.0034510Z 2025-12-04T12:25:17.0035560Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_True SKIPPED [0.0040s] (Skip non-critical tests to save resources.) [ 2%] 2025-12-04T12:25:17.0036617Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_True SKIPPED [0.0029s] (Skip non-critical tests to save resources.) [ 4%] 2025-12-04T12:25:17.0037661Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_True_initial_xblock_1_add_1dim_False SKIPPED [0.0036s] (Skip non-critical tests to save resources.) [ 7%] 2025-12-04T12:25:17.0038704Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_True_initial_xblock_2_add_1dim_True SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 9%] 2025-12-04T12:25:17.0039625Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6587s] [ 11%] 2025-12-04T12:25:17.0040589Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1636s] [ 11%] 2025-12-04T12:25:17.0041435Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1579s] [ 11%] 2025-12-04T12:25:17.0041441Z 2025-12-04T12:25:17.0041596Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0042171Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0042293Z Traceback (most recent call last): 2025-12-04T12:25:17.0042779Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0042980Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0043207Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0043217Z 2025-12-04T12:25:17.0043432Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0044393Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0044399Z 2025-12-04T12:25:17.0044675Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0052693Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0052901Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0053033Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0053401Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0053634Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0053741Z graph_break [] 2025-12-04T12:25:17.0053967Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0054950Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0055054Z warnings.warn( 2025-12-04T12:25:17.0055631Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0055764Z Traceback (most recent call last): 2025-12-04T12:25:17.0056237Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0056446Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0056661Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0056669Z 2025-12-04T12:25:17.0056983Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0057965Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0057975Z 2025-12-04T12:25:17.0058244Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0058474Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0058585Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0058699Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0059054Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0059273Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0059371Z graph_break [] 2025-12-04T12:25:17.0059637Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0060369Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0060485Z warnings.warn( 2025-12-04T12:25:17.0060702Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0060812Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0060929Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0061147Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0061486Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0061589Z graph_break [] 2025-12-04T12:25:17.0061802Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0062544Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0062643Z warnings.warn( 2025-12-04T12:25:17.0062792Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0063376Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0063498Z Traceback (most recent call last): 2025-12-04T12:25:17.0063977Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0064176Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0064388Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0064395Z 2025-12-04T12:25:17.0064616Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0065582Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0065648Z 2025-12-04T12:25:17.0065926Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0066174Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0066288Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0066409Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0066749Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0066963Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0067064Z graph_break [] 2025-12-04T12:25:17.0067279Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0068049Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0068148Z warnings.warn( 2025-12-04T12:25:17.0068369Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0068482Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0068595Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0068817Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0069156Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0069249Z graph_break [] 2025-12-04T12:25:17.0069464Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0070188Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0070287Z warnings.warn( 2025-12-04T12:25:17.0070506Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0070645Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0070765Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0070989Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0071325Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0071428Z graph_break [] 2025-12-04T12:25:17.0071639Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0072366Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0072471Z warnings.warn( 2025-12-04T12:25:17.0073288Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e15b12282035b26d.xml - 2025-12-04T12:25:17.0073456Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0074568Z FAILED [0.1579s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0074582Z 2025-12-04T12:25:17.0074794Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0075763Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0075769Z 2025-12-04T12:25:17.0076028Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0076219Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0076435Z ============ 1 failed, 4 skipped, 163 deselected, 2 rerun in 5.05s ============= 2025-12-04T12:25:17.0076532Z Got exit code 1 2025-12-04T12:25:17.0076725Z Retrying single test... 2025-12-04T12:25:17.0077369Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e08a7bc39929d142.xml 2025-12-04T12:25:17.0077563Z ============================= test session starts ============================== 2025-12-04T12:25:17.0077921Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0078027Z cachedir: .pytest_cache 2025-12-04T12:25:17.0078556Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0078677Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0078783Z configfile: pytest.ini 2025-12-04T12:25:17.0079517Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0079756Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.0080999Z stepcurrent: skipping 167 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0081117Z Running 1 items in this shard 2025-12-04T12:25:17.0081122Z 2025-12-04T12:25:17.0082050Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6469s] [100%] 2025-12-04T12:25:17.0082976Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1643s] [100%] 2025-12-04T12:25:17.0083865Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1598s] [100%] 2025-12-04T12:25:17.0083876Z 2025-12-04T12:25:17.0084025Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0084599Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0084720Z Traceback (most recent call last): 2025-12-04T12:25:17.0085199Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0085396Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0085615Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0085620Z 2025-12-04T12:25:17.0085833Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0086796Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0086817Z 2025-12-04T12:25:17.0087086Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0087306Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0087426Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0087539Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0087876Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0088105Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0088202Z graph_break [] 2025-12-04T12:25:17.0088413Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0089157Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0089325Z warnings.warn( 2025-12-04T12:25:17.0089901Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0090059Z Traceback (most recent call last): 2025-12-04T12:25:17.0090527Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0090726Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0090935Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0090941Z 2025-12-04T12:25:17.0091158Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0092119Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0092128Z 2025-12-04T12:25:17.0092391Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0092617Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0092731Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0092853Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0093194Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0093410Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0093513Z graph_break [] 2025-12-04T12:25:17.0093726Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0094445Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0094587Z warnings.warn( 2025-12-04T12:25:17.0094803Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0094925Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0095038Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0095255Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0095602Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0095698Z graph_break [] 2025-12-04T12:25:17.0095906Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0096822Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0097003Z warnings.warn( 2025-12-04T12:25:17.0097155Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0097737Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0097860Z Traceback (most recent call last): 2025-12-04T12:25:17.0098339Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0098541Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0098749Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0098767Z 2025-12-04T12:25:17.0098979Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0099939Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0099946Z 2025-12-04T12:25:17.0100216Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0100434Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0100674Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0100789Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0101175Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0101403Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0101502Z graph_break [] 2025-12-04T12:25:17.0101712Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0102458Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0102557Z warnings.warn( 2025-12-04T12:25:17.0102776Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0102882Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0103001Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0103231Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0103565Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0103660Z graph_break [] 2025-12-04T12:25:17.0103885Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0104607Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0104703Z warnings.warn( 2025-12-04T12:25:17.0104926Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0105034Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0105155Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0105373Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0105749Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0105857Z graph_break [] 2025-12-04T12:25:17.0106069Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0106797Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0106908Z warnings.warn( 2025-12-04T12:25:17.0107732Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e08a7bc39929d142.xml - 2025-12-04T12:25:17.0107906Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0109003Z FAILED [0.1598s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0109013Z 2025-12-04T12:25:17.0109230Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0110196Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0110202Z 2025-12-04T12:25:17.0110466Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0110649Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0110849Z ================== 1 failed, 204 deselected, 2 rerun in 5.03s ================== 2025-12-04T12:25:17.0110956Z Got exit code 1 2025-12-04T12:25:17.0111062Z Retrying single test... 2025-12-04T12:25:17.0111713Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-13b16b357405c710.xml 2025-12-04T12:25:17.0111916Z ============================= test session starts ============================== 2025-12-04T12:25:17.0112296Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0112434Z cachedir: .pytest_cache 2025-12-04T12:25:17.0112968Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0113091Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0113207Z configfile: pytest.ini 2025-12-04T12:25:17.0113799Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0114023Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.0115096Z stepcurrent: skipping 167 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0115215Z Running 1 items in this shard 2025-12-04T12:25:17.0115220Z 2025-12-04T12:25:17.0116152Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6806s] [100%] 2025-12-04T12:25:17.0117075Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1667s] [100%] 2025-12-04T12:25:17.0117914Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1610s] [100%] 2025-12-04T12:25:17.0117930Z 2025-12-04T12:25:17.0118102Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0118676Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0118803Z Traceback (most recent call last): 2025-12-04T12:25:17.0119274Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0119473Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0119693Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0119698Z 2025-12-04T12:25:17.0119913Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0120882Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0120887Z 2025-12-04T12:25:17.0121157Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0121378Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0121502Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0121617Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0121969Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0122189Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0122284Z graph_break [] 2025-12-04T12:25:17.0122508Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0123245Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0123343Z warnings.warn( 2025-12-04T12:25:17.0123931Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0124129Z Traceback (most recent call last): 2025-12-04T12:25:17.0124612Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0124845Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0125057Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0125062Z 2025-12-04T12:25:17.0125284Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0126240Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0126246Z 2025-12-04T12:25:17.0126517Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0126739Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0126849Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0126975Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0127316Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0127544Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0127638Z graph_break [] 2025-12-04T12:25:17.0127847Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0128586Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0128683Z warnings.warn( 2025-12-04T12:25:17.0128890Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0129009Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0129117Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0129423Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0129774Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0129867Z graph_break [] 2025-12-04T12:25:17.0130092Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0130817Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0130911Z warnings.warn( 2025-12-04T12:25:17.0131066Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0131635Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0131766Z Traceback (most recent call last): 2025-12-04T12:25:17.0132236Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0132430Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0132647Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0132653Z 2025-12-04T12:25:17.0132864Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0133831Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0133837Z 2025-12-04T12:25:17.0134099Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0134308Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0134428Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0134542Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0134877Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0135163Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0135258Z graph_break [] 2025-12-04T12:25:17.0135585Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0136320Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0136416Z warnings.warn( 2025-12-04T12:25:17.0136641Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0136747Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0136943Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0137177Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0137512Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0137618Z graph_break [] 2025-12-04T12:25:17.0137833Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0138564Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0138674Z warnings.warn( 2025-12-04T12:25:17.0138884Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0138989Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0139113Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0139327Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0139672Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0139769Z graph_break [] 2025-12-04T12:25:17.0140014Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0140749Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0140853Z warnings.warn( 2025-12-04T12:25:17.0141675Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-13b16b357405c710.xml - 2025-12-04T12:25:17.0141855Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0142954Z FAILED [0.1610s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0142961Z 2025-12-04T12:25:17.0143184Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0144151Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0144160Z 2025-12-04T12:25:17.0144434Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0144613Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0144807Z ================== 1 failed, 204 deselected, 2 rerun in 5.06s ================== 2025-12-04T12:25:17.0144911Z Got exit code 1 2025-12-04T12:25:17.0145781Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0146195Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:17.0146834Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9275f55235cebf56.xml 2025-12-04T12:25:17.0147061Z ============================= test session starts ============================== 2025-12-04T12:25:17.0147449Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0147559Z cachedir: .pytest_cache 2025-12-04T12:25:17.0148080Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0148211Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0148315Z configfile: pytest.ini 2025-12-04T12:25:17.0148917Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0149144Z collecting ... collected 380 items / 168 deselected / 212 selected 2025-12-04T12:25:17.0149288Z stepcurrent: skipping 168 already run items. 2025-12-04T12:25:17.0149414Z Running 37 items in this shard 2025-12-04T12:25:17.0149425Z 2025-12-04T12:25:17.0150343Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [4.6790s] [ 2%] 2025-12-04T12:25:17.0151270Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [0.1680s] [ 2%] 2025-12-04T12:25:17.0152108Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True FAILED [0.1665s] [ 2%] 2025-12-04T12:25:17.0152116Z 2025-12-04T12:25:17.0152252Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0153011Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:17.0153142Z Traceback (most recent call last): 2025-12-04T12:25:17.0153623Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0153823Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0154032Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0154038Z 2025-12-04T12:25:17.0154258Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0155209Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:17.0155214Z 2025-12-04T12:25:17.0155490Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0155711Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0155823Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0155949Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0156289Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0156517Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0156613Z graph_break [] 2025-12-04T12:25:17.0156824Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0159552Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0159722Z return x.grad, w.grad 2025-12-04T12:25:17.0160502Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0160604Z warnings.warn( 2025-12-04T12:25:17.0163309Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0163421Z return x.grad, w.grad 2025-12-04T12:25:17.0163994Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:17.0164124Z Traceback (most recent call last): 2025-12-04T12:25:17.0164589Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0164794Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0165005Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0165011Z 2025-12-04T12:25:17.0165221Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0166252Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:17.0166258Z 2025-12-04T12:25:17.0166533Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0166755Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0166864Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0166980Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0167326Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0167541Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0167638Z graph_break [] 2025-12-04T12:25:17.0167855Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0170563Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0170683Z return x.grad, w.grad 2025-12-04T12:25:17.0171413Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0171520Z warnings.warn( 2025-12-04T12:25:17.0174197Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0174375Z return x.grad, w.grad 2025-12-04T12:25:17.0174619Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0174728Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0174850Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0175068Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0175404Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0175507Z graph_break [] 2025-12-04T12:25:17.0175717Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0178508Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0178617Z return x.grad, w.grad 2025-12-04T12:25:17.0179353Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0179451Z warnings.warn( 2025-12-04T12:25:17.0182179Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0182301Z return x.grad, w.grad 2025-12-04T12:25:17.0182443Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0183020Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:17.0183142Z Traceback (most recent call last): 2025-12-04T12:25:17.0183605Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0183815Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0184025Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0184032Z 2025-12-04T12:25:17.0184262Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0185224Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:17.0185229Z 2025-12-04T12:25:17.0185495Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0185725Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0185839Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0185968Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0186310Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0186532Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0186706Z graph_break [] 2025-12-04T12:25:17.0186924Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0189679Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0189793Z return x.grad, w.grad 2025-12-04T12:25:17.0190523Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0190645Z warnings.warn( 2025-12-04T12:25:17.0193344Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0193467Z return x.grad, w.grad 2025-12-04T12:25:17.0193685Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0193809Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0193956Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0194179Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0194536Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0194636Z graph_break [] 2025-12-04T12:25:17.0194854Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0197723Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0197834Z return x.grad, w.grad 2025-12-04T12:25:17.0198588Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0198689Z warnings.warn( 2025-12-04T12:25:17.0201389Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0201494Z return x.grad, w.grad 2025-12-04T12:25:17.0201728Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0201943Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0202058Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0202295Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0202695Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0202794Z graph_break [] 2025-12-04T12:25:17.0203026Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0203763Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0203872Z warnings.warn( 2025-12-04T12:25:17.0206580Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0206702Z return x.grad, w.grad 2025-12-04T12:25:17.0207522Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9275f55235cebf56.xml - 2025-12-04T12:25:17.0207694Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0208848Z FAILED [0.1665s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0208855Z 2025-12-04T12:25:17.0209077Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0210050Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:17.0210056Z 2025-12-04T12:25:17.0210321Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0210498Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0210708Z ================== 1 failed, 168 deselected, 2 rerun in 5.07s ================== 2025-12-04T12:25:17.0210805Z Got exit code 1 2025-12-04T12:25:17.0210923Z Retrying single test... 2025-12-04T12:25:17.0211566Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a0875a42ac68f671.xml 2025-12-04T12:25:17.0211730Z ============================= test session starts ============================== 2025-12-04T12:25:17.0212093Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0212199Z cachedir: .pytest_cache 2025-12-04T12:25:17.0212723Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0212855Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0212961Z configfile: pytest.ini 2025-12-04T12:25:17.0213559Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0213784Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.0214833Z stepcurrent: skipping 168 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:17.0215016Z Running 1 items in this shard 2025-12-04T12:25:17.0215021Z 2025-12-04T12:25:17.0215994Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [4.6845s] [100%] 2025-12-04T12:25:17.0216982Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [0.1697s] [100%] 2025-12-04T12:25:17.0217822Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True FAILED [0.1661s] [100%] 2025-12-04T12:25:17.0217829Z 2025-12-04T12:25:17.0217981Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0218561Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:17.0218687Z Traceback (most recent call last): 2025-12-04T12:25:17.0219171Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0219369Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0219580Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0219597Z 2025-12-04T12:25:17.0219813Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0220768Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:17.0220773Z 2025-12-04T12:25:17.0221088Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0221306Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0221431Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0221545Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0221887Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0222115Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0222211Z graph_break [] 2025-12-04T12:25:17.0222428Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0225148Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0225257Z return x.grad, w.grad 2025-12-04T12:25:17.0225998Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0226096Z warnings.warn( 2025-12-04T12:25:17.0228796Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0228959Z return x.grad, w.grad 2025-12-04T12:25:17.0229554Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:17.0229690Z Traceback (most recent call last): 2025-12-04T12:25:17.0230155Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0230363Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0230573Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0230578Z 2025-12-04T12:25:17.0230788Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0231757Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:17.0231766Z 2025-12-04T12:25:17.0232031Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0232260Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0232370Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0232482Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0232836Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0233052Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0233163Z graph_break [] 2025-12-04T12:25:17.0233379Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0236110Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0236235Z return x.grad, w.grad 2025-12-04T12:25:17.0236963Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0237077Z warnings.warn( 2025-12-04T12:25:17.0239790Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0239910Z return x.grad, w.grad 2025-12-04T12:25:17.0240126Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0240234Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0240359Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0240579Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0240919Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0241026Z graph_break [] 2025-12-04T12:25:17.0241239Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0243975Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0244141Z return x.grad, w.grad 2025-12-04T12:25:17.0244879Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0244979Z warnings.warn( 2025-12-04T12:25:17.0247666Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0247784Z return x.grad, w.grad 2025-12-04T12:25:17.0247929Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0248511Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:17.0248633Z Traceback (most recent call last): 2025-12-04T12:25:17.0249133Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0249337Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0249552Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0249557Z 2025-12-04T12:25:17.0249783Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0250743Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:17.0250748Z 2025-12-04T12:25:17.0251028Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0251245Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0251355Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0251482Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0251821Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0252042Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0252150Z graph_break [] 2025-12-04T12:25:17.0252365Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0255076Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0255181Z return x.grad, w.grad 2025-12-04T12:25:17.0255923Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0256080Z warnings.warn( 2025-12-04T12:25:17.0258902Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0259026Z return x.grad, w.grad 2025-12-04T12:25:17.0259338Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0259464Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0259574Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0259804Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0260157Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0260256Z graph_break [] 2025-12-04T12:25:17.0260471Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0263215Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0263325Z return x.grad, w.grad 2025-12-04T12:25:17.0264064Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0264167Z warnings.warn( 2025-12-04T12:25:17.0266871Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0266976Z return x.grad, w.grad 2025-12-04T12:25:17.0267205Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0267316Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0267428Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0267663Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0267998Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0268095Z graph_break [] 2025-12-04T12:25:17.0268321Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0269045Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0269158Z warnings.warn( 2025-12-04T12:25:17.0271879Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0272073Z return x.grad, w.grad 2025-12-04T12:25:17.0272892Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a0875a42ac68f671.xml - 2025-12-04T12:25:17.0273059Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0274175Z FAILED [0.1661s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0274185Z 2025-12-04T12:25:17.0274399Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0275371Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:17.0275377Z 2025-12-04T12:25:17.0275642Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0275822Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0276032Z ================== 1 failed, 204 deselected, 2 rerun in 5.08s ================== 2025-12-04T12:25:17.0276125Z Got exit code 1 2025-12-04T12:25:17.0276243Z Retrying single test... 2025-12-04T12:25:17.0276915Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-56c85a19db7b6fad.xml 2025-12-04T12:25:17.0277083Z ============================= test session starts ============================== 2025-12-04T12:25:17.0277443Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0277552Z cachedir: .pytest_cache 2025-12-04T12:25:17.0278073Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0278209Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0278316Z configfile: pytest.ini 2025-12-04T12:25:17.0278916Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0279143Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.0280192Z stepcurrent: skipping 168 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:17.0280324Z Running 1 items in this shard 2025-12-04T12:25:17.0280329Z 2025-12-04T12:25:17.0281258Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [4.6835s] [100%] 2025-12-04T12:25:17.0282186Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True ('RERUN', {'yellow': True}) [0.1733s] [100%] 2025-12-04T12:25:17.0283020Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True FAILED [0.1698s] [100%] 2025-12-04T12:25:17.0283027Z 2025-12-04T12:25:17.0283181Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0283814Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:17.0283937Z Traceback (most recent call last): 2025-12-04T12:25:17.0284449Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0284645Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0284871Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0284876Z 2025-12-04T12:25:17.0285091Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0286052Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:17.0286060Z 2025-12-04T12:25:17.0286342Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0286563Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0286684Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0286801Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0287139Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0287370Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0287467Z graph_break [] 2025-12-04T12:25:17.0287683Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0290440Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0290556Z return x.grad, w.grad 2025-12-04T12:25:17.0291305Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0291408Z warnings.warn( 2025-12-04T12:25:17.0294119Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0294229Z return x.grad, w.grad 2025-12-04T12:25:17.0294816Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:17.0294936Z Traceback (most recent call last): 2025-12-04T12:25:17.0295407Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0295616Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0295826Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0295832Z 2025-12-04T12:25:17.0296348Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0297404Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:17.0297524Z 2025-12-04T12:25:17.0297793Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0298066Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0298179Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0298296Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0298649Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0298865Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0298980Z graph_break [] 2025-12-04T12:25:17.0299198Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0301906Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0302028Z return x.grad, w.grad 2025-12-04T12:25:17.0302754Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0302867Z warnings.warn( 2025-12-04T12:25:17.0305847Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0305977Z return x.grad, w.grad 2025-12-04T12:25:17.0306200Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0306307Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0306431Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0306648Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0307001Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0307095Z graph_break [] 2025-12-04T12:25:17.0307309Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0310015Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0310120Z return x.grad, w.grad 2025-12-04T12:25:17.0310858Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0310960Z warnings.warn( 2025-12-04T12:25:17.0313697Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0313860Z return x.grad, w.grad 2025-12-04T12:25:17.0314007Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0314589Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True _ 2025-12-04T12:25:17.0314709Z Traceback (most recent call last): 2025-12-04T12:25:17.0315191Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0315391Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0315597Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0315602Z 2025-12-04T12:25:17.0315827Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0316787Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:17.0316793Z 2025-12-04T12:25:17.0317071Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0317287Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0317396Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0317522Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0317890Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0318116Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0318223Z graph_break [] 2025-12-04T12:25:17.0318439Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0321152Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0321260Z return x.grad, w.grad 2025-12-04T12:25:17.0321998Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0322101Z warnings.warn( 2025-12-04T12:25:17.0324793Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0324911Z return x.grad, w.grad 2025-12-04T12:25:17.0325127Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0325327Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0325441Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0325664Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0326044Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0326143Z graph_break [] 2025-12-04T12:25:17.0326372Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0329064Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0329185Z return x.grad, w.grad 2025-12-04T12:25:17.0329915Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0330013Z warnings.warn( 2025-12-04T12:25:17.0332757Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0332867Z return x.grad, w.grad 2025-12-04T12:25:17.0333092Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0333202Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0333318Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0333552Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0333890Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0333990Z graph_break [] 2025-12-04T12:25:17.0334215Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0334944Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0335053Z warnings.warn( 2025-12-04T12:25:17.0337854Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0337982Z return x.grad, w.grad 2025-12-04T12:25:17.0338812Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-56c85a19db7b6fad.xml - 2025-12-04T12:25:17.0338985Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0340102Z FAILED [0.1698s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0340179Z 2025-12-04T12:25:17.0340401Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0341400Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:17.0341407Z 2025-12-04T12:25:17.0341671Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0341861Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0342058Z ================== 1 failed, 204 deselected, 2 rerun in 5.08s ================== 2025-12-04T12:25:17.0342156Z Got exit code 1 2025-12-04T12:25:17.0343037Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True 2025-12-04T12:25:17.0343452Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:17.0344097Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b5990bb8b6e78b98.xml 2025-12-04T12:25:17.0344269Z ============================= test session starts ============================== 2025-12-04T12:25:17.0344622Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0344744Z cachedir: .pytest_cache 2025-12-04T12:25:17.0345262Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0345384Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0345535Z configfile: pytest.ini 2025-12-04T12:25:17.0346127Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0346365Z collecting ... collected 380 items / 169 deselected / 211 selected 2025-12-04T12:25:17.0346511Z stepcurrent: skipping 169 already run items. 2025-12-04T12:25:17.0346623Z Running 36 items in this shard 2025-12-04T12:25:17.0346629Z 2025-12-04T12:25:17.0347570Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6658s] [ 2%] 2025-12-04T12:25:17.0348487Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1688s] [ 2%] 2025-12-04T12:25:17.0349345Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1648s] [ 2%] 2025-12-04T12:25:17.0349355Z 2025-12-04T12:25:17.0349493Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0350066Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.0350198Z Traceback (most recent call last): 2025-12-04T12:25:17.0350663Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0350871Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0351080Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0351086Z 2025-12-04T12:25:17.0351296Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0352269Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0352402Z 2025-12-04T12:25:17.0352667Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0352928Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0353040Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0353153Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0353508Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0353725Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0353823Z graph_break [] 2025-12-04T12:25:17.0354050Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0354784Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0354903Z warnings.warn( 2025-12-04T12:25:17.0355472Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.0355595Z Traceback (most recent call last): 2025-12-04T12:25:17.0356078Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0356274Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0356495Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0356500Z 2025-12-04T12:25:17.0356713Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0357700Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0357709Z 2025-12-04T12:25:17.0357989Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0358205Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0358329Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0358447Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0358786Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0359019Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0359116Z graph_break [] 2025-12-04T12:25:17.0359330Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0360074Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0360174Z warnings.warn( 2025-12-04T12:25:17.0360398Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0360510Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0360619Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0360849Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0361183Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0361279Z graph_break [] 2025-12-04T12:25:17.0361502Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0362226Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0362336Z warnings.warn( 2025-12-04T12:25:17.0362476Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0363058Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.0363249Z Traceback (most recent call last): 2025-12-04T12:25:17.0363728Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0363956Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0364182Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0364187Z 2025-12-04T12:25:17.0364400Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0365369Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0365374Z 2025-12-04T12:25:17.0365642Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0365859Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0365991Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0366106Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0366464Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0366683Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0366782Z graph_break [] 2025-12-04T12:25:17.0367012Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0367746Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0367847Z warnings.warn( 2025-12-04T12:25:17.0368077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0368189Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0368365Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0368593Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0368927Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0369041Z graph_break [] 2025-12-04T12:25:17.0369259Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0369991Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0370107Z warnings.warn( 2025-12-04T12:25:17.0370322Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0370446Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0370562Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0370781Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0371132Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0371232Z graph_break [] 2025-12-04T12:25:17.0371445Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0372185Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0372285Z warnings.warn( 2025-12-04T12:25:17.0373121Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b5990bb8b6e78b98.xml - 2025-12-04T12:25:17.0373291Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0374391Z FAILED [0.1648s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0374463Z 2025-12-04T12:25:17.0374679Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0375667Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0375673Z 2025-12-04T12:25:17.0375952Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0376129Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0376343Z ================== 1 failed, 169 deselected, 2 rerun in 5.05s ================== 2025-12-04T12:25:17.0376443Z Got exit code 1 2025-12-04T12:25:17.0376550Z Retrying single test... 2025-12-04T12:25:17.0377281Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9e1ed788d5899650.xml 2025-12-04T12:25:17.0377445Z ============================= test session starts ============================== 2025-12-04T12:25:17.0377795Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0377916Z cachedir: .pytest_cache 2025-12-04T12:25:17.0378437Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0378575Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0378682Z configfile: pytest.ini 2025-12-04T12:25:17.0379273Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0379513Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.0380607Z stepcurrent: skipping 169 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0380728Z Running 1 items in this shard 2025-12-04T12:25:17.0380745Z 2025-12-04T12:25:17.0381671Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6601s] [100%] 2025-12-04T12:25:17.0382596Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1610s] [100%] 2025-12-04T12:25:17.0383454Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1572s] [100%] 2025-12-04T12:25:17.0383460Z 2025-12-04T12:25:17.0383602Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0384192Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.0384315Z Traceback (most recent call last): 2025-12-04T12:25:17.0384788Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0385000Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0385211Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0385217Z 2025-12-04T12:25:17.0385445Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0386412Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0386418Z 2025-12-04T12:25:17.0386684Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0386972Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0387086Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0387216Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0387588Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0387809Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0387919Z graph_break [] 2025-12-04T12:25:17.0388134Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0388870Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0388983Z warnings.warn( 2025-12-04T12:25:17.0389560Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.0389699Z Traceback (most recent call last): 2025-12-04T12:25:17.0390167Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0390365Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0390589Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0390594Z 2025-12-04T12:25:17.0390804Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0391770Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0391776Z 2025-12-04T12:25:17.0392039Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0392284Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0392414Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0392526Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0392864Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0393097Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0393194Z graph_break [] 2025-12-04T12:25:17.0393418Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0394151Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0394252Z warnings.warn( 2025-12-04T12:25:17.0394477Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0394588Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0394701Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0394932Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0395276Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0395388Z graph_break [] 2025-12-04T12:25:17.0395604Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0396529Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0396645Z warnings.warn( 2025-12-04T12:25:17.0396787Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0397377Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.0397499Z Traceback (most recent call last): 2025-12-04T12:25:17.0397972Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0398295Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0398507Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0398512Z 2025-12-04T12:25:17.0398774Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0399753Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0399759Z 2025-12-04T12:25:17.0400027Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0400258Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0400372Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0400486Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0400838Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0401061Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0401172Z graph_break [] 2025-12-04T12:25:17.0401390Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0402124Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0402238Z warnings.warn( 2025-12-04T12:25:17.0402453Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0402561Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0402687Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0402906Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0403300Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0403398Z graph_break [] 2025-12-04T12:25:17.0403616Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0404357Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0404461Z warnings.warn( 2025-12-04T12:25:17.0404674Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0404796Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0404910Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0405143Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0405480Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0405577Z graph_break [] 2025-12-04T12:25:17.0405805Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0406529Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0406634Z warnings.warn( 2025-12-04T12:25:17.0407532Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9e1ed788d5899650.xml - 2025-12-04T12:25:17.0407754Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0409033Z FAILED [0.1572s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0409041Z 2025-12-04T12:25:17.0409258Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0410224Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0410320Z 2025-12-04T12:25:17.0410587Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0410826Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0411040Z ================== 1 failed, 204 deselected, 2 rerun in 5.03s ================== 2025-12-04T12:25:17.0411139Z Got exit code 1 2025-12-04T12:25:17.0411245Z Retrying single test... 2025-12-04T12:25:17.0411898Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e97cc37d8f662932.xml 2025-12-04T12:25:17.0412059Z ============================= test session starts ============================== 2025-12-04T12:25:17.0412420Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0412526Z cachedir: .pytest_cache 2025-12-04T12:25:17.0413054Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0413195Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0413305Z configfile: pytest.ini 2025-12-04T12:25:17.0413899Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0414137Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.0415182Z stepcurrent: skipping 169 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0415307Z Running 1 items in this shard 2025-12-04T12:25:17.0415313Z 2025-12-04T12:25:17.0416272Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6779s] [100%] 2025-12-04T12:25:17.0417286Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1642s] [100%] 2025-12-04T12:25:17.0418126Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1623s] [100%] 2025-12-04T12:25:17.0418132Z 2025-12-04T12:25:17.0418272Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0418860Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.0418984Z Traceback (most recent call last): 2025-12-04T12:25:17.0419469Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0419671Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0419886Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0419892Z 2025-12-04T12:25:17.0420118Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0421073Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0421079Z 2025-12-04T12:25:17.0421359Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0421576Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0421688Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0421814Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0422224Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0422452Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0422581Z graph_break [] 2025-12-04T12:25:17.0422799Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0423548Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0423650Z warnings.warn( 2025-12-04T12:25:17.0424225Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.0424360Z Traceback (most recent call last): 2025-12-04T12:25:17.0424828Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0425037Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0425255Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0425261Z 2025-12-04T12:25:17.0425477Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0426450Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0426456Z 2025-12-04T12:25:17.0426722Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0426952Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0427067Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0427183Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0427568Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0427794Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0427893Z graph_break [] 2025-12-04T12:25:17.0428124Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0428858Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0428980Z warnings.warn( 2025-12-04T12:25:17.0429201Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0429314Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0429447Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0429668Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0430009Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0430129Z graph_break [] 2025-12-04T12:25:17.0430345Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0431102Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0431204Z warnings.warn( 2025-12-04T12:25:17.0431348Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0431938Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.0432060Z Traceback (most recent call last): 2025-12-04T12:25:17.0432530Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0432746Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0432961Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0432966Z 2025-12-04T12:25:17.0433257Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0434255Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0434262Z 2025-12-04T12:25:17.0434543Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0434759Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0434871Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0434997Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0435336Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0435554Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0435665Z graph_break [] 2025-12-04T12:25:17.0435882Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0436619Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0436735Z warnings.warn( 2025-12-04T12:25:17.0436950Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0437075Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0437188Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0437407Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0437753Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0437849Z graph_break [] 2025-12-04T12:25:17.0438062Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0438834Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0438941Z warnings.warn( 2025-12-04T12:25:17.0439163Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0439274Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0439387Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0439615Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0439951Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0440047Z graph_break [] 2025-12-04T12:25:17.0440269Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0440999Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0441110Z warnings.warn( 2025-12-04T12:25:17.0441929Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e97cc37d8f662932.xml - 2025-12-04T12:25:17.0442102Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0443203Z FAILED [0.1623s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0443209Z 2025-12-04T12:25:17.0443423Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0444396Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0444402Z 2025-12-04T12:25:17.0444673Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0444927Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0445124Z ================== 1 failed, 204 deselected, 2 rerun in 5.06s ================== 2025-12-04T12:25:17.0445251Z Got exit code 1 2025-12-04T12:25:17.0446141Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0446549Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:17.0447197Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d35e9b6eebdaac7b.xml 2025-12-04T12:25:17.0447371Z ============================= test session starts ============================== 2025-12-04T12:25:17.0447723Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0447851Z cachedir: .pytest_cache 2025-12-04T12:25:17.0448373Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0448498Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0448620Z configfile: pytest.ini 2025-12-04T12:25:17.0449209Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0449450Z collecting ... collected 380 items / 170 deselected / 210 selected 2025-12-04T12:25:17.0449597Z stepcurrent: skipping 170 already run items. 2025-12-04T12:25:17.0449711Z Running 35 items in this shard 2025-12-04T12:25:17.0449716Z 2025-12-04T12:25:17.0450680Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [4.6718s] [ 2%] 2025-12-04T12:25:17.0451605Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [0.1659s] [ 2%] 2025-12-04T12:25:17.0452456Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True FAILED [0.1639s] [ 2%] 2025-12-04T12:25:17.0452462Z 2025-12-04T12:25:17.0452599Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0453170Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:17.0453304Z Traceback (most recent call last): 2025-12-04T12:25:17.0453774Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0453983Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0454197Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0454202Z 2025-12-04T12:25:17.0454416Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0455389Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:17.0455394Z 2025-12-04T12:25:17.0455659Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0455888Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0455998Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0456112Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0456470Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0456761Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0456939Z graph_break [] 2025-12-04T12:25:17.0457172Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0459916Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0460040Z return x.grad, w.grad 2025-12-04T12:25:17.0460772Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0460894Z warnings.warn( 2025-12-04T12:25:17.0463616Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0463742Z return x.grad, w.grad 2025-12-04T12:25:17.0464346Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:17.0464471Z Traceback (most recent call last): 2025-12-04T12:25:17.0464957Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0465155Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0465365Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0465381Z 2025-12-04T12:25:17.0465594Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0466552Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:17.0466557Z 2025-12-04T12:25:17.0466833Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0467056Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0467178Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0467296Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0467637Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0467871Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0467967Z graph_break [] 2025-12-04T12:25:17.0468183Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0470899Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0471066Z return x.grad, w.grad 2025-12-04T12:25:17.0471837Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0471938Z warnings.warn( 2025-12-04T12:25:17.0474666Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0474774Z return x.grad, w.grad 2025-12-04T12:25:17.0474996Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0475117Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0475229Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0475465Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0475804Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0475900Z graph_break [] 2025-12-04T12:25:17.0476128Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0478855Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0478979Z return x.grad, w.grad 2025-12-04T12:25:17.0479706Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0479805Z warnings.warn( 2025-12-04T12:25:17.0482523Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0482633Z return x.grad, w.grad 2025-12-04T12:25:17.0482790Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0483361Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:17.0483493Z Traceback (most recent call last): 2025-12-04T12:25:17.0483966Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0484163Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0484384Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0484389Z 2025-12-04T12:25:17.0484604Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0485578Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:17.0485640Z 2025-12-04T12:25:17.0485937Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0486154Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0486276Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0486389Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0486725Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0486954Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0487050Z graph_break [] 2025-12-04T12:25:17.0487276Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0489981Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0490108Z return x.grad, w.grad 2025-12-04T12:25:17.0490840Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0490940Z warnings.warn( 2025-12-04T12:25:17.0493679Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0493791Z return x.grad, w.grad 2025-12-04T12:25:17.0494019Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0494128Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0494240Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0494476Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0494815Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0494933Z graph_break [] 2025-12-04T12:25:17.0495150Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0498194Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0498320Z return x.grad, w.grad 2025-12-04T12:25:17.0499050Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0499166Z warnings.warn( 2025-12-04T12:25:17.0502032Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0502152Z return x.grad, w.grad 2025-12-04T12:25:17.0502371Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0502480Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0502608Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0502831Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0503183Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0503284Z graph_break [] 2025-12-04T12:25:17.0503500Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0504245Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0504344Z warnings.warn( 2025-12-04T12:25:17.0507095Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0507205Z return x.grad, w.grad 2025-12-04T12:25:17.0508041Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d35e9b6eebdaac7b.xml - 2025-12-04T12:25:17.0508215Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0509311Z FAILED [0.1639s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0509318Z 2025-12-04T12:25:17.0509546Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0510511Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:17.0510521Z 2025-12-04T12:25:17.0510799Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0510979Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0511178Z ================== 1 failed, 170 deselected, 2 rerun in 5.06s ================== 2025-12-04T12:25:17.0511287Z Got exit code 1 2025-12-04T12:25:17.0511393Z Retrying single test... 2025-12-04T12:25:17.0512059Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f6ccc2d46fa5c9ea.xml 2025-12-04T12:25:17.0512221Z ============================= test session starts ============================== 2025-12-04T12:25:17.0512568Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0512694Z cachedir: .pytest_cache 2025-12-04T12:25:17.0513293Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0513417Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0513541Z configfile: pytest.ini 2025-12-04T12:25:17.0514161Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0514402Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.0515446Z stepcurrent: skipping 170 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:17.0515558Z Running 1 items in this shard 2025-12-04T12:25:17.0515563Z 2025-12-04T12:25:17.0516493Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [4.6799s] [100%] 2025-12-04T12:25:17.0517416Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [0.1698s] [100%] 2025-12-04T12:25:17.0518266Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True FAILED [0.1673s] [100%] 2025-12-04T12:25:17.0518272Z 2025-12-04T12:25:17.0518412Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0518997Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:17.0519117Z Traceback (most recent call last): 2025-12-04T12:25:17.0519616Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0519830Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0520042Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0520049Z 2025-12-04T12:25:17.0520277Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0521237Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:17.0521242Z 2025-12-04T12:25:17.0521514Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0521743Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0521855Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0521987Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0522331Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0522555Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0522662Z graph_break [] 2025-12-04T12:25:17.0522878Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0525572Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0525692Z return x.grad, w.grad 2025-12-04T12:25:17.0526482Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0526595Z warnings.warn( 2025-12-04T12:25:17.0529310Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0529432Z return x.grad, w.grad 2025-12-04T12:25:17.0530003Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:17.0530141Z Traceback (most recent call last): 2025-12-04T12:25:17.0530609Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0530803Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0531024Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0531029Z 2025-12-04T12:25:17.0531239Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0532196Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:17.0532212Z 2025-12-04T12:25:17.0532478Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0532720Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0532847Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0532958Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0533301Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0533524Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0533620Z graph_break [] 2025-12-04T12:25:17.0533848Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0536552Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0536673Z return x.grad, w.grad 2025-12-04T12:25:17.0537460Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0537561Z warnings.warn( 2025-12-04T12:25:17.0540273Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0540437Z return x.grad, w.grad 2025-12-04T12:25:17.0540664Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0540802Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0540916Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0541153Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0541488Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0541585Z graph_break [] 2025-12-04T12:25:17.0541813Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0544524Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0544645Z return x.grad, w.grad 2025-12-04T12:25:17.0545377Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0545487Z warnings.warn( 2025-12-04T12:25:17.0548206Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0548324Z return x.grad, w.grad 2025-12-04T12:25:17.0548473Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0549038Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:17.0549169Z Traceback (most recent call last): 2025-12-04T12:25:17.0549638Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0549844Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0550055Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0550060Z 2025-12-04T12:25:17.0550275Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0551250Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:17.0551255Z 2025-12-04T12:25:17.0551522Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0551749Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0551858Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0551972Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0552315Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0552537Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0552634Z graph_break [] 2025-12-04T12:25:17.0552863Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0555723Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0555840Z return x.grad, w.grad 2025-12-04T12:25:17.0556570Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0556683Z warnings.warn( 2025-12-04T12:25:17.0559375Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0559500Z return x.grad, w.grad 2025-12-04T12:25:17.0559714Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0559824Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0559950Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0560171Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0560550Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0560660Z graph_break [] 2025-12-04T12:25:17.0560875Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0563581Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0563685Z return x.grad, w.grad 2025-12-04T12:25:17.0564420Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0564533Z warnings.warn( 2025-12-04T12:25:17.0567215Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0567332Z return x.grad, w.grad 2025-12-04T12:25:17.0567546Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0567665Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0567778Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0568003Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0568413Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0568510Z graph_break [] 2025-12-04T12:25:17.0568753Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0569498Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0569597Z warnings.warn( 2025-12-04T12:25:17.0572297Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0572407Z return x.grad, w.grad 2025-12-04T12:25:17.0573245Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f6ccc2d46fa5c9ea.xml - 2025-12-04T12:25:17.0573413Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0574503Z FAILED [0.1673s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0574521Z 2025-12-04T12:25:17.0574737Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0575715Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:17.0575726Z 2025-12-04T12:25:17.0576006Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0576184Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0576392Z ================== 1 failed, 204 deselected, 2 rerun in 5.07s ================== 2025-12-04T12:25:17.0576489Z Got exit code 1 2025-12-04T12:25:17.0576592Z Retrying single test... 2025-12-04T12:25:17.0577310Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dac88d9039f141ad.xml 2025-12-04T12:25:17.0577473Z ============================= test session starts ============================== 2025-12-04T12:25:17.0577823Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0577946Z cachedir: .pytest_cache 2025-12-04T12:25:17.0578463Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0578600Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0578706Z configfile: pytest.ini 2025-12-04T12:25:17.0579297Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0579533Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.0580574Z stepcurrent: skipping 170 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:17.0580683Z Running 1 items in this shard 2025-12-04T12:25:17.0580703Z 2025-12-04T12:25:17.0581622Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [4.6448s] [100%] 2025-12-04T12:25:17.0582626Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True ('RERUN', {'yellow': True}) [0.1658s] [100%] 2025-12-04T12:25:17.0583477Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True FAILED [0.1628s] [100%] 2025-12-04T12:25:17.0583483Z 2025-12-04T12:25:17.0583625Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0584204Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:17.0584325Z Traceback (most recent call last): 2025-12-04T12:25:17.0584799Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0585007Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0585216Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0585222Z 2025-12-04T12:25:17.0585448Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0586407Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:17.0586413Z 2025-12-04T12:25:17.0586674Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0586901Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0587042Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0587169Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0587513Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0587729Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0587843Z graph_break [] 2025-12-04T12:25:17.0588054Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0590782Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0590891Z return x.grad, w.grad 2025-12-04T12:25:17.0591617Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0591730Z warnings.warn( 2025-12-04T12:25:17.0594418Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0594538Z return x.grad, w.grad 2025-12-04T12:25:17.0595105Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:17.0595408Z Traceback (most recent call last): 2025-12-04T12:25:17.0595938Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0596588Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0596840Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0596846Z 2025-12-04T12:25:17.0597149Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0598146Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:17.0598152Z 2025-12-04T12:25:17.0598473Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0598823Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0598991Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0599194Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0599583Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0599852Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0600016Z graph_break [] 2025-12-04T12:25:17.0600304Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0603161Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0603319Z return x.grad, w.grad 2025-12-04T12:25:17.0604163Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0604354Z warnings.warn( 2025-12-04T12:25:17.0607060Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0607321Z return x.grad, w.grad 2025-12-04T12:25:17.0607580Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0607793Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0607945Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0608207Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0608648Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0608802Z graph_break [] 2025-12-04T12:25:17.0609114Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0611900Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0612183Z return x.grad, w.grad 2025-12-04T12:25:17.0612929Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0613107Z warnings.warn( 2025-12-04T12:25:17.0615921Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0616072Z return x.grad, w.grad 2025-12-04T12:25:17.0616304Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0616983Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True _ 2025-12-04T12:25:17.0617125Z Traceback (most recent call last): 2025-12-04T12:25:17.0617750Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0618051Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0618390Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0618401Z 2025-12-04T12:25:17.0618656Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0619653Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:17.0619688Z 2025-12-04T12:25:17.0620043Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0620321Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0620519Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0620671Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0621048Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0621335Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0621536Z graph_break [] 2025-12-04T12:25:17.0621861Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0624610Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0624805Z return x.grad, w.grad 2025-12-04T12:25:17.0625572Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0625699Z warnings.warn( 2025-12-04T12:25:17.0628616Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0628790Z return x.grad, w.grad 2025-12-04T12:25:17.0629093Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0629254Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0629407Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0629799Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0630196Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0630387Z graph_break [] 2025-12-04T12:25:17.0630649Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0633404Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0633577Z return x.grad, w.grad 2025-12-04T12:25:17.0634413Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0634624Z warnings.warn( 2025-12-04T12:25:17.0637371Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0637564Z return x.grad, w.grad 2025-12-04T12:25:17.0637823Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0638036Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0638213Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0638478Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0638912Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0639049Z graph_break [] 2025-12-04T12:25:17.0639302Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0640127Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0640285Z warnings.warn( 2025-12-04T12:25:17.0643076Z /var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py:315: UserWarning: The .grad attribute of a Tensor that is not a leaf Tensor is being accessed. Its .grad attribute won't be populated during autograd.backward(). If you indeed want the .grad field to be populated for a non-leaf Tensor, use .retain_grad() on the non-leaf Tensor. If you access the non-leaf Tensor by mistake, make sure you access the leaf Tensor instead. See github.com/pytorch/pytorch/pull/30531 for more information. (Triggered internally at /var/lib/jenkins/workspace/build/aten/src/ATen/core/TensorBody.h:492.) 2025-12-04T12:25:17.0643840Z return x.grad, w.grad 2025-12-04T12:25:17.0644821Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dac88d9039f141ad.xml - 2025-12-04T12:25:17.0645007Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0646175Z FAILED [0.1628s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0646182Z 2025-12-04T12:25:17.0646521Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0647516Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:17.0647524Z 2025-12-04T12:25:17.0647880Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0648098Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0648410Z ================== 1 failed, 204 deselected, 2 rerun in 5.03s ================== 2025-12-04T12:25:17.0648567Z Got exit code 1 2025-12-04T12:25:17.0649478Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True 2025-12-04T12:25:17.0650010Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:17.0650699Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-1d81442a0ab3908c.xml 2025-12-04T12:25:17.0650902Z ============================= test session starts ============================== 2025-12-04T12:25:17.0651371Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0651539Z cachedir: .pytest_cache 2025-12-04T12:25:17.0652149Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0652310Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0652456Z configfile: pytest.ini 2025-12-04T12:25:17.0653125Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0653436Z collecting ... collected 380 items / 171 deselected / 209 selected 2025-12-04T12:25:17.0653696Z stepcurrent: skipping 171 already run items. 2025-12-04T12:25:17.0653849Z Running 34 items in this shard 2025-12-04T12:25:17.0653854Z 2025-12-04T12:25:17.0655002Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_True_initial_xblock_1_add_1dim_False SKIPPED [0.0041s] (Skip non-critical tests to save resources.) [ 2%] 2025-12-04T12:25:17.0656141Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_True_initial_xblock_2_add_1dim_False SKIPPED [0.0029s] (Skip non-critical tests to save resources.) [ 5%] 2025-12-04T12:25:17.0657290Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_True_initial_xblock_2_add_1dim_True SKIPPED [0.0035s] (Skip non-critical tests to save resources.) [ 8%] 2025-12-04T12:25:17.0658488Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_True SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 11%] 2025-12-04T12:25:17.0659667Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape1_max_autotune_True_initial_xblock_1_add_1dim_False SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 14%] 2025-12-04T12:25:17.0660792Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape1_max_autotune_True_initial_xblock_1_add_1dim_True SKIPPED [0.0026s] (Skip non-critical tests to save resources.) [ 17%] 2025-12-04T12:25:17.0661858Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape1_max_autotune_True_initial_xblock_2_add_1dim_True SKIPPED [0.0026s] (Skip non-critical tests to save resources.) [ 20%] 2025-12-04T12:25:17.0662839Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6411s] [ 23%] 2025-12-04T12:25:17.0663854Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1616s] [ 23%] 2025-12-04T12:25:17.0664802Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1585s] [ 23%] 2025-12-04T12:25:17.0664809Z 2025-12-04T12:25:17.0664997Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0665638Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0665850Z Traceback (most recent call last): 2025-12-04T12:25:17.0666340Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0666685Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0666951Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0666957Z 2025-12-04T12:25:17.0667209Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0668266Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0668271Z 2025-12-04T12:25:17.0668575Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0668903Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0669071Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0669307Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0669739Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0670000Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0670137Z graph_break [] 2025-12-04T12:25:17.0670464Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0671272Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0671463Z warnings.warn( 2025-12-04T12:25:17.0672076Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0672236Z Traceback (most recent call last): 2025-12-04T12:25:17.0672769Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0673114Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0673431Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0673438Z 2025-12-04T12:25:17.0673719Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0674726Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0674731Z 2025-12-04T12:25:17.0675087Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0675328Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0675587Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0675739Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0676122Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0676429Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0676575Z graph_break [] 2025-12-04T12:25:17.0676960Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0677757Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0677896Z warnings.warn( 2025-12-04T12:25:17.0678200Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0678351Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0678513Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0678827Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0679254Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0679444Z graph_break [] 2025-12-04T12:25:17.0679694Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0680477Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0680643Z warnings.warn( 2025-12-04T12:25:17.0680866Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0681496Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0681705Z Traceback (most recent call last): 2025-12-04T12:25:17.0682227Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0682515Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0682743Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0682753Z 2025-12-04T12:25:17.0683040Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0684116Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0684122Z 2025-12-04T12:25:17.0684428Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0684744Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0684894Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0685082Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0685565Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0685827Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0686022Z graph_break [] 2025-12-04T12:25:17.0686346Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0687144Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0687354Z warnings.warn( 2025-12-04T12:25:17.0687637Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0687785Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0687986Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0688245Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0688645Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0688822Z graph_break [] 2025-12-04T12:25:17.0689105Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0689925Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0690071Z warnings.warn( 2025-12-04T12:25:17.0690374Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0690500Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0690694Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0691031Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0691405Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0691540Z graph_break [] 2025-12-04T12:25:17.0691910Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0692705Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0692995Z warnings.warn( 2025-12-04T12:25:17.0693855Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-1d81442a0ab3908c.xml - 2025-12-04T12:25:17.0694066Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0695250Z FAILED [0.1585s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0695257Z 2025-12-04T12:25:17.0695510Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0696784Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0696791Z 2025-12-04T12:25:17.0697223Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0697494Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0697761Z ============ 1 failed, 7 skipped, 171 deselected, 2 rerun in 5.05s ============= 2025-12-04T12:25:17.0697899Z Got exit code 1 2025-12-04T12:25:17.0698084Z Retrying single test... 2025-12-04T12:25:17.0698807Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-773b425f2e921837.xml 2025-12-04T12:25:17.0699026Z ============================= test session starts ============================== 2025-12-04T12:25:17.0699464Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0699613Z cachedir: .pytest_cache 2025-12-04T12:25:17.0700233Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0700482Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0700667Z configfile: pytest.ini 2025-12-04T12:25:17.0701476Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0701744Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.0702899Z stepcurrent: skipping 178 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0703051Z Running 1 items in this shard 2025-12-04T12:25:17.0703057Z 2025-12-04T12:25:17.0704001Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6810s] [100%] 2025-12-04T12:25:17.0705071Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1657s] [100%] 2025-12-04T12:25:17.0705969Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1610s] [100%] 2025-12-04T12:25:17.0705975Z 2025-12-04T12:25:17.0706212Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0706827Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0707015Z Traceback (most recent call last): 2025-12-04T12:25:17.0707557Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0707854Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0708171Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0708177Z 2025-12-04T12:25:17.0708427Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0709479Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0709485Z 2025-12-04T12:25:17.0709765Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0710055Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0710284Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0710436Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0710819Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0711127Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0711244Z graph_break [] 2025-12-04T12:25:17.0711683Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0712457Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0712600Z warnings.warn( 2025-12-04T12:25:17.0713263Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0713422Z Traceback (most recent call last): 2025-12-04T12:25:17.0714005Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0714270Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0714521Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0714526Z 2025-12-04T12:25:17.0714890Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0715924Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0715929Z 2025-12-04T12:25:17.0716273Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0716575Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0716745Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0716948Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0717328Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0717645Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0717760Z graph_break [] 2025-12-04T12:25:17.0718054Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0718899Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0719039Z warnings.warn( 2025-12-04T12:25:17.0719289Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0719497Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0719625Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0720055Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0720432Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0720567Z graph_break [] 2025-12-04T12:25:17.0720881Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0721683Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0721805Z warnings.warn( 2025-12-04T12:25:17.0722119Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0722734Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0722952Z Traceback (most recent call last): 2025-12-04T12:25:17.0723463Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0723698Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0724018Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0724023Z 2025-12-04T12:25:17.0724294Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0725353Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0725363Z 2025-12-04T12:25:17.0725679Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0725930Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0726105Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0726293Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0726737Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0727001Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0727135Z graph_break [] 2025-12-04T12:25:17.0727505Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0728262Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0728574Z warnings.warn( 2025-12-04T12:25:17.0728841Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0729020Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0729224Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0729484Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0729835Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0730089Z graph_break [] 2025-12-04T12:25:17.0730340Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0731158Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0731299Z warnings.warn( 2025-12-04T12:25:17.0731551Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0731767Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0731946Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0732205Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0732626Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0732762Z graph_break [] 2025-12-04T12:25:17.0733043Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0733849Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0734005Z warnings.warn( 2025-12-04T12:25:17.0734944Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-773b425f2e921837.xml - 2025-12-04T12:25:17.0735251Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0736675Z FAILED [0.1610s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0736683Z 2025-12-04T12:25:17.0737014Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0738098Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0738172Z 2025-12-04T12:25:17.0738477Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0738697Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0738984Z ================== 1 failed, 204 deselected, 2 rerun in 5.06s ================== 2025-12-04T12:25:17.0739138Z Got exit code 1 2025-12-04T12:25:17.0739262Z Retrying single test... 2025-12-04T12:25:17.0740052Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-09fb0d74c7a00548.xml 2025-12-04T12:25:17.0740253Z ============================= test session starts ============================== 2025-12-04T12:25:17.0740692Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0740840Z cachedir: .pytest_cache 2025-12-04T12:25:17.0741408Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0741632Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0741797Z configfile: pytest.ini 2025-12-04T12:25:17.0742431Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0742838Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.0743967Z stepcurrent: skipping 178 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0744148Z Running 1 items in this shard 2025-12-04T12:25:17.0744154Z 2025-12-04T12:25:17.0745161Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6441s] [100%] 2025-12-04T12:25:17.0746263Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1601s] [100%] 2025-12-04T12:25:17.0747160Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1590s] [100%] 2025-12-04T12:25:17.0747168Z 2025-12-04T12:25:17.0747398Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0748042Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0748180Z Traceback (most recent call last): 2025-12-04T12:25:17.0748794Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0749031Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0749291Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0749347Z 2025-12-04T12:25:17.0749638Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0750645Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0750654Z 2025-12-04T12:25:17.0751021Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0751296Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0751507Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0751663Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0752043Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0752325Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0752493Z graph_break [] 2025-12-04T12:25:17.0752779Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0753601Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0753744Z warnings.warn( 2025-12-04T12:25:17.0754404Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0754540Z Traceback (most recent call last): 2025-12-04T12:25:17.0755078Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0755460Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0755709Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0755714Z 2025-12-04T12:25:17.0756012Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0757009Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0757078Z 2025-12-04T12:25:17.0757360Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0757763Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0757913Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0758113Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0758493Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0758762Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0758973Z graph_break [] 2025-12-04T12:25:17.0759244Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0760014Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0760211Z warnings.warn( 2025-12-04T12:25:17.0760463Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0760647Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0760841Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0761119Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0761549Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0761683Z graph_break [] 2025-12-04T12:25:17.0761944Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0762798Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0768398Z warnings.warn( 2025-12-04T12:25:17.0768687Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0769282Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0769411Z Traceback (most recent call last): 2025-12-04T12:25:17.0769890Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0770097Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0770313Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0770320Z 2025-12-04T12:25:17.0770530Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0771504Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0771510Z 2025-12-04T12:25:17.0771777Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0772010Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0772120Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0772232Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0772582Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0772799Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0772891Z graph_break [] 2025-12-04T12:25:17.0773115Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0773850Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0773954Z warnings.warn( 2025-12-04T12:25:17.0774169Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0774356Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0774478Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0774691Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0775057Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0775165Z graph_break [] 2025-12-04T12:25:17.0775372Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0776110Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0776206Z warnings.warn( 2025-12-04T12:25:17.0776413Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0776529Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0776637Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0776980Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0777331Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0777428Z graph_break [] 2025-12-04T12:25:17.0777651Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0778378Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0778473Z warnings.warn( 2025-12-04T12:25:17.0779303Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-09fb0d74c7a00548.xml - 2025-12-04T12:25:17.0779465Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0780603Z FAILED [0.1590s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0780615Z 2025-12-04T12:25:17.0780831Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0781792Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0781808Z 2025-12-04T12:25:17.0782070Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0782246Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0782451Z ================== 1 failed, 204 deselected, 2 rerun in 5.02s ================== 2025-12-04T12:25:17.0782544Z Got exit code 1 2025-12-04T12:25:17.0783420Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0783841Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:17.0784477Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2f000bd6c93ab369.xml 2025-12-04T12:25:17.0784646Z ============================= test session starts ============================== 2025-12-04T12:25:17.0784990Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0785096Z cachedir: .pytest_cache 2025-12-04T12:25:17.0785623Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0785742Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0785854Z configfile: pytest.ini 2025-12-04T12:25:17.0786443Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0786728Z collecting ... collected 380 items / 179 deselected / 201 selected 2025-12-04T12:25:17.0786911Z stepcurrent: skipping 179 already run items. 2025-12-04T12:25:17.0787019Z Running 26 items in this shard 2025-12-04T12:25:17.0787024Z 2025-12-04T12:25:17.0788065Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_True SKIPPED [0.0041s] (Skip non-critical tests to save resources.) [ 3%] 2025-12-04T12:25:17.0788994Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6512s] [ 7%] 2025-12-04T12:25:17.0789914Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1624s] [ 7%] 2025-12-04T12:25:17.0790762Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1607s] [ 7%] 2025-12-04T12:25:17.0790768Z 2025-12-04T12:25:17.0790902Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0791478Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.0791595Z Traceback (most recent call last): 2025-12-04T12:25:17.0792061Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0792260Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0792495Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0792505Z 2025-12-04T12:25:17.0792722Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0793681Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0793686Z 2025-12-04T12:25:17.0793946Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0794168Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0794273Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0794390Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0794724Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0794936Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0795036Z graph_break [] 2025-12-04T12:25:17.0795253Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0796157Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0796264Z warnings.warn( 2025-12-04T12:25:17.0796837Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.0796961Z Traceback (most recent call last): 2025-12-04T12:25:17.0797428Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0797621Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0797836Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0797842Z 2025-12-04T12:25:17.0798055Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0799145Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0799191Z 2025-12-04T12:25:17.0799456Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0799670Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0799787Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0799899Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0800236Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0800463Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0800558Z graph_break [] 2025-12-04T12:25:17.0800781Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0801514Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0801617Z warnings.warn( 2025-12-04T12:25:17.0801840Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0801945Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0802052Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0802278Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0802608Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0802709Z graph_break [] 2025-12-04T12:25:17.0802916Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0803675Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0803786Z warnings.warn( 2025-12-04T12:25:17.0803926Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0804509Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.0804626Z Traceback (most recent call last): 2025-12-04T12:25:17.0805090Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0805293Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0805501Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0805506Z 2025-12-04T12:25:17.0805715Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0806680Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0806690Z 2025-12-04T12:25:17.0806949Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0807169Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0807275Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0807387Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0807733Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0807948Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0808054Z graph_break [] 2025-12-04T12:25:17.0808260Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0808988Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0809091Z warnings.warn( 2025-12-04T12:25:17.0809372Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0809482Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0809605Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0809847Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0810186Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0810281Z graph_break [] 2025-12-04T12:25:17.0810489Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0811226Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0811323Z warnings.warn( 2025-12-04T12:25:17.0811534Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0811650Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0811764Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0811988Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0812321Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0812413Z graph_break [] 2025-12-04T12:25:17.0812629Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0813349Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0813445Z warnings.warn( 2025-12-04T12:25:17.0814272Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2f000bd6c93ab369.xml - 2025-12-04T12:25:17.0814466Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0815572Z FAILED [0.1607s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0815578Z 2025-12-04T12:25:17.0815790Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0816744Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0816756Z 2025-12-04T12:25:17.0817083Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0817263Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0817492Z ============ 1 failed, 1 skipped, 179 deselected, 2 rerun in 5.03s ============= 2025-12-04T12:25:17.0817584Z Got exit code 1 2025-12-04T12:25:17.0817690Z Retrying single test... 2025-12-04T12:25:17.0818344Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-64628aab5ce6aeb9.xml 2025-12-04T12:25:17.0818502Z ============================= test session starts ============================== 2025-12-04T12:25:17.0818854Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0818959Z cachedir: .pytest_cache 2025-12-04T12:25:17.0819475Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0819601Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0819703Z configfile: pytest.ini 2025-12-04T12:25:17.0820293Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0820521Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.0821661Z stepcurrent: skipping 180 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0821780Z Running 1 items in this shard 2025-12-04T12:25:17.0821785Z 2025-12-04T12:25:17.0822700Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6317s] [100%] 2025-12-04T12:25:17.0823622Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1659s] [100%] 2025-12-04T12:25:17.0824462Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1597s] [100%] 2025-12-04T12:25:17.0824472Z 2025-12-04T12:25:17.0824610Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0825188Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.0825307Z Traceback (most recent call last): 2025-12-04T12:25:17.0825779Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0825972Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0826182Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0826187Z 2025-12-04T12:25:17.0826407Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0827399Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0827409Z 2025-12-04T12:25:17.0827681Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0827893Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0827998Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0828112Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0828449Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0828666Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0828758Z graph_break [] 2025-12-04T12:25:17.0828966Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0829713Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0829815Z warnings.warn( 2025-12-04T12:25:17.0830382Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.0830506Z Traceback (most recent call last): 2025-12-04T12:25:17.0830969Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0831172Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0831380Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0831385Z 2025-12-04T12:25:17.0831598Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0832561Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0832626Z 2025-12-04T12:25:17.0832886Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0833099Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0833235Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0833346Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0833688Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0833902Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0833993Z graph_break [] 2025-12-04T12:25:17.0834215Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0834946Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0835059Z warnings.warn( 2025-12-04T12:25:17.0835268Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0835378Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0835494Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0835710Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0836044Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0836144Z graph_break [] 2025-12-04T12:25:17.0836352Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0837075Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0837168Z warnings.warn( 2025-12-04T12:25:17.0837304Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0837911Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.0838029Z Traceback (most recent call last): 2025-12-04T12:25:17.0838496Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0838695Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0838897Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0838902Z 2025-12-04T12:25:17.0839114Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0840072Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0840078Z 2025-12-04T12:25:17.0840346Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0840557Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0840667Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0840784Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0841118Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0841331Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0841431Z graph_break [] 2025-12-04T12:25:17.0841638Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0842370Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0842466Z warnings.warn( 2025-12-04T12:25:17.0842672Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0842791Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0842899Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0843187Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0843525Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0843671Z graph_break [] 2025-12-04T12:25:17.0843882Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0844608Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0844702Z warnings.warn( 2025-12-04T12:25:17.0844913Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0845016Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0845123Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0845347Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0845681Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0845776Z graph_break [] 2025-12-04T12:25:17.0845994Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0846721Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0846823Z warnings.warn( 2025-12-04T12:25:17.0847637Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-64628aab5ce6aeb9.xml - 2025-12-04T12:25:17.0847802Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0848934Z FAILED [0.1597s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0848945Z 2025-12-04T12:25:17.0849156Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0850121Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0850127Z 2025-12-04T12:25:17.0850386Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0850569Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0850762Z ================== 1 failed, 204 deselected, 2 rerun in 5.01s ================== 2025-12-04T12:25:17.0850854Z Got exit code 1 2025-12-04T12:25:17.0850960Z Retrying single test... 2025-12-04T12:25:17.0851601Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9fedf51ad965eb7d.xml 2025-12-04T12:25:17.0851760Z ============================= test session starts ============================== 2025-12-04T12:25:17.0852110Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0852213Z cachedir: .pytest_cache 2025-12-04T12:25:17.0852731Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0852847Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0852949Z configfile: pytest.ini 2025-12-04T12:25:17.0853543Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0853763Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.0854804Z stepcurrent: skipping 180 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0854975Z Running 1 items in this shard 2025-12-04T12:25:17.0854980Z 2025-12-04T12:25:17.0855934Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6353s] [100%] 2025-12-04T12:25:17.0856937Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1632s] [100%] 2025-12-04T12:25:17.0857782Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1614s] [100%] 2025-12-04T12:25:17.0857788Z 2025-12-04T12:25:17.0857934Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0858503Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.0858619Z Traceback (most recent call last): 2025-12-04T12:25:17.0859095Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0859289Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0859513Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0859518Z 2025-12-04T12:25:17.0859725Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0860687Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0860701Z 2025-12-04T12:25:17.0860994Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0861212Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0861328Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0861441Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0861774Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0861994Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0862086Z graph_break [] 2025-12-04T12:25:17.0862296Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0863032Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0863130Z warnings.warn( 2025-12-04T12:25:17.0863707Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.0863830Z Traceback (most recent call last): 2025-12-04T12:25:17.0864298Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0864501Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0864709Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0864714Z 2025-12-04T12:25:17.0864929Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0865884Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0865889Z 2025-12-04T12:25:17.0866149Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0866372Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0866539Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0866658Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0866992Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0867231Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0867332Z graph_break [] 2025-12-04T12:25:17.0867542Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0868265Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0868367Z warnings.warn( 2025-12-04T12:25:17.0868576Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0868687Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0868798Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0869012Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0869358Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0869454Z graph_break [] 2025-12-04T12:25:17.0869662Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0870401Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0870498Z warnings.warn( 2025-12-04T12:25:17.0870645Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0871214Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.0871328Z Traceback (most recent call last): 2025-12-04T12:25:17.0871833Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0872029Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0872235Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0872240Z 2025-12-04T12:25:17.0872460Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0873413Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0873419Z 2025-12-04T12:25:17.0873686Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0873896Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0874000Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0874115Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0874451Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0874677Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0874771Z graph_break [] 2025-12-04T12:25:17.0874981Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0875711Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0875807Z warnings.warn( 2025-12-04T12:25:17.0876017Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0876132Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0876239Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0876461Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0876794Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0876948Z graph_break [] 2025-12-04T12:25:17.0877163Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0877912Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0878005Z warnings.warn( 2025-12-04T12:25:17.0878224Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0878328Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0878444Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0878654Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0878981Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0879079Z graph_break [] 2025-12-04T12:25:17.0879289Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0880007Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0880117Z warnings.warn( 2025-12-04T12:25:17.0880943Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9fedf51ad965eb7d.xml - 2025-12-04T12:25:17.0881115Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0882207Z FAILED [0.1614s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0882213Z 2025-12-04T12:25:17.0882421Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0883433Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0883443Z 2025-12-04T12:25:17.0883707Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0883887Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0884079Z ================== 1 failed, 204 deselected, 2 rerun in 5.01s ================== 2025-12-04T12:25:17.0884170Z Got exit code 1 2025-12-04T12:25:17.0885048Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.0885451Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:17.0886100Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c5a3e4c540d200e7.xml 2025-12-04T12:25:17.0886262Z ============================= test session starts ============================== 2025-12-04T12:25:17.0886603Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0886714Z cachedir: .pytest_cache 2025-12-04T12:25:17.0887229Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0887368Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0887475Z configfile: pytest.ini 2025-12-04T12:25:17.0888066Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0888303Z collecting ... collected 380 items / 181 deselected / 199 selected 2025-12-04T12:25:17.0888453Z stepcurrent: skipping 181 already run items. 2025-12-04T12:25:17.0888565Z Running 24 items in this shard 2025-12-04T12:25:17.0888642Z 2025-12-04T12:25:17.0889723Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_True_initial_xblock_1_add_1dim_True SKIPPED [0.0040s] (Skip non-critical tests to save resources.) [ 4%] 2025-12-04T12:25:17.0890768Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_True_initial_xblock_2_add_1dim_False SKIPPED [0.0029s] (Skip non-critical tests to save resources.) [ 8%] 2025-12-04T12:25:17.0891817Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_2_add_1dim_True SKIPPED [0.0035s] (Skip non-critical tests to save resources.) [ 12%] 2025-12-04T12:25:17.0892846Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_True_initial_xblock_1_add_1dim_True SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 16%] 2025-12-04T12:25:17.0893776Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6342s] [ 20%] 2025-12-04T12:25:17.0894694Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1622s] [ 20%] 2025-12-04T12:25:17.0895535Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1589s] [ 20%] 2025-12-04T12:25:17.0895540Z 2025-12-04T12:25:17.0895681Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0896470Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0896608Z Traceback (most recent call last): 2025-12-04T12:25:17.0897135Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0897347Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0897557Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0897563Z 2025-12-04T12:25:17.0897775Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0898744Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0898749Z 2025-12-04T12:25:17.0899019Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0899248Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0899363Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0899477Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0899832Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0900050Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0900163Z graph_break [] 2025-12-04T12:25:17.0900378Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0901111Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0901224Z warnings.warn( 2025-12-04T12:25:17.0901795Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0901918Z Traceback (most recent call last): 2025-12-04T12:25:17.0902488Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0902685Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0902951Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0902956Z 2025-12-04T12:25:17.0903171Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0904126Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0904132Z 2025-12-04T12:25:17.0904410Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0904627Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0904751Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0904870Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0905208Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0905438Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0905537Z graph_break [] 2025-12-04T12:25:17.0905753Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0906490Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0906589Z warnings.warn( 2025-12-04T12:25:17.0906809Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0906915Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0907027Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0907288Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0907624Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0907722Z graph_break [] 2025-12-04T12:25:17.0907940Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0908667Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0908777Z warnings.warn( 2025-12-04T12:25:17.0908920Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0909489Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0909616Z Traceback (most recent call last): 2025-12-04T12:25:17.0910086Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0910277Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0910500Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0910505Z 2025-12-04T12:25:17.0910714Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0911691Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0911697Z 2025-12-04T12:25:17.0911960Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0912183Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0912293Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0912406Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0912756Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0913009Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0913137Z graph_break [] 2025-12-04T12:25:17.0913358Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0914118Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0914226Z warnings.warn( 2025-12-04T12:25:17.0914441Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0914549Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0914669Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0914887Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0915224Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0915334Z graph_break [] 2025-12-04T12:25:17.0915546Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0916277Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0916389Z warnings.warn( 2025-12-04T12:25:17.0916602Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0916723Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0916833Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0917049Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0917394Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0917489Z graph_break [] 2025-12-04T12:25:17.0917699Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0918469Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0918573Z warnings.warn( 2025-12-04T12:25:17.0919405Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c5a3e4c540d200e7.xml - 2025-12-04T12:25:17.0919574Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0920663Z FAILED [0.1589s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0920680Z 2025-12-04T12:25:17.0920894Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0921852Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0921862Z 2025-12-04T12:25:17.0922133Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0922314Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0922538Z ============ 1 failed, 4 skipped, 181 deselected, 2 rerun in 5.03s ============= 2025-12-04T12:25:17.0922635Z Got exit code 1 2025-12-04T12:25:17.0922742Z Retrying single test... 2025-12-04T12:25:17.0923395Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ab8ed79c5f08a6c3.xml 2025-12-04T12:25:17.0923556Z ============================= test session starts ============================== 2025-12-04T12:25:17.0923905Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0924024Z cachedir: .pytest_cache 2025-12-04T12:25:17.0924548Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0924743Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0924852Z configfile: pytest.ini 2025-12-04T12:25:17.0925493Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0925729Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.0926781Z stepcurrent: skipping 185 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0926906Z Running 1 items in this shard 2025-12-04T12:25:17.0926912Z 2025-12-04T12:25:17.0927834Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6609s] [100%] 2025-12-04T12:25:17.0928757Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1650s] [100%] 2025-12-04T12:25:17.0929601Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1588s] [100%] 2025-12-04T12:25:17.0929606Z 2025-12-04T12:25:17.0929745Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0930324Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0930446Z Traceback (most recent call last): 2025-12-04T12:25:17.0930948Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0931161Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0931373Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0931378Z 2025-12-04T12:25:17.0931602Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0932556Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0932561Z 2025-12-04T12:25:17.0932826Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0933052Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0933162Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0933286Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0933630Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0933850Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0933958Z graph_break [] 2025-12-04T12:25:17.0934172Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0934909Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0935019Z warnings.warn( 2025-12-04T12:25:17.0935584Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0935713Z Traceback (most recent call last): 2025-12-04T12:25:17.0936184Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0936380Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0936658Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0936664Z 2025-12-04T12:25:17.0936942Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0937950Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0937955Z 2025-12-04T12:25:17.0938221Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0938438Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0938560Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0938673Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0939023Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0939240Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0939339Z graph_break [] 2025-12-04T12:25:17.0939562Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0940294Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0940396Z warnings.warn( 2025-12-04T12:25:17.0940617Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0940725Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0940839Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0941068Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0941406Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0941512Z graph_break [] 2025-12-04T12:25:17.0941755Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0942491Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0942599Z warnings.warn( 2025-12-04T12:25:17.0942744Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0943322Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0943443Z Traceback (most recent call last): 2025-12-04T12:25:17.0943916Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0944120Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0944330Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0944335Z 2025-12-04T12:25:17.0944546Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0945526Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0945533Z 2025-12-04T12:25:17.0945797Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0946021Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0946130Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0946242Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0946592Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0946806Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0946911Z graph_break [] 2025-12-04T12:25:17.0947126Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0947963Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0948107Z warnings.warn( 2025-12-04T12:25:17.0948352Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0948462Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0948588Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0948806Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0949154Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0949252Z graph_break [] 2025-12-04T12:25:17.0949466Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0950208Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0950314Z warnings.warn( 2025-12-04T12:25:17.0950528Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0950651Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0950768Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0951001Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0951344Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0951442Z graph_break [] 2025-12-04T12:25:17.0951667Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0952398Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0952497Z warnings.warn( 2025-12-04T12:25:17.0953364Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ab8ed79c5f08a6c3.xml - 2025-12-04T12:25:17.0953543Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0954644Z FAILED [0.1588s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0954650Z 2025-12-04T12:25:17.0954868Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0955833Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0955839Z 2025-12-04T12:25:17.0956105Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0956285Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0956502Z ================== 1 failed, 204 deselected, 2 rerun in 5.04s ================== 2025-12-04T12:25:17.0956601Z Got exit code 1 2025-12-04T12:25:17.0956710Z Retrying single test... 2025-12-04T12:25:17.0957367Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-43cb3996049bee78.xml 2025-12-04T12:25:17.0957529Z ============================= test session starts ============================== 2025-12-04T12:25:17.0957888Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0957998Z cachedir: .pytest_cache 2025-12-04T12:25:17.0958520Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0958658Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0958768Z configfile: pytest.ini 2025-12-04T12:25:17.0959423Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0959687Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.0960737Z stepcurrent: skipping 185 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0960858Z Running 1 items in this shard 2025-12-04T12:25:17.0960863Z 2025-12-04T12:25:17.0961777Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6760s] [100%] 2025-12-04T12:25:17.0962704Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1673s] [100%] 2025-12-04T12:25:17.0963544Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1621s] [100%] 2025-12-04T12:25:17.0963549Z 2025-12-04T12:25:17.0963689Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0964264Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0964386Z Traceback (most recent call last): 2025-12-04T12:25:17.0964868Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0965064Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0965311Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0965316Z 2025-12-04T12:25:17.0965543Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0966506Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0966511Z 2025-12-04T12:25:17.0966788Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0967002Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0967113Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0967237Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0967580Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0967810Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0967907Z graph_break [] 2025-12-04T12:25:17.0968122Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0968869Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0968968Z warnings.warn( 2025-12-04T12:25:17.0969536Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0969666Z Traceback (most recent call last): 2025-12-04T12:25:17.0970132Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0970339Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0970550Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0970555Z 2025-12-04T12:25:17.0970769Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0971744Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0971834Z 2025-12-04T12:25:17.0972130Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0972358Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0972468Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0972583Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0972934Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0973150Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0973246Z graph_break [] 2025-12-04T12:25:17.0973468Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0974198Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0974310Z warnings.warn( 2025-12-04T12:25:17.0974522Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0974632Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0974755Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0974973Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0975309Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0975416Z graph_break [] 2025-12-04T12:25:17.0975630Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0976395Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0976494Z warnings.warn( 2025-12-04T12:25:17.0976642Z =================================== FAILURES =================================== 2025-12-04T12:25:17.0977296Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.0977419Z Traceback (most recent call last): 2025-12-04T12:25:17.0977885Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0978093Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0978302Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0978308Z 2025-12-04T12:25:17.0978526Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0979489Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0979498Z 2025-12-04T12:25:17.0979772Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0979991Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0980104Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0980231Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0980569Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0980786Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0980894Z graph_break [] 2025-12-04T12:25:17.0981106Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0981842Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0981942Z warnings.warn( 2025-12-04T12:25:17.0982154Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0982338Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0982452Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0982701Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0983052Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0983145Z graph_break [] 2025-12-04T12:25:17.0983356Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0984090Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0984189Z warnings.warn( 2025-12-04T12:25:17.0984410Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.0984518Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.0984630Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.0984867Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.0985200Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.0985299Z graph_break [] 2025-12-04T12:25:17.0985522Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.0986248Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.0986361Z warnings.warn( 2025-12-04T12:25:17.0987181Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-43cb3996049bee78.xml - 2025-12-04T12:25:17.0987377Z =========================== short test summary info ============================ 2025-12-04T12:25:17.0988479Z FAILED [0.1621s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0988488Z 2025-12-04T12:25:17.0988702Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.0989666Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0989672Z 2025-12-04T12:25:17.0989932Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.0990119Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.0990315Z ================== 1 failed, 204 deselected, 2 rerun in 5.06s ================== 2025-12-04T12:25:17.0990418Z Got exit code 1 2025-12-04T12:25:17.0991304Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.0991717Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:17.0992358Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d3a2df1a150e9085.xml 2025-12-04T12:25:17.0992527Z ============================= test session starts ============================== 2025-12-04T12:25:17.0992874Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.0992994Z cachedir: .pytest_cache 2025-12-04T12:25:17.0993511Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.0993635Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.0993814Z configfile: pytest.ini 2025-12-04T12:25:17.0994406Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.0994674Z collecting ... collected 380 items / 186 deselected / 194 selected 2025-12-04T12:25:17.0994825Z stepcurrent: skipping 186 already run items. 2025-12-04T12:25:17.0994938Z Running 19 items in this shard 2025-12-04T12:25:17.0994943Z 2025-12-04T12:25:17.0995870Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6757s] [ 5%] 2025-12-04T12:25:17.0997109Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1629s] [ 5%] 2025-12-04T12:25:17.0997966Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1604s] [ 5%] 2025-12-04T12:25:17.0997976Z 2025-12-04T12:25:17.0998116Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.0998684Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.0998814Z Traceback (most recent call last): 2025-12-04T12:25:17.0999284Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.0999492Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.0999709Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.0999715Z 2025-12-04T12:25:17.1000011Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1000985Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.1000995Z 2025-12-04T12:25:17.1001261Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1001490Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1001602Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1001718Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1002076Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1002296Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1002393Z graph_break [] 2025-12-04T12:25:17.1002618Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1003355Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1003468Z warnings.warn( 2025-12-04T12:25:17.1004039Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.1004160Z Traceback (most recent call last): 2025-12-04T12:25:17.1004638Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1004834Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1005054Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1005059Z 2025-12-04T12:25:17.1005270Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1006232Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.1006325Z 2025-12-04T12:25:17.1006600Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1006858Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1006984Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1007098Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1007439Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1007664Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1007761Z graph_break [] 2025-12-04T12:25:17.1007976Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1008725Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1008831Z warnings.warn( 2025-12-04T12:25:17.1009054Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1009163Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1009273Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1009506Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1009841Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1009933Z graph_break [] 2025-12-04T12:25:17.1010158Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1010893Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1011008Z warnings.warn( 2025-12-04T12:25:17.1011179Z =================================== FAILURES =================================== 2025-12-04T12:25:17.1011751Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.1011891Z Traceback (most recent call last): 2025-12-04T12:25:17.1012363Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1012560Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1012785Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1012791Z 2025-12-04T12:25:17.1013004Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1013974Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.1013979Z 2025-12-04T12:25:17.1014250Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1014470Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1014593Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1014706Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1015062Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1015284Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1015383Z graph_break [] 2025-12-04T12:25:17.1015614Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1016349Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1016450Z warnings.warn( 2025-12-04T12:25:17.1016677Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1016790Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1017073Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1017293Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1017663Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1017777Z graph_break [] 2025-12-04T12:25:17.1017994Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1018722Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1018836Z warnings.warn( 2025-12-04T12:25:17.1019054Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1019180Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1019293Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1019516Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1019871Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1019968Z graph_break [] 2025-12-04T12:25:17.1020184Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1020928Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1021028Z warnings.warn( 2025-12-04T12:25:17.1021860Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d3a2df1a150e9085.xml - 2025-12-04T12:25:17.1022031Z =========================== short test summary info ============================ 2025-12-04T12:25:17.1023161Z FAILED [0.1604s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1023186Z 2025-12-04T12:25:17.1023404Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1024361Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.1024367Z 2025-12-04T12:25:17.1024642Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1024821Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.1025034Z ================== 1 failed, 186 deselected, 2 rerun in 5.05s ================== 2025-12-04T12:25:17.1025131Z Got exit code 1 2025-12-04T12:25:17.1025236Z Retrying single test... 2025-12-04T12:25:17.1025899Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2b2d696e555e6fc9.xml 2025-12-04T12:25:17.1026064Z ============================= test session starts ============================== 2025-12-04T12:25:17.1026415Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.1026535Z cachedir: .pytest_cache 2025-12-04T12:25:17.1027057Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.1027192Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.1027301Z configfile: pytest.ini 2025-12-04T12:25:17.1027891Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.1028130Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.1029187Z stepcurrent: skipping 186 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.1029368Z Running 1 items in this shard 2025-12-04T12:25:17.1029385Z 2025-12-04T12:25:17.1030346Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6523s] [100%] 2025-12-04T12:25:17.1031268Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1653s] [100%] 2025-12-04T12:25:17.1032114Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1600s] [100%] 2025-12-04T12:25:17.1032122Z 2025-12-04T12:25:17.1032260Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.1032840Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.1032963Z Traceback (most recent call last): 2025-12-04T12:25:17.1033434Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1033639Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1033853Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1033858Z 2025-12-04T12:25:17.1034084Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1035073Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.1035078Z 2025-12-04T12:25:17.1035346Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1035577Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1035689Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1035817Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1036157Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1036375Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1036483Z graph_break [] 2025-12-04T12:25:17.1036696Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1037428Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1037541Z warnings.warn( 2025-12-04T12:25:17.1038111Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.1038247Z Traceback (most recent call last): 2025-12-04T12:25:17.1038718Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1038916Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1039139Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1039144Z 2025-12-04T12:25:17.1039353Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1040322Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.1040327Z 2025-12-04T12:25:17.1040591Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1040806Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1040989Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1041104Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1041476Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1041708Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1041806Z graph_break [] 2025-12-04T12:25:17.1042032Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1042761Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1042862Z warnings.warn( 2025-12-04T12:25:17.1043088Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1043200Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1043313Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1043547Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1043885Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1043996Z graph_break [] 2025-12-04T12:25:17.1044209Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1044932Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1045044Z warnings.warn( 2025-12-04T12:25:17.1045186Z =================================== FAILURES =================================== 2025-12-04T12:25:17.1045760Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.1045911Z Traceback (most recent call last): 2025-12-04T12:25:17.1046384Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1046594Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1046808Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1046813Z 2025-12-04T12:25:17.1047025Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1047996Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.1048001Z 2025-12-04T12:25:17.1048266Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1048489Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1048599Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1048714Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1049070Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1049290Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1049397Z graph_break [] 2025-12-04T12:25:17.1049615Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1050345Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1050458Z warnings.warn( 2025-12-04T12:25:17.1050670Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1050801Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1050930Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1051150Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1051500Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1051661Z graph_break [] 2025-12-04T12:25:17.1051873Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1052644Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1052745Z warnings.warn( 2025-12-04T12:25:17.1052957Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1053077Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1053189Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1053419Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1053752Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1053845Z graph_break [] 2025-12-04T12:25:17.1054073Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1054804Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1054903Z warnings.warn( 2025-12-04T12:25:17.1055733Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2b2d696e555e6fc9.xml - 2025-12-04T12:25:17.1055902Z =========================== short test summary info ============================ 2025-12-04T12:25:17.1057080Z FAILED [0.1600s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1057088Z 2025-12-04T12:25:17.1057356Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1058311Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.1058338Z 2025-12-04T12:25:17.1058605Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1058782Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.1058995Z ================== 1 failed, 204 deselected, 2 rerun in 5.03s ================== 2025-12-04T12:25:17.1059090Z Got exit code 1 2025-12-04T12:25:17.1059193Z Retrying single test... 2025-12-04T12:25:17.1059858Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-449e4fae7bbefa8d.xml 2025-12-04T12:25:17.1060017Z ============================= test session starts ============================== 2025-12-04T12:25:17.1060380Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.1060494Z cachedir: .pytest_cache 2025-12-04T12:25:17.1061014Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.1061149Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.1061254Z configfile: pytest.ini 2025-12-04T12:25:17.1061848Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.1062085Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.1063133Z stepcurrent: skipping 186 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.1063260Z Running 1 items in this shard 2025-12-04T12:25:17.1063265Z 2025-12-04T12:25:17.1064209Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [4.6850s] [100%] 2025-12-04T12:25:17.1065204Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False ('RERUN', {'yellow': True}) [0.1668s] [100%] 2025-12-04T12:25:17.1066041Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False FAILED [0.1628s] [100%] 2025-12-04T12:25:17.1066046Z 2025-12-04T12:25:17.1066184Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.1066768Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.1066889Z Traceback (most recent call last): 2025-12-04T12:25:17.1067372Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1067570Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1067784Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1067789Z 2025-12-04T12:25:17.1068012Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1068976Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.1068981Z 2025-12-04T12:25:17.1069257Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1069502Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1069614Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1069743Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1070081Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1070313Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1070408Z graph_break [] 2025-12-04T12:25:17.1070626Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1071374Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1071477Z warnings.warn( 2025-12-04T12:25:17.1072044Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.1072176Z Traceback (most recent call last): 2025-12-04T12:25:17.1072647Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1072863Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1073073Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1073079Z 2025-12-04T12:25:17.1073337Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1074469Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.1074475Z 2025-12-04T12:25:17.1074835Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1075066Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1075179Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1075292Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1075649Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1075948Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1076045Z graph_break [] 2025-12-04T12:25:17.1076308Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1077046Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1077160Z warnings.warn( 2025-12-04T12:25:17.1077376Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1077486Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1077614Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1077832Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1078170Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1078289Z graph_break [] 2025-12-04T12:25:17.1078502Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1079249Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1079352Z warnings.warn( 2025-12-04T12:25:17.1079499Z =================================== FAILURES =================================== 2025-12-04T12:25:17.1080085Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False _ 2025-12-04T12:25:17.1080206Z Traceback (most recent call last): 2025-12-04T12:25:17.1080676Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1080890Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1081133Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1081143Z 2025-12-04T12:25:17.1081374Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1082334Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.1082340Z 2025-12-04T12:25:17.1082621Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1082838Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1082948Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1083076Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1083420Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1083642Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1083756Z graph_break [] 2025-12-04T12:25:17.1083977Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1084721Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1084822Z warnings.warn( 2025-12-04T12:25:17.1085038Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1085158Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1085269Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1085490Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1085835Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1085931Z graph_break [] 2025-12-04T12:25:17.1086143Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1086881Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1087044Z warnings.warn( 2025-12-04T12:25:17.1087295Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1087407Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1087519Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1087744Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1088083Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1088180Z graph_break [] 2025-12-04T12:25:17.1088405Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1089129Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1089240Z warnings.warn( 2025-12-04T12:25:17.1090069Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-449e4fae7bbefa8d.xml - 2025-12-04T12:25:17.1090242Z =========================== short test summary info ============================ 2025-12-04T12:25:17.1091345Z FAILED [0.1628s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1091351Z 2025-12-04T12:25:17.1091565Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1092569Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.1092574Z 2025-12-04T12:25:17.1092844Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1093036Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.1093236Z ================== 1 failed, 204 deselected, 2 rerun in 5.07s ================== 2025-12-04T12:25:17.1093332Z Got exit code 1 2025-12-04T12:25:17.1094212Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False 2025-12-04T12:25:17.1094623Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:17.1095264Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4631de08a3dd09e1.xml 2025-12-04T12:25:17.1095439Z ============================= test session starts ============================== 2025-12-04T12:25:17.1095786Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.1095909Z cachedir: .pytest_cache 2025-12-04T12:25:17.1096592Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.1096716Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.1096907Z configfile: pytest.ini 2025-12-04T12:25:17.1097506Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.1097742Z collecting ... collected 380 items / 187 deselected / 193 selected 2025-12-04T12:25:17.1097891Z stepcurrent: skipping 187 already run items. 2025-12-04T12:25:17.1098005Z Running 18 items in this shard 2025-12-04T12:25:17.1098010Z 2025-12-04T12:25:17.1099066Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_True_initial_xblock_1_add_1dim_False SKIPPED [0.0041s] (Skip non-critical tests to save resources.) [ 5%] 2025-12-04T12:25:17.1100263Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_True_initial_xblock_2_add_1dim_False SKIPPED [0.0030s] (Skip non-critical tests to save resources.) [ 11%] 2025-12-04T12:25:17.1101296Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_True_initial_xblock_2_add_1dim_True SKIPPED [0.0036s] (Skip non-critical tests to save resources.) [ 16%] 2025-12-04T12:25:17.1102215Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6342s] [ 22%] 2025-12-04T12:25:17.1103144Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1598s] [ 22%] 2025-12-04T12:25:17.1103988Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1560s] [ 22%] 2025-12-04T12:25:17.1103993Z 2025-12-04T12:25:17.1104132Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.1104716Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.1104838Z Traceback (most recent call last): 2025-12-04T12:25:17.1105323Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1105523Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1105780Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1105791Z 2025-12-04T12:25:17.1106023Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1106987Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1106993Z 2025-12-04T12:25:17.1107272Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1107490Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1107601Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1107727Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1108068Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1108283Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1108396Z graph_break [] 2025-12-04T12:25:17.1108611Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1109370Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1109470Z warnings.warn( 2025-12-04T12:25:17.1110041Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.1110173Z Traceback (most recent call last): 2025-12-04T12:25:17.1110640Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1110846Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1111054Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1111060Z 2025-12-04T12:25:17.1111271Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1112273Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1112307Z 2025-12-04T12:25:17.1112604Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1112833Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1112944Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1113058Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1113410Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1113627Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1113725Z graph_break [] 2025-12-04T12:25:17.1113951Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1114683Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1114798Z warnings.warn( 2025-12-04T12:25:17.1115011Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1115121Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1115251Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1115470Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1115811Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1115915Z graph_break [] 2025-12-04T12:25:17.1116129Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1116900Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1116998Z warnings.warn( 2025-12-04T12:25:17.1117142Z =================================== FAILURES =================================== 2025-12-04T12:25:17.1117721Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.1117842Z Traceback (most recent call last): 2025-12-04T12:25:17.1118314Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1118518Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1118726Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1118731Z 2025-12-04T12:25:17.1118952Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1119909Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1119918Z 2025-12-04T12:25:17.1120180Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1120409Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1120519Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1120643Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1120979Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1121197Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1121304Z graph_break [] 2025-12-04T12:25:17.1121516Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1122245Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1122352Z warnings.warn( 2025-12-04T12:25:17.1122631Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1122753Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1122867Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1123116Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1123463Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1123564Z graph_break [] 2025-12-04T12:25:17.1123781Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1124522Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1124622Z warnings.warn( 2025-12-04T12:25:17.1124848Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1124959Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1125077Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1125304Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1125643Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1125740Z graph_break [] 2025-12-04T12:25:17.1125963Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1126688Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1126799Z warnings.warn( 2025-12-04T12:25:17.1127614Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4631de08a3dd09e1.xml - 2025-12-04T12:25:17.1127876Z =========================== short test summary info ============================ 2025-12-04T12:25:17.1128977Z FAILED [0.1560s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1128989Z 2025-12-04T12:25:17.1129203Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1130166Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1130171Z 2025-12-04T12:25:17.1130434Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1130611Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.1130845Z ============ 1 failed, 3 skipped, 187 deselected, 2 rerun in 5.02s ============= 2025-12-04T12:25:17.1130944Z Got exit code 1 2025-12-04T12:25:17.1131065Z Retrying single test... 2025-12-04T12:25:17.1131705Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-7cb60f50c6534a73.xml 2025-12-04T12:25:17.1131871Z ============================= test session starts ============================== 2025-12-04T12:25:17.1132231Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.1132346Z cachedir: .pytest_cache 2025-12-04T12:25:17.1132873Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.1133011Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.1133120Z configfile: pytest.ini 2025-12-04T12:25:17.1133728Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.1133959Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.1135098Z stepcurrent: skipping 190 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1135230Z Running 1 items in this shard 2025-12-04T12:25:17.1135236Z 2025-12-04T12:25:17.1136158Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6457s] [100%] 2025-12-04T12:25:17.1137172Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1629s] [100%] 2025-12-04T12:25:17.1138013Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1599s] [100%] 2025-12-04T12:25:17.1138023Z 2025-12-04T12:25:17.1138178Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.1138752Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.1138875Z Traceback (most recent call last): 2025-12-04T12:25:17.1139363Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1139565Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1139793Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1139798Z 2025-12-04T12:25:17.1140013Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1141012Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1141025Z 2025-12-04T12:25:17.1141305Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1141528Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1141656Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1141773Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1142113Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1142345Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1142444Z graph_break [] 2025-12-04T12:25:17.1142664Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1143420Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1143525Z warnings.warn( 2025-12-04T12:25:17.1144108Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.1144231Z Traceback (most recent call last): 2025-12-04T12:25:17.1144697Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1144910Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1145121Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1145127Z 2025-12-04T12:25:17.1145355Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1146311Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1146317Z 2025-12-04T12:25:17.1146661Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1146896Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1147037Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1147156Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1147513Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1147731Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1147848Z graph_break [] 2025-12-04T12:25:17.1148063Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1148798Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1148915Z warnings.warn( 2025-12-04T12:25:17.1149130Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1149244Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1149368Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1149589Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1149940Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1150037Z graph_break [] 2025-12-04T12:25:17.1150251Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1150989Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1151089Z warnings.warn( 2025-12-04T12:25:17.1151231Z =================================== FAILURES =================================== 2025-12-04T12:25:17.1151846Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.1151972Z Traceback (most recent call last): 2025-12-04T12:25:17.1152453Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1152652Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1152861Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1152866Z 2025-12-04T12:25:17.1153095Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1154052Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1154058Z 2025-12-04T12:25:17.1154333Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1154550Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1154665Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1154793Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1155135Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1155362Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1155460Z graph_break [] 2025-12-04T12:25:17.1155674Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1156421Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1156519Z warnings.warn( 2025-12-04T12:25:17.1156734Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1156855Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1156968Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1157185Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1157592Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1157690Z graph_break [] 2025-12-04T12:25:17.1157942Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1158668Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1158767Z warnings.warn( 2025-12-04T12:25:17.1158992Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1159100Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1159212Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1159440Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1159780Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1159896Z graph_break [] 2025-12-04T12:25:17.1160111Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1160834Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1160946Z warnings.warn( 2025-12-04T12:25:17.1161760Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-7cb60f50c6534a73.xml - 2025-12-04T12:25:17.1161944Z =========================== short test summary info ============================ 2025-12-04T12:25:17.1163061Z FAILED [0.1599s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1163072Z 2025-12-04T12:25:17.1163290Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1164257Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1164262Z 2025-12-04T12:25:17.1164528Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1164719Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.1164919Z ================== 1 failed, 204 deselected, 2 rerun in 5.02s ================== 2025-12-04T12:25:17.1165016Z Got exit code 1 2025-12-04T12:25:17.1165137Z Retrying single test... 2025-12-04T12:25:17.1165783Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-bbe75c180c40946f.xml 2025-12-04T12:25:17.1165956Z ============================= test session starts ============================== 2025-12-04T12:25:17.1166307Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.1166417Z cachedir: .pytest_cache 2025-12-04T12:25:17.1166955Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.1167078Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.1167188Z configfile: pytest.ini 2025-12-04T12:25:17.1167794Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.1168020Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.1169087Z stepcurrent: skipping 190 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1169260Z Running 1 items in this shard 2025-12-04T12:25:17.1169265Z 2025-12-04T12:25:17.1170214Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6339s] [100%] 2025-12-04T12:25:17.1171147Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1660s] [100%] 2025-12-04T12:25:17.1171980Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1618s] [100%] 2025-12-04T12:25:17.1171986Z 2025-12-04T12:25:17.1172142Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.1172709Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.1172848Z Traceback (most recent call last): 2025-12-04T12:25:17.1173321Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1173518Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1173740Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1173745Z 2025-12-04T12:25:17.1173960Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1174927Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1174933Z 2025-12-04T12:25:17.1175229Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1175450Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1175573Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1175688Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1176033Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1176264Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1176361Z graph_break [] 2025-12-04T12:25:17.1176590Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1177400Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1177505Z warnings.warn( 2025-12-04T12:25:17.1178092Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.1178217Z Traceback (most recent call last): 2025-12-04T12:25:17.1178704Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1178904Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1179114Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1179119Z 2025-12-04T12:25:17.1179345Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1180297Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1180302Z 2025-12-04T12:25:17.1180580Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1180799Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1180909Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1181103Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1181445Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1181690Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1181800Z graph_break [] 2025-12-04T12:25:17.1182015Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1182762Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1182864Z warnings.warn( 2025-12-04T12:25:17.1183077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1183207Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1183322Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1183541Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1183893Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1183993Z graph_break [] 2025-12-04T12:25:17.1184224Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1184950Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1185051Z warnings.warn( 2025-12-04T12:25:17.1185207Z =================================== FAILURES =================================== 2025-12-04T12:25:17.1185770Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.1185892Z Traceback (most recent call last): 2025-12-04T12:25:17.1186406Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1186611Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1186837Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1186842Z 2025-12-04T12:25:17.1187057Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1188010Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1188029Z 2025-12-04T12:25:17.1188293Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1188507Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1188632Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1188746Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1189085Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1189318Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1189416Z graph_break [] 2025-12-04T12:25:17.1189631Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1190378Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1190476Z warnings.warn( 2025-12-04T12:25:17.1190701Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1190810Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1190924Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1191153Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1191492Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1191588Z graph_break [] 2025-12-04T12:25:17.1191900Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1192660Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1192775Z warnings.warn( 2025-12-04T12:25:17.1192991Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1193104Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1193232Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1193453Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1193789Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1193902Z graph_break [] 2025-12-04T12:25:17.1194113Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1194863Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1194968Z warnings.warn( 2025-12-04T12:25:17.1195789Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-bbe75c180c40946f.xml - 2025-12-04T12:25:17.1196134Z =========================== short test summary info ============================ 2025-12-04T12:25:17.1197227Z FAILED [0.1618s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1197233Z 2025-12-04T12:25:17.1197462Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1198476Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1198486Z 2025-12-04T12:25:17.1198753Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1198949Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.1199148Z ================== 1 failed, 204 deselected, 2 rerun in 5.02s ================== 2025-12-04T12:25:17.1199264Z Got exit code 1 2025-12-04T12:25:17.1200140Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1200551Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:17.1201216Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-19b98f6e1bce5088.xml 2025-12-04T12:25:17.1201384Z ============================= test session starts ============================== 2025-12-04T12:25:17.1201748Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.1201864Z cachedir: .pytest_cache 2025-12-04T12:25:17.1202387Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.1202527Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.1202636Z configfile: pytest.ini 2025-12-04T12:25:17.1203233Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.1203476Z collecting ... collected 380 items / 191 deselected / 189 selected 2025-12-04T12:25:17.1203626Z stepcurrent: skipping 191 already run items. 2025-12-04T12:25:17.1203759Z Running 14 items in this shard 2025-12-04T12:25:17.1203764Z 2025-12-04T12:25:17.1204925Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_True SKIPPED [0.0040s] (Skip non-critical tests to save resources.) [ 7%] 2025-12-04T12:25:17.1205973Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_True_initial_xblock_2_add_1dim_False SKIPPED [0.0030s] (Skip non-critical tests to save resources.) [ 14%] 2025-12-04T12:25:17.1207000Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_True_initial_xblock_2_add_1dim_True SKIPPED [0.0037s] (Skip non-critical tests to save resources.) [ 21%] 2025-12-04T12:25:17.1208030Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape2_max_autotune_True_initial_xblock_1_add_1dim_False SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 28%] 2025-12-04T12:25:17.1209075Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape2_max_autotune_True_initial_xblock_1_add_1dim_True SKIPPED [0.0028s] (Skip non-critical tests to save resources.) [ 35%] 2025-12-04T12:25:17.1210102Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape2_max_autotune_True_initial_xblock_2_add_1dim_False SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 42%] 2025-12-04T12:25:17.1211136Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape2_max_autotune_True_initial_xblock_2_add_1dim_True SKIPPED [0.0027s] (Skip non-critical tests to save resources.) [ 50%] 2025-12-04T12:25:17.1212084Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6456s] [ 57%] 2025-12-04T12:25:17.1213022Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1628s] [ 57%] 2025-12-04T12:25:17.1213858Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1575s] [ 57%] 2025-12-04T12:25:17.1213864Z 2025-12-04T12:25:17.1214015Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.1214585Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.1214709Z Traceback (most recent call last): 2025-12-04T12:25:17.1215190Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1215391Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1215610Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1215616Z 2025-12-04T12:25:17.1215828Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1216786Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1216792Z 2025-12-04T12:25:17.1217129Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1217350Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1217474Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1217586Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1217927Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1218224Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1218321Z graph_break [] 2025-12-04T12:25:17.1218575Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1219321Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1219422Z warnings.warn( 2025-12-04T12:25:17.1220007Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.1220129Z Traceback (most recent call last): 2025-12-04T12:25:17.1220596Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1220807Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1221012Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1221022Z 2025-12-04T12:25:17.1221232Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1222203Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1222209Z 2025-12-04T12:25:17.1222477Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1222705Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1222815Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1222930Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1223277Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1223527Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1223642Z graph_break [] 2025-12-04T12:25:17.1223859Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1224600Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1224713Z warnings.warn( 2025-12-04T12:25:17.1224925Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1225036Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1225160Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1225375Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1225724Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1225821Z graph_break [] 2025-12-04T12:25:17.1226037Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1226776Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1226879Z warnings.warn( 2025-12-04T12:25:17.1227022Z =================================== FAILURES =================================== 2025-12-04T12:25:17.1227601Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.1227721Z Traceback (most recent call last): 2025-12-04T12:25:17.1228205Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1228442Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1228736Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1228742Z 2025-12-04T12:25:17.1228973Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1230001Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1230006Z 2025-12-04T12:25:17.1230311Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1230529Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1230641Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1230769Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1231110Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1231342Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1231442Z graph_break [] 2025-12-04T12:25:17.1231661Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1232410Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1232516Z warnings.warn( 2025-12-04T12:25:17.1232731Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1232853Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1232967Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1233186Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1233532Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1233631Z graph_break [] 2025-12-04T12:25:17.1233858Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1234627Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1234733Z warnings.warn( 2025-12-04T12:25:17.1234960Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1235069Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1235182Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1235413Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1235751Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1235861Z graph_break [] 2025-12-04T12:25:17.1236074Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1236802Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1236914Z warnings.warn( 2025-12-04T12:25:17.1237744Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-19b98f6e1bce5088.xml - 2025-12-04T12:25:17.1237931Z =========================== short test summary info ============================ 2025-12-04T12:25:17.1239021Z FAILED [0.1575s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1239027Z 2025-12-04T12:25:17.1239242Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1240211Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1240217Z 2025-12-04T12:25:17.1240483Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1240677Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.1240953Z ============ 1 failed, 7 skipped, 191 deselected, 2 rerun in 5.05s ============= 2025-12-04T12:25:17.1241050Z Got exit code 1 2025-12-04T12:25:17.1241170Z Retrying single test... 2025-12-04T12:25:17.1241845Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-fd7621eb24f73b8c.xml 2025-12-04T12:25:17.1242018Z ============================= test session starts ============================== 2025-12-04T12:25:17.1242365Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.1242474Z cachedir: .pytest_cache 2025-12-04T12:25:17.1243006Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.1243129Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.1243238Z configfile: pytest.ini 2025-12-04T12:25:17.1243839Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.1244068Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.1245122Z stepcurrent: skipping 198 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1245235Z Running 1 items in this shard 2025-12-04T12:25:17.1245240Z 2025-12-04T12:25:17.1246154Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6248s] [100%] 2025-12-04T12:25:17.1247119Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1596s] [100%] 2025-12-04T12:25:17.1247968Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1577s] [100%] 2025-12-04T12:25:17.1247974Z 2025-12-04T12:25:17.1248127Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.1248690Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.1248822Z Traceback (most recent call last): 2025-12-04T12:25:17.1249293Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1249488Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1249713Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1249719Z 2025-12-04T12:25:17.1249930Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1250904Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1250909Z 2025-12-04T12:25:17.1251173Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1251390Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1251514Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1251630Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1251973Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1252207Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1252306Z graph_break [] 2025-12-04T12:25:17.1252534Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1253332Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1253462Z warnings.warn( 2025-12-04T12:25:17.1254042Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.1254162Z Traceback (most recent call last): 2025-12-04T12:25:17.1254646Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1254843Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1255052Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1255058Z 2025-12-04T12:25:17.1255279Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1256232Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1256242Z 2025-12-04T12:25:17.1256521Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1256733Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1256923Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1257055Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1257397Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1257617Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1257727Z graph_break [] 2025-12-04T12:25:17.1257940Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1258723Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1258831Z warnings.warn( 2025-12-04T12:25:17.1259046Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1259179Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1259292Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1259510Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1259857Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1259954Z graph_break [] 2025-12-04T12:25:17.1260166Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1260906Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1261007Z warnings.warn( 2025-12-04T12:25:17.1261160Z =================================== FAILURES =================================== 2025-12-04T12:25:17.1261732Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.1261858Z Traceback (most recent call last): 2025-12-04T12:25:17.1262340Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1262536Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1262757Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1262762Z 2025-12-04T12:25:17.1262973Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1263930Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1263936Z 2025-12-04T12:25:17.1264277Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1264492Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1264649Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1264764Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1265099Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1265329Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1265427Z graph_break [] 2025-12-04T12:25:17.1265643Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1266389Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1266491Z warnings.warn( 2025-12-04T12:25:17.1266719Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1266835Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1266949Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1267188Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1267526Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1267623Z graph_break [] 2025-12-04T12:25:17.1267847Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1268575Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1268686Z warnings.warn( 2025-12-04T12:25:17.1268898Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1269039Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1269164Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1269385Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1269720Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1269830Z graph_break [] 2025-12-04T12:25:17.1270043Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1270777Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1270875Z warnings.warn( 2025-12-04T12:25:17.1271699Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-fd7621eb24f73b8c.xml - 2025-12-04T12:25:17.1271881Z =========================== short test summary info ============================ 2025-12-04T12:25:17.1272968Z FAILED [0.1577s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1272978Z 2025-12-04T12:25:17.1273200Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1274151Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1274156Z 2025-12-04T12:25:17.1274420Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1274606Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.1274801Z ================== 1 failed, 204 deselected, 2 rerun in 5.00s ================== 2025-12-04T12:25:17.1274908Z Got exit code 1 2025-12-04T12:25:17.1275015Z Retrying single test... 2025-12-04T12:25:17.1275719Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e11a9d6da1493a65.xml 2025-12-04T12:25:17.1275935Z ============================= test session starts ============================== 2025-12-04T12:25:17.1276279Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.1276400Z cachedir: .pytest_cache 2025-12-04T12:25:17.1276930Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.1277053Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.1277169Z configfile: pytest.ini 2025-12-04T12:25:17.1277759Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.1277985Z collecting ... collected 380 items / 204 deselected / 176 selected 2025-12-04T12:25:17.1279050Z stepcurrent: skipping 198 already run items. Running only test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1279166Z Running 1 items in this shard 2025-12-04T12:25:17.1279171Z 2025-12-04T12:25:17.1280094Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [4.6671s] [100%] 2025-12-04T12:25:17.1281010Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False ('RERUN', {'yellow': True}) [0.1641s] [100%] 2025-12-04T12:25:17.1281892Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False FAILED [0.1632s] [100%] 2025-12-04T12:25:17.1281903Z 2025-12-04T12:25:17.1282043Z ==================================== RERUNS ==================================== 2025-12-04T12:25:17.1282613Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.1282747Z Traceback (most recent call last): 2025-12-04T12:25:17.1283218Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1283424Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1283632Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1283637Z 2025-12-04T12:25:17.1283847Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1284815Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1284825Z 2025-12-04T12:25:17.1285089Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1285315Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1285425Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1285541Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1285894Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1286112Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1286207Z graph_break [] 2025-12-04T12:25:17.1286433Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1287167Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1287312Z warnings.warn( 2025-12-04T12:25:17.1287904Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.1288025Z Traceback (most recent call last): 2025-12-04T12:25:17.1288529Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1288725Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1288932Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1288946Z 2025-12-04T12:25:17.1289156Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1290116Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1290124Z 2025-12-04T12:25:17.1290397Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1290615Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1290724Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1290849Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1291187Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1291415Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1291511Z graph_break [] 2025-12-04T12:25:17.1291726Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1292471Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1292570Z warnings.warn( 2025-12-04T12:25:17.1292817Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1292943Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1293055Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1293281Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1293615Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1293709Z graph_break [] 2025-12-04T12:25:17.1293928Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1294653Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1294751Z warnings.warn( 2025-12-04T12:25:17.1294900Z =================================== FAILURES =================================== 2025-12-04T12:25:17.1295467Z _ NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False _ 2025-12-04T12:25:17.1295600Z Traceback (most recent call last): 2025-12-04T12:25:17.1296371Z File "/var/lib/jenkins/workspace/test/inductor/test_mix_order_reduction.py", line 346, in test_rms_norm_bwd 2025-12-04T12:25:17.1296573Z act, (_, bwd_wrapper) = utils.run_and_get_code(fwd_bwd, opt_f) 2025-12-04T12:25:17.1296796Z ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1296802Z 2025-12-04T12:25:17.1297101Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1298073Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1298079Z 2025-12-04T12:25:17.1298342Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1298561Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1298685Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1298907Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1299262Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1299521Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1299618Z graph_break [] 2025-12-04T12:25:17.1299846Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1300786Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1300888Z warnings.warn( 2025-12-04T12:25:17.1301110Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1301219Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1301347Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1301571Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1301911Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1302022Z graph_break [] 2025-12-04T12:25:17.1302235Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1302957Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1303069Z warnings.warn( 2025-12-04T12:25:17.1303281Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T12:25:17.1303399Z frames [('total', 1), ('ok', 1)] 2025-12-04T12:25:17.1303511Z stats [('calls_captured', 10)] 2025-12-04T12:25:17.1303729Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('not_ok', 1)] 2025-12-04T12:25:17.1304123Z inductor [('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_miss', 1)] 2025-12-04T12:25:17.1304221Z graph_break [] 2025-12-04T12:25:17.1304437Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T12:25:17.1305175Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/compile_fx.py:2891: UserWarning: Tesla T4 does not support bfloat16 compilation natively, skipping 2025-12-04T12:25:17.1305276Z warnings.warn( 2025-12-04T12:25:17.1306104Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e11a9d6da1493a65.xml - 2025-12-04T12:25:17.1306272Z =========================== short test summary info ============================ 2025-12-04T12:25:17.1307369Z FAILED [0.1632s] inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False - ValueError: not enough values to unpack (expected 2, got 0) 2025-12-04T12:25:17.1307377Z 2025-12-04T12:25:17.1307606Z To execute this test, run the following from the base repo dir: 2025-12-04T12:25:17.1308562Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/inductor/test_mix_order_reduction.py NoMixOrderReductionTest.test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1308568Z 2025-12-04T12:25:17.1308844Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:25:17.1309022Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:25:17.1309217Z ================== 1 failed, 204 deselected, 2 rerun in 5.05s ================== 2025-12-04T12:25:17.1309330Z Got exit code 1 2025-12-04T12:25:17.1310201Z FAILED CONSISTENTLY: test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False 2025-12-04T12:25:17.1310629Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:25:17.1311333Z Test results will be stored in test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d245bc4e7ed2a74d.xml 2025-12-04T12:25:17.1311593Z ============================= test session starts ============================== 2025-12-04T12:25:17.1311961Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:25:17.1312071Z cachedir: .pytest_cache 2025-12-04T12:25:17.1312600Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:25:17.1312722Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:25:17.1312831Z configfile: pytest.ini 2025-12-04T12:25:17.1313436Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:25:17.1313660Z collecting ... collected 380 items / 199 deselected / 181 selected 2025-12-04T12:25:17.1313813Z stepcurrent: skipping 199 already run items. 2025-12-04T12:25:17.1313941Z Running 6 items in this shard 2025-12-04T12:25:17.1313946Z 2025-12-04T12:25:17.1314991Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_True SKIPPED [0.0041s] (Skip non-critical tests to save resources.) [ 16%] 2025-12-04T12:25:17.1316032Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_True_initial_xblock_2_add_1dim_False SKIPPED [0.0029s] (Skip non-critical tests to save resources.) [ 33%] 2025-12-04T12:25:17.1317084Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_True_initial_xblock_2_add_1dim_True SKIPPED [0.0037s] (Skip non-critical tests to save resources.) [ 50%] 2025-12-04T12:25:17.1323055Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_with_dynamic_shape_dynamic_dims0 SKIPPED [0.0028s] (Mix order reduction not enabled) [ 66%] 2025-12-04T12:25:17.1323883Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_with_dynamic_shape_dynamic_dims2 SKIPPED [0.0027s] (Mix order reduction not enabled) [ 83%] 2025-12-04T12:25:17.1324447Z inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_xmask SKIPPED [0.0027s] (Mix order reduction not enabled) [100%] 2025-12-04T12:25:17.1324453Z 2025-12-04T12:25:17.1325287Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d245bc4e7ed2a74d.xml - 2025-12-04T12:25:17.1325474Z ====================== 6 skipped, 199 deselected in 0.08s ====================== 2025-12-04T12:25:17.1363156Z The following tests failed consistently: ['test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0', 'test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape1', 'test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0', 'test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1', 'test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True', 'test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_True', 'test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True', 'test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True', 'test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::MixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_2_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_False_shape0', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape0', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_layer_norm_bwd_with_bias_bfloat16_split_reductions_True_shape1', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_1_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape1_max_autotune_False_initial_xblock_2_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_False_shape3_max_autotune_False_initial_xblock_1_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_True', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape1_max_autotune_False_initial_xblock_2_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape2_max_autotune_False_initial_xblock_2_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_bfloat16_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_1_add_1dim_True', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape0_max_autotune_False_initial_xblock_2_add_1dim_True', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_1_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_False_shape2_max_autotune_False_initial_xblock_2_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_1_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape0_max_autotune_False_initial_xblock_2_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape1_max_autotune_False_initial_xblock_1_add_1dim_False', 'test/inductor/test_mix_order_reduction.py::NoMixOrderReductionTest::test_rms_norm_bwd_float32_split_reductions_True_shape3_max_autotune_False_initial_xblock_1_add_1dim_False'] 2025-12-04T12:25:17.1363337Z 2025-12-04T12:25:17.1363957Z FINISHED PRINTING LOG FILE of inductor/test_mix_order_reduction 2/2 (test/test-reports/inductor.test_mix_order_reduction_2.2_7b65ea7a55b42f88_.log) 2025-12-04T12:25:17.1363963Z 2025-12-04T12:25:17.1364348Z Finished inductor/test_mix_order_reduction 2/2 ... [2025-12-04 12:25:16.462535][11500.15291547], took 44.35min 2025-12-04T12:25:17.1365237Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-fb2ec7b0e0c81fea.xml 2025-12-04T12:25:17.1366150Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6ae6fdb299843870.xml 2025-12-04T12:25:17.1367025Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2362ba98ed8b8900.xml 2025-12-04T12:25:17.1367883Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-52334a8157857cd5.xml 2025-12-04T12:25:17.1368755Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0b589120e3bef198.xml 2025-12-04T12:25:17.1369682Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f9457e8d87a8d0e5.xml 2025-12-04T12:25:17.1370571Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e4f574242b541525.xml 2025-12-04T12:25:17.1371443Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-19baa3ae62072b2f.xml 2025-12-04T12:25:17.1372307Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-debf0ef50304edda.xml 2025-12-04T12:25:17.1373182Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a0af7423e85581ef.xml 2025-12-04T12:25:17.1374055Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6e37efe6d0d7b5f7.xml 2025-12-04T12:25:17.1374924Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-64b9c8b908c45f51.xml 2025-12-04T12:25:17.1375798Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3343e9ff31c5255a.xml 2025-12-04T12:25:17.1376657Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ae334ed847078f28.xml 2025-12-04T12:25:17.1377614Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0eab24e125a2cf7d.xml 2025-12-04T12:25:17.1378485Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6060ecdf9a2577f9.xml 2025-12-04T12:25:17.1379355Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0eb661cc20714db1.xml 2025-12-04T12:25:17.1440941Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-5774b4beb84c0975.xml 2025-12-04T12:25:17.1811974Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-74651aae35dd9844.xml 2025-12-04T12:25:17.2143088Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-7a5a8b9ba5d60164.xml 2025-12-04T12:25:17.2466729Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-26faac4e438f6204.xml 2025-12-04T12:25:17.2784027Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-aa20002b977e7ab4.xml 2025-12-04T12:25:17.3150528Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6316dfd7aad65a93.xml 2025-12-04T12:25:17.3664760Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d1f50caa5a44c0ee.xml 2025-12-04T12:25:17.4008650Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-8edf4f7e001e74f9.xml 2025-12-04T12:25:17.4491376Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-defc51812af7e0b0.xml 2025-12-04T12:25:17.4902622Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0bc167e05b941aff.xml 2025-12-04T12:25:17.5263283Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-98b78c8ed19e0643.xml 2025-12-04T12:25:17.5646591Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-cc69c969909d9707.xml 2025-12-04T12:25:17.6010172Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9977f3a62a4ac5d4.xml 2025-12-04T12:25:17.6362879Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ec39337b11cd8a4d.xml 2025-12-04T12:25:17.6726566Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f9d493fa40deaa3c.xml 2025-12-04T12:25:17.7071987Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-492f33e438c88cbb.xml 2025-12-04T12:25:17.7500237Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-cfe26d7f4e2f142c.xml 2025-12-04T12:25:17.7845565Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-85dd835c0392b89d.xml 2025-12-04T12:25:17.8203293Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d745daffa2211db3.xml 2025-12-04T12:25:17.8565529Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-29d661e244421a0d.xml 2025-12-04T12:25:17.8880229Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6a080fc2c141366f.xml 2025-12-04T12:25:17.9305062Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b727a3d5cb43a100.xml 2025-12-04T12:25:17.9609414Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9820049c2fc9b56d.xml 2025-12-04T12:25:18.0243652Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-bcfbf334a5fb1b19.xml 2025-12-04T12:25:18.0660507Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-55cd5c19b34efcf0.xml 2025-12-04T12:25:18.1040733Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3f17acc66e50cb98.xml 2025-12-04T12:25:18.1370008Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a7cd72d0cd16fd7c.xml 2025-12-04T12:25:18.1700072Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-45e1674e77992a96.xml 2025-12-04T12:25:18.2157594Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-384ea3a6b9159dd5.xml 2025-12-04T12:25:18.2509432Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-40502088a31692e3.xml 2025-12-04T12:25:18.2795292Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-1a1d3c182ee91d42.xml 2025-12-04T12:25:18.3150064Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9bd3fb708dd83987.xml 2025-12-04T12:25:18.3538149Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-991a5e87e1946535.xml 2025-12-04T12:25:18.3917540Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-61bd448d788d4193.xml 2025-12-04T12:25:18.4222294Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0c87f6ca176b74e4.xml 2025-12-04T12:25:18.4668073Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-00981048468ad3fe.xml 2025-12-04T12:25:18.4992366Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-7f9686778c56ad66.xml 2025-12-04T12:25:18.5326143Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dc76d780af365e97.xml 2025-12-04T12:25:18.5662365Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3d712c264776fdc2.xml 2025-12-04T12:25:18.5954066Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4da37e7aafa09519.xml 2025-12-04T12:25:18.6284205Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4f9e658df3b9ce6d.xml 2025-12-04T12:25:18.6595889Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3d80a9d76d3be932.xml 2025-12-04T12:25:18.6912638Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ede94f48d428153a.xml 2025-12-04T12:25:18.7233878Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6ae027693c5dbbf4.xml 2025-12-04T12:25:18.7559120Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-48db12bdca4a53d6.xml 2025-12-04T12:25:18.8084598Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0c861fbac71335e0.xml 2025-12-04T12:25:18.8357791Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ffbb324af5af0672.xml 2025-12-04T12:25:18.8682902Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-042dc549bdd77a45.xml 2025-12-04T12:25:18.9076149Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0942d088a0709320.xml 2025-12-04T12:25:18.9369909Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c3fbd4a060fe6236.xml 2025-12-04T12:25:19.0129809Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f308a9dbba7be53a.xml 2025-12-04T12:25:19.0488764Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-bd2ede9c3a5cd1d1.xml 2025-12-04T12:25:19.0865409Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d27c0d840d6fe564.xml 2025-12-04T12:25:19.1178314Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2f600ca197676d4b.xml 2025-12-04T12:25:19.1532920Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ecad5133d6245a6a.xml 2025-12-04T12:25:19.1863501Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d38b619605c8f606.xml 2025-12-04T12:25:19.2440384Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ca04900b7cf26510.xml 2025-12-04T12:25:19.2767173Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-8a1dc4f258524121.xml 2025-12-04T12:25:19.3078209Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a0112c740677bec4.xml 2025-12-04T12:25:19.3384201Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-67764f47d32c222e.xml 2025-12-04T12:25:19.3712660Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c39fdfd87cf64578.xml 2025-12-04T12:25:19.4023222Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-48a46382e3216972.xml 2025-12-04T12:25:19.4336709Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d40d195c41f12813.xml 2025-12-04T12:25:19.4656129Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-35b201d0b4c8de0a.xml 2025-12-04T12:25:19.4991709Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e10bae22111bfdec.xml 2025-12-04T12:25:19.5337209Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-5822c2fc7d137525.xml 2025-12-04T12:25:19.5658907Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a6ba20f02ccefade.xml 2025-12-04T12:25:19.6001210Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dc63b1e627261305.xml 2025-12-04T12:25:19.6418168Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9dda8854f56bbc5e.xml 2025-12-04T12:25:19.6760137Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d35f802df30bd39d.xml 2025-12-04T12:25:19.7193413Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d1473b4b18944240.xml 2025-12-04T12:25:19.7536668Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c84649193ffdba8a.xml 2025-12-04T12:25:19.7846670Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b127809f84c30c37.xml 2025-12-04T12:25:19.8222827Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-1a7c4cad159549d9.xml 2025-12-04T12:25:19.8536606Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2f388a73f046ed87.xml 2025-12-04T12:25:19.8835416Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d68d80f1ba61c89f.xml 2025-12-04T12:25:19.9151528Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b8a30a3f316b7f0b.xml 2025-12-04T12:25:19.9491097Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d3917b026aff3e64.xml 2025-12-04T12:25:19.9835316Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c0c657ac1df627f9.xml 2025-12-04T12:25:20.0146489Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-175aa275ea059250.xml 2025-12-04T12:25:20.0444222Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-5ce17ea59d8fd893.xml 2025-12-04T12:25:20.0783443Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4874d0c887fabf8b.xml 2025-12-04T12:25:20.1362878Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a28d30f3a161bea7.xml 2025-12-04T12:25:20.1700362Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-23cf59059171b3d3.xml 2025-12-04T12:25:20.2460179Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-14e4b4194810001e.xml 2025-12-04T12:25:20.2715836Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4b58d6ae0af928ee.xml 2025-12-04T12:25:20.2991515Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9605f50d5f418e61.xml 2025-12-04T12:25:20.3286403Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2668b63c8efdb7a5.xml 2025-12-04T12:25:20.3582202Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-8d527629568e40ed.xml 2025-12-04T12:25:20.3881646Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a273900f3029c317.xml 2025-12-04T12:25:20.4190764Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dcefced86a4c35d8.xml 2025-12-04T12:25:20.4514471Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e15b12282035b26d.xml 2025-12-04T12:25:20.4959579Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e08a7bc39929d142.xml 2025-12-04T12:25:20.5244091Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-13b16b357405c710.xml 2025-12-04T12:25:20.5557850Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9275f55235cebf56.xml 2025-12-04T12:25:20.5814740Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a0875a42ac68f671.xml 2025-12-04T12:25:20.6105966Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-56c85a19db7b6fad.xml 2025-12-04T12:25:20.6381751Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b5990bb8b6e78b98.xml 2025-12-04T12:25:20.6707161Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9e1ed788d5899650.xml 2025-12-04T12:25:20.6992180Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e97cc37d8f662932.xml 2025-12-04T12:25:20.7266756Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d35e9b6eebdaac7b.xml 2025-12-04T12:25:20.7534220Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f6ccc2d46fa5c9ea.xml 2025-12-04T12:25:20.7820976Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dac88d9039f141ad.xml 2025-12-04T12:25:20.8327887Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-1d81442a0ab3908c.xml 2025-12-04T12:25:20.8634663Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-773b425f2e921837.xml 2025-12-04T12:25:20.8941270Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-09fb0d74c7a00548.xml 2025-12-04T12:25:20.9235704Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2f000bd6c93ab369.xml 2025-12-04T12:25:20.9549373Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-64628aab5ce6aeb9.xml 2025-12-04T12:25:20.9816296Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9fedf51ad965eb7d.xml 2025-12-04T12:25:21.0101360Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c5a3e4c540d200e7.xml 2025-12-04T12:25:21.0370601Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ab8ed79c5f08a6c3.xml 2025-12-04T12:25:21.0650264Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-43cb3996049bee78.xml 2025-12-04T12:25:21.0951512Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d3a2df1a150e9085.xml 2025-12-04T12:25:21.1254749Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2b2d696e555e6fc9.xml 2025-12-04T12:25:21.1845434Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-449e4fae7bbefa8d.xml 2025-12-04T12:25:21.2121504Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4631de08a3dd09e1.xml 2025-12-04T12:25:21.2435476Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-7cb60f50c6534a73.xml 2025-12-04T12:25:21.2743960Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-bbe75c180c40946f.xml 2025-12-04T12:25:21.3033003Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-19b98f6e1bce5088.xml 2025-12-04T12:25:21.3305426Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-fd7621eb24f73b8c.xml 2025-12-04T12:25:21.3618378Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e11a9d6da1493a65.xml 2025-12-04T12:25:21.3908447Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d245bc4e7ed2a74d.xml 2025-12-04T12:25:23.1100386Z Uploading logs for 57119749282 to S3 2025-12-04T12:25:23.4116357Z Uploading artifacts took 1.97 seconds 2025-12-04T12:25:23.4116846Z inductor/test_mix_order_reduction 2/2 failed! 2025-12-04T12:25:23.4121597Z Running test_matmul_cuda 1/1 ... [2025-12-04 12:25:23.411979][11507.102368589] 2025-12-04T12:25:23.4122451Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T12:25:23.4127182Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_matmul_cuda.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 12:25:23.412466] 2025-12-04T12:38:28.1928097Z 2025-12-04T12:38:28.1930816Z PRINTING LOG FILE of test_matmul_cuda 1/1 (test/test-reports/test_matmul_cuda_1.1_e0f80f5015105a39_.log) 2025-12-04T12:38:28.1932074Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-d7ddbde7cf40ab13.xml 2025-12-04T12:38:28.1933083Z ============================= test session starts ============================== 2025-12-04T12:38:28.1934251Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.1935113Z cachedir: .pytest_cache 2025-12-04T12:38:28.1936131Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.1937406Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.1937928Z configfile: pytest.ini 2025-12-04T12:38:28.1939194Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.1940490Z collecting ... collected 1584 items 2025-12-04T12:38:28.1940925Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T12:38:28.3097614Z Running 1584 items in this shard: test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_alignment_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_bias_shapes_size_128_backend_cublas_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_bias_shapes_size_128_backend_cublas_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_bias_shapes_size_128_backend_cublas_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_bias_shapes_size_128_backend_cublaslt_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_bias_shapes_size_128_backend_cublaslt_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_bias_shapes_size_128_backend_cublaslt_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_no_reduced_precision_small_size_4_size_32768_backend_cublas_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_no_reduced_precision_small_size_4_size_32768_backend_cublaslt_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_no_reduced_precision_small_size_8_size_32768_backend_cublas_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_no_reduced_precision_small_size_8_size_32768_backend_cublaslt_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_10000_backend_cublas_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_10000_backend_cublas_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_10000_backend_cublaslt_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_10000_backend_cublaslt_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_1000_backend_cublas_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_1000_backend_cublas_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_1000_backend_cublaslt_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_1000_backend_cublaslt_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_100_backend_cublas_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_100_backend_cublas_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_100_backend_cublaslt_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_100_backend_cublaslt_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_10000_backend_cublas_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_10000_backend_cublas_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_10000_backend_cublaslt_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_10000_backend_cublaslt_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_1000_backend_cublas_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_1000_backend_cublas_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_1000_backend_cublaslt_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_1000_backend_cublaslt_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_100_backend_cublas_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_100_backend_cublas_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_100_backend_cublaslt_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_100_backend_cublaslt_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_10000_backend_cublas_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_10000_backend_cublas_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_10000_backend_cublas_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_10000_backend_cublaslt_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_10000_backend_cublaslt_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_10000_backend_cublaslt_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_1000_backend_cublas_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_1000_backend_cublas_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_1000_backend_cublas_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_1000_backend_cublaslt_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_1000_backend_cublaslt_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_1000_backend_cublaslt_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_100_backend_cublas_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_100_backend_cublas_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_100_backend_cublas_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_100_backend_cublaslt_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_100_backend_cublaslt_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_100_backend_cublaslt_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_and_lt_reduced_precision_fp16_accumulate_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_1_10000_10000_10000_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_1_10000_10000_10000_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_1_10000_10000_10000_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_1_10000_1000_10000_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_1_10000_1000_10000_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_1_10000_1000_10000_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_2_1000_1000_1000_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_2_1000_1000_1000_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_2_1000_1000_1000_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_2_100_100_100_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_2_100_100_100_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_2_100_100_100_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_batch_invariance_blackwell_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_batch_invariance_blackwell_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_1024_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_1024_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_1024_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_128_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_128_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_128_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_2048_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_2048_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_2048_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_256_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_256_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_256_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_32_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_32_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_32_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_4096_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_4096_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_4096_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_512_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_512_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_512_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_64_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_64_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_64_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_8192_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_8192_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_8192_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_fp16_accum_and_fp32_out_failure_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_fp16_accum_and_fp32_out_failure_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_fp16_accum_and_fp32_out_failure_batch_size_32_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_fp16_accum_and_fp32_out_failure_batch_size_32_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_greencontext_carveout_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_False_b_row_major_False_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_False_b_row_major_False_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_False_b_row_major_False_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_False_b_row_major_True_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_False_b_row_major_True_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_False_b_row_major_True_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_True_b_row_major_False_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_True_b_row_major_False_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_True_b_row_major_False_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_True_b_row_major_True_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_True_b_row_major_True_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_True_b_row_major_True_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_False_b_row_major_False_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_False_b_row_major_False_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_False_b_row_major_False_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_False_b_row_major_True_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_False_b_row_major_True_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_False_b_row_major_True_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_True_b_row_major_False_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_True_b_row_major_False_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_True_b_row_major_False_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_True_b_row_major_True_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_True_b_row_major_True_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_True_b_row_major_True_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_False_b_row_major_False_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_False_b_row_major_False_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_False_b_row_major_False_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_False_b_row_major_True_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_False_b_row_major_True_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_False_b_row_major_True_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_True_b_row_major_False_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_True_b_row_major_False_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_True_b_row_major_False_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_True_b_row_major_True_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_True_b_row_major_True_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_True_b_row_major_True_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_False_b_row_major_False_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_False_b_row_major_False_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_False_b_row_major_False_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_False_b_row_major_True_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_False_b_row_major_True_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_False_b_row_major_True_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_True_b_row_major_False_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_True_b_row_major_False_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_True_b_row_major_False_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_True_b_row_major_True_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_True_b_row_major_True_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_True_b_row_major_True_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_False_b_row_major_False_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_False_b_row_major_False_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_False_b_row_major_False_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_False_b_row_major_True_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_False_b_row_major_True_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_False_b_row_major_True_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_True_b_row_major_False_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_True_b_row_major_False_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_True_b_row_major_False_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_True_b_row_major_True_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_True_b_row_major_True_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_True_b_row_major_True_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_False_b_row_major_False_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_False_b_row_major_False_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_False_b_row_major_False_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_False_b_row_major_True_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_False_b_row_major_True_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_False_b_row_major_True_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_True_b_row_major_False_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_True_b_row_major_False_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_True_b_row_major_False_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_True_b_row_major_True_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_True_b_row_major_True_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_True_b_row_major_True_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_False_b_row_major_False_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_False_b_row_major_False_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_False_b_row_major_False_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_False_b_row_major_True_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_False_b_row_major_True_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_False_b_row_major_True_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_True_b_row_major_False_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_True_b_row_major_False_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_True_b_row_major_False_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_True_b_row_major_True_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_True_b_row_major_True_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_True_b_row_major_True_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_False_b_row_major_False_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_False_b_row_major_False_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_False_b_row_major_False_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_False_b_row_major_True_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_False_b_row_major_True_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_False_b_row_major_True_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_True_b_row_major_False_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_True_b_row_major_False_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_True_b_row_major_False_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_True_b_row_major_True_cuda_bfloat16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_True_b_row_major_True_cuda_float16, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_True_b_row_major_True_cuda_float32, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/2d_a_row_major_False_b_row_major_False_max_autotune_False_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/2d_a_row_major_False_b_row_major_False_max_autotune_True_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/2d_a_row_major_False_b_row_major_True_max_autotune_False_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/2d_a_row_major_False_b_row_major_True_max_autotune_True_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/2d_a_row_major_True_b_row_major_False_max_autotune_False_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/2d_a_row_major_True_b_row_major_False_max_autotune_True_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/2d_a_row_major_True_b_row_major_True_max_autotune_False_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/2d_a_row_major_True_b_row_major_True_max_autotune_True_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/3d_a_row_major_False_b_row_major_False_max_autotune_False_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/3d_a_row_major_False_b_row_major_False_max_autotune_True_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/3d_a_row_major_False_b_row_major_True_max_autotune_False_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/3d_a_row_major_False_b_row_major_True_max_autotune_True_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/3d_a_row_major_True_b_row_major_False_max_autotune_False_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/3d_a_row_major_True_b_row_major_False_max_autotune_True_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/3d_a_row_major_True_b_row_major_True_max_autotune_False_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/3d_a_row_major_True_b_row_major_True_max_autotune_True_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/2d_a_row_major_False_b_row_major_False_max_autotune_False_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/2d_a_row_major_False_b_row_major_False_max_autotune_True_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/2d_a_row_major_False_b_row_major_True_max_autotune_False_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/2d_a_row_major_False_b_row_major_True_max_autotune_True_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/2d_a_row_major_True_b_row_major_False_max_autotune_False_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/2d_a_row_major_True_b_row_major_False_max_autotune_True_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/2d_a_row_major_True_b_row_major_True_max_autotune_False_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/2d_a_row_major_True_b_row_major_True_max_autotune_True_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/3d_a_row_major_False_b_row_major_False_max_autotune_False_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/3d_a_row_major_False_b_row_major_False_max_autotune_True_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/3d_a_row_major_False_b_row_major_True_max_autotune_False_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/3d_a_row_major_False_b_row_major_True_max_autotune_True_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/3d_a_row_major_True_b_row_major_False_max_autotune_False_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/3d_a_row_major_True_b_row_major_False_max_autotune_True_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/3d_a_row_major_True_b_row_major_True_max_autotune_False_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/3d_a_row_major_True_b_row_major_True_max_autotune_True_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_input_dimension_checking_out_dtype_ops0_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_input_dimension_checking_out_dtype_ops1_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_input_dimension_checking_out_dtype_ops2_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_input_dimension_checking_out_dtype_ops3_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_16_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_16_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_backend_cublas_cuda, test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_backend_cublaslt_cuda, test/test_matmul_cuda.py::TestMixedDtypesLinearCudaCUDA::test_mixed_dtypes_linear_cuda_bfloat16, test/test_matmul_cuda.py::TestMixedDtypesLinearCudaCUDA::test_mixed_dtypes_linear_cuda_float16 2025-12-04T12:38:28.4228178Z 2025-12-04T12:38:28.4229104Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0095s] [ 0%] 2025-12-04T12:38:28.4231909Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda [W1204 12:25:34.982767278 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.4233991Z PASSED [0.0044s] [ 0%] 2025-12-04T12:38:28.4235168Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0038s] [ 0%] 2025-12-04T12:38:28.4236967Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0038s] [ 0%] 2025-12-04T12:38:28.4238752Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0036s] [ 0%] 2025-12-04T12:38:28.4240529Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0036s] [ 0%] 2025-12-04T12:38:28.4242308Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0037s] [ 0%] 2025-12-04T12:38:28.4244075Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0035s] [ 0%] 2025-12-04T12:38:28.4245850Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0036s] [ 0%] 2025-12-04T12:38:28.4247647Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0036s] [ 0%] 2025-12-04T12:38:28.4249424Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0038s] [ 0%] 2025-12-04T12:38:28.4251207Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0035s] [ 0%] 2025-12-04T12:38:28.4253229Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0035s] [ 0%] 2025-12-04T12:38:28.4254987Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0035s] [ 0%] 2025-12-04T12:38:28.4256753Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0038s] [ 0%] 2025-12-04T12:38:28.4258596Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0035s] [ 1%] 2025-12-04T12:38:28.4260371Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0036s] [ 1%] 2025-12-04T12:38:28.4262169Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0035s] [ 1%] 2025-12-04T12:38:28.4263940Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0036s] [ 1%] 2025-12-04T12:38:28.4265727Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0038s] [ 1%] 2025-12-04T12:38:28.4267550Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0035s] [ 1%] 2025-12-04T12:38:28.4269339Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0036s] [ 1%] 2025-12-04T12:38:28.4271120Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0036s] [ 1%] 2025-12-04T12:38:28.4272882Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0035s] [ 1%] 2025-12-04T12:38:28.4274644Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0038s] [ 1%] 2025-12-04T12:38:28.4276436Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0035s] [ 1%] 2025-12-04T12:38:28.4278221Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0035s] [ 1%] 2025-12-04T12:38:28.4279990Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0035s] [ 1%] 2025-12-04T12:38:28.4281760Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0035s] [ 1%] 2025-12-04T12:38:28.4283523Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0043s] [ 1%] 2025-12-04T12:38:28.4285388Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0035s] [ 1%] 2025-12-04T12:38:28.4287163Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0035s] [ 2%] 2025-12-04T12:38:28.4288936Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 2%] 2025-12-04T12:38:28.4290722Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0037s] [ 2%] 2025-12-04T12:38:28.4292506Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 2%] 2025-12-04T12:38:28.4294285Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0037s] [ 2%] 2025-12-04T12:38:28.4296264Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 2%] 2025-12-04T12:38:28.4298106Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 2%] 2025-12-04T12:38:28.4299941Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 2%] 2025-12-04T12:38:28.4301705Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 2%] 2025-12-04T12:38:28.4303486Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 2%] 2025-12-04T12:38:28.4305288Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0038s] [ 2%] 2025-12-04T12:38:28.4307079Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 2%] 2025-12-04T12:38:28.4308867Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 2%] 2025-12-04T12:38:28.4310655Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 2%] 2025-12-04T12:38:28.4312442Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 2%] 2025-12-04T12:38:28.4314201Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 2%] 2025-12-04T12:38:28.4315975Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0038s] [ 3%] 2025-12-04T12:38:28.4317897Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.4319687Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.4321484Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.4323255Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.4325045Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.4326832Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0037s] [ 3%] 2025-12-04T12:38:28.4328616Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0033s] [ 3%] 2025-12-04T12:38:28.4330380Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0033s] [ 3%] 2025-12-04T12:38:28.4332180Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.4333977Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.4335768Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.4337619Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0037s] [ 3%] 2025-12-04T12:38:28.4339409Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.4341193Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.4342956Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0035s] [ 3%] 2025-12-04T12:38:28.4344732Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.4346518Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.4348314Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0037s] [ 4%] 2025-12-04T12:38:28.4350202Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.4351963Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0037s] [ 4%] 2025-12-04T12:38:28.4353743Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.4355528Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.4357302Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.4359077Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0037s] [ 4%] 2025-12-04T12:38:28.4360847Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.4362646Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.4364474Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.4366269Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.4368052Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.4369834Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0037s] [ 4%] 2025-12-04T12:38:28.4371621Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.4373395Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 5%] 2025-12-04T12:38:28.4375180Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 5%] 2025-12-04T12:38:28.4377026Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 5%] 2025-12-04T12:38:28.4378823Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0037s] [ 5%] 2025-12-04T12:38:28.4380591Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 5%] 2025-12-04T12:38:28.4382496Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 5%] 2025-12-04T12:38:28.4384287Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 5%] 2025-12-04T12:38:28.4386075Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 5%] 2025-12-04T12:38:28.4387861Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 5%] 2025-12-04T12:38:28.4389651Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0037s] [ 5%] 2025-12-04T12:38:28.4392544Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 5%] 2025-12-04T12:38:28.4394337Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 5%] 2025-12-04T12:38:28.4396367Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 5%] 2025-12-04T12:38:28.4398171Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 5%] 2025-12-04T12:38:28.4399959Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 5%] 2025-12-04T12:38:28.4401744Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0037s] [ 5%] 2025-12-04T12:38:28.4403519Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0033s] [ 6%] 2025-12-04T12:38:28.4405312Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 6%] 2025-12-04T12:38:28.4407118Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 6%] 2025-12-04T12:38:28.4408900Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 6%] 2025-12-04T12:38:28.4410663Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 6%] 2025-12-04T12:38:28.4412430Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0041s] [ 6%] 2025-12-04T12:38:28.4414215Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0037s] [ 6%] 2025-12-04T12:38:28.4416123Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 6%] 2025-12-04T12:38:28.4417971Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 6%] 2025-12-04T12:38:28.4419736Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 6%] 2025-12-04T12:38:28.4421529Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 6%] 2025-12-04T12:38:28.4423323Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0038s] [ 6%] 2025-12-04T12:38:28.4425093Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 6%] 2025-12-04T12:38:28.4426867Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 6%] 2025-12-04T12:38:28.4428674Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 6%] 2025-12-04T12:38:28.4430446Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 7%] 2025-12-04T12:38:28.4432215Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 7%] 2025-12-04T12:38:28.4434005Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0037s] [ 7%] 2025-12-04T12:38:28.4435801Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 7%] 2025-12-04T12:38:28.4437584Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 7%] 2025-12-04T12:38:28.4439367Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 7%] 2025-12-04T12:38:28.4441147Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0035s] [ 7%] 2025-12-04T12:38:28.4442931Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 7%] 2025-12-04T12:38:28.4444711Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0037s] [ 7%] 2025-12-04T12:38:28.4446537Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 7%] 2025-12-04T12:38:28.4448348Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 7%] 2025-12-04T12:38:28.4450145Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 7%] 2025-12-04T12:38:28.4451934Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 7%] 2025-12-04T12:38:28.4453715Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 7%] 2025-12-04T12:38:28.4455505Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0037s] [ 7%] 2025-12-04T12:38:28.4457337Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 7%] 2025-12-04T12:38:28.4459122Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 8%] 2025-12-04T12:38:28.4461002Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 8%] 2025-12-04T12:38:28.4462804Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 8%] 2025-12-04T12:38:28.4464613Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 8%] 2025-12-04T12:38:28.4466393Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0037s] [ 8%] 2025-12-04T12:38:28.4468181Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 8%] 2025-12-04T12:38:28.4469973Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 8%] 2025-12-04T12:38:28.4471767Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 8%] 2025-12-04T12:38:28.4473550Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 8%] 2025-12-04T12:38:28.4475318Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0037s] [ 8%] 2025-12-04T12:38:28.4477117Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0037s] [ 8%] 2025-12-04T12:38:28.4478981Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 8%] 2025-12-04T12:38:28.4480816Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 8%] 2025-12-04T12:38:28.4482612Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 8%] 2025-12-04T12:38:28.4484388Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 8%] 2025-12-04T12:38:28.4486180Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 8%] 2025-12-04T12:38:28.4487963Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0037s] [ 9%] 2025-12-04T12:38:28.4489740Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 9%] 2025-12-04T12:38:28.4491529Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 9%] 2025-12-04T12:38:28.4493362Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 9%] 2025-12-04T12:38:28.4495147Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 9%] 2025-12-04T12:38:28.4497132Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 9%] 2025-12-04T12:38:28.4498923Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0038s] [ 9%] 2025-12-04T12:38:28.4500703Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 9%] 2025-12-04T12:38:28.4502487Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 9%] 2025-12-04T12:38:28.4504261Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 9%] 2025-12-04T12:38:28.4506054Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 9%] 2025-12-04T12:38:28.4507851Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 9%] 2025-12-04T12:38:28.4509643Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0038s] [ 9%] 2025-12-04T12:38:28.4511587Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 9%] 2025-12-04T12:38:28.4513365Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 9%] 2025-12-04T12:38:28.4515145Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 9%] 2025-12-04T12:38:28.4517013Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 10%] 2025-12-04T12:38:28.4518793Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0038s] [ 10%] 2025-12-04T12:38:28.4520578Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 10%] 2025-12-04T12:38:28.4522375Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 10%] 2025-12-04T12:38:28.4524170Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 10%] 2025-12-04T12:38:28.4526021Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 10%] 2025-12-04T12:38:28.4527921Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8697s] [ 10%] 2025-12-04T12:38:28.4529885Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1761s] [ 10%] 2025-12-04T12:38:28.4531757Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda FAILED [0.1638s] [ 10%] 2025-12-04T12:38:28.4532709Z 2025-12-04T12:38:28.4532854Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.4533806Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.4534732Z Traceback (most recent call last): 2025-12-04T12:38:28.4535427Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4536141Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4536722Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4537415Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4538052Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4538699Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4539127Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4539385Z 2025-12-04T12:38:28.4539623Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4540919Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.4542080Z 2025-12-04T12:38:28.4542382Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4543451Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.4544371Z Traceback (most recent call last): 2025-12-04T12:38:28.4545037Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4545758Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4546334Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4546969Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4547604Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4548249Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4548667Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4548928Z 2025-12-04T12:38:28.4549144Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4550449Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.4551535Z 2025-12-04T12:38:28.4551803Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4552364Z =================================== FAILURES =================================== 2025-12-04T12:38:28.4553334Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.4554255Z Traceback (most recent call last): 2025-12-04T12:38:28.4554933Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4555660Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4556223Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4556855Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4557506Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4558124Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4558539Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4558809Z 2025-12-04T12:38:28.4559024Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4560323Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.4561400Z 2025-12-04T12:38:28.4561665Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4562725Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-d7ddbde7cf40ab13.xml - 2025-12-04T12:38:28.4563674Z =========================== short test summary info ============================ 2025-12-04T12:38:28.4564992Z FAILED [0.1638s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4566127Z 2025-12-04T12:38:28.4566356Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4567642Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.4568806Z 2025-12-04T12:38:28.4569104Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4569703Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.4570222Z ==================== 1 failed, 164 passed, 2 rerun in 2.13s ==================== 2025-12-04T12:38:28.4570649Z Got exit code 1 2025-12-04T12:38:28.4570918Z Retrying single test... 2025-12-04T12:38:28.4571565Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-63a23e8fc1ef8137.xml 2025-12-04T12:38:28.4572322Z ============================= test session starts ============================== 2025-12-04T12:38:28.4572983Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.4573591Z cachedir: .pytest_cache 2025-12-04T12:38:28.4574310Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.4575090Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.4575452Z configfile: pytest.ini 2025-12-04T12:38:28.4576237Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.4577266Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.4578674Z stepcurrent: skipping 164 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.4579968Z Running 1 items in this shard 2025-12-04T12:38:28.4580179Z 2025-12-04T12:38:28.4581156Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8696s] [100%] 2025-12-04T12:38:28.4583128Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1704s] [100%] 2025-12-04T12:38:28.4585006Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda FAILED [0.1628s] [100%] 2025-12-04T12:38:28.4585977Z 2025-12-04T12:38:28.4586120Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.4587064Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.4587985Z Traceback (most recent call last): 2025-12-04T12:38:28.4588669Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4589398Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4589985Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4590622Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4591259Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4591901Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4592322Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4592578Z 2025-12-04T12:38:28.4592796Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4594094Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.4595252Z 2025-12-04T12:38:28.4595517Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4596826Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.4597734Z Traceback (most recent call last): 2025-12-04T12:38:28.4598417Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4599135Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4599709Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4600337Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4600982Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4601617Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4602022Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4602287Z 2025-12-04T12:38:28.4602500Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4603805Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.4604884Z 2025-12-04T12:38:28.4605167Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4605715Z =================================== FAILURES =================================== 2025-12-04T12:38:28.4606660Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.4607578Z Traceback (most recent call last): 2025-12-04T12:38:28.4608327Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4609043Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4609618Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4610254Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4610900Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4611520Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4611937Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4612191Z 2025-12-04T12:38:28.4612415Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4613705Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.4614788Z 2025-12-04T12:38:28.4615057Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4616104Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-63a23e8fc1ef8137.xml - 2025-12-04T12:38:28.4617114Z =========================== short test summary info ============================ 2025-12-04T12:38:28.4618428Z FAILED [0.1628s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4619560Z 2025-12-04T12:38:28.4619777Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4621082Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.4622170Z 2025-12-04T12:38:28.4622542Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4623136Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.4623689Z ================= 1 failed, 1583 deselected, 2 rerun in 1.32s ================== 2025-12-04T12:38:28.4624147Z Got exit code 1 2025-12-04T12:38:28.4624424Z Retrying single test... 2025-12-04T12:38:28.4625050Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-11f9c4926783dfef.xml 2025-12-04T12:38:28.4625817Z ============================= test session starts ============================== 2025-12-04T12:38:28.4626481Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.4627082Z cachedir: .pytest_cache 2025-12-04T12:38:28.4627786Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.4628579Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.4628934Z configfile: pytest.ini 2025-12-04T12:38:28.4629704Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.4630664Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.4632066Z stepcurrent: skipping 164 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.4633366Z Running 1 items in this shard 2025-12-04T12:38:28.4633579Z 2025-12-04T12:38:28.4634542Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8764s] [100%] 2025-12-04T12:38:28.4636497Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1835s] [100%] 2025-12-04T12:38:28.4638371Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda FAILED [0.1705s] [100%] 2025-12-04T12:38:28.4639336Z 2025-12-04T12:38:28.4639478Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.4640426Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.4641325Z Traceback (most recent call last): 2025-12-04T12:38:28.4642009Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4642730Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4643313Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4643930Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4644583Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4645222Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4645627Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4645899Z 2025-12-04T12:38:28.4646114Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4647413Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.4648487Z 2025-12-04T12:38:28.4648766Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4649913Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.4650821Z Traceback (most recent call last): 2025-12-04T12:38:28.4671344Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4672088Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4672663Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4673301Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4673946Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4674560Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4674976Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4675234Z 2025-12-04T12:38:28.4675471Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4676783Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.4677860Z 2025-12-04T12:38:28.4678128Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4678686Z =================================== FAILURES =================================== 2025-12-04T12:38:28.4679630Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.4680538Z Traceback (most recent call last): 2025-12-04T12:38:28.4681202Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4681969Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4682538Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4683158Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4683803Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4684432Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4684842Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4685096Z 2025-12-04T12:38:28.4685306Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4686597Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.4687680Z 2025-12-04T12:38:28.4687946Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4688981Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-11f9c4926783dfef.xml - 2025-12-04T12:38:28.4689906Z =========================== short test summary info ============================ 2025-12-04T12:38:28.4691228Z FAILED [0.1705s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4692369Z 2025-12-04T12:38:28.4692578Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4693862Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.4694922Z 2025-12-04T12:38:28.4695196Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4695766Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.4696558Z ================= 1 failed, 1583 deselected, 2 rerun in 1.35s ================== 2025-12-04T12:38:28.4697067Z Got exit code 1 2025-12-04T12:38:28.4698147Z FAILED CONSISTENTLY: test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.4699551Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:38:28.4700546Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-737659f69f9145f2.xml 2025-12-04T12:38:28.4701294Z ============================= test session starts ============================== 2025-12-04T12:38:28.4701938Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.4702529Z cachedir: .pytest_cache 2025-12-04T12:38:28.4703232Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.4704020Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.4704355Z configfile: pytest.ini 2025-12-04T12:38:28.4705123Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.4706077Z collecting ... collected 1584 items / 165 deselected / 1419 selected 2025-12-04T12:38:28.4706571Z stepcurrent: skipping 165 already run items. 2025-12-04T12:38:28.4706952Z Running 1419 items in this shard 2025-12-04T12:38:28.4707163Z 2025-12-04T12:38:28.4709054Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda [W1204 12:26:15.704160635 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.4711133Z ('RERUN', {'yellow': True}) [0.8732s] [ 0%] 2025-12-04T12:38:28.4712306Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1721s] [ 0%] 2025-12-04T12:38:28.4714186Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda FAILED [0.1679s] [ 0%] 2025-12-04T12:38:28.4715165Z 2025-12-04T12:38:28.4715303Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.4716240Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.4717141Z Traceback (most recent call last): 2025-12-04T12:38:28.4717794Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4718494Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4719064Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4719691Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4720317Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4720942Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4721342Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4721590Z 2025-12-04T12:38:28.4721797Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4723098Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.4724277Z 2025-12-04T12:38:28.4724539Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4725659Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.4726572Z Traceback (most recent call last): 2025-12-04T12:38:28.4727251Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4727961Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4728523Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4729134Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4729774Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4730399Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4730800Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4731057Z 2025-12-04T12:38:28.4731273Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4732564Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.4733639Z 2025-12-04T12:38:28.4733910Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4734455Z =================================== FAILURES =================================== 2025-12-04T12:38:28.4735387Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.4736332Z Traceback (most recent call last): 2025-12-04T12:38:28.4737083Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4737787Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4738358Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4738985Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4739620Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4740229Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4740634Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4740886Z 2025-12-04T12:38:28.4741107Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4742399Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.4743496Z 2025-12-04T12:38:28.4743761Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4744788Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-737659f69f9145f2.xml - 2025-12-04T12:38:28.4745710Z =========================== short test summary info ============================ 2025-12-04T12:38:28.4747024Z FAILED [0.1679s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4748161Z 2025-12-04T12:38:28.4748374Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4749673Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.4750821Z 2025-12-04T12:38:28.4751085Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4751705Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.4752218Z ================== 1 failed, 165 deselected, 2 rerun in 1.34s ================== 2025-12-04T12:38:28.4752657Z Got exit code 1 2025-12-04T12:38:28.4752921Z Retrying single test... 2025-12-04T12:38:28.4753540Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2099d82b2931fb99.xml 2025-12-04T12:38:28.4754297Z ============================= test session starts ============================== 2025-12-04T12:38:28.4754948Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.4755550Z cachedir: .pytest_cache 2025-12-04T12:38:28.4756250Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.4757033Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.4757379Z configfile: pytest.ini 2025-12-04T12:38:28.4758149Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.4759091Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.4760491Z stepcurrent: skipping 165 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.4761771Z Running 1 items in this shard 2025-12-04T12:38:28.4761979Z 2025-12-04T12:38:28.4763833Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda [W1204 12:26:28.244236483 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.4765897Z ('RERUN', {'yellow': True}) [0.8705s] [100%] 2025-12-04T12:38:28.4767079Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1781s] [100%] 2025-12-04T12:38:28.4768953Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda FAILED [0.1739s] [100%] 2025-12-04T12:38:28.4769920Z 2025-12-04T12:38:28.4770061Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.4771000Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.4771917Z Traceback (most recent call last): 2025-12-04T12:38:28.4772592Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4773302Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4773865Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4774560Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4775212Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4775837Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4776232Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4776491Z 2025-12-04T12:38:28.4776698Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4778057Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.4779228Z 2025-12-04T12:38:28.4779531Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4780583Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.4781495Z Traceback (most recent call last): 2025-12-04T12:38:28.4782167Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4782868Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4783423Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4784047Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4784679Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4785295Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4785699Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4785963Z 2025-12-04T12:38:28.4786173Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4787467Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.4788543Z 2025-12-04T12:38:28.4788805Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4789349Z =================================== FAILURES =================================== 2025-12-04T12:38:28.4790330Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.4791250Z Traceback (most recent call last): 2025-12-04T12:38:28.4791912Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4792625Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4793190Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4793811Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4794435Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4795057Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4795465Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4795716Z 2025-12-04T12:38:28.4795925Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4797381Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.4798486Z 2025-12-04T12:38:28.4798747Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4799787Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2099d82b2931fb99.xml - 2025-12-04T12:38:28.4800710Z =========================== short test summary info ============================ 2025-12-04T12:38:28.4802022Z FAILED [0.1739s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4803162Z 2025-12-04T12:38:28.4803373Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4804675Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.4805866Z 2025-12-04T12:38:28.4806185Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4806761Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.4807288Z ================= 1 failed, 1583 deselected, 2 rerun in 1.34s ================== 2025-12-04T12:38:28.4807731Z Got exit code 1 2025-12-04T12:38:28.4807981Z Retrying single test... 2025-12-04T12:38:28.4808615Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-5f0933bf702563bb.xml 2025-12-04T12:38:28.4809371Z ============================= test session starts ============================== 2025-12-04T12:38:28.4810035Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.4810614Z cachedir: .pytest_cache 2025-12-04T12:38:28.4811322Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.4812096Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.4812428Z configfile: pytest.ini 2025-12-04T12:38:28.4813192Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.4814143Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.4815534Z stepcurrent: skipping 165 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.4816798Z Running 1 items in this shard 2025-12-04T12:38:28.4817072Z 2025-12-04T12:38:28.4818941Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda [W1204 12:26:41.594525283 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.4821017Z ('RERUN', {'yellow': True}) [0.8614s] [100%] 2025-12-04T12:38:28.4822186Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1711s] [100%] 2025-12-04T12:38:28.4824078Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda FAILED [0.1620s] [100%] 2025-12-04T12:38:28.4825038Z 2025-12-04T12:38:28.4825183Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.4826118Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.4827038Z Traceback (most recent call last): 2025-12-04T12:38:28.4827710Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4828410Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4828980Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4829604Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4830241Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4830855Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4831262Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4831514Z 2025-12-04T12:38:28.4831753Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4833182Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.4834269Z 2025-12-04T12:38:28.4834535Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4835615Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.4836542Z Traceback (most recent call last): 2025-12-04T12:38:28.4837229Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4837936Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4838518Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4839159Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4839798Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4840445Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4840869Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4841127Z 2025-12-04T12:38:28.4841357Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4842655Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.4843756Z 2025-12-04T12:38:28.4844024Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4844614Z =================================== FAILURES =================================== 2025-12-04T12:38:28.4845569Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.4846487Z Traceback (most recent call last): 2025-12-04T12:38:28.4847171Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4847890Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4848450Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4849079Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4849723Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4850366Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4850767Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4851035Z 2025-12-04T12:38:28.4851249Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4852555Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.4853633Z 2025-12-04T12:38:28.4853909Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4854936Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-5f0933bf702563bb.xml - 2025-12-04T12:38:28.4855882Z =========================== short test summary info ============================ 2025-12-04T12:38:28.4857268Z FAILED [0.1620s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4858393Z 2025-12-04T12:38:28.4858626Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4860015Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.4861118Z 2025-12-04T12:38:28.4861385Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4861980Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.4862511Z ================= 1 failed, 1583 deselected, 2 rerun in 1.32s ================== 2025-12-04T12:38:28.4862946Z Got exit code 1 2025-12-04T12:38:28.4863980Z FAILED CONSISTENTLY: test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.4865404Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:38:28.4866412Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-ed1b6e1c95269b63.xml 2025-12-04T12:38:28.4867161Z ============================= test session starts ============================== 2025-12-04T12:38:28.4867830Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.4868431Z cachedir: .pytest_cache 2025-12-04T12:38:28.4869141Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.4869913Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.4870268Z configfile: pytest.ini 2025-12-04T12:38:28.4871048Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.4872032Z collecting ... collected 1584 items / 166 deselected / 1418 selected 2025-12-04T12:38:28.4872553Z stepcurrent: skipping 166 already run items. 2025-12-04T12:38:28.4872950Z Running 1418 items in this shard 2025-12-04T12:38:28.4873170Z 2025-12-04T12:38:28.4874101Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8731s] [ 0%] 2025-12-04T12:38:28.4876049Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1642s] [ 0%] 2025-12-04T12:38:28.4877901Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda FAILED [0.1627s] [ 0%] 2025-12-04T12:38:28.4878856Z 2025-12-04T12:38:28.4878998Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.4879933Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.4880840Z Traceback (most recent call last): 2025-12-04T12:38:28.4881515Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4882230Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4882809Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4883428Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4884074Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4884709Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4885133Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4885385Z 2025-12-04T12:38:28.4885600Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4887001Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.4888078Z 2025-12-04T12:38:28.4888340Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4889387Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.4890277Z Traceback (most recent call last): 2025-12-04T12:38:28.4890946Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4891655Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4892217Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4892821Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4893450Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4894063Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4894459Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4894710Z 2025-12-04T12:38:28.4894916Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4896376Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.4897497Z 2025-12-04T12:38:28.4897771Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4898303Z =================================== FAILURES =================================== 2025-12-04T12:38:28.4899296Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.4900191Z Traceback (most recent call last): 2025-12-04T12:38:28.4900854Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4901549Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4902105Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4902720Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4903342Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4903956Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4904347Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4904594Z 2025-12-04T12:38:28.4904811Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4906084Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.4907160Z 2025-12-04T12:38:28.4907423Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4908450Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-ed1b6e1c95269b63.xml - 2025-12-04T12:38:28.4909371Z =========================== short test summary info ============================ 2025-12-04T12:38:28.4910645Z FAILED [0.1627s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4911758Z 2025-12-04T12:38:28.4911970Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4913244Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.4914403Z 2025-12-04T12:38:28.4914781Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4915359Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.4915862Z ================== 1 failed, 166 deselected, 2 rerun in 1.32s ================== 2025-12-04T12:38:28.4916291Z Got exit code 1 2025-12-04T12:38:28.4916543Z Retrying single test... 2025-12-04T12:38:28.4917160Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-ccb90d6b004c5139.xml 2025-12-04T12:38:28.4917910Z ============================= test session starts ============================== 2025-12-04T12:38:28.4918563Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.4919160Z cachedir: .pytest_cache 2025-12-04T12:38:28.4919846Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.4920623Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.4920966Z configfile: pytest.ini 2025-12-04T12:38:28.4921720Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.4922671Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.4924056Z stepcurrent: skipping 166 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.4925317Z Running 1 items in this shard 2025-12-04T12:38:28.4925557Z 2025-12-04T12:38:28.4926468Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8801s] [100%] 2025-12-04T12:38:28.4928420Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1652s] [100%] 2025-12-04T12:38:28.4930270Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda FAILED [0.1612s] [100%] 2025-12-04T12:38:28.4931226Z 2025-12-04T12:38:28.4931368Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.4932290Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.4933184Z Traceback (most recent call last): 2025-12-04T12:38:28.4933859Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4934567Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4935130Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4935740Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4936370Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4937083Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4937479Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4937741Z 2025-12-04T12:38:28.4937948Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4939231Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.4940394Z 2025-12-04T12:38:28.4940663Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4941729Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.4942633Z Traceback (most recent call last): 2025-12-04T12:38:28.4943301Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4944002Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4944555Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4945171Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4945802Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4946418Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4946818Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4947073Z 2025-12-04T12:38:28.4947280Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4948555Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.4949616Z 2025-12-04T12:38:28.4949888Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4950430Z =================================== FAILURES =================================== 2025-12-04T12:38:28.4951367Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.4952306Z Traceback (most recent call last): 2025-12-04T12:38:28.4952973Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4953695Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4954272Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4954903Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4955532Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4956160Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4956571Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4956825Z 2025-12-04T12:38:28.4957037Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4958333Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.4959412Z 2025-12-04T12:38:28.4959679Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4960728Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-ccb90d6b004c5139.xml - 2025-12-04T12:38:28.4961673Z =========================== short test summary info ============================ 2025-12-04T12:38:28.4962962Z FAILED [0.1612s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4964087Z 2025-12-04T12:38:28.4964306Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4965594Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.4966724Z 2025-12-04T12:38:28.4967004Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4967580Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.4968140Z ================= 1 failed, 1583 deselected, 2 rerun in 1.33s ================== 2025-12-04T12:38:28.4968592Z Got exit code 1 2025-12-04T12:38:28.4968862Z Retrying single test... 2025-12-04T12:38:28.4969494Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a3b2bb4f2a94fb32.xml 2025-12-04T12:38:28.4970254Z ============================= test session starts ============================== 2025-12-04T12:38:28.4970908Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.4971490Z cachedir: .pytest_cache 2025-12-04T12:38:28.4972200Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.4972982Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.4973334Z configfile: pytest.ini 2025-12-04T12:38:28.4974106Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.4975059Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.4976451Z stepcurrent: skipping 166 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.4977799Z Running 1 items in this shard 2025-12-04T12:38:28.4978010Z 2025-12-04T12:38:28.4978962Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8596s] [100%] 2025-12-04T12:38:28.4980915Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1717s] [100%] 2025-12-04T12:38:28.4982770Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda FAILED [0.1664s] [100%] 2025-12-04T12:38:28.4983721Z 2025-12-04T12:38:28.4983875Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.4984800Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.4985705Z Traceback (most recent call last): 2025-12-04T12:38:28.4986387Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4987110Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4987674Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4988306Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4988951Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4989580Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4989991Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4990258Z 2025-12-04T12:38:28.4990477Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.4991776Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.4992853Z 2025-12-04T12:38:28.4993137Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.4994261Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.4995204Z Traceback (most recent call last): 2025-12-04T12:38:28.4995897Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.4996785Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.4997364Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.4998003Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.4998651Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.4999281Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.4999704Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.4999963Z 2025-12-04T12:38:28.5000192Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5001501Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5002571Z 2025-12-04T12:38:28.5002835Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5003401Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5004348Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.5005265Z Traceback (most recent call last): 2025-12-04T12:38:28.5005939Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5006726Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5007308Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5007929Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5008583Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5009220Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5009635Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5009889Z 2025-12-04T12:38:28.5010102Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5011396Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5012464Z 2025-12-04T12:38:28.5012742Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5013782Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a3b2bb4f2a94fb32.xml - 2025-12-04T12:38:28.5014715Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5016027Z FAILED [0.1664s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5017220Z 2025-12-04T12:38:28.5017436Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5018731Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5019807Z 2025-12-04T12:38:28.5020089Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5020677Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5021314Z ================= 1 failed, 1583 deselected, 2 rerun in 1.32s ================== 2025-12-04T12:38:28.5021763Z Got exit code 1 2025-12-04T12:38:28.5022813Z FAILED CONSISTENTLY: test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5024212Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:38:28.5025220Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-581c94c41301ad68.xml 2025-12-04T12:38:28.5025978Z ============================= test session starts ============================== 2025-12-04T12:38:28.5026626Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5027231Z cachedir: .pytest_cache 2025-12-04T12:38:28.5027946Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5028737Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5029077Z configfile: pytest.ini 2025-12-04T12:38:28.5029858Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5030831Z collecting ... collected 1584 items / 167 deselected / 1417 selected 2025-12-04T12:38:28.5031339Z stepcurrent: skipping 167 already run items. 2025-12-04T12:38:28.5031734Z Running 1417 items in this shard 2025-12-04T12:38:28.5031950Z 2025-12-04T12:38:28.5033846Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda [W1204 12:27:35.146779987 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.5035926Z ('RERUN', {'yellow': True}) [0.8751s] [ 0%] 2025-12-04T12:38:28.5037115Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1664s] [ 0%] 2025-12-04T12:38:28.5038995Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda FAILED [0.1604s] [ 0%] 2025-12-04T12:38:28.5039963Z 2025-12-04T12:38:28.5040104Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5041055Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5041974Z Traceback (most recent call last): 2025-12-04T12:38:28.5042646Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5043365Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5043941Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5044559Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5045203Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5045841Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5046261Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5046515Z 2025-12-04T12:38:28.5046729Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5048037Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5049201Z 2025-12-04T12:38:28.5049466Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5050566Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5051475Z Traceback (most recent call last): 2025-12-04T12:38:28.5052162Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5052883Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5053456Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5054076Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5054719Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5055357Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5055762Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5056029Z 2025-12-04T12:38:28.5056245Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5057625Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5058704Z 2025-12-04T12:38:28.5058984Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5059531Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5060490Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5061454Z Traceback (most recent call last): 2025-12-04T12:38:28.5062144Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5062850Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5063426Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5064059Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5064705Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5065326Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5065741Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5065996Z 2025-12-04T12:38:28.5066219Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5067503Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5068600Z 2025-12-04T12:38:28.5068864Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5069899Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-581c94c41301ad68.xml - 2025-12-04T12:38:28.5070841Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5072137Z FAILED [0.1604s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5073274Z 2025-12-04T12:38:28.5073492Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5074799Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5075953Z 2025-12-04T12:38:28.5076222Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5076847Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5077364Z ================== 1 failed, 167 deselected, 2 rerun in 1.32s ================== 2025-12-04T12:38:28.5077814Z Got exit code 1 2025-12-04T12:38:28.5078089Z Retrying single test... 2025-12-04T12:38:28.5078723Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-809b8dfa72183a04.xml 2025-12-04T12:38:28.5079489Z ============================= test session starts ============================== 2025-12-04T12:38:28.5080152Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5080757Z cachedir: .pytest_cache 2025-12-04T12:38:28.5081467Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5082268Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5082622Z configfile: pytest.ini 2025-12-04T12:38:28.5083392Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5084360Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5085772Z stepcurrent: skipping 167 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5087058Z Running 1 items in this shard 2025-12-04T12:38:28.5087268Z 2025-12-04T12:38:28.5089122Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda [W1204 12:27:48.312573391 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.5091197Z ('RERUN', {'yellow': True}) [0.8584s] [100%] 2025-12-04T12:38:28.5092391Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1695s] [100%] 2025-12-04T12:38:28.5094276Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda FAILED [0.1632s] [100%] 2025-12-04T12:38:28.5095227Z 2025-12-04T12:38:28.5095379Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5096470Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5097454Z Traceback (most recent call last): 2025-12-04T12:38:28.5098140Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5098858Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5099421Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5100056Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5100704Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5101337Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5101747Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5102014Z 2025-12-04T12:38:28.5102230Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5103634Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5104842Z 2025-12-04T12:38:28.5105170Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5106230Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5107157Z Traceback (most recent call last): 2025-12-04T12:38:28.5107843Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5108563Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5109128Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5109761Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5110411Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5111033Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5111451Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5111710Z 2025-12-04T12:38:28.5111937Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5113237Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5114310Z 2025-12-04T12:38:28.5114772Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5115360Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5116379Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5117308Z Traceback (most recent call last): 2025-12-04T12:38:28.5117979Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5118704Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5119285Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5119903Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5120552Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5121195Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5121616Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5121871Z 2025-12-04T12:38:28.5122089Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5123396Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5124488Z 2025-12-04T12:38:28.5124756Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5125798Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-809b8dfa72183a04.xml - 2025-12-04T12:38:28.5126729Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5128041Z FAILED [0.1632s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5129181Z 2025-12-04T12:38:28.5129399Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5130705Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5131853Z 2025-12-04T12:38:28.5132184Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5132761Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5133297Z ================= 1 failed, 1583 deselected, 2 rerun in 1.31s ================== 2025-12-04T12:38:28.5133744Z Got exit code 1 2025-12-04T12:38:28.5134001Z Retrying single test... 2025-12-04T12:38:28.5134639Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-59f3dc780871f4a4.xml 2025-12-04T12:38:28.5135397Z ============================= test session starts ============================== 2025-12-04T12:38:28.5136056Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5136646Z cachedir: .pytest_cache 2025-12-04T12:38:28.5137427Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5138217Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5138563Z configfile: pytest.ini 2025-12-04T12:38:28.5139340Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5140299Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5141707Z stepcurrent: skipping 167 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5142975Z Running 1 items in this shard 2025-12-04T12:38:28.5143199Z 2025-12-04T12:38:28.5145057Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda [W1204 12:28:02.845169608 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.5147134Z ('RERUN', {'yellow': True}) [0.8758s] [100%] 2025-12-04T12:38:28.5148306Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1821s] [100%] 2025-12-04T12:38:28.5150185Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda FAILED [0.1842s] [100%] 2025-12-04T12:38:28.5151154Z 2025-12-04T12:38:28.5151296Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5152246Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5153164Z Traceback (most recent call last): 2025-12-04T12:38:28.5153845Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5154554Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5155138Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5155771Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5156421Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5157047Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5157469Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5157728Z 2025-12-04T12:38:28.5157959Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5159360Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5160454Z 2025-12-04T12:38:28.5160723Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5161798Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5162727Z Traceback (most recent call last): 2025-12-04T12:38:28.5163400Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5164123Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5164705Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5165350Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5165990Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5166631Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5167053Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5167312Z 2025-12-04T12:38:28.5167542Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5168834Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5169930Z 2025-12-04T12:38:28.5170198Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5170798Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5171758Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5172669Z Traceback (most recent call last): 2025-12-04T12:38:28.5173358Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5174076Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5174640Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5175272Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5175916Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5176547Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5177031Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5177297Z 2025-12-04T12:38:28.5177516Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5178820Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5179897Z 2025-12-04T12:38:28.5180179Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5181205Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-59f3dc780871f4a4.xml - 2025-12-04T12:38:28.5182151Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5183469Z FAILED [0.1842s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5184591Z 2025-12-04T12:38:28.5184819Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5186183Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5187314Z 2025-12-04T12:38:28.5187582Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5188173Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5188703Z ================= 1 failed, 1583 deselected, 2 rerun in 1.36s ================== 2025-12-04T12:38:28.5189142Z Got exit code 1 2025-12-04T12:38:28.5190170Z FAILED CONSISTENTLY: test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5191585Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:38:28.5192593Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2648ef692c69d7ff.xml 2025-12-04T12:38:28.5193340Z ============================= test session starts ============================== 2025-12-04T12:38:28.5194001Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5194605Z cachedir: .pytest_cache 2025-12-04T12:38:28.5195315Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5196270Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5196632Z configfile: pytest.ini 2025-12-04T12:38:28.5197417Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5198439Z collecting ... collected 1584 items / 168 deselected / 1416 selected 2025-12-04T12:38:28.5198971Z stepcurrent: skipping 168 already run items. 2025-12-04T12:38:28.5199376Z Running 1416 items in this shard 2025-12-04T12:38:28.5199598Z 2025-12-04T12:38:28.5200451Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0093s] [ 0%] 2025-12-04T12:38:28.5203235Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda [W1204 12:28:15.091838441 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.5205285Z PASSED [0.0043s] [ 0%] 2025-12-04T12:38:28.5206299Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0038s] [ 0%] 2025-12-04T12:38:28.5208106Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0038s] [ 0%] 2025-12-04T12:38:28.5209889Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0036s] [ 0%] 2025-12-04T12:38:28.5211662Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0036s] [ 0%] 2025-12-04T12:38:28.5213440Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0037s] [ 0%] 2025-12-04T12:38:28.5215269Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0035s] [ 0%] 2025-12-04T12:38:28.5217212Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0037s] [ 0%] 2025-12-04T12:38:28.5219017Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0036s] [ 0%] 2025-12-04T12:38:28.5220812Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0038s] [ 0%] 2025-12-04T12:38:28.5222597Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0036s] [ 0%] 2025-12-04T12:38:28.5224484Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8561s] [ 0%] 2025-12-04T12:38:28.5226452Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1630s] [ 0%] 2025-12-04T12:38:28.5228324Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda FAILED [0.1563s] [ 0%] 2025-12-04T12:38:28.5229276Z 2025-12-04T12:38:28.5229462Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5230398Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5231325Z Traceback (most recent call last): 2025-12-04T12:38:28.5232015Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5232735Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5233298Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5233928Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5234575Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5235195Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5235617Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5235874Z 2025-12-04T12:38:28.5236106Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5237413Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5238489Z 2025-12-04T12:38:28.5238759Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5239819Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5240735Z Traceback (most recent call last): 2025-12-04T12:38:28.5241424Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5242129Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5242704Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5243337Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5244052Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5244686Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5245135Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5245395Z 2025-12-04T12:38:28.5245625Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5246924Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5248022Z 2025-12-04T12:38:28.5248291Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5248862Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5249814Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5250728Z Traceback (most recent call last): 2025-12-04T12:38:28.5251420Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5252146Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5252727Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5253352Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5254002Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5254641Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5255049Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5255322Z 2025-12-04T12:38:28.5255537Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5256951Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5258042Z 2025-12-04T12:38:28.5258331Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5259358Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2648ef692c69d7ff.xml - 2025-12-04T12:38:28.5260307Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5261625Z FAILED [0.1563s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5262750Z 2025-12-04T12:38:28.5262983Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5264287Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5265367Z 2025-12-04T12:38:28.5265636Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5266233Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5266787Z ============ 1 failed, 12 passed, 168 deselected, 2 rerun in 1.36s ============= 2025-12-04T12:38:28.5267251Z Got exit code 1 2025-12-04T12:38:28.5267503Z Retrying single test... 2025-12-04T12:38:28.5268140Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-1e4f023d2e76e67f.xml 2025-12-04T12:38:28.5268904Z ============================= test session starts ============================== 2025-12-04T12:38:28.5269553Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5270215Z cachedir: .pytest_cache 2025-12-04T12:38:28.5270925Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5271713Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5272081Z configfile: pytest.ini 2025-12-04T12:38:28.5272860Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5273816Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5275197Z stepcurrent: skipping 180 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5276491Z Running 1 items in this shard 2025-12-04T12:38:28.5276715Z 2025-12-04T12:38:28.5277630Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8636s] [100%] 2025-12-04T12:38:28.5279601Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1799s] [100%] 2025-12-04T12:38:28.5281465Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda FAILED [0.1754s] [100%] 2025-12-04T12:38:28.5282419Z 2025-12-04T12:38:28.5282559Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5283535Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5284453Z Traceback (most recent call last): 2025-12-04T12:38:28.5285142Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5285851Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5286431Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5287063Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5287713Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5288340Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5288759Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5289016Z 2025-12-04T12:38:28.5289244Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5290528Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5291629Z 2025-12-04T12:38:28.5291896Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5292957Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5293868Z Traceback (most recent call last): 2025-12-04T12:38:28.5294533Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5295250Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5295822Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5296622Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5297323Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5297966Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5298495Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5298750Z 2025-12-04T12:38:28.5298963Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5300307Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5301392Z 2025-12-04T12:38:28.5301659Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5302218Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5303150Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5303287Z Traceback (most recent call last): 2025-12-04T12:38:28.5303731Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5303886Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5304195Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5304373Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5304722Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5304871Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5305004Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5305021Z 2025-12-04T12:38:28.5305233Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5306216Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5306223Z 2025-12-04T12:38:28.5306510Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5307139Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-1e4f023d2e76e67f.xml - 2025-12-04T12:38:28.5307323Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5308317Z FAILED [0.1754s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5308323Z 2025-12-04T12:38:28.5308536Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5309491Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5309500Z 2025-12-04T12:38:28.5309763Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5309953Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5310158Z ================= 1 failed, 1583 deselected, 2 rerun in 1.34s ================== 2025-12-04T12:38:28.5310256Z Got exit code 1 2025-12-04T12:38:28.5310377Z Retrying single test... 2025-12-04T12:38:28.5310826Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-88a75603bd167dcb.xml 2025-12-04T12:38:28.5311001Z ============================= test session starts ============================== 2025-12-04T12:38:28.5311351Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5311464Z cachedir: .pytest_cache 2025-12-04T12:38:28.5312007Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5312203Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5312318Z configfile: pytest.ini 2025-12-04T12:38:28.5312959Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5313192Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5314239Z stepcurrent: skipping 180 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5314354Z Running 1 items in this shard 2025-12-04T12:38:28.5314358Z 2025-12-04T12:38:28.5315274Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8667s] [100%] 2025-12-04T12:38:28.5316198Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1667s] [100%] 2025-12-04T12:38:28.5317018Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda FAILED [0.1615s] [100%] 2025-12-04T12:38:28.5317024Z 2025-12-04T12:38:28.5317177Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5317830Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5317963Z Traceback (most recent call last): 2025-12-04T12:38:28.5318432Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5318573Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5318896Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5319071Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5319409Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5319573Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5319710Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5319715Z 2025-12-04T12:38:28.5319943Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5320889Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5320895Z 2025-12-04T12:38:28.5321175Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5321828Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5321949Z Traceback (most recent call last): 2025-12-04T12:38:28.5322408Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5322543Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5322857Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5323045Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5323379Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5323546Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5323681Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5323688Z 2025-12-04T12:38:28.5323905Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5325056Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5325061Z 2025-12-04T12:38:28.5325331Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5325493Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5326145Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5326268Z Traceback (most recent call last): 2025-12-04T12:38:28.5326722Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5326860Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5327164Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5327359Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5327696Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5327860Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5327994Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5327999Z 2025-12-04T12:38:28.5328215Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5329180Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5329185Z 2025-12-04T12:38:28.5329452Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5330126Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-88a75603bd167dcb.xml - 2025-12-04T12:38:28.5330305Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5331297Z FAILED [0.1615s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5331316Z 2025-12-04T12:38:28.5331532Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5332477Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5332483Z 2025-12-04T12:38:28.5332767Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5332948Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5333171Z ================= 1 failed, 1583 deselected, 2 rerun in 1.32s ================== 2025-12-04T12:38:28.5333272Z Got exit code 1 2025-12-04T12:38:28.5334135Z FAILED CONSISTENTLY: test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5334563Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:38:28.5335018Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-58d14c6ca3ad4514.xml 2025-12-04T12:38:28.5335197Z ============================= test session starts ============================== 2025-12-04T12:38:28.5335551Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5335659Z cachedir: .pytest_cache 2025-12-04T12:38:28.5336252Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5336374Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5336515Z configfile: pytest.ini 2025-12-04T12:38:28.5337189Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5337426Z collecting ... collected 1584 items / 181 deselected / 1403 selected 2025-12-04T12:38:28.5337589Z stepcurrent: skipping 181 already run items. 2025-12-04T12:38:28.5337705Z Running 1403 items in this shard 2025-12-04T12:38:28.5337711Z 2025-12-04T12:38:28.5339544Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda [W1204 12:28:55.781587514 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.5339696Z ('RERUN', {'yellow': True}) [0.8759s] [ 0%] 2025-12-04T12:38:28.5340621Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1760s] [ 0%] 2025-12-04T12:38:28.5341466Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda FAILED [0.1697s] [ 0%] 2025-12-04T12:38:28.5341473Z 2025-12-04T12:38:28.5341610Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5342324Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5342451Z Traceback (most recent call last): 2025-12-04T12:38:28.5342893Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5343043Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5343352Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5343528Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5343878Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5344026Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5344171Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5344176Z 2025-12-04T12:38:28.5344395Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5345346Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5345355Z 2025-12-04T12:38:28.5345638Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5346297Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5346426Z Traceback (most recent call last): 2025-12-04T12:38:28.5346865Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5346997Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5347313Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5347488Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5347833Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5348047Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5348183Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5348187Z 2025-12-04T12:38:28.5348441Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5349391Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5349396Z 2025-12-04T12:38:28.5349675Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5349822Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5350475Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5350612Z Traceback (most recent call last): 2025-12-04T12:38:28.5351061Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5351195Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5351509Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5351684Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5352030Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5352181Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5352314Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5352319Z 2025-12-04T12:38:28.5352545Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5353527Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5353537Z 2025-12-04T12:38:28.5353818Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5354448Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-58d14c6ca3ad4514.xml - 2025-12-04T12:38:28.5354618Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5355636Z FAILED [0.1697s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5355642Z 2025-12-04T12:38:28.5355855Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5356817Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5356826Z 2025-12-04T12:38:28.5357089Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5357284Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5357484Z ================== 1 failed, 181 deselected, 2 rerun in 1.34s ================== 2025-12-04T12:38:28.5357581Z Got exit code 1 2025-12-04T12:38:28.5357704Z Retrying single test... 2025-12-04T12:38:28.5358152Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-4358034553dc4d55.xml 2025-12-04T12:38:28.5358315Z ============================= test session starts ============================== 2025-12-04T12:38:28.5358676Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5358785Z cachedir: .pytest_cache 2025-12-04T12:38:28.5359321Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5359504Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5359614Z configfile: pytest.ini 2025-12-04T12:38:28.5360255Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5360485Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5361526Z stepcurrent: skipping 181 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5361655Z Running 1 items in this shard 2025-12-04T12:38:28.5361660Z 2025-12-04T12:38:28.5363478Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda [W1204 12:29:08.169055018 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.5363627Z ('RERUN', {'yellow': True}) [0.8557s] [100%] 2025-12-04T12:38:28.5364562Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1641s] [100%] 2025-12-04T12:38:28.5365406Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda FAILED [0.1653s] [100%] 2025-12-04T12:38:28.5365411Z 2025-12-04T12:38:28.5365550Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5366279Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5366406Z Traceback (most recent call last): 2025-12-04T12:38:28.5366853Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5367000Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5367309Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5367486Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5367832Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5367984Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5368128Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5368133Z 2025-12-04T12:38:28.5368350Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5369312Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5369321Z 2025-12-04T12:38:28.5369605Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5370267Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5370403Z Traceback (most recent call last): 2025-12-04T12:38:28.5370843Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5370978Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5371297Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5371474Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5371866Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5372028Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5372160Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5372194Z 2025-12-04T12:38:28.5372423Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5373374Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5373380Z 2025-12-04T12:38:28.5373646Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5373802Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5374465Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5374603Z Traceback (most recent call last): 2025-12-04T12:38:28.5375040Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5375176Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5375489Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5375664Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5376007Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5376155Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5376285Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5376290Z 2025-12-04T12:38:28.5376516Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5377564Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5377576Z 2025-12-04T12:38:28.5377861Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5378486Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-4358034553dc4d55.xml - 2025-12-04T12:38:28.5378657Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5379676Z FAILED [0.1653s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5379681Z 2025-12-04T12:38:28.5379894Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5380859Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5380868Z 2025-12-04T12:38:28.5381135Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5381317Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5381532Z ================= 1 failed, 1583 deselected, 2 rerun in 1.31s ================== 2025-12-04T12:38:28.5381631Z Got exit code 1 2025-12-04T12:38:28.5381758Z Retrying single test... 2025-12-04T12:38:28.5382213Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-e604dacd305a4e36.xml 2025-12-04T12:38:28.5382374Z ============================= test session starts ============================== 2025-12-04T12:38:28.5382740Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5382913Z cachedir: .pytest_cache 2025-12-04T12:38:28.5383435Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5383574Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5383711Z configfile: pytest.ini 2025-12-04T12:38:28.5384316Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5384546Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5385586Z stepcurrent: skipping 181 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5385713Z Running 1 items in this shard 2025-12-04T12:38:28.5385718Z 2025-12-04T12:38:28.5387538Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda [W1204 12:29:21.496685994 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.5387686Z ('RERUN', {'yellow': True}) [0.8612s] [100%] 2025-12-04T12:38:28.5396677Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1753s] [100%] 2025-12-04T12:38:28.5397665Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda FAILED [0.1695s] [100%] 2025-12-04T12:38:28.5397824Z 2025-12-04T12:38:28.5397974Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5398656Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5398799Z Traceback (most recent call last): 2025-12-04T12:38:28.5399254Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5399437Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5399816Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5399995Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5400348Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5400500Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5400637Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5400643Z 2025-12-04T12:38:28.5400894Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5401859Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5401867Z 2025-12-04T12:38:28.5402154Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5402822Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5402946Z Traceback (most recent call last): 2025-12-04T12:38:28.5403404Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5403541Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5403863Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5404158Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5404492Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5404697Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5404829Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5404835Z 2025-12-04T12:38:28.5405049Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5406017Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5406023Z 2025-12-04T12:38:28.5406287Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5406446Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5407111Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5407249Z Traceback (most recent call last): 2025-12-04T12:38:28.5407693Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5407827Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5408149Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5408323Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5408657Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5408819Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5408950Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5408956Z 2025-12-04T12:38:28.5409213Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5410163Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5410170Z 2025-12-04T12:38:28.5410440Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5411081Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-e604dacd305a4e36.xml - 2025-12-04T12:38:28.5411255Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5412280Z FAILED [0.1695s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5412289Z 2025-12-04T12:38:28.5412502Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5413460Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5413478Z 2025-12-04T12:38:28.5413745Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5413924Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5414146Z ================= 1 failed, 1583 deselected, 2 rerun in 1.33s ================== 2025-12-04T12:38:28.5414244Z Got exit code 1 2025-12-04T12:38:28.5415115Z FAILED CONSISTENTLY: test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5415542Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:38:28.5416058Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-29c18a70d3fdf25d.xml 2025-12-04T12:38:28.5416280Z ============================= test session starts ============================== 2025-12-04T12:38:28.5416633Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5416741Z cachedir: .pytest_cache 2025-12-04T12:38:28.5417354Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5417481Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5417602Z configfile: pytest.ini 2025-12-04T12:38:28.5418198Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5418433Z collecting ... collected 1584 items / 182 deselected / 1402 selected 2025-12-04T12:38:28.5418600Z stepcurrent: skipping 182 already run items. 2025-12-04T12:38:28.5418717Z Running 1402 items in this shard 2025-12-04T12:38:28.5418722Z 2025-12-04T12:38:28.5419630Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8572s] [ 0%] 2025-12-04T12:38:28.5420554Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1712s] [ 0%] 2025-12-04T12:38:28.5421367Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda FAILED [0.1623s] [ 0%] 2025-12-04T12:38:28.5421372Z 2025-12-04T12:38:28.5421563Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5422218Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.5422355Z Traceback (most recent call last): 2025-12-04T12:38:28.5422807Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5422941Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5423261Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5423436Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5423783Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5423932Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5424068Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5424075Z 2025-12-04T12:38:28.5424303Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5425253Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5425259Z 2025-12-04T12:38:28.5425538Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5426182Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.5426304Z Traceback (most recent call last): 2025-12-04T12:38:28.5426756Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5426890Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5427200Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5427419Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5427783Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5427944Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5428105Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5428110Z 2025-12-04T12:38:28.5428321Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5429281Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5429287Z 2025-12-04T12:38:28.5429551Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5429711Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5430361Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.5430487Z Traceback (most recent call last): 2025-12-04T12:38:28.5430944Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5431080Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5431393Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5431566Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5431899Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5432060Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5432190Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5432195Z 2025-12-04T12:38:28.5432438Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5433382Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5433392Z 2025-12-04T12:38:28.5433658Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5434298Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-29c18a70d3fdf25d.xml - 2025-12-04T12:38:28.5434471Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5435456Z FAILED [0.1623s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5435473Z 2025-12-04T12:38:28.5435693Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5436634Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5436641Z 2025-12-04T12:38:28.5436921Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5437101Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5437314Z ================== 1 failed, 182 deselected, 2 rerun in 1.31s ================== 2025-12-04T12:38:28.5437412Z Got exit code 1 2025-12-04T12:38:28.5437520Z Retrying single test... 2025-12-04T12:38:28.5437985Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2a489f89b00449c3.xml 2025-12-04T12:38:28.5438147Z ============================= test session starts ============================== 2025-12-04T12:38:28.5438494Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5438680Z cachedir: .pytest_cache 2025-12-04T12:38:28.5439204Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5439370Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5439479Z configfile: pytest.ini 2025-12-04T12:38:28.5440066Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5440305Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5441338Z stepcurrent: skipping 182 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5441471Z Running 1 items in this shard 2025-12-04T12:38:28.5441476Z 2025-12-04T12:38:28.5442390Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8560s] [100%] 2025-12-04T12:38:28.5443299Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1638s] [100%] 2025-12-04T12:38:28.5444119Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda FAILED [0.1554s] [100%] 2025-12-04T12:38:28.5444124Z 2025-12-04T12:38:28.5444262Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5444950Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.5445078Z Traceback (most recent call last): 2025-12-04T12:38:28.5445524Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5445673Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5445983Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5446166Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5446503Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5446652Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5446795Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5446801Z 2025-12-04T12:38:28.5447013Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5447973Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5447982Z 2025-12-04T12:38:28.5448246Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5448892Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.5449024Z Traceback (most recent call last): 2025-12-04T12:38:28.5449465Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5449613Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5449918Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5450092Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5450440Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5450651Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5450784Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5450789Z 2025-12-04T12:38:28.5451016Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5451981Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5451987Z 2025-12-04T12:38:28.5452267Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5452412Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5453056Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.5453193Z Traceback (most recent call last): 2025-12-04T12:38:28.5453638Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5453784Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5454090Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5454263Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5454608Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5454757Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5454891Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5454908Z 2025-12-04T12:38:28.5455122Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5456092Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5456102Z 2025-12-04T12:38:28.5456381Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5457083Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2a489f89b00449c3.xml - 2025-12-04T12:38:28.5457274Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5458259Z FAILED [0.1554s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5458265Z 2025-12-04T12:38:28.5458479Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5459443Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5459452Z 2025-12-04T12:38:28.5459720Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5459918Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5460121Z ================= 1 failed, 1583 deselected, 2 rerun in 1.29s ================== 2025-12-04T12:38:28.5460221Z Got exit code 1 2025-12-04T12:38:28.5460345Z Retrying single test... 2025-12-04T12:38:28.5460803Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-1e3f7f3e1ddc719e.xml 2025-12-04T12:38:28.5460982Z ============================= test session starts ============================== 2025-12-04T12:38:28.5461331Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5461439Z cachedir: .pytest_cache 2025-12-04T12:38:28.5461980Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5462187Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5462294Z configfile: pytest.ini 2025-12-04T12:38:28.5462931Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5463164Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5464210Z stepcurrent: skipping 182 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5464327Z Running 1 items in this shard 2025-12-04T12:38:28.5464333Z 2025-12-04T12:38:28.5465252Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8610s] [100%] 2025-12-04T12:38:28.5466181Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1870s] [100%] 2025-12-04T12:38:28.5466995Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda FAILED [0.1825s] [100%] 2025-12-04T12:38:28.5467001Z 2025-12-04T12:38:28.5467155Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5467802Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.5467941Z Traceback (most recent call last): 2025-12-04T12:38:28.5468419Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5468563Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5468891Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5469072Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5469411Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5469574Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5469708Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5469712Z 2025-12-04T12:38:28.5469942Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5470889Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5470896Z 2025-12-04T12:38:28.5471162Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5471817Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.5471939Z Traceback (most recent call last): 2025-12-04T12:38:28.5472391Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5472525Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5472835Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5473022Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5473356Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5473517Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5473651Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5473656Z 2025-12-04T12:38:28.5473902Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5474915Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5474921Z 2025-12-04T12:38:28.5475188Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5475347Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5475987Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.5476108Z Traceback (most recent call last): 2025-12-04T12:38:28.5476558Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5476696Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5477008Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5477196Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5477535Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5477702Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5477833Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5477838Z 2025-12-04T12:38:28.5478055Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5479001Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5479006Z 2025-12-04T12:38:28.5479301Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5479944Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-1e3f7f3e1ddc719e.xml - 2025-12-04T12:38:28.5480122Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5481107Z FAILED [0.1825s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5481124Z 2025-12-04T12:38:28.5481337Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5482276Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5482280Z 2025-12-04T12:38:28.5482557Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5482739Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5482958Z ================= 1 failed, 1583 deselected, 2 rerun in 1.35s ================== 2025-12-04T12:38:28.5483057Z Got exit code 1 2025-12-04T12:38:28.5483917Z FAILED CONSISTENTLY: test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5484340Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:38:28.5484789Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-0a2928546dd5ddd2.xml 2025-12-04T12:38:28.5484952Z ============================= test session starts ============================== 2025-12-04T12:38:28.5485315Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5485425Z cachedir: .pytest_cache 2025-12-04T12:38:28.5486024Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5486147Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5486289Z configfile: pytest.ini 2025-12-04T12:38:28.5486893Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5487127Z collecting ... collected 1584 items / 183 deselected / 1401 selected 2025-12-04T12:38:28.5487288Z stepcurrent: skipping 183 already run items. 2025-12-04T12:38:28.5487404Z Running 1401 items in this shard 2025-12-04T12:38:28.5487409Z 2025-12-04T12:38:28.5489235Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda [W1204 12:30:14.477747404 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.5489382Z ('RERUN', {'yellow': True}) [0.8594s] [ 0%] 2025-12-04T12:38:28.5490299Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1661s] [ 0%] 2025-12-04T12:38:28.5491139Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda FAILED [0.1653s] [ 0%] 2025-12-04T12:38:28.5491144Z 2025-12-04T12:38:28.5491285Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5491985Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5492111Z Traceback (most recent call last): 2025-12-04T12:38:28.5492556Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5492705Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5493013Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5493188Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5493531Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5493680Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5493825Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5493830Z 2025-12-04T12:38:28.5494044Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5494988Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5494998Z 2025-12-04T12:38:28.5495276Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5495932Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5496279Z Traceback (most recent call last): 2025-12-04T12:38:28.5496723Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5496913Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5497240Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5497417Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5497757Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5498031Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5498164Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5498169Z 2025-12-04T12:38:28.5498441Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5499394Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5499399Z 2025-12-04T12:38:28.5499663Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5499826Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5500486Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5500619Z Traceback (most recent call last): 2025-12-04T12:38:28.5501065Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5501197Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5501527Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5501701Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5502048Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5502197Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5502329Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5502335Z 2025-12-04T12:38:28.5502563Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5503546Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5503557Z 2025-12-04T12:38:28.5503834Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5504469Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-0a2928546dd5ddd2.xml - 2025-12-04T12:38:28.5504642Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5505653Z FAILED [0.1653s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5505658Z 2025-12-04T12:38:28.5505881Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5506837Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5506847Z 2025-12-04T12:38:28.5507111Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5507293Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5507507Z ================== 1 failed, 183 deselected, 2 rerun in 1.31s ================== 2025-12-04T12:38:28.5507606Z Got exit code 1 2025-12-04T12:38:28.5507724Z Retrying single test... 2025-12-04T12:38:28.5508177Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-36be7a929b5f8cea.xml 2025-12-04T12:38:28.5508338Z ============================= test session starts ============================== 2025-12-04T12:38:28.5508702Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5508812Z cachedir: .pytest_cache 2025-12-04T12:38:28.5509336Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5509560Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5509671Z configfile: pytest.ini 2025-12-04T12:38:28.5510300Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5510530Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5511573Z stepcurrent: skipping 183 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5511700Z Running 1 items in this shard 2025-12-04T12:38:28.5511705Z 2025-12-04T12:38:28.5513532Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda [W1204 12:30:28.802618169 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.5513682Z ('RERUN', {'yellow': True}) [0.8676s] [100%] 2025-12-04T12:38:28.5514606Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1723s] [100%] 2025-12-04T12:38:28.5515438Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda FAILED [0.1666s] [100%] 2025-12-04T12:38:28.5515444Z 2025-12-04T12:38:28.5515612Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5516265Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5516405Z Traceback (most recent call last): 2025-12-04T12:38:28.5516852Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5516996Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5517309Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5517537Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5517889Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5518052Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5518186Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5518203Z 2025-12-04T12:38:28.5518423Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5519375Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5519382Z 2025-12-04T12:38:28.5519663Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5520317Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5520451Z Traceback (most recent call last): 2025-12-04T12:38:28.5520893Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5521026Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5521342Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5521520Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5521988Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5522153Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5522321Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5522326Z 2025-12-04T12:38:28.5522556Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5523506Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5523511Z 2025-12-04T12:38:28.5523782Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5523947Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5524608Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5524750Z Traceback (most recent call last): 2025-12-04T12:38:28.5525192Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5525330Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5525650Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5525824Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5526169Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5526318Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5526448Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5526453Z 2025-12-04T12:38:28.5526678Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5527655Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5527665Z 2025-12-04T12:38:28.5527944Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5528576Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-36be7a929b5f8cea.xml - 2025-12-04T12:38:28.5528748Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5529757Z FAILED [0.1666s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5529763Z 2025-12-04T12:38:28.5529979Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5530934Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5530944Z 2025-12-04T12:38:28.5531210Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5531389Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5531608Z ================= 1 failed, 1583 deselected, 2 rerun in 1.33s ================== 2025-12-04T12:38:28.5531707Z Got exit code 1 2025-12-04T12:38:28.5531822Z Retrying single test... 2025-12-04T12:38:28.5532274Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-db4ab888748348ef.xml 2025-12-04T12:38:28.5532442Z ============================= test session starts ============================== 2025-12-04T12:38:28.5532805Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5532975Z cachedir: .pytest_cache 2025-12-04T12:38:28.5533506Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5533659Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5533768Z configfile: pytest.ini 2025-12-04T12:38:28.5534372Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5534599Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5535644Z stepcurrent: skipping 183 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5535762Z Running 1 items in this shard 2025-12-04T12:38:28.5535767Z 2025-12-04T12:38:28.5537688Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda [W1204 12:30:41.232389340 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.5537841Z ('RERUN', {'yellow': True}) [0.8674s] [100%] 2025-12-04T12:38:28.5538774Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1623s] [100%] 2025-12-04T12:38:28.5539609Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda FAILED [0.1625s] [100%] 2025-12-04T12:38:28.5539656Z 2025-12-04T12:38:28.5539796Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5540466Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5540587Z Traceback (most recent call last): 2025-12-04T12:38:28.5541037Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5541187Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5541497Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5541675Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5542027Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5542177Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5542328Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5542333Z 2025-12-04T12:38:28.5542552Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5543501Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5543506Z 2025-12-04T12:38:28.5543792Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5544443Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5544578Z Traceback (most recent call last): 2025-12-04T12:38:28.5545018Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5545148Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5545461Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5546326Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5546674Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5546856Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5546986Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5546992Z 2025-12-04T12:38:28.5547214Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5548169Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5548174Z 2025-12-04T12:38:28.5548449Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5548598Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5549258Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5549395Z Traceback (most recent call last): 2025-12-04T12:38:28.5549837Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5549967Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5550282Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5550456Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5550796Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5550946Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5551075Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5551080Z 2025-12-04T12:38:28.5551349Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5552300Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5552308Z 2025-12-04T12:38:28.5552584Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5553206Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-db4ab888748348ef.xml - 2025-12-04T12:38:28.5553378Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5554377Z FAILED [0.1625s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5554384Z 2025-12-04T12:38:28.5554595Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5555552Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5555557Z 2025-12-04T12:38:28.5555818Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5556004Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5556203Z ================= 1 failed, 1583 deselected, 2 rerun in 1.31s ================== 2025-12-04T12:38:28.5556300Z Got exit code 1 2025-12-04T12:38:28.5557167Z FAILED CONSISTENTLY: test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5557574Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:38:28.5558087Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a61e284ca72dd15d.xml 2025-12-04T12:38:28.5558291Z ============================= test session starts ============================== 2025-12-04T12:38:28.5558641Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5558764Z cachedir: .pytest_cache 2025-12-04T12:38:28.5559286Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5559407Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5559520Z configfile: pytest.ini 2025-12-04T12:38:28.5560115Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5560360Z collecting ... collected 1584 items / 184 deselected / 1400 selected 2025-12-04T12:38:28.5560508Z stepcurrent: skipping 184 already run items. 2025-12-04T12:38:28.5560621Z Running 1400 items in this shard 2025-12-04T12:38:28.5560626Z 2025-12-04T12:38:28.5561473Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0094s] [ 0%] 2025-12-04T12:38:28.5563297Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda [W1204 12:30:54.428561797 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.5563410Z PASSED [0.0042s] [ 0%] 2025-12-04T12:38:28.5564258Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0037s] [ 0%] 2025-12-04T12:38:28.5565089Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0038s] [ 0%] 2025-12-04T12:38:28.5565915Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0035s] [ 0%] 2025-12-04T12:38:28.5566744Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0035s] [ 0%] 2025-12-04T12:38:28.5567573Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0037s] [ 0%] 2025-12-04T12:38:28.5568393Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0035s] [ 0%] 2025-12-04T12:38:28.5569223Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0036s] [ 0%] 2025-12-04T12:38:28.5570047Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0035s] [ 0%] 2025-12-04T12:38:28.5570861Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0038s] [ 0%] 2025-12-04T12:38:28.5571694Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0035s] [ 0%] 2025-12-04T12:38:28.5572598Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0035s] [ 0%] 2025-12-04T12:38:28.5573427Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0036s] [ 1%] 2025-12-04T12:38:28.5574230Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0038s] [ 1%] 2025-12-04T12:38:28.5575057Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0035s] [ 1%] 2025-12-04T12:38:28.5575879Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0035s] [ 1%] 2025-12-04T12:38:28.5576717Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0035s] [ 1%] 2025-12-04T12:38:28.5577604Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0036s] [ 1%] 2025-12-04T12:38:28.5578439Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0038s] [ 1%] 2025-12-04T12:38:28.5579284Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0035s] [ 1%] 2025-12-04T12:38:28.5580116Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0035s] [ 1%] 2025-12-04T12:38:28.5580935Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0035s] [ 1%] 2025-12-04T12:38:28.5581744Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0035s] [ 1%] 2025-12-04T12:38:28.5582571Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0038s] [ 1%] 2025-12-04T12:38:28.5583405Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0035s] [ 1%] 2025-12-04T12:38:28.5584229Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0035s] [ 1%] 2025-12-04T12:38:28.5585051Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0035s] [ 2%] 2025-12-04T12:38:28.5585860Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0036s] [ 2%] 2025-12-04T12:38:28.5586692Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0043s] [ 2%] 2025-12-04T12:38:28.5587586Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0036s] [ 2%] 2025-12-04T12:38:28.5588412Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0035s] [ 2%] 2025-12-04T12:38:28.5589226Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 2%] 2025-12-04T12:38:28.5590066Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 2%] 2025-12-04T12:38:28.5590883Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0037s] [ 2%] 2025-12-04T12:38:28.5591717Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0037s] [ 2%] 2025-12-04T12:38:28.5592527Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 2%] 2025-12-04T12:38:28.5593346Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 2%] 2025-12-04T12:38:28.5594201Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 2%] 2025-12-04T12:38:28.5595027Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 2%] 2025-12-04T12:38:28.5595859Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 2%] 2025-12-04T12:38:28.5596854Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0037s] [ 3%] 2025-12-04T12:38:28.5597684Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.5598515Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.5599325Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.5600159Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.5600968Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.5601802Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0037s] [ 3%] 2025-12-04T12:38:28.5602780Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.5603619Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.5604429Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.5605269Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.5606084Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.5606902Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0037s] [ 3%] 2025-12-04T12:38:28.5607722Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 3%] 2025-12-04T12:38:28.5608576Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.5609405Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.5610235Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.5611064Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.5611885Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0037s] [ 4%] 2025-12-04T12:38:28.5612705Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.5613531Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.5614334Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.5615165Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.5615983Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0035s] [ 4%] 2025-12-04T12:38:28.5616930Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0037s] [ 4%] 2025-12-04T12:38:28.5617779Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.5618618Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 4%] 2025-12-04T12:38:28.5619431Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0037s] [ 4%] 2025-12-04T12:38:28.5620257Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 5%] 2025-12-04T12:38:28.5621081Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 5%] 2025-12-04T12:38:28.5621896Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0037s] [ 5%] 2025-12-04T12:38:28.5622734Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0034s] [ 5%] 2025-12-04T12:38:28.5623597Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0034s] [ 5%] 2025-12-04T12:38:28.5624430Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0034s] [ 5%] 2025-12-04T12:38:28.5625248Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0034s] [ 5%] 2025-12-04T12:38:28.5626169Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8515s] [ 5%] 2025-12-04T12:38:28.5627077Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1758s] [ 5%] 2025-12-04T12:38:28.5627882Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda FAILED [0.1661s] [ 5%] 2025-12-04T12:38:28.5627903Z 2025-12-04T12:38:28.5628045Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5628697Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5628833Z Traceback (most recent call last): 2025-12-04T12:38:28.5629281Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5629412Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5629736Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5629915Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5630262Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5630472Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5630604Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5630609Z 2025-12-04T12:38:28.5630862Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5631819Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5631825Z 2025-12-04T12:38:28.5632100Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5632750Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5632875Z Traceback (most recent call last): 2025-12-04T12:38:28.5633326Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5633466Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5633783Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5633959Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5634296Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5634455Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5634585Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5634590Z 2025-12-04T12:38:28.5634805Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5635795Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5635805Z 2025-12-04T12:38:28.5636073Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5636226Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5636876Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5636998Z Traceback (most recent call last): 2025-12-04T12:38:28.5637447Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5637581Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5637900Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5638074Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5638410Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5638577Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5638710Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5638716Z 2025-12-04T12:38:28.5638931Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5639888Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5639895Z 2025-12-04T12:38:28.5640162Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5640806Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a61e284ca72dd15d.xml - 2025-12-04T12:38:28.5640974Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5641969Z FAILED [0.1661s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5642055Z 2025-12-04T12:38:28.5642299Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5643235Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5643254Z 2025-12-04T12:38:28.5643522Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5643700Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5643929Z ============ 1 failed, 76 passed, 184 deselected, 2 rerun in 1.69s ============= 2025-12-04T12:38:28.5644030Z Got exit code 1 2025-12-04T12:38:28.5644139Z Retrying single test... 2025-12-04T12:38:28.5644607Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-8da3a511cea25a62.xml 2025-12-04T12:38:28.5644774Z ============================= test session starts ============================== 2025-12-04T12:38:28.5645135Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5645244Z cachedir: .pytest_cache 2025-12-04T12:38:28.5645771Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5645906Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5646015Z configfile: pytest.ini 2025-12-04T12:38:28.5646602Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5646847Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5647919Z stepcurrent: skipping 260 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5648050Z Running 1 items in this shard 2025-12-04T12:38:28.5648057Z 2025-12-04T12:38:28.5648969Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8752s] [100%] 2025-12-04T12:38:28.5649877Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1651s] [100%] 2025-12-04T12:38:28.5650710Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda FAILED [0.1587s] [100%] 2025-12-04T12:38:28.5650719Z 2025-12-04T12:38:28.5650859Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5651521Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5651642Z Traceback (most recent call last): 2025-12-04T12:38:28.5652088Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5652224Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5652534Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5652715Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5653048Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5653198Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5653334Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5653417Z 2025-12-04T12:38:28.5653631Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5654617Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5654622Z 2025-12-04T12:38:28.5654893Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5655547Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5655675Z Traceback (most recent call last): 2025-12-04T12:38:28.5656118Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5656264Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5656572Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5656744Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5657158Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5657308Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5657436Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5657455Z 2025-12-04T12:38:28.5657667Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5658608Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5658613Z 2025-12-04T12:38:28.5658932Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5659075Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5659739Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5659863Z Traceback (most recent call last): 2025-12-04T12:38:28.5660304Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5660450Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5660754Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5660929Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5661280Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5661424Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5661571Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5661576Z 2025-12-04T12:38:28.5661791Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5662742Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5662748Z 2025-12-04T12:38:28.5663027Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5663652Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-8da3a511cea25a62.xml - 2025-12-04T12:38:28.5663836Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5664826Z FAILED [0.1587s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5664889Z 2025-12-04T12:38:28.5665103Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5666084Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5666090Z 2025-12-04T12:38:28.5666356Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5666544Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5666745Z ================= 1 failed, 1583 deselected, 2 rerun in 1.32s ================== 2025-12-04T12:38:28.5666842Z Got exit code 1 2025-12-04T12:38:28.5666960Z Retrying single test... 2025-12-04T12:38:28.5667416Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-aaebada1c5676813.xml 2025-12-04T12:38:28.5667590Z ============================= test session starts ============================== 2025-12-04T12:38:28.5667942Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5668048Z cachedir: .pytest_cache 2025-12-04T12:38:28.5668584Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5668709Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5668813Z configfile: pytest.ini 2025-12-04T12:38:28.5669412Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5669642Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5670707Z stepcurrent: skipping 260 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5670828Z Running 1 items in this shard 2025-12-04T12:38:28.5670833Z 2025-12-04T12:38:28.5671767Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8916s] [100%] 2025-12-04T12:38:28.5672679Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1752s] [100%] 2025-12-04T12:38:28.5673498Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda FAILED [0.1695s] [100%] 2025-12-04T12:38:28.5673512Z 2025-12-04T12:38:28.5673653Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5674294Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5674434Z Traceback (most recent call last): 2025-12-04T12:38:28.5674876Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5675011Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5675330Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5675507Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5675850Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5675999Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5676131Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5676136Z 2025-12-04T12:38:28.5676364Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5677369Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5677404Z 2025-12-04T12:38:28.5677689Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5678340Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5678461Z Traceback (most recent call last): 2025-12-04T12:38:28.5678911Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5679043Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5679346Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5679536Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5679871Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5680031Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5680163Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5680168Z 2025-12-04T12:38:28.5680377Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5681333Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5681338Z 2025-12-04T12:38:28.5681603Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5681761Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5682439Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5682564Z Traceback (most recent call last): 2025-12-04T12:38:28.5683015Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5683150Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5683461Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5683634Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5683962Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5684125Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5684255Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5684259Z 2025-12-04T12:38:28.5684482Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5685525Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5685539Z 2025-12-04T12:38:28.5685803Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5686449Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-aaebada1c5676813.xml - 2025-12-04T12:38:28.5686618Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5687624Z FAILED [0.1695s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5687630Z 2025-12-04T12:38:28.5687845Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5688787Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5688860Z 2025-12-04T12:38:28.5689170Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5689348Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5689565Z ================= 1 failed, 1583 deselected, 2 rerun in 1.36s ================== 2025-12-04T12:38:28.5689663Z Got exit code 1 2025-12-04T12:38:28.5690529Z FAILED CONSISTENTLY: test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5690950Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:38:28.5691406Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-c9e9611bbdf13a2b.xml 2025-12-04T12:38:28.5691586Z ============================= test session starts ============================== 2025-12-04T12:38:28.5691940Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5692050Z cachedir: .pytest_cache 2025-12-04T12:38:28.5692591Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5692714Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5692817Z configfile: pytest.ini 2025-12-04T12:38:28.5693421Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5693651Z collecting ... collected 1584 items / 261 deselected / 1323 selected 2025-12-04T12:38:28.5693838Z stepcurrent: skipping 261 already run items. 2025-12-04T12:38:28.5693954Z Running 1323 items in this shard 2025-12-04T12:38:28.5693964Z 2025-12-04T12:38:28.5695797Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda [W1204 12:31:34.490645327 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.5695942Z ('RERUN', {'yellow': True}) [0.8597s] [ 0%] 2025-12-04T12:38:28.5697071Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1759s] [ 0%] 2025-12-04T12:38:28.5697928Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda FAILED [0.1710s] [ 0%] 2025-12-04T12:38:28.5697937Z 2025-12-04T12:38:28.5698077Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5698749Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5698872Z Traceback (most recent call last): 2025-12-04T12:38:28.5699314Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5699461Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5699772Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5699944Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5700296Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5700445Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5700716Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5700721Z 2025-12-04T12:38:28.5700936Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5701929Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5701951Z 2025-12-04T12:38:28.5702218Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5702877Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5703010Z Traceback (most recent call last): 2025-12-04T12:38:28.5703453Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5703586Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5703919Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5704096Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5704445Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5704595Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5704729Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5704735Z 2025-12-04T12:38:28.5704967Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5705917Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5705923Z 2025-12-04T12:38:28.5706245Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5706398Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5707068Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5707209Z Traceback (most recent call last): 2025-12-04T12:38:28.5707654Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5707803Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5708110Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5708287Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5708637Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5708793Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5708926Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5708935Z 2025-12-04T12:38:28.5709166Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5710117Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5710122Z 2025-12-04T12:38:28.5710405Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5711036Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-c9e9611bbdf13a2b.xml - 2025-12-04T12:38:28.5711210Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5712233Z FAILED [0.1710s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5712304Z 2025-12-04T12:38:28.5712519Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5713507Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5713513Z 2025-12-04T12:38:28.5713779Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5713973Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5714172Z ================== 1 failed, 261 deselected, 2 rerun in 1.33s ================== 2025-12-04T12:38:28.5714273Z Got exit code 1 2025-12-04T12:38:28.5714397Z Retrying single test... 2025-12-04T12:38:28.5714853Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-963810090dd7c38b.xml 2025-12-04T12:38:28.5715021Z ============================= test session starts ============================== 2025-12-04T12:38:28.5715387Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5715500Z cachedir: .pytest_cache 2025-12-04T12:38:28.5716034Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5716153Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5716260Z configfile: pytest.ini 2025-12-04T12:38:28.5716868Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5717096Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5718251Z stepcurrent: skipping 261 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5718372Z Running 1 items in this shard 2025-12-04T12:38:28.5718378Z 2025-12-04T12:38:28.5720210Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda [W1204 12:31:48.049863097 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.5720359Z ('RERUN', {'yellow': True}) [0.8848s] [100%] 2025-12-04T12:38:28.5721282Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1635s] [100%] 2025-12-04T12:38:28.5722127Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda FAILED [0.1556s] [100%] 2025-12-04T12:38:28.5722137Z 2025-12-04T12:38:28.5722280Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5722950Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5723071Z Traceback (most recent call last): 2025-12-04T12:38:28.5723515Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5723660Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5723970Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5724144Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5724490Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5724700Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5724845Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5724850Z 2025-12-04T12:38:28.5725092Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5726043Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5726049Z 2025-12-04T12:38:28.5726323Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5726984Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5727118Z Traceback (most recent call last): 2025-12-04T12:38:28.5727557Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5727696Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5728015Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5728190Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5728535Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5728683Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5728816Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5728820Z 2025-12-04T12:38:28.5729043Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5730022Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5730031Z 2025-12-04T12:38:28.5730314Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5730460Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5731119Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5731256Z Traceback (most recent call last): 2025-12-04T12:38:28.5731693Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5731828Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5732141Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5732317Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5732664Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5732816Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5732947Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5732952Z 2025-12-04T12:38:28.5733176Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5734128Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5734134Z 2025-12-04T12:38:28.5734409Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5735031Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-963810090dd7c38b.xml - 2025-12-04T12:38:28.5735201Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5736218Z FAILED [0.1556s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5736279Z 2025-12-04T12:38:28.5736526Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5737573Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5737580Z 2025-12-04T12:38:28.5737849Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5738042Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5738248Z ================= 1 failed, 1583 deselected, 2 rerun in 1.32s ================== 2025-12-04T12:38:28.5738347Z Got exit code 1 2025-12-04T12:38:28.5738473Z Retrying single test... 2025-12-04T12:38:28.5738933Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-202f1cfa5f6f2e59.xml 2025-12-04T12:38:28.5739099Z ============================= test session starts ============================== 2025-12-04T12:38:28.5739469Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5739580Z cachedir: .pytest_cache 2025-12-04T12:38:28.5740118Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5740244Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5740355Z configfile: pytest.ini 2025-12-04T12:38:28.5740967Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5741197Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5742296Z stepcurrent: skipping 261 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5742430Z Running 1 items in this shard 2025-12-04T12:38:28.5742437Z 2025-12-04T12:38:28.5744260Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda [W1204 12:32:01.172246944 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.5744406Z ('RERUN', {'yellow': True}) [0.8626s] [100%] 2025-12-04T12:38:28.5745336Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1888s] [100%] 2025-12-04T12:38:28.5746183Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda FAILED [0.1846s] [100%] 2025-12-04T12:38:28.5746188Z 2025-12-04T12:38:28.5746328Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5746997Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5747116Z Traceback (most recent call last): 2025-12-04T12:38:28.5747560Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5747708Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5748027Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5748204Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5748613Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5748764Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5748938Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5748943Z 2025-12-04T12:38:28.5749161Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5750110Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5750115Z 2025-12-04T12:38:28.5750394Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5751060Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5751197Z Traceback (most recent call last): 2025-12-04T12:38:28.5751639Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5751775Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5752091Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5752266Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5752601Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5752764Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5752896Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5752901Z 2025-12-04T12:38:28.5753128Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5754119Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5754130Z 2025-12-04T12:38:28.5754397Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5754554Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5755211Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5755341Z Traceback (most recent call last): 2025-12-04T12:38:28.5755780Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5755914Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5756229Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5756405Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5756758Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5756907Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5757041Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5757048Z 2025-12-04T12:38:28.5757273Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5758220Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5758225Z 2025-12-04T12:38:28.5758502Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5759127Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-202f1cfa5f6f2e59.xml - 2025-12-04T12:38:28.5759300Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5760405Z FAILED [0.1846s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5760411Z 2025-12-04T12:38:28.5760626Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5761589Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5761595Z 2025-12-04T12:38:28.5761861Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5762038Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5762256Z ================= 1 failed, 1583 deselected, 2 rerun in 1.36s ================== 2025-12-04T12:38:28.5762358Z Got exit code 1 2025-12-04T12:38:28.5763242Z FAILED CONSISTENTLY: test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.5763656Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:38:28.5764112Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a5f1ba0194e0f361.xml 2025-12-04T12:38:28.5764285Z ============================= test session starts ============================== 2025-12-04T12:38:28.5764636Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5764756Z cachedir: .pytest_cache 2025-12-04T12:38:28.5765307Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5765433Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5765559Z configfile: pytest.ini 2025-12-04T12:38:28.5766155Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5766385Z collecting ... collected 1584 items / 262 deselected / 1322 selected 2025-12-04T12:38:28.5766544Z stepcurrent: skipping 262 already run items. 2025-12-04T12:38:28.5766658Z Running 1322 items in this shard 2025-12-04T12:38:28.5766663Z 2025-12-04T12:38:28.5767593Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8677s] [ 0%] 2025-12-04T12:38:28.5768504Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1746s] [ 0%] 2025-12-04T12:38:28.5769327Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda FAILED [0.1673s] [ 0%] 2025-12-04T12:38:28.5769332Z 2025-12-04T12:38:28.5769473Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5770120Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.5770254Z Traceback (most recent call last): 2025-12-04T12:38:28.5770693Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5770841Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5771152Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5771332Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5771739Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5771890Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5772051Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5772057Z 2025-12-04T12:38:28.5772284Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5773224Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5773232Z 2025-12-04T12:38:28.5773510Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5774159Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.5774281Z Traceback (most recent call last): 2025-12-04T12:38:28.5774746Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5774884Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5775210Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5775385Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5775721Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5775889Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5776022Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5776027Z 2025-12-04T12:38:28.5776244Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5777323Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5777334Z 2025-12-04T12:38:28.5777607Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5777771Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5778416Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.5778553Z Traceback (most recent call last): 2025-12-04T12:38:28.5778993Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5779128Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5779448Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5779622Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5779957Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5780127Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5780258Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5780263Z 2025-12-04T12:38:28.5780496Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5781439Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5781444Z 2025-12-04T12:38:28.5781710Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5782352Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a5f1ba0194e0f361.xml - 2025-12-04T12:38:28.5782526Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5783538Z FAILED [0.1673s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5783602Z 2025-12-04T12:38:28.5783864Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5784820Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5784825Z 2025-12-04T12:38:28.5785093Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5785274Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5785489Z ================== 1 failed, 262 deselected, 2 rerun in 1.33s ================== 2025-12-04T12:38:28.5785591Z Got exit code 1 2025-12-04T12:38:28.5785700Z Retrying single test... 2025-12-04T12:38:28.5786174Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-784ae621331a35c7.xml 2025-12-04T12:38:28.5786338Z ============================= test session starts ============================== 2025-12-04T12:38:28.5786705Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5786814Z cachedir: .pytest_cache 2025-12-04T12:38:28.5787341Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5787479Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5787588Z configfile: pytest.ini 2025-12-04T12:38:28.5788187Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5788456Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5789494Z stepcurrent: skipping 262 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5789623Z Running 1 items in this shard 2025-12-04T12:38:28.5789628Z 2025-12-04T12:38:28.5790530Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8730s] [100%] 2025-12-04T12:38:28.5791447Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1767s] [100%] 2025-12-04T12:38:28.5792266Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda FAILED [0.1694s] [100%] 2025-12-04T12:38:28.5792276Z 2025-12-04T12:38:28.5792417Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5793072Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.5793192Z Traceback (most recent call last): 2025-12-04T12:38:28.5793641Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5793776Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5794085Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5794270Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5794607Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5794767Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5794963Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5794968Z 2025-12-04T12:38:28.5795182Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5796362Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5796368Z 2025-12-04T12:38:28.5796637Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5797295Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.5797415Z Traceback (most recent call last): 2025-12-04T12:38:28.5797859Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5798006Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5798316Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5798493Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5798842Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5798990Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5799137Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5799142Z 2025-12-04T12:38:28.5799356Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5800292Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5800297Z 2025-12-04T12:38:28.5800619Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5800770Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5801432Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.5801552Z Traceback (most recent call last): 2025-12-04T12:38:28.5801996Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5802143Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5802444Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5802629Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5802961Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5803110Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5803257Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5803266Z 2025-12-04T12:38:28.5803482Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5804423Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5804440Z 2025-12-04T12:38:28.5804705Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5805331Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-784ae621331a35c7.xml - 2025-12-04T12:38:28.5805515Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5806504Z FAILED [0.1694s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5806599Z 2025-12-04T12:38:28.5806828Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5807808Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5807814Z 2025-12-04T12:38:28.5808085Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5808280Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5808484Z ================= 1 failed, 1583 deselected, 2 rerun in 1.34s ================== 2025-12-04T12:38:28.5808597Z Got exit code 1 2025-12-04T12:38:28.5808705Z Retrying single test... 2025-12-04T12:38:28.5809168Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2c9da68c717b465c.xml 2025-12-04T12:38:28.5809347Z ============================= test session starts ============================== 2025-12-04T12:38:28.5809701Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5809813Z cachedir: .pytest_cache 2025-12-04T12:38:28.5810347Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5810472Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5810593Z configfile: pytest.ini 2025-12-04T12:38:28.5811184Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5811411Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5812477Z stepcurrent: skipping 262 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5812596Z Running 1 items in this shard 2025-12-04T12:38:28.5812601Z 2025-12-04T12:38:28.5813527Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8651s] [100%] 2025-12-04T12:38:28.5814428Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1649s] [100%] 2025-12-04T12:38:28.5815237Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda FAILED [0.1560s] [100%] 2025-12-04T12:38:28.5815254Z 2025-12-04T12:38:28.5815394Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5816040Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.5816175Z Traceback (most recent call last): 2025-12-04T12:38:28.5816622Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5816758Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5817143Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5817323Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5817671Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5817822Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5817952Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5817957Z 2025-12-04T12:38:28.5818189Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5819227Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5819234Z 2025-12-04T12:38:28.5819516Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5820163Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.5820286Z Traceback (most recent call last): 2025-12-04T12:38:28.5820741Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5820880Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5821201Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5821375Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5821711Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5821876Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5822008Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5822014Z 2025-12-04T12:38:28.5822228Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5823177Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5823182Z 2025-12-04T12:38:28.5823448Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5823607Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5824278Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.5824404Z Traceback (most recent call last): 2025-12-04T12:38:28.5824858Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5824993Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5825310Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5825483Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5825818Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5825980Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5826112Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5826117Z 2025-12-04T12:38:28.5826347Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5827296Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5827306Z 2025-12-04T12:38:28.5827576Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5828217Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2c9da68c717b465c.xml - 2025-12-04T12:38:28.5828390Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5829379Z FAILED [0.1560s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5829385Z 2025-12-04T12:38:28.5829601Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5830538Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5830627Z 2025-12-04T12:38:28.5830923Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5831105Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5831321Z ================= 1 failed, 1583 deselected, 2 rerun in 1.31s ================== 2025-12-04T12:38:28.5831421Z Got exit code 1 2025-12-04T12:38:28.5832274Z FAILED CONSISTENTLY: test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.5832699Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:38:28.5833149Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a06aff42aa9d7210.xml 2025-12-04T12:38:28.5833329Z ============================= test session starts ============================== 2025-12-04T12:38:28.5833681Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5833789Z cachedir: .pytest_cache 2025-12-04T12:38:28.5834320Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5834444Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5834566Z configfile: pytest.ini 2025-12-04T12:38:28.5835157Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5835388Z collecting ... collected 1584 items / 263 deselected / 1321 selected 2025-12-04T12:38:28.5835577Z stepcurrent: skipping 263 already run items. 2025-12-04T12:38:28.5835696Z Running 1321 items in this shard 2025-12-04T12:38:28.5835702Z 2025-12-04T12:38:28.5837531Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda [W1204 12:32:54.175179017 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.5837661Z ('RERUN', {'yellow': True}) [0.8625s] [ 0%] 2025-12-04T12:38:28.5838578Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1722s] [ 0%] 2025-12-04T12:38:28.5839413Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda FAILED [0.1716s] [ 0%] 2025-12-04T12:38:28.5839422Z 2025-12-04T12:38:28.5839561Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5840227Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5845632Z Traceback (most recent call last): 2025-12-04T12:38:28.5846095Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5846252Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5846569Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5846747Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5847108Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5847259Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5847471Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5847493Z 2025-12-04T12:38:28.5847713Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5848725Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5848770Z 2025-12-04T12:38:28.5849042Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5849713Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5849836Z Traceback (most recent call last): 2025-12-04T12:38:28.5850278Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5850432Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5850741Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5850934Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5851273Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5851424Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5851572Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5851577Z 2025-12-04T12:38:28.5851792Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5852743Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5852763Z 2025-12-04T12:38:28.5853070Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5853220Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5853892Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5854014Z Traceback (most recent call last): 2025-12-04T12:38:28.5854454Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5854606Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5854913Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5855102Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5855435Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5855590Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5855738Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5855746Z 2025-12-04T12:38:28.5855961Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5857026Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5857104Z 2025-12-04T12:38:28.5857378Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5858012Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a06aff42aa9d7210.xml - 2025-12-04T12:38:28.5858202Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5859209Z FAILED [0.1716s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5859248Z 2025-12-04T12:38:28.5859540Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5860919Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5860931Z 2025-12-04T12:38:28.5861224Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5861406Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5861610Z ================== 1 failed, 263 deselected, 2 rerun in 1.33s ================== 2025-12-04T12:38:28.5861724Z Got exit code 1 2025-12-04T12:38:28.5861833Z Retrying single test... 2025-12-04T12:38:28.5862294Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-66f0187c8b5c56f0.xml 2025-12-04T12:38:28.5862478Z ============================= test session starts ============================== 2025-12-04T12:38:28.5862830Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5862958Z cachedir: .pytest_cache 2025-12-04T12:38:28.5863485Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5863611Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5863733Z configfile: pytest.ini 2025-12-04T12:38:28.5864328Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5864558Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5865659Z stepcurrent: skipping 263 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5865780Z Running 1 items in this shard 2025-12-04T12:38:28.5865787Z 2025-12-04T12:38:28.5867624Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda [W1204 12:33:07.360715797 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.5867760Z ('RERUN', {'yellow': True}) [0.8552s] [100%] 2025-12-04T12:38:28.5868696Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1646s] [100%] 2025-12-04T12:38:28.5869525Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda FAILED [0.1746s] [100%] 2025-12-04T12:38:28.5869532Z 2025-12-04T12:38:28.5869689Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5870349Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5870530Z Traceback (most recent call last): 2025-12-04T12:38:28.5870991Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5871126Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5871437Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5871628Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5871964Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5872161Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5872294Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5872300Z 2025-12-04T12:38:28.5872545Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5873510Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5873519Z 2025-12-04T12:38:28.5873787Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5874455Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5874579Z Traceback (most recent call last): 2025-12-04T12:38:28.5875021Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5875170Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5875478Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5875667Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5876003Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5876155Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5876299Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5876304Z 2025-12-04T12:38:28.5876548Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5877552Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5877575Z 2025-12-04T12:38:28.5877842Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5877989Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5878659Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5878782Z Traceback (most recent call last): 2025-12-04T12:38:28.5879223Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5879374Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5879680Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5879866Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5880204Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5880355Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5880501Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5880506Z 2025-12-04T12:38:28.5880720Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5881686Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5881728Z 2025-12-04T12:38:28.5882000Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5882630Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-66f0187c8b5c56f0.xml - 2025-12-04T12:38:28.5882812Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5883809Z FAILED [0.1746s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5883846Z 2025-12-04T12:38:28.5884103Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5885053Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5885061Z 2025-12-04T12:38:28.5885339Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5885520Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5885724Z ================= 1 failed, 1583 deselected, 2 rerun in 1.32s ================== 2025-12-04T12:38:28.5885836Z Got exit code 1 2025-12-04T12:38:28.5885946Z Retrying single test... 2025-12-04T12:38:28.5886402Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a93a155b666a377d.xml 2025-12-04T12:38:28.5886581Z ============================= test session starts ============================== 2025-12-04T12:38:28.5886934Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5887057Z cachedir: .pytest_cache 2025-12-04T12:38:28.5887581Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5887709Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5887832Z configfile: pytest.ini 2025-12-04T12:38:28.5888424Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5888655Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5889739Z stepcurrent: skipping 263 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5889859Z Running 1 items in this shard 2025-12-04T12:38:28.5889864Z 2025-12-04T12:38:28.5891692Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda [W1204 12:33:21.734790869 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.5891828Z ('RERUN', {'yellow': True}) [0.8661s] [100%] 2025-12-04T12:38:28.5892760Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1685s] [100%] 2025-12-04T12:38:28.5893590Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda FAILED [0.1649s] [100%] 2025-12-04T12:38:28.5893596Z 2025-12-04T12:38:28.5893735Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5894438Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5894561Z Traceback (most recent call last): 2025-12-04T12:38:28.5895016Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5895151Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5895461Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5895650Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5896255Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5896424Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5896629Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5896634Z 2025-12-04T12:38:28.5896921Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5897893Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5897898Z 2025-12-04T12:38:28.5898164Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5898843Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5898967Z Traceback (most recent call last): 2025-12-04T12:38:28.5899410Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5899562Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5899867Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5900045Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5900398Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5900548Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5900698Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5900703Z 2025-12-04T12:38:28.5900916Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5901912Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5901920Z 2025-12-04T12:38:28.5902206Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5902354Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5903025Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5903150Z Traceback (most recent call last): 2025-12-04T12:38:28.5903591Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5903738Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5904044Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5904235Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5904565Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5904716Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5904865Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5904872Z 2025-12-04T12:38:28.5905085Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5906030Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5906107Z 2025-12-04T12:38:28.5906377Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5907006Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a93a155b666a377d.xml - 2025-12-04T12:38:28.5907192Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5908344Z FAILED [0.1649s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5908351Z 2025-12-04T12:38:28.5908583Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5909525Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5909530Z 2025-12-04T12:38:28.5909795Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5909987Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5910192Z ================= 1 failed, 1583 deselected, 2 rerun in 1.32s ================== 2025-12-04T12:38:28.5910307Z Got exit code 1 2025-12-04T12:38:28.5911171Z FAILED CONSISTENTLY: test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.5911583Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:38:28.5912049Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-f6900cb9a4d4e17a.xml 2025-12-04T12:38:28.5912214Z ============================= test session starts ============================== 2025-12-04T12:38:28.5912577Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5912688Z cachedir: .pytest_cache 2025-12-04T12:38:28.5913240Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5913377Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5913489Z configfile: pytest.ini 2025-12-04T12:38:28.5914084Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5914334Z collecting ... collected 1584 items / 264 deselected / 1320 selected 2025-12-04T12:38:28.5914482Z stepcurrent: skipping 264 already run items. 2025-12-04T12:38:28.5914614Z Running 1320 items in this shard 2025-12-04T12:38:28.5914619Z 2025-12-04T12:38:28.5915457Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0092s] [ 0%] 2025-12-04T12:38:28.5917294Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda [W1204 12:33:34.070656560 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.5917402Z PASSED [0.0042s] [ 0%] 2025-12-04T12:38:28.5918229Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0037s] [ 0%] 2025-12-04T12:38:28.5919114Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0037s] [ 0%] 2025-12-04T12:38:28.5919928Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0035s] [ 0%] 2025-12-04T12:38:28.5920776Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0036s] [ 0%] 2025-12-04T12:38:28.5921672Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0037s] [ 0%] 2025-12-04T12:38:28.5922508Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0035s] [ 0%] 2025-12-04T12:38:28.5923331Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0036s] [ 0%] 2025-12-04T12:38:28.5924170Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0036s] [ 0%] 2025-12-04T12:38:28.5925007Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0037s] [ 0%] 2025-12-04T12:38:28.5925839Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0035s] [ 0%] 2025-12-04T12:38:28.5926768Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8634s] [ 0%] 2025-12-04T12:38:28.5927703Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1717s] [ 0%] 2025-12-04T12:38:28.5928542Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda FAILED [0.1587s] [ 0%] 2025-12-04T12:38:28.5928550Z 2025-12-04T12:38:28.5928693Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5929341Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5929480Z Traceback (most recent call last): 2025-12-04T12:38:28.5929924Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5930075Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5930384Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5930562Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5930913Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5931068Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5931214Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5931222Z 2025-12-04T12:38:28.5931437Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5932384Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5932424Z 2025-12-04T12:38:28.5932708Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5933358Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5933497Z Traceback (most recent call last): 2025-12-04T12:38:28.5933970Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5934105Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5934456Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5934633Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5934970Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5935134Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5935266Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5935271Z 2025-12-04T12:38:28.5935500Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5936448Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5936456Z 2025-12-04T12:38:28.5936723Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5936951Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5937605Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5937744Z Traceback (most recent call last): 2025-12-04T12:38:28.5938187Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5938323Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5938643Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5938820Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5939209Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5939363Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5939498Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5939504Z 2025-12-04T12:38:28.5939733Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5940670Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5940677Z 2025-12-04T12:38:28.5940957Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5941589Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-f6900cb9a4d4e17a.xml - 2025-12-04T12:38:28.5941763Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5942769Z FAILED [0.1587s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5942776Z 2025-12-04T12:38:28.5942994Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5943944Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5943986Z 2025-12-04T12:38:28.5944252Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5944431Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5944661Z ============ 1 failed, 12 passed, 264 deselected, 2 rerun in 1.38s ============= 2025-12-04T12:38:28.5944763Z Got exit code 1 2025-12-04T12:38:28.5944884Z Retrying single test... 2025-12-04T12:38:28.5945369Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-b4de48ce2713ebfd.xml 2025-12-04T12:38:28.5945530Z ============================= test session starts ============================== 2025-12-04T12:38:28.5945923Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5946035Z cachedir: .pytest_cache 2025-12-04T12:38:28.5946555Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5946690Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5946798Z configfile: pytest.ini 2025-12-04T12:38:28.5947400Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5947629Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5948659Z stepcurrent: skipping 276 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5948788Z Running 1 items in this shard 2025-12-04T12:38:28.5948793Z 2025-12-04T12:38:28.5949706Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8751s] [100%] 2025-12-04T12:38:28.5950631Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1674s] [100%] 2025-12-04T12:38:28.5951492Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda FAILED [0.1646s] [100%] 2025-12-04T12:38:28.5951500Z 2025-12-04T12:38:28.5951655Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5952305Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5952426Z Traceback (most recent call last): 2025-12-04T12:38:28.5952889Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5953028Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5953341Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5953531Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5953867Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5954035Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5954170Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5954175Z 2025-12-04T12:38:28.5954392Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5955349Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5955389Z 2025-12-04T12:38:28.5955662Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5956321Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5956443Z Traceback (most recent call last): 2025-12-04T12:38:28.5956885Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5957033Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5957372Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5957560Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5957923Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5958075Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5958221Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5958225Z 2025-12-04T12:38:28.5958435Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5959376Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5959393Z 2025-12-04T12:38:28.5959661Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5959805Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5960469Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5960591Z Traceback (most recent call last): 2025-12-04T12:38:28.5961043Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5961181Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5961485Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5961673Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5962008Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5962157Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5962336Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5962344Z 2025-12-04T12:38:28.5962559Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5963628Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5963636Z 2025-12-04T12:38:28.5963905Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5964541Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-b4de48ce2713ebfd.xml - 2025-12-04T12:38:28.5964727Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5965721Z FAILED [0.1646s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5965729Z 2025-12-04T12:38:28.5965957Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5966901Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5966906Z 2025-12-04T12:38:28.5967238Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5967417Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5967622Z ================= 1 failed, 1583 deselected, 2 rerun in 1.33s ================== 2025-12-04T12:38:28.5967737Z Got exit code 1 2025-12-04T12:38:28.5967845Z Retrying single test... 2025-12-04T12:38:28.5968308Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-9d9a1cc8efaa0b76.xml 2025-12-04T12:38:28.5968484Z ============================= test session starts ============================== 2025-12-04T12:38:28.5968883Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5969004Z cachedir: .pytest_cache 2025-12-04T12:38:28.5969561Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5969688Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5969808Z configfile: pytest.ini 2025-12-04T12:38:28.5970403Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5970630Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.5971681Z stepcurrent: skipping 276 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5971798Z Running 1 items in this shard 2025-12-04T12:38:28.5971804Z 2025-12-04T12:38:28.5972725Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8655s] [100%] 2025-12-04T12:38:28.5973643Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1845s] [100%] 2025-12-04T12:38:28.5974477Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda FAILED [0.1753s] [100%] 2025-12-04T12:38:28.5974483Z 2025-12-04T12:38:28.5974653Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5975306Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5975440Z Traceback (most recent call last): 2025-12-04T12:38:28.5975885Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5976036Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5976347Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5976526Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5976940Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5977093Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5977226Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5977245Z 2025-12-04T12:38:28.5977467Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5978414Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5978422Z 2025-12-04T12:38:28.5978704Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5979394Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5979532Z Traceback (most recent call last): 2025-12-04T12:38:28.5979973Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5980108Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5980431Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5980606Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5980970Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5981131Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5981294Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5981300Z 2025-12-04T12:38:28.5981526Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5982478Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5982484Z 2025-12-04T12:38:28.5982749Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5982904Z =================================== FAILURES =================================== 2025-12-04T12:38:28.5983556Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda _ 2025-12-04T12:38:28.5983692Z Traceback (most recent call last): 2025-12-04T12:38:28.5984133Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.5984269Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.5984590Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.5984766Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.5985103Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.5985268Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.5985398Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5985403Z 2025-12-04T12:38:28.5985629Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5986597Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5986604Z 2025-12-04T12:38:28.5986885Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5987514Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-9d9a1cc8efaa0b76.xml - 2025-12-04T12:38:28.5987689Z =========================== short test summary info ============================ 2025-12-04T12:38:28.5988688Z FAILED [0.1753s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.5988694Z 2025-12-04T12:38:28.5988914Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.5989855Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5989862Z 2025-12-04T12:38:28.5990139Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.5990332Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.5990569Z ================= 1 failed, 1583 deselected, 2 rerun in 1.35s ================== 2025-12-04T12:38:28.5990682Z Got exit code 1 2025-12-04T12:38:28.5991550Z FAILED CONSISTENTLY: test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda 2025-12-04T12:38:28.5991966Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:38:28.5992438Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-f2c6d1655dbd05d5.xml 2025-12-04T12:38:28.5992633Z ============================= test session starts ============================== 2025-12-04T12:38:28.5993025Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.5993137Z cachedir: .pytest_cache 2025-12-04T12:38:28.5993661Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.5993803Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.5993914Z configfile: pytest.ini 2025-12-04T12:38:28.5994512Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.5994763Z collecting ... collected 1584 items / 277 deselected / 1307 selected 2025-12-04T12:38:28.5994916Z stepcurrent: skipping 277 already run items. 2025-12-04T12:38:28.5995049Z Running 1307 items in this shard 2025-12-04T12:38:28.5995054Z 2025-12-04T12:38:28.5997055Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda [W1204 12:34:14.045338970 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.5997191Z ('RERUN', {'yellow': True}) [0.8628s] [ 0%] 2025-12-04T12:38:28.5998128Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1688s] [ 0%] 2025-12-04T12:38:28.5999017Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda FAILED [0.1612s] [ 0%] 2025-12-04T12:38:28.5999026Z 2025-12-04T12:38:28.5999180Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.5999842Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.5999981Z Traceback (most recent call last): 2025-12-04T12:38:28.6000427Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6000567Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6000894Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6001073Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6001412Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6001580Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6001720Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6001725Z 2025-12-04T12:38:28.6001956Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6002914Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.6002972Z 2025-12-04T12:38:28.6003246Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6003922Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.6004043Z Traceback (most recent call last): 2025-12-04T12:38:28.6004506Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6004685Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6004990Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6005182Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6005554Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6005722Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6005856Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6005861Z 2025-12-04T12:38:28.6006077Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6007047Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.6007052Z 2025-12-04T12:38:28.6007318Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6007475Z =================================== FAILURES =================================== 2025-12-04T12:38:28.6008133Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.6008256Z Traceback (most recent call last): 2025-12-04T12:38:28.6008705Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6008840Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6009144Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6009332Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6009666Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6009859Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6009991Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6009998Z 2025-12-04T12:38:28.6010212Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6011172Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.6011180Z 2025-12-04T12:38:28.6011450Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6012095Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-f2c6d1655dbd05d5.xml - 2025-12-04T12:38:28.6012269Z =========================== short test summary info ============================ 2025-12-04T12:38:28.6013275Z FAILED [0.1612s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6013293Z 2025-12-04T12:38:28.6013508Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6014455Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.6014514Z 2025-12-04T12:38:28.6014796Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6014977Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.6015193Z ================== 1 failed, 277 deselected, 2 rerun in 1.32s ================== 2025-12-04T12:38:28.6015291Z Got exit code 1 2025-12-04T12:38:28.6015398Z Retrying single test... 2025-12-04T12:38:28.6015873Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-30cf0cd6fabb5398.xml 2025-12-04T12:38:28.6016071Z ============================= test session starts ============================== 2025-12-04T12:38:28.6016416Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.6016566Z cachedir: .pytest_cache 2025-12-04T12:38:28.6017159Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.6017305Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.6017414Z configfile: pytest.ini 2025-12-04T12:38:28.6018011Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.6018252Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.6019305Z stepcurrent: skipping 277 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.6019423Z Running 1 items in this shard 2025-12-04T12:38:28.6019443Z 2025-12-04T12:38:28.6021267Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda [W1204 12:34:27.244772257 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.6021399Z ('RERUN', {'yellow': True}) [0.8523s] [100%] 2025-12-04T12:38:28.6022377Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1638s] [100%] 2025-12-04T12:38:28.6023207Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda FAILED [0.1591s] [100%] 2025-12-04T12:38:28.6023214Z 2025-12-04T12:38:28.6023370Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.6024033Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.6024170Z Traceback (most recent call last): 2025-12-04T12:38:28.6024616Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6024750Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6025071Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6025252Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6025587Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6025754Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6025886Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6025891Z 2025-12-04T12:38:28.6026121Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6027074Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.6027114Z 2025-12-04T12:38:28.6027385Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6028055Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.6028178Z Traceback (most recent call last): 2025-12-04T12:38:28.6028630Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6028799Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6029136Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6029325Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6029658Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6029810Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6029957Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6029962Z 2025-12-04T12:38:28.6030175Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6031137Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.6031144Z 2025-12-04T12:38:28.6031411Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6031557Z =================================== FAILURES =================================== 2025-12-04T12:38:28.6032228Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.6032353Z Traceback (most recent call last): 2025-12-04T12:38:28.6032807Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6032941Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6033245Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6033431Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6033802Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6033968Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6034100Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6034106Z 2025-12-04T12:38:28.6034322Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6035283Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.6035290Z 2025-12-04T12:38:28.6035559Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6036205Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-30cf0cd6fabb5398.xml - 2025-12-04T12:38:28.6036378Z =========================== short test summary info ============================ 2025-12-04T12:38:28.6037380Z FAILED [0.1591s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6037387Z 2025-12-04T12:38:28.6037611Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6038557Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.6038601Z 2025-12-04T12:38:28.6038881Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6039060Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.6039265Z ================= 1 failed, 1583 deselected, 2 rerun in 1.30s ================== 2025-12-04T12:38:28.6039380Z Got exit code 1 2025-12-04T12:38:28.6039490Z Retrying single test... 2025-12-04T12:38:28.6039989Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-29e9b8d6f2f6950a.xml 2025-12-04T12:38:28.6040150Z ============================= test session starts ============================== 2025-12-04T12:38:28.6040525Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.6040645Z cachedir: .pytest_cache 2025-12-04T12:38:28.6041169Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.6041293Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.6041411Z configfile: pytest.ini 2025-12-04T12:38:28.6042002Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.6042246Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.6043278Z stepcurrent: skipping 277 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.6043396Z Running 1 items in this shard 2025-12-04T12:38:28.6043401Z 2025-12-04T12:38:28.6045233Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda [W1204 12:34:40.624526546 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.6045367Z ('RERUN', {'yellow': True}) [0.8636s] [100%] 2025-12-04T12:38:28.6046340Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1715s] [100%] 2025-12-04T12:38:28.6047181Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda FAILED [0.1718s] [100%] 2025-12-04T12:38:28.6047186Z 2025-12-04T12:38:28.6047339Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.6048002Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.6048125Z Traceback (most recent call last): 2025-12-04T12:38:28.6048581Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6048720Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6049051Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6049229Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6049565Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6049731Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6049865Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6049871Z 2025-12-04T12:38:28.6050105Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6051115Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.6051122Z 2025-12-04T12:38:28.6051392Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6052068Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.6052224Z Traceback (most recent call last): 2025-12-04T12:38:28.6052680Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6052848Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6053157Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6053348Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6053683Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6053833Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6053979Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6053984Z 2025-12-04T12:38:28.6054197Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6055163Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.6055171Z 2025-12-04T12:38:28.6055735Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6055946Z =================================== FAILURES =================================== 2025-12-04T12:38:28.6056652Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda _ 2025-12-04T12:38:28.6056921Z Traceback (most recent call last): 2025-12-04T12:38:28.6064030Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6064235Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6064568Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6064860Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6065214Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6065373Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6065529Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6065537Z 2025-12-04T12:38:28.6065755Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6066726Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.6066745Z 2025-12-04T12:38:28.6067017Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6067648Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-29e9b8d6f2f6950a.xml - 2025-12-04T12:38:28.6067838Z =========================== short test summary info ============================ 2025-12-04T12:38:28.6068849Z FAILED [0.1718s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6068857Z 2025-12-04T12:38:28.6069084Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6070080Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.6070085Z 2025-12-04T12:38:28.6070354Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6070549Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.6070758Z ================= 1 failed, 1583 deselected, 2 rerun in 1.33s ================== 2025-12-04T12:38:28.6070910Z Got exit code 1 2025-12-04T12:38:28.6071896Z FAILED CONSISTENTLY: test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda 2025-12-04T12:38:28.6072311Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:38:28.6072785Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-0ebf80986aa75a9d.xml 2025-12-04T12:38:28.6072952Z ============================= test session starts ============================== 2025-12-04T12:38:28.6073321Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.6073432Z cachedir: .pytest_cache 2025-12-04T12:38:28.6073957Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.6074097Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.6074206Z configfile: pytest.ini 2025-12-04T12:38:28.6074806Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.6075053Z collecting ... collected 1584 items / 278 deselected / 1306 selected 2025-12-04T12:38:28.6075200Z stepcurrent: skipping 278 already run items. 2025-12-04T12:38:28.6075333Z Running 1306 items in this shard 2025-12-04T12:38:28.6075338Z 2025-12-04T12:38:28.6076247Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8651s] [ 0%] 2025-12-04T12:38:28.6077181Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1674s] [ 0%] 2025-12-04T12:38:28.6078009Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda FAILED [0.1568s] [ 0%] 2025-12-04T12:38:28.6078016Z 2025-12-04T12:38:28.6078157Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.6078817Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.6078938Z Traceback (most recent call last): 2025-12-04T12:38:28.6079384Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6079532Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6079843Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6080031Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6080369Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6080517Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6080663Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6080668Z 2025-12-04T12:38:28.6080884Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6081884Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.6081890Z 2025-12-04T12:38:28.6082156Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6082806Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.6082938Z Traceback (most recent call last): 2025-12-04T12:38:28.6083412Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6083556Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6083891Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6084066Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6084416Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6084566Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6084697Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6084702Z 2025-12-04T12:38:28.6084931Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6085872Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.6085880Z 2025-12-04T12:38:28.6086158Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6086307Z =================================== FAILURES =================================== 2025-12-04T12:38:28.6086955Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.6087092Z Traceback (most recent call last): 2025-12-04T12:38:28.6087535Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6087681Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6087987Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6088191Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6088539Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6088694Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6088829Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6088848Z 2025-12-04T12:38:28.6089063Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6090000Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.6090007Z 2025-12-04T12:38:28.6090285Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6090916Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-0ebf80986aa75a9d.xml - 2025-12-04T12:38:28.6091102Z =========================== short test summary info ============================ 2025-12-04T12:38:28.6092084Z FAILED [0.1568s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6092092Z 2025-12-04T12:38:28.6092306Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6093252Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.6093287Z 2025-12-04T12:38:28.6093554Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6093746Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.6093950Z ================== 1 failed, 278 deselected, 2 rerun in 1.31s ================== 2025-12-04T12:38:28.6094050Z Got exit code 1 2025-12-04T12:38:28.6094201Z Retrying single test... 2025-12-04T12:38:28.6094653Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-ed818918883fb5a6.xml 2025-12-04T12:38:28.6094858Z ============================= test session starts ============================== 2025-12-04T12:38:28.6095209Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.6095321Z cachedir: .pytest_cache 2025-12-04T12:38:28.6095856Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.6096217Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.6096328Z configfile: pytest.ini 2025-12-04T12:38:28.6097004Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.6097242Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.6098299Z stepcurrent: skipping 278 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.6098417Z Running 1 items in this shard 2025-12-04T12:38:28.6098422Z 2025-12-04T12:38:28.6099336Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8834s] [100%] 2025-12-04T12:38:28.6100261Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1720s] [100%] 2025-12-04T12:38:28.6101154Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda FAILED [0.1665s] [100%] 2025-12-04T12:38:28.6101163Z 2025-12-04T12:38:28.6101320Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.6101968Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.6102109Z Traceback (most recent call last): 2025-12-04T12:38:28.6102554Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6102692Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6103020Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6103200Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6103539Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6103704Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6103840Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6103845Z 2025-12-04T12:38:28.6104075Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6105018Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.6105071Z 2025-12-04T12:38:28.6105346Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6106007Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.6106127Z Traceback (most recent call last): 2025-12-04T12:38:28.6106581Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6106757Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6107062Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6107255Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6107630Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6107795Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6107929Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6107934Z 2025-12-04T12:38:28.6108152Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6109110Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.6109115Z 2025-12-04T12:38:28.6109381Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6109542Z =================================== FAILURES =================================== 2025-12-04T12:38:28.6110187Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.6110310Z Traceback (most recent call last): 2025-12-04T12:38:28.6110762Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6110898Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6111205Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6111394Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6111728Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6111923Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6112054Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6112062Z 2025-12-04T12:38:28.6112277Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6113228Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.6113235Z 2025-12-04T12:38:28.6113500Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6114139Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-ed818918883fb5a6.xml - 2025-12-04T12:38:28.6114311Z =========================== short test summary info ============================ 2025-12-04T12:38:28.6115298Z FAILED [0.1665s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6115318Z 2025-12-04T12:38:28.6115530Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6116470Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.6116506Z 2025-12-04T12:38:28.6116784Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6116964Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.6117181Z ================= 1 failed, 1583 deselected, 2 rerun in 1.34s ================== 2025-12-04T12:38:28.6117282Z Got exit code 1 2025-12-04T12:38:28.6117391Z Retrying single test... 2025-12-04T12:38:28.6117863Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-688ec0aa50c9a8c6.xml 2025-12-04T12:38:28.6118057Z ============================= test session starts ============================== 2025-12-04T12:38:28.6118410Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.6118563Z cachedir: .pytest_cache 2025-12-04T12:38:28.6119093Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.6119231Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.6119339Z configfile: pytest.ini 2025-12-04T12:38:28.6119929Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.6120168Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.6121200Z stepcurrent: skipping 278 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.6121320Z Running 1 items in this shard 2025-12-04T12:38:28.6121339Z 2025-12-04T12:38:28.6122251Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.8657s] [100%] 2025-12-04T12:38:28.6123163Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda ('RERUN', {'yellow': True}) [0.1717s] [100%] 2025-12-04T12:38:28.6123988Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda FAILED [0.1683s] [100%] 2025-12-04T12:38:28.6123993Z 2025-12-04T12:38:28.6124184Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.6124850Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.6124975Z Traceback (most recent call last): 2025-12-04T12:38:28.6125421Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6125570Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6125881Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6126070Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6126403Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6126557Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6126702Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6126709Z 2025-12-04T12:38:28.6126924Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6127883Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.6127889Z 2025-12-04T12:38:28.6128157Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6128840Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.6128975Z Traceback (most recent call last): 2025-12-04T12:38:28.6129414Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6129561Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6129868Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6130079Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6130426Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6130608Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6130741Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6130747Z 2025-12-04T12:38:28.6130976Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6131917Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.6131922Z 2025-12-04T12:38:28.6132203Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6132350Z =================================== FAILURES =================================== 2025-12-04T12:38:28.6132993Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda _ 2025-12-04T12:38:28.6133130Z Traceback (most recent call last): 2025-12-04T12:38:28.6133573Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6133720Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6134025Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6134205Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6134552Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6134700Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6134836Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6134842Z 2025-12-04T12:38:28.6135099Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6136035Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.6136043Z 2025-12-04T12:38:28.6136324Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6137049Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-688ec0aa50c9a8c6.xml - 2025-12-04T12:38:28.6137242Z =========================== short test summary info ============================ 2025-12-04T12:38:28.6138227Z FAILED [0.1683s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6138232Z 2025-12-04T12:38:28.6138450Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6139404Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.6139411Z 2025-12-04T12:38:28.6139676Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6139870Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.6140116Z ================= 1 failed, 1583 deselected, 2 rerun in 1.33s ================== 2025-12-04T12:38:28.6140217Z Got exit code 1 2025-12-04T12:38:28.6141097Z FAILED CONSISTENTLY: test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda 2025-12-04T12:38:28.6141510Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:38:28.6141975Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-6fda2bf2ae878b46.xml 2025-12-04T12:38:28.6142172Z ============================= test session starts ============================== 2025-12-04T12:38:28.6142554Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.6142682Z cachedir: .pytest_cache 2025-12-04T12:38:28.6143206Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.6143328Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.6143449Z configfile: pytest.ini 2025-12-04T12:38:28.6144044Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.6144292Z collecting ... collected 1584 items / 279 deselected / 1305 selected 2025-12-04T12:38:28.6144440Z stepcurrent: skipping 279 already run items. 2025-12-04T12:38:28.6144556Z Running 1305 items in this shard 2025-12-04T12:38:28.6144562Z 2025-12-04T12:38:28.6146402Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda [W1204 12:35:34.770506577 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.6146536Z ('RERUN', {'yellow': True}) [0.8709s] [ 0%] 2025-12-04T12:38:28.6147464Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1731s] [ 0%] 2025-12-04T12:38:28.6148320Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda FAILED [0.1696s] [ 0%] 2025-12-04T12:38:28.6148328Z 2025-12-04T12:38:28.6148482Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.6149139Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.6149263Z Traceback (most recent call last): 2025-12-04T12:38:28.6149718Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6149852Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6150174Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6150349Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6150685Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6150848Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6150982Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6150987Z 2025-12-04T12:38:28.6151204Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6152167Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.6152207Z 2025-12-04T12:38:28.6152475Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6153150Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.6153271Z Traceback (most recent call last): 2025-12-04T12:38:28.6153712Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6153888Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6154193Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6154417Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6154752Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6154906Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6155050Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6155055Z 2025-12-04T12:38:28.6155269Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6156230Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.6156236Z 2025-12-04T12:38:28.6156504Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6156653Z =================================== FAILURES =================================== 2025-12-04T12:38:28.6157327Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.6157448Z Traceback (most recent call last): 2025-12-04T12:38:28.6157904Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6158039Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6158347Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6158532Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6158863Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6159042Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6159191Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6159196Z 2025-12-04T12:38:28.6159409Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6160361Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.6160369Z 2025-12-04T12:38:28.6160634Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6161263Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-6fda2bf2ae878b46.xml - 2025-12-04T12:38:28.6161447Z =========================== short test summary info ============================ 2025-12-04T12:38:28.6162447Z FAILED [0.1696s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6162454Z 2025-12-04T12:38:28.6162681Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6163630Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.6163667Z 2025-12-04T12:38:28.6163946Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6164127Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.6164330Z ================== 1 failed, 279 deselected, 2 rerun in 1.34s ================== 2025-12-04T12:38:28.6164443Z Got exit code 1 2025-12-04T12:38:28.6164549Z Retrying single test... 2025-12-04T12:38:28.6165006Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-0857a7a11662fc06.xml 2025-12-04T12:38:28.6165211Z ============================= test session starts ============================== 2025-12-04T12:38:28.6165615Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.6165743Z cachedir: .pytest_cache 2025-12-04T12:38:28.6166264Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.6166390Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.6166512Z configfile: pytest.ini 2025-12-04T12:38:28.6167173Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.6167403Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.6168452Z stepcurrent: skipping 279 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.6168569Z Running 1 items in this shard 2025-12-04T12:38:28.6168575Z 2025-12-04T12:38:28.6170405Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda [W1204 12:35:47.064317387 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.6170541Z ('RERUN', {'yellow': True}) [0.8589s] [100%] 2025-12-04T12:38:28.6171506Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1748s] [100%] 2025-12-04T12:38:28.6172328Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda FAILED [0.1590s] [100%] 2025-12-04T12:38:28.6172336Z 2025-12-04T12:38:28.6172490Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.6173146Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.6173270Z Traceback (most recent call last): 2025-12-04T12:38:28.6173728Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6173863Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6174171Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6174361Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6174699Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6174863Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6174994Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6175002Z 2025-12-04T12:38:28.6175215Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6176176Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.6176213Z 2025-12-04T12:38:28.6176481Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6177224Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.6177351Z Traceback (most recent call last): 2025-12-04T12:38:28.6177830Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6177980Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6178319Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6178508Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6178844Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6178996Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6179146Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6179151Z 2025-12-04T12:38:28.6179367Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6180318Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.6180338Z 2025-12-04T12:38:28.6180602Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6180746Z =================================== FAILURES =================================== 2025-12-04T12:38:28.6181413Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.6181536Z Traceback (most recent call last): 2025-12-04T12:38:28.6181973Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6182118Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6182423Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6182608Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6182974Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6183129Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6183272Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6183278Z 2025-12-04T12:38:28.6183492Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6184461Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.6184469Z 2025-12-04T12:38:28.6184737Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6185364Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-0857a7a11662fc06.xml - 2025-12-04T12:38:28.6185548Z =========================== short test summary info ============================ 2025-12-04T12:38:28.6186549Z FAILED [0.1590s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6186557Z 2025-12-04T12:38:28.6186785Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6187729Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.6187768Z 2025-12-04T12:38:28.6188050Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6188231Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.6188435Z ================= 1 failed, 1583 deselected, 2 rerun in 1.31s ================== 2025-12-04T12:38:28.6188546Z Got exit code 1 2025-12-04T12:38:28.6188653Z Retrying single test... 2025-12-04T12:38:28.6189142Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a3e03d3f7a7bbb81.xml 2025-12-04T12:38:28.6189318Z ============================= test session starts ============================== 2025-12-04T12:38:28.6189695Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.6189816Z cachedir: .pytest_cache 2025-12-04T12:38:28.6190344Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.6190465Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.6190586Z configfile: pytest.ini 2025-12-04T12:38:28.6191178Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.6191408Z collecting ... collected 1584 items / 1583 deselected / 1 selected 2025-12-04T12:38:28.6192453Z stepcurrent: skipping 279 already run items. Running only test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.6192572Z Running 1 items in this shard 2025-12-04T12:38:28.6192578Z 2025-12-04T12:38:28.6194394Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda [W1204 12:36:00.331622111 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.6194529Z ('RERUN', {'yellow': True}) [0.8718s] [100%] 2025-12-04T12:38:28.6195497Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda ('RERUN', {'yellow': True}) [0.1732s] [100%] 2025-12-04T12:38:28.6196515Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda FAILED [0.1770s] [100%] 2025-12-04T12:38:28.6196522Z 2025-12-04T12:38:28.6196676Z ==================================== RERUNS ==================================== 2025-12-04T12:38:28.6197331Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.6197452Z Traceback (most recent call last): 2025-12-04T12:38:28.6197911Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6198047Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6198358Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6198552Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6198889Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6199055Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6199187Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6199193Z 2025-12-04T12:38:28.6199414Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6200440Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.6200445Z 2025-12-04T12:38:28.6200731Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6201390Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.6201570Z Traceback (most recent call last): 2025-12-04T12:38:28.6202010Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6202185Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6202504Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6202684Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6203019Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6203183Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6203315Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6203320Z 2025-12-04T12:38:28.6203546Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6204500Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.6204507Z 2025-12-04T12:38:28.6204773Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6204934Z =================================== FAILURES =================================== 2025-12-04T12:38:28.6205593Z _ TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda _ 2025-12-04T12:38:28.6205731Z Traceback (most recent call last): 2025-12-04T12:38:28.6206172Z File "/var/lib/jenkins/workspace/test/test_matmul_cuda.py", line 805, in test_addmm_baddmm_dtype_overload 2025-12-04T12:38:28.6206305Z with self.assertRaises(RuntimeError): 2025-12-04T12:38:28.6206625Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 226, in __exit__ 2025-12-04T12:38:28.6206841Z self._raiseFailure("{} not raised".format(exc_name)) 2025-12-04T12:38:28.6207178Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 163, in _raiseFailure 2025-12-04T12:38:28.6207345Z raise self.test_case.failureException(msg) 2025-12-04T12:38:28.6207480Z AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6207486Z 2025-12-04T12:38:28.6207715Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6208661Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.6208666Z 2025-12-04T12:38:28.6208930Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6209575Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a3e03d3f7a7bbb81.xml - 2025-12-04T12:38:28.6209751Z =========================== short test summary info ============================ 2025-12-04T12:38:28.6210766Z FAILED [0.1770s] test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda - AssertionError: RuntimeError not raised 2025-12-04T12:38:28.6210772Z 2025-12-04T12:38:28.6210985Z To execute this test, run the following from the base repo dir: 2025-12-04T12:38:28.6211973Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/test_matmul_cuda.py TestMatmulCudaCUDA.test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.6211978Z 2025-12-04T12:38:28.6212242Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T12:38:28.6212419Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T12:38:28.6212637Z ================= 1 failed, 1583 deselected, 2 rerun in 1.34s ================== 2025-12-04T12:38:28.6212779Z Got exit code 1 2025-12-04T12:38:28.6213685Z FAILED CONSISTENTLY: test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda 2025-12-04T12:38:28.6214096Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T12:38:28.6214550Z Test results will be stored in test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-829800aaf6d2511a.xml 2025-12-04T12:38:28.6214726Z ============================= test session starts ============================== 2025-12-04T12:38:28.6215076Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T12:38:28.6215186Z cachedir: .pytest_cache 2025-12-04T12:38:28.6215731Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T12:38:28.6215856Z rootdir: /var/lib/jenkins/workspace 2025-12-04T12:38:28.6215975Z configfile: pytest.ini 2025-12-04T12:38:28.6216570Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T12:38:28.6216802Z collecting ... collected 1584 items / 280 deselected / 1304 selected 2025-12-04T12:38:28.6217021Z stepcurrent: skipping 280 already run items. 2025-12-04T12:38:28.6217136Z Running 1304 items in this shard 2025-12-04T12:38:28.6217141Z 2025-12-04T12:38:28.6217984Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0093s] [ 0%] 2025-12-04T12:38:28.6219843Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda [W1204 12:36:13.680486261 Context.cpp:525] Warning: torch.backends.cuda.preferred_blas_library is an experimental feature. If you see any error or unexpected behavior when this flag is set please file an issue on GitHub. (function operator()) 2025-12-04T12:38:28.6219963Z PASSED [0.0042s] [ 0%] 2025-12-04T12:38:28.6220788Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0038s] [ 0%] 2025-12-04T12:38:28.6221623Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0038s] [ 0%] 2025-12-04T12:38:28.6222450Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0035s] [ 0%] 2025-12-04T12:38:28.6223275Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0036s] [ 0%] 2025-12-04T12:38:28.6224101Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0037s] [ 0%] 2025-12-04T12:38:28.6224923Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0035s] [ 0%] 2025-12-04T12:38:28.6225779Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.9664s] [ 0%] 2025-12-04T12:38:28.6226599Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0061s] [ 0%] 2025-12-04T12:38:28.6227474Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0053s] [ 0%] 2025-12-04T12:38:28.6228296Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0050s] [ 0%] 2025-12-04T12:38:28.6229101Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0049s] [ 0%] 2025-12-04T12:38:28.6229930Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0051s] [ 1%] 2025-12-04T12:38:28.6230727Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0048s] [ 1%] 2025-12-04T12:38:28.6231552Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0049s] [ 1%] 2025-12-04T12:38:28.6232361Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0049s] [ 1%] 2025-12-04T12:38:28.6233183Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0048s] [ 1%] 2025-12-04T12:38:28.6234030Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0051s] [ 1%] 2025-12-04T12:38:28.6234851Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0048s] [ 1%] 2025-12-04T12:38:28.6235671Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0048s] [ 1%] 2025-12-04T12:38:28.6236484Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0048s] [ 1%] 2025-12-04T12:38:28.6237292Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0048s] [ 1%] 2025-12-04T12:38:28.6238103Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0057s] [ 1%] 2025-12-04T12:38:28.6238932Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0049s] [ 1%] 2025-12-04T12:38:28.6239750Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0049s] [ 1%] 2025-12-04T12:38:28.6240591Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0049s] [ 2%] 2025-12-04T12:38:28.6241420Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0049s] [ 2%] 2025-12-04T12:38:28.6242254Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0048s] [ 2%] 2025-12-04T12:38:28.6243108Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0052s] [ 2%] 2025-12-04T12:38:28.6243913Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0048s] [ 2%] 2025-12-04T12:38:28.6244732Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0049s] [ 2%] 2025-12-04T12:38:28.6245544Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0048s] [ 2%] 2025-12-04T12:38:28.6246366Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0048s] [ 2%] 2025-12-04T12:38:28.6247179Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0048s] [ 2%] 2025-12-04T12:38:28.6247993Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0051s] [ 2%] 2025-12-04T12:38:28.6248847Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0047s] [ 2%] 2025-12-04T12:38:28.6249667Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0047s] [ 2%] 2025-12-04T12:38:28.6250477Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0047s] [ 2%] 2025-12-04T12:38:28.6251287Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0048s] [ 3%] 2025-12-04T12:38:28.6252107Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0049s] [ 3%] 2025-12-04T12:38:28.6252928Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0049s] [ 3%] 2025-12-04T12:38:28.6253738Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 3%] 2025-12-04T12:38:28.6254566Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 3%] 2025-12-04T12:38:28.6255417Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0046s] [ 3%] 2025-12-04T12:38:28.6256244Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 3%] 2025-12-04T12:38:28.6257137Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0049s] [ 3%] 2025-12-04T12:38:28.6257997Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 3%] 2025-12-04T12:38:28.6258809Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 3%] 2025-12-04T12:38:28.6259646Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 3%] 2025-12-04T12:38:28.6260447Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 3%] 2025-12-04T12:38:28.6261268Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 3%] 2025-12-04T12:38:28.6262084Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0049s] [ 4%] 2025-12-04T12:38:28.6262898Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 4%] 2025-12-04T12:38:28.6263740Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 4%] 2025-12-04T12:38:28.6264551Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0045s] [ 4%] 2025-12-04T12:38:28.6265372Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0047s] [ 4%] 2025-12-04T12:38:28.6266195Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 4%] 2025-12-04T12:38:28.6266996Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0049s] [ 4%] 2025-12-04T12:38:28.6267820Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 4%] 2025-12-04T12:38:28.6268627Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 4%] 2025-12-04T12:38:28.6269451Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 4%] 2025-12-04T12:38:28.6270303Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 4%] 2025-12-04T12:38:28.6271240Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 4%] 2025-12-04T12:38:28.6272172Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0049s] [ 4%] 2025-12-04T12:38:28.6273039Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 5%] 2025-12-04T12:38:28.6273846Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 5%] 2025-12-04T12:38:28.6274664Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0045s] [ 5%] 2025-12-04T12:38:28.6275483Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 5%] 2025-12-04T12:38:28.6276298Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 5%] 2025-12-04T12:38:28.6277114Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0049s] [ 5%] 2025-12-04T12:38:28.6277928Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0045s] [ 5%] 2025-12-04T12:38:28.6278781Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0046s] [ 5%] 2025-12-04T12:38:28.6279610Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 5%] 2025-12-04T12:38:28.6280416Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 5%] 2025-12-04T12:38:28.6281243Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0050s] [ 5%] 2025-12-04T12:38:28.6282051Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0051s] [ 5%] 2025-12-04T12:38:28.6282879Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 5%] 2025-12-04T12:38:28.6283687Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 6%] 2025-12-04T12:38:28.6284513Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 6%] 2025-12-04T12:38:28.6285369Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0046s] [ 6%] 2025-12-04T12:38:28.6286206Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 6%] 2025-12-04T12:38:28.6287084Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0049s] [ 6%] 2025-12-04T12:38:28.6287899Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 6%] 2025-12-04T12:38:28.6288723Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 6%] 2025-12-04T12:38:28.6289536Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 6%] 2025-12-04T12:38:28.6290353Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 6%] 2025-12-04T12:38:28.6291177Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 6%] 2025-12-04T12:38:28.6292003Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0049s] [ 6%] 2025-12-04T12:38:28.6292829Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 6%] 2025-12-04T12:38:28.6293677Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 6%] 2025-12-04T12:38:28.6294498Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 7%] 2025-12-04T12:38:28.6295301Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 7%] 2025-12-04T12:38:28.6296272Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 7%] 2025-12-04T12:38:28.6297138Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0052s] [ 7%] 2025-12-04T12:38:28.6297965Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 7%] 2025-12-04T12:38:28.6298780Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 7%] 2025-12-04T12:38:28.6299683Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 7%] 2025-12-04T12:38:28.6300493Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 7%] 2025-12-04T12:38:28.6301313Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0045s] [ 7%] 2025-12-04T12:38:28.6302222Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0048s] [ 7%] 2025-12-04T12:38:28.6303043Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 7%] 2025-12-04T12:38:28.6303866Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 7%] 2025-12-04T12:38:28.6304680Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 7%] 2025-12-04T12:38:28.6305504Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 8%] 2025-12-04T12:38:28.6306334Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 8%] 2025-12-04T12:38:28.6307151Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0049s] [ 8%] 2025-12-04T12:38:28.6307968Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0045s] [ 8%] 2025-12-04T12:38:28.6308807Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 8%] 2025-12-04T12:38:28.6309639Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0048s] [ 8%] 2025-12-04T12:38:28.6310434Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 8%] 2025-12-04T12:38:28.6311251Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0045s] [ 8%] 2025-12-04T12:38:28.6312061Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0049s] [ 8%] 2025-12-04T12:38:28.6312890Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 8%] 2025-12-04T12:38:28.6313696Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 8%] 2025-12-04T12:38:28.6314541Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 8%] 2025-12-04T12:38:28.6315362Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 8%] 2025-12-04T12:38:28.6316176Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 9%] 2025-12-04T12:38:28.6317049Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0049s] [ 9%] 2025-12-04T12:38:28.6317859Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0045s] [ 9%] 2025-12-04T12:38:28.6318684Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0047s] [ 9%] 2025-12-04T12:38:28.6319507Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 9%] 2025-12-04T12:38:28.6320328Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 9%] 2025-12-04T12:38:28.6321151Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 9%] 2025-12-04T12:38:28.6321958Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0049s] [ 9%] 2025-12-04T12:38:28.6322789Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 9%] 2025-12-04T12:38:28.6323636Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 9%] 2025-12-04T12:38:28.6324459Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 9%] 2025-12-04T12:38:28.6325276Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0046s] [ 9%] 2025-12-04T12:38:28.6326113Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 9%] 2025-12-04T12:38:28.6326925Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0049s] [ 10%] 2025-12-04T12:38:28.6327758Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0045s] [ 10%] 2025-12-04T12:38:28.6328571Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 10%] 2025-12-04T12:38:28.6329423Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 10%] 2025-12-04T12:38:28.6330242Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 10%] 2025-12-04T12:38:28.6331062Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 10%] 2025-12-04T12:38:28.6331951Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0050s] [ 10%] 2025-12-04T12:38:28.6332782Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 10%] 2025-12-04T12:38:28.6333605Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 10%] 2025-12-04T12:38:28.6334426Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 10%] 2025-12-04T12:38:28.6335233Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 10%] 2025-12-04T12:38:28.6336069Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 10%] 2025-12-04T12:38:28.6336947Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0050s] [ 10%] 2025-12-04T12:38:28.6337776Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 11%] 2025-12-04T12:38:28.6338623Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0047s] [ 11%] 2025-12-04T12:38:28.6339464Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 11%] 2025-12-04T12:38:28.6340273Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 11%] 2025-12-04T12:38:28.6341108Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 11%] 2025-12-04T12:38:28.6341918Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0049s] [ 11%] 2025-12-04T12:38:28.6342734Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 11%] 2025-12-04T12:38:28.6343553Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 11%] 2025-12-04T12:38:28.6344405Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 11%] 2025-12-04T12:38:28.6345227Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0046s] [ 11%] 2025-12-04T12:38:28.6346051Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0050s] [ 11%] 2025-12-04T12:38:28.6346941Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0050s] [ 11%] 2025-12-04T12:38:28.6347762Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 11%] 2025-12-04T12:38:28.6348574Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 12%] 2025-12-04T12:38:28.6349412Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 12%] 2025-12-04T12:38:28.6350215Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 12%] 2025-12-04T12:38:28.6351039Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 12%] 2025-12-04T12:38:28.6351857Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0050s] [ 12%] 2025-12-04T12:38:28.6352695Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 12%] 2025-12-04T12:38:28.6353532Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 12%] 2025-12-04T12:38:28.6354372Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 12%] 2025-12-04T12:38:28.6355177Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0046s] [ 12%] 2025-12-04T12:38:28.6355992Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 12%] 2025-12-04T12:38:28.6356811Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0052s] [ 12%] 2025-12-04T12:38:28.6357624Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 12%] 2025-12-04T12:38:28.6358450Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0047s] [ 12%] 2025-12-04T12:38:28.6359310Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 13%] 2025-12-04T12:38:28.6360131Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 13%] 2025-12-04T12:38:28.6360952Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 13%] 2025-12-04T12:38:28.6361833Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0049s] [ 13%] 2025-12-04T12:38:28.6362654Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 13%] 2025-12-04T12:38:28.6363458Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 13%] 2025-12-04T12:38:28.6364288Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0045s] [ 13%] 2025-12-04T12:38:28.6365111Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0046s] [ 13%] 2025-12-04T12:38:28.6365944Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 13%] 2025-12-04T12:38:28.6366753Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0050s] [ 13%] 2025-12-04T12:38:28.6367586Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 13%] 2025-12-04T12:38:28.6368427Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 13%] 2025-12-04T12:38:28.6369249Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 13%] 2025-12-04T12:38:28.6370064Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 14%] 2025-12-04T12:38:28.6370876Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 14%] 2025-12-04T12:38:28.6371706Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0050s] [ 14%] 2025-12-04T12:38:28.6372533Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 14%] 2025-12-04T12:38:28.6373361Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 14%] 2025-12-04T12:38:28.6374212Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0049s] [ 14%] 2025-12-04T12:38:28.6375036Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 14%] 2025-12-04T12:38:28.6375854Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 14%] 2025-12-04T12:38:28.6376734Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0049s] [ 14%] 2025-12-04T12:38:28.6377633Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0045s] [ 14%] 2025-12-04T12:38:28.6378456Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 14%] 2025-12-04T12:38:28.6379296Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 14%] 2025-12-04T12:38:28.6380115Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 14%] 2025-12-04T12:38:28.6380944Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 15%] 2025-12-04T12:38:28.6381760Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0050s] [ 15%] 2025-12-04T12:38:28.6382620Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 15%] 2025-12-04T12:38:28.6383438Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 15%] 2025-12-04T12:38:28.6384255Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 15%] 2025-12-04T12:38:28.6385080Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0046s] [ 15%] 2025-12-04T12:38:28.6385899Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 15%] 2025-12-04T12:38:28.6386724Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0049s] [ 15%] 2025-12-04T12:38:28.6387549Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 15%] 2025-12-04T12:38:28.6388369Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 15%] 2025-12-04T12:38:28.6389214Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 15%] 2025-12-04T12:38:28.6390012Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 15%] 2025-12-04T12:38:28.6390835Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 15%] 2025-12-04T12:38:28.6391709Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0049s] [ 16%] 2025-12-04T12:38:28.6392541Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 16%] 2025-12-04T12:38:28.6393345Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 16%] 2025-12-04T12:38:28.6394174Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 16%] 2025-12-04T12:38:28.6394979Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 16%] 2025-12-04T12:38:28.6395806Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 16%] 2025-12-04T12:38:28.6396783Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0049s] [ 16%] 2025-12-04T12:38:28.6397655Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0045s] [ 16%] 2025-12-04T12:38:28.6398488Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0046s] [ 16%] 2025-12-04T12:38:28.6399314Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 16%] 2025-12-04T12:38:28.6400141Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 16%] 2025-12-04T12:38:28.6400956Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 16%] 2025-12-04T12:38:28.6401780Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0049s] [ 16%] 2025-12-04T12:38:28.6402606Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0048s] [ 17%] 2025-12-04T12:38:28.6403407Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 17%] 2025-12-04T12:38:28.6404269Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0045s] [ 17%] 2025-12-04T12:38:28.6405088Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0046s] [ 17%] 2025-12-04T12:38:28.6405923Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 17%] 2025-12-04T12:38:28.6406818Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0049s] [ 17%] 2025-12-04T12:38:28.6407657Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 17%] 2025-12-04T12:38:28.6408464Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 17%] 2025-12-04T12:38:28.6409289Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 17%] 2025-12-04T12:38:28.6410093Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 17%] 2025-12-04T12:38:28.6410904Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 17%] 2025-12-04T12:38:28.6411731Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0050s] [ 17%] 2025-12-04T12:38:28.6412591Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 17%] 2025-12-04T12:38:28.6413414Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 18%] 2025-12-04T12:38:28.6414239Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 18%] 2025-12-04T12:38:28.6415065Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 18%] 2025-12-04T12:38:28.6415881Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 18%] 2025-12-04T12:38:28.6416688Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0051s] [ 18%] 2025-12-04T12:38:28.6417584Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 18%] 2025-12-04T12:38:28.6418403Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 18%] 2025-12-04T12:38:28.6419272Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 18%] 2025-12-04T12:38:28.6420081Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 18%] 2025-12-04T12:38:28.6420908Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 18%] 2025-12-04T12:38:28.6421775Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0049s] [ 18%] 2025-12-04T12:38:28.6422613Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 18%] 2025-12-04T12:38:28.6423410Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 18%] 2025-12-04T12:38:28.6424222Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 19%] 2025-12-04T12:38:28.6425051Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0046s] [ 19%] 2025-12-04T12:38:28.6425874Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 19%] 2025-12-04T12:38:28.6426700Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0049s] [ 19%] 2025-12-04T12:38:28.6427548Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 19%] 2025-12-04T12:38:28.6428371Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 19%] 2025-12-04T12:38:28.6429193Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 19%] 2025-12-04T12:38:28.6430009Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 19%] 2025-12-04T12:38:28.6430824Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0045s] [ 19%] 2025-12-04T12:38:28.6431640Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0049s] [ 19%] 2025-12-04T12:38:28.6432480Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 19%] 2025-12-04T12:38:28.6433290Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 19%] 2025-12-04T12:38:28.6434149Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0045s] [ 19%] 2025-12-04T12:38:28.6434960Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 20%] 2025-12-04T12:38:28.6435839Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 20%] 2025-12-04T12:38:28.6436674Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0049s] [ 20%] 2025-12-04T12:38:28.6437487Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 20%] 2025-12-04T12:38:28.6438309Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0048s] [ 20%] 2025-12-04T12:38:28.6439137Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0047s] [ 20%] 2025-12-04T12:38:28.6439961Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 20%] 2025-12-04T12:38:28.6440778Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 20%] 2025-12-04T12:38:28.6441601Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0050s] [ 20%] 2025-12-04T12:38:28.6442451Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 20%] 2025-12-04T12:38:28.6443272Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 20%] 2025-12-04T12:38:28.6444095Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 20%] 2025-12-04T12:38:28.6444907Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 20%] 2025-12-04T12:38:28.6445743Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 21%] 2025-12-04T12:38:28.6446555Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0050s] [ 21%] 2025-12-04T12:38:28.6447386Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 21%] 2025-12-04T12:38:28.6448196Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0046s] [ 21%] 2025-12-04T12:38:28.6449060Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 21%] 2025-12-04T12:38:28.6449864Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 21%] 2025-12-04T12:38:28.6450712Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0050s] [ 21%] 2025-12-04T12:38:28.6451571Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0051s] [ 21%] 2025-12-04T12:38:28.6452404Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0047s] [ 21%] 2025-12-04T12:38:28.6453221Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 21%] 2025-12-04T12:38:28.6454042Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 21%] 2025-12-04T12:38:28.6454870Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0046s] [ 21%] 2025-12-04T12:38:28.6455692Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 21%] 2025-12-04T12:38:28.6456508Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0050s] [ 22%] 2025-12-04T12:38:28.6457410Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0045s] [ 22%] 2025-12-04T12:38:28.6458230Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0046s] [ 22%] 2025-12-04T12:38:28.6459065Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0045s] [ 22%] 2025-12-04T12:38:28.6459877Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 22%] 2025-12-04T12:38:28.6460711Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 22%] 2025-12-04T12:38:28.6461524Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0050s] [ 22%] 2025-12-04T12:38:28.6462354Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0046s] [ 22%] 2025-12-04T12:38:28.6463158Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 22%] 2025-12-04T12:38:28.6464017Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0046s] [ 22%] 2025-12-04T12:38:28.6464822Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0040s] [ 22%] 2025-12-04T12:38:28.6465722Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 22%] 2025-12-04T12:38:28.6466536Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0044s] [ 22%] 2025-12-04T12:38:28.6467349Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 23%] 2025-12-04T12:38:28.6468159Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 23%] 2025-12-04T12:38:28.6468971Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 23%] 2025-12-04T12:38:28.6469777Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 23%] 2025-12-04T12:38:28.6470577Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 23%] 2025-12-04T12:38:28.6471384Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0043s] [ 23%] 2025-12-04T12:38:28.6472246Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 23%] 2025-12-04T12:38:28.6473068Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 23%] 2025-12-04T12:38:28.6473897Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 23%] 2025-12-04T12:38:28.6474700Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0040s] [ 23%] 2025-12-04T12:38:28.6475529Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 23%] 2025-12-04T12:38:28.6476332Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 23%] 2025-12-04T12:38:28.6477161Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 23%] 2025-12-04T12:38:28.6477974Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 24%] 2025-12-04T12:38:28.6478830Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0043s] [ 24%] 2025-12-04T12:38:28.6479651Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0040s] [ 24%] 2025-12-04T12:38:28.6480500Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 24%] 2025-12-04T12:38:28.6481349Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0044s] [ 24%] 2025-12-04T12:38:28.6482162Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 24%] 2025-12-04T12:38:28.6482978Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 24%] 2025-12-04T12:38:28.6483791Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 24%] 2025-12-04T12:38:28.6484606Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 24%] 2025-12-04T12:38:28.6485449Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 24%] 2025-12-04T12:38:28.6486257Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0043s] [ 24%] 2025-12-04T12:38:28.6487116Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 24%] 2025-12-04T12:38:28.6487928Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 24%] 2025-12-04T12:38:28.6488760Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 25%] 2025-12-04T12:38:28.6489560Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 25%] 2025-12-04T12:38:28.6490383Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 25%] 2025-12-04T12:38:28.6491197Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0043s] [ 25%] 2025-12-04T12:38:28.6492025Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 25%] 2025-12-04T12:38:28.6492846Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 25%] 2025-12-04T12:38:28.6493726Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 25%] 2025-12-04T12:38:28.6494550Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 25%] 2025-12-04T12:38:28.6495397Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 25%] 2025-12-04T12:38:28.6496422Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0048s] [ 25%] 2025-12-04T12:38:28.6497287Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0043s] [ 25%] 2025-12-04T12:38:28.6498102Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0042s] [ 25%] 2025-12-04T12:38:28.6498939Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 25%] 2025-12-04T12:38:28.6499748Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0041s] [ 25%] 2025-12-04T12:38:28.6500587Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0042s] [ 26%] 2025-12-04T12:38:28.6501393Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0049s] [ 26%] 2025-12-04T12:38:28.6502270Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0041s] [ 26%] 2025-12-04T12:38:28.6503074Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0040s] [ 26%] 2025-12-04T12:38:28.6503904Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 26%] 2025-12-04T12:38:28.6504715Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0040s] [ 26%] 2025-12-04T12:38:28.6505541Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 26%] 2025-12-04T12:38:28.6506363Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0044s] [ 26%] 2025-12-04T12:38:28.6507183Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0043s] [ 26%] 2025-12-04T12:38:28.6507999Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 26%] 2025-12-04T12:38:28.6508935Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 26%] 2025-12-04T12:38:28.6509750Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 26%] 2025-12-04T12:38:28.6510598Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 26%] 2025-12-04T12:38:28.6511452Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 27%] 2025-12-04T12:38:28.6512278Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 27%] 2025-12-04T12:38:28.6513083Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 27%] 2025-12-04T12:38:28.6513907Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 27%] 2025-12-04T12:38:28.6514715Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 27%] 2025-12-04T12:38:28.6515538Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 27%] 2025-12-04T12:38:28.6516338Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0043s] [ 27%] 2025-12-04T12:38:28.6517194Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 27%] 2025-12-04T12:38:28.6518008Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 27%] 2025-12-04T12:38:28.6518833Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 27%] 2025-12-04T12:38:28.6519649Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0040s] [ 27%] 2025-12-04T12:38:28.6520465Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 27%] 2025-12-04T12:38:28.6521287Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0044s] [ 27%] 2025-12-04T12:38:28.6522102Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 28%] 2025-12-04T12:38:28.6522922Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 28%] 2025-12-04T12:38:28.6523766Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 28%] 2025-12-04T12:38:28.6524593Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0040s] [ 28%] 2025-12-04T12:38:28.6525478Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 28%] 2025-12-04T12:38:28.6526286Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0044s] [ 28%] 2025-12-04T12:38:28.6527118Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 28%] 2025-12-04T12:38:28.6527919Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0040s] [ 28%] 2025-12-04T12:38:28.6528750Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 28%] 2025-12-04T12:38:28.6529557Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 28%] 2025-12-04T12:38:28.6530382Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 28%] 2025-12-04T12:38:28.6531194Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0044s] [ 28%] 2025-12-04T12:38:28.6532047Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 28%] 2025-12-04T12:38:28.6532875Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 29%] 2025-12-04T12:38:28.6533688Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 29%] 2025-12-04T12:38:28.6534510Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0040s] [ 29%] 2025-12-04T12:38:28.6535322Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0043s] [ 29%] 2025-12-04T12:38:28.6536133Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 29%] 2025-12-04T12:38:28.6537006Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 29%] 2025-12-04T12:38:28.6537835Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 29%] 2025-12-04T12:38:28.6538698Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 29%] 2025-12-04T12:38:28.6539508Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 29%] 2025-12-04T12:38:28.6540404Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 29%] 2025-12-04T12:38:28.6541214Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0043s] [ 29%] 2025-12-04T12:38:28.6542050Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 29%] 2025-12-04T12:38:28.6542853Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 29%] 2025-12-04T12:38:28.6543678Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 30%] 2025-12-04T12:38:28.6544495Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 30%] 2025-12-04T12:38:28.6545331Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 30%] 2025-12-04T12:38:28.6546135Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0043s] [ 30%] 2025-12-04T12:38:28.6546997Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 30%] 2025-12-04T12:38:28.6547822Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 30%] 2025-12-04T12:38:28.6548639Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 30%] 2025-12-04T12:38:28.6549454Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 30%] 2025-12-04T12:38:28.6550266Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 30%] 2025-12-04T12:38:28.6551091Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0044s] [ 30%] 2025-12-04T12:38:28.6551916Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 30%] 2025-12-04T12:38:28.6552756Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 30%] 2025-12-04T12:38:28.6553583Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 30%] 2025-12-04T12:38:28.6554389Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 31%] 2025-12-04T12:38:28.6555864Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 31%] 2025-12-04T12:38:28.6556672Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0043s] [ 31%] 2025-12-04T12:38:28.6557503Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 31%] 2025-12-04T12:38:28.6558312Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 31%] 2025-12-04T12:38:28.6559145Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 31%] 2025-12-04T12:38:28.6559954Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0042s] [ 31%] 2025-12-04T12:38:28.6560775Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 31%] 2025-12-04T12:38:28.6561599Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 31%] 2025-12-04T12:38:28.6562452Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 31%] 2025-12-04T12:38:28.6563270Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0040s] [ 31%] 2025-12-04T12:38:28.6564082Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0044s] [ 31%] 2025-12-04T12:38:28.6564905Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0040s] [ 31%] 2025-12-04T12:38:28.6565730Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0041s] [ 32%] 2025-12-04T12:38:28.6566535Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 32%] 2025-12-04T12:38:28.6567369Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 32%] 2025-12-04T12:38:28.6568208Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0040s] [ 32%] 2025-12-04T12:38:28.6569090Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 32%] 2025-12-04T12:38:28.6569932Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 32%] 2025-12-04T12:38:28.6570828Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 32%] 2025-12-04T12:38:28.6571643Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 32%] 2025-12-04T12:38:28.6572478Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0041s] [ 32%] 2025-12-04T12:38:28.6573285Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0040s] [ 32%] 2025-12-04T12:38:28.6574107Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0041s] [ 32%] 2025-12-04T12:38:28.6574934Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 32%] 2025-12-04T12:38:28.6575757Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 32%] 2025-12-04T12:38:28.6576574Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0044s] [ 33%] 2025-12-04T12:38:28.6577496Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 33%] 2025-12-04T12:38:28.6578336Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 33%] 2025-12-04T12:38:28.6579157Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 33%] 2025-12-04T12:38:28.6579977Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 33%] 2025-12-04T12:38:28.6580796Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 33%] 2025-12-04T12:38:28.6581606Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0044s] [ 33%] 2025-12-04T12:38:28.6582440Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 33%] 2025-12-04T12:38:28.6583277Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 33%] 2025-12-04T12:38:28.6584101Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 33%] 2025-12-04T12:38:28.6584920Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 33%] 2025-12-04T12:38:28.6585817Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 33%] 2025-12-04T12:38:28.6586626Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0044s] [ 33%] 2025-12-04T12:38:28.6587448Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 34%] 2025-12-04T12:38:28.6588267Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 34%] 2025-12-04T12:38:28.6589087Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 34%] 2025-12-04T12:38:28.6589904Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 34%] 2025-12-04T12:38:28.6590719Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 34%] 2025-12-04T12:38:28.6591545Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0043s] [ 34%] 2025-12-04T12:38:28.6592406Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 34%] 2025-12-04T12:38:28.6593233Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 34%] 2025-12-04T12:38:28.6594051Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 34%] 2025-12-04T12:38:28.6594859Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 34%] 2025-12-04T12:38:28.6595684Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 34%] 2025-12-04T12:38:28.6596671Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 34%] 2025-12-04T12:38:28.6597502Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 34%] 2025-12-04T12:38:28.6598393Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 35%] 2025-12-04T12:38:28.6599232Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 35%] 2025-12-04T12:38:28.6600046Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 35%] 2025-12-04T12:38:28.6600959Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 35%] 2025-12-04T12:38:28.6601784Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0046s] [ 35%] 2025-12-04T12:38:28.6602609Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 35%] 2025-12-04T12:38:28.6603426Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0038s] [ 35%] 2025-12-04T12:38:28.6604243Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 35%] 2025-12-04T12:38:28.6605069Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0041s] [ 35%] 2025-12-04T12:38:28.6605891Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0041s] [ 35%] 2025-12-04T12:38:28.6606710Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 35%] 2025-12-04T12:38:28.6607582Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 35%] 2025-12-04T12:38:28.6608402Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0040s] [ 35%] 2025-12-04T12:38:28.6609231Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 36%] 2025-12-04T12:38:28.6610039Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 36%] 2025-12-04T12:38:28.6610863Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 36%] 2025-12-04T12:38:28.6611680Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0044s] [ 36%] 2025-12-04T12:38:28.6612516Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 36%] 2025-12-04T12:38:28.6613365Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 36%] 2025-12-04T12:38:28.6614200Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 36%] 2025-12-04T12:38:28.6615014Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 36%] 2025-12-04T12:38:28.6615901Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 36%] 2025-12-04T12:38:28.6616716Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0044s] [ 36%] 2025-12-04T12:38:28.6617588Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 36%] 2025-12-04T12:38:28.6618428Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 36%] 2025-12-04T12:38:28.6619257Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 36%] 2025-12-04T12:38:28.6620086Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 37%] 2025-12-04T12:38:28.6620913Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 37%] 2025-12-04T12:38:28.6621720Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0044s] [ 37%] 2025-12-04T12:38:28.6622606Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 37%] 2025-12-04T12:38:28.6623418Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 37%] 2025-12-04T12:38:28.6624253Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 37%] 2025-12-04T12:38:28.6625069Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 37%] 2025-12-04T12:38:28.6625909Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 37%] 2025-12-04T12:38:28.6626719Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0044s] [ 37%] 2025-12-04T12:38:28.6627551Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 37%] 2025-12-04T12:38:28.6628394Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 37%] 2025-12-04T12:38:28.6629209Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 37%] 2025-12-04T12:38:28.6630029Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0042s] [ 37%] 2025-12-04T12:38:28.6630906Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 38%] 2025-12-04T12:38:28.6631726Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0043s] [ 38%] 2025-12-04T12:38:28.6632549Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 38%] 2025-12-04T12:38:28.6633369Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 38%] 2025-12-04T12:38:28.6634190Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 38%] 2025-12-04T12:38:28.6634999Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0040s] [ 38%] 2025-12-04T12:38:28.6635830Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 38%] 2025-12-04T12:38:28.6636628Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0044s] [ 38%] 2025-12-04T12:38:28.6637488Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 38%] 2025-12-04T12:38:28.6638307Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0040s] [ 38%] 2025-12-04T12:38:28.6639152Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 38%] 2025-12-04T12:38:28.6639962Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0040s] [ 38%] 2025-12-04T12:38:28.6640797Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 38%] 2025-12-04T12:38:28.6641603Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0044s] [ 39%] 2025-12-04T12:38:28.6642424Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 39%] 2025-12-04T12:38:28.6643282Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 39%] 2025-12-04T12:38:28.6644093Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 39%] 2025-12-04T12:38:28.6644925Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0040s] [ 39%] 2025-12-04T12:38:28.6645820Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 39%] 2025-12-04T12:38:28.6646642Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0044s] [ 39%] 2025-12-04T12:38:28.6647460Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 39%] 2025-12-04T12:38:28.6648284Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 39%] 2025-12-04T12:38:28.6649109Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 39%] 2025-12-04T12:38:28.6649912Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 39%] 2025-12-04T12:38:28.6650745Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 39%] 2025-12-04T12:38:28.6651588Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0044s] [ 39%] 2025-12-04T12:38:28.6652427Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0042s] [ 40%] 2025-12-04T12:38:28.6653238Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0041s] [ 40%] 2025-12-04T12:38:28.6654071Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0042s] [ 40%] 2025-12-04T12:38:28.6654875Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0040s] [ 40%] 2025-12-04T12:38:28.6655698Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 40%] 2025-12-04T12:38:28.6656513Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0047s] [ 40%] 2025-12-04T12:38:28.6657397Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 40%] 2025-12-04T12:38:28.6658272Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0043s] [ 40%] 2025-12-04T12:38:28.6659095Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 40%] 2025-12-04T12:38:28.6659915Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 40%] 2025-12-04T12:38:28.6660813Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 40%] 2025-12-04T12:38:28.6661639Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0043s] [ 40%] 2025-12-04T12:38:28.6662451Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 40%] 2025-12-04T12:38:28.6663259Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0040s] [ 41%] 2025-12-04T12:38:28.6664087Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 41%] 2025-12-04T12:38:28.6664901Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0041s] [ 41%] 2025-12-04T12:38:28.6665735Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0041s] [ 41%] 2025-12-04T12:38:28.6666579Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0046s] [ 41%] 2025-12-04T12:38:28.6667412Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0041s] [ 41%] 2025-12-04T12:38:28.6668221Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0040s] [ 41%] 2025-12-04T12:38:28.6669039Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 41%] 2025-12-04T12:38:28.6669847Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0040s] [ 41%] 2025-12-04T12:38:28.6670659Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 41%] 2025-12-04T12:38:28.6671489Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 41%] 2025-12-04T12:38:28.6672311Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 41%] 2025-12-04T12:38:28.6673173Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 41%] 2025-12-04T12:38:28.6673996Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 42%] 2025-12-04T12:38:28.6674821Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 42%] 2025-12-04T12:38:28.6675694Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 42%] 2025-12-04T12:38:28.6676504Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0045s] [ 42%] 2025-12-04T12:38:28.6677331Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 42%] 2025-12-04T12:38:28.6678149Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0040s] [ 42%] 2025-12-04T12:38:28.6678988Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 42%] 2025-12-04T12:38:28.6679790Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 42%] 2025-12-04T12:38:28.6680623Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 42%] 2025-12-04T12:38:28.6681460Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0044s] [ 42%] 2025-12-04T12:38:28.6682297Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 42%] 2025-12-04T12:38:28.6683106Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 42%] 2025-12-04T12:38:28.6683923Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 42%] 2025-12-04T12:38:28.6684753Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0040s] [ 43%] 2025-12-04T12:38:28.6685581Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 43%] 2025-12-04T12:38:28.6686407Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0048s] [ 43%] 2025-12-04T12:38:28.6687228Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 43%] 2025-12-04T12:38:28.6688086Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 43%] 2025-12-04T12:38:28.6688901Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 43%] 2025-12-04T12:38:28.6689736Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0040s] [ 43%] 2025-12-04T12:38:28.6690590Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0040s] [ 43%] 2025-12-04T12:38:28.6691412Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0045s] [ 43%] 2025-12-04T12:38:28.6692252Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 43%] 2025-12-04T12:38:28.6693058Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 43%] 2025-12-04T12:38:28.6693893Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 43%] 2025-12-04T12:38:28.6694703Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 43%] 2025-12-04T12:38:28.6695537Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 44%] 2025-12-04T12:38:28.6696588Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0043s] [ 44%] 2025-12-04T12:38:28.6697481Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 44%] 2025-12-04T12:38:28.6698318Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublas_cuda PASSED [0.0039s] [ 44%] 2025-12-04T12:38:28.6699147Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0039s] [ 44%] 2025-12-04T12:38:28.6699970Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublas_cuda PASSED [0.0039s] [ 44%] 2025-12-04T12:38:28.6700801Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_broadcast_self_False_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0039s] [ 44%] 2025-12-04T12:38:28.6701633Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublas_cuda PASSED [0.0044s] [ 44%] 2025-12-04T12:38:28.6702451Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda PASSED [0.0040s] [ 44%] 2025-12-04T12:38:28.6703327Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublas_cuda PASSED [0.0040s] [ 44%] 2025-12-04T12:38:28.6704144Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda PASSED [0.0041s] [ 44%] 2025-12-04T12:38:28.6704599Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_alignment_cuda_float16 PASSED [0.0103s] [ 44%] 2025-12-04T12:38:28.6705189Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_bias_shapes_size_128_backend_cublas_cuda_bfloat16 PASSED [0.0090s] [ 44%] 2025-12-04T12:38:28.6705710Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_bias_shapes_size_128_backend_cublas_cuda_float16 PASSED [0.0073s] [ 45%] 2025-12-04T12:38:28.6706242Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_bias_shapes_size_128_backend_cublas_cuda_float32 PASSED [0.0069s] [ 45%] 2025-12-04T12:38:28.6706780Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_bias_shapes_size_128_backend_cublaslt_cuda_bfloat16 PASSED [0.0072s] [ 45%] 2025-12-04T12:38:28.6707312Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_bias_shapes_size_128_backend_cublaslt_cuda_float16 PASSED [0.0071s] [ 45%] 2025-12-04T12:38:28.6707855Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_bias_shapes_size_128_backend_cublaslt_cuda_float32 PASSED [0.0066s] [ 45%] 2025-12-04T12:38:28.6708492Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_no_reduced_precision_small_size_4_size_32768_backend_cublas_cuda_float16 PASSED [0.0045s] [ 45%] 2025-12-04T12:38:28.6709145Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_no_reduced_precision_small_size_4_size_32768_backend_cublaslt_cuda_float16 PASSED [0.0036s] [ 45%] 2025-12-04T12:38:28.6709777Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_no_reduced_precision_small_size_8_size_32768_backend_cublas_cuda_float16 PASSED [0.0039s] [ 45%] 2025-12-04T12:38:28.6710420Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_no_reduced_precision_small_size_8_size_32768_backend_cublaslt_cuda_float16 PASSED [0.0037s] [ 45%] 2025-12-04T12:38:28.6711115Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_10000_backend_cublas_cuda_bfloat16 PASSED [6.1450s] [ 45%] 2025-12-04T12:38:28.6711760Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_10000_backend_cublas_cuda_float16 PASSED [5.4090s] [ 45%] 2025-12-04T12:38:28.6712434Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_10000_backend_cublaslt_cuda_bfloat16 PASSED [6.0835s] [ 45%] 2025-12-04T12:38:28.6713093Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_10000_backend_cublaslt_cuda_float16 PASSED [5.4251s] [ 45%] 2025-12-04T12:38:28.6713746Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_1000_backend_cublas_cuda_bfloat16 PASSED [0.0387s] [ 46%] 2025-12-04T12:38:28.6714383Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_1000_backend_cublas_cuda_float16 PASSED [0.0343s] [ 46%] 2025-12-04T12:38:28.6715037Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_1000_backend_cublaslt_cuda_bfloat16 PASSED [0.0342s] [ 46%] 2025-12-04T12:38:28.6715701Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_1000_backend_cublaslt_cuda_float16 PASSED [0.0336s] [ 46%] 2025-12-04T12:38:28.6716337Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_100_backend_cublas_cuda_bfloat16 PASSED [0.0054s] [ 46%] 2025-12-04T12:38:28.6717024Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_100_backend_cublas_cuda_float16 PASSED [0.0046s] [ 46%] 2025-12-04T12:38:28.6717673Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_100_backend_cublaslt_cuda_bfloat16 PASSED [0.0044s] [ 46%] 2025-12-04T12:38:28.6718316Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_fp16_accumulate_size_100_backend_cublaslt_cuda_float16 PASSED [0.0045s] [ 46%] 2025-12-04T12:38:28.6718934Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_10000_backend_cublas_cuda_bfloat16 PASSED [6.0405s] [ 46%] 2025-12-04T12:38:28.6719532Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_10000_backend_cublas_cuda_float16 PASSED [5.3207s] [ 46%] 2025-12-04T12:38:28.6720127Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_10000_backend_cublaslt_cuda_bfloat16 PASSED [6.0662s] [ 46%] 2025-12-04T12:38:28.6720707Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_10000_backend_cublaslt_cuda_float16 PASSED [5.3272s] [ 46%] 2025-12-04T12:38:28.6721284Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_1000_backend_cublas_cuda_bfloat16 PASSED [0.0347s] [ 46%] 2025-12-04T12:38:28.6721845Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_1000_backend_cublas_cuda_float16 PASSED [0.0333s] [ 47%] 2025-12-04T12:38:28.6722425Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_1000_backend_cublaslt_cuda_bfloat16 PASSED [0.0334s] [ 47%] 2025-12-04T12:38:28.6723010Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_1000_backend_cublaslt_cuda_float16 PASSED [0.0330s] [ 47%] 2025-12-04T12:38:28.6723567Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_100_backend_cublas_cuda_bfloat16 PASSED [0.0056s] [ 47%] 2025-12-04T12:38:28.6724133Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_100_backend_cublas_cuda_float16 PASSED [0.0046s] [ 47%] 2025-12-04T12:38:28.6724707Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_100_backend_cublaslt_cuda_bfloat16 PASSED [0.0045s] [ 47%] 2025-12-04T12:38:28.6725305Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_reduced_precision_size_100_backend_cublaslt_cuda_float16 PASSED [0.0044s] [ 47%] 2025-12-04T12:38:28.6725808Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_10000_backend_cublas_cuda_bfloat16 PASSED [6.0432s] [ 47%] 2025-12-04T12:38:28.6726281Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_10000_backend_cublas_cuda_float16 PASSED [5.3420s] [ 47%] 2025-12-04T12:38:28.6726769Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_10000_backend_cublas_cuda_float32 PASSED [5.4199s] [ 47%] 2025-12-04T12:38:28.6727268Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_10000_backend_cublaslt_cuda_bfloat16 PASSED [6.0379s] [ 47%] 2025-12-04T12:38:28.6727755Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_10000_backend_cublaslt_cuda_float16 PASSED [5.3254s] [ 47%] 2025-12-04T12:38:28.6728251Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_10000_backend_cublaslt_cuda_float32 PASSED [5.4828s] [ 47%] 2025-12-04T12:38:28.6728728Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_1000_backend_cublas_cuda_bfloat16 PASSED [0.0336s] [ 48%] 2025-12-04T12:38:28.6729212Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_1000_backend_cublas_cuda_float16 PASSED [0.0329s] [ 48%] 2025-12-04T12:38:28.6729682Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_1000_backend_cublas_cuda_float32 PASSED [0.0291s] [ 48%] 2025-12-04T12:38:28.6730167Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_1000_backend_cublaslt_cuda_bfloat16 PASSED [0.0334s] [ 48%] 2025-12-04T12:38:28.6730711Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_1000_backend_cublaslt_cuda_float16 PASSED [0.0330s] [ 48%] 2025-12-04T12:38:28.6731188Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_1000_backend_cublaslt_cuda_float32 PASSED [0.0269s] [ 48%] 2025-12-04T12:38:28.6731669Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_100_backend_cublas_cuda_bfloat16 PASSED [0.0051s] [ 48%] 2025-12-04T12:38:28.6732134Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_100_backend_cublas_cuda_float16 PASSED [0.0048s] [ 48%] 2025-12-04T12:38:28.6732634Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_100_backend_cublas_cuda_float32 PASSED [0.0049s] [ 48%] 2025-12-04T12:38:28.6733159Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_100_backend_cublaslt_cuda_bfloat16 PASSED [0.0044s] [ 48%] 2025-12-04T12:38:28.6733633Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_100_backend_cublaslt_cuda_float16 PASSED [0.0045s] [ 48%] 2025-12-04T12:38:28.6734126Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_addmm_size_100_backend_cublaslt_cuda_float32 PASSED [0.0042s] [ 48%] 2025-12-04T12:38:28.6734615Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_and_lt_reduced_precision_fp16_accumulate_cuda PASSED [0.2523s] [ 48%] 2025-12-04T12:38:28.6735137Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_1_10000_10000_10000_cuda_bfloat16 PASSED [10.1221s] [ 49%] 2025-12-04T12:38:28.6735658Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_1_10000_10000_10000_cuda_float16 PASSED [7.1838s] [ 49%] 2025-12-04T12:38:28.6736173Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_1_10000_10000_10000_cuda_float32 PASSED [10.1423s] [ 49%] 2025-12-04T12:38:28.6736685Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_1_10000_1000_10000_cuda_bfloat16 PASSED [2.4227s] [ 49%] 2025-12-04T12:38:28.6737281Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_1_10000_1000_10000_cuda_float16 PASSED [2.2202s] [ 49%] 2025-12-04T12:38:28.6737781Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_1_10000_1000_10000_cuda_float32 PASSED [3.0301s] [ 49%] 2025-12-04T12:38:28.6738291Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_2_1000_1000_1000_cuda_bfloat16 PASSED [0.0419s] [ 49%] 2025-12-04T12:38:28.6738899Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_2_1000_1000_1000_cuda_float16 PASSED [0.0319s] [ 49%] 2025-12-04T12:38:28.6739406Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_2_1000_1000_1000_cuda_float32 PASSED [0.0453s] [ 49%] 2025-12-04T12:38:28.6739891Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_2_100_100_100_cuda_bfloat16 PASSED [0.0079s] [ 49%] 2025-12-04T12:38:28.6740368Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_2_100_100_100_cuda_float16 PASSED [0.0108s] [ 49%] 2025-12-04T12:38:28.6740863Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_baddbmm_large_input_2_100_100_100_cuda_float32 PASSED [0.0066s] [ 49%] 2025-12-04T12:38:28.6741647Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_batch_invariance_blackwell_cuda_bfloat16 SKIPPED [0.0003s] (cuBLAS integration for batch invariance is only on Blackwell) [ 49%] 2025-12-04T12:38:28.6742435Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_batch_invariance_blackwell_cuda_float16 SKIPPED [0.0002s] (cuBLAS integration for batch invariance is only on Blackwell) [ 50%] 2025-12-04T12:38:28.6742903Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_1024_cuda_bfloat16 PASSED [0.0200s] [ 50%] 2025-12-04T12:38:28.6743373Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_1024_cuda_float16 PASSED [0.0066s] [ 50%] 2025-12-04T12:38:28.6743827Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_1024_cuda_float32 PASSED [0.0133s] [ 50%] 2025-12-04T12:38:28.6744319Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_128_cuda_bfloat16 PASSED [0.0056s] [ 50%] 2025-12-04T12:38:28.6744781Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_128_cuda_float16 PASSED [0.0056s] [ 50%] 2025-12-04T12:38:28.6745230Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_128_cuda_float32 PASSED [0.0061s] [ 50%] 2025-12-04T12:38:28.6745706Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_2048_cuda_bfloat16 PASSED [0.1133s] [ 50%] 2025-12-04T12:38:28.6746193Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_2048_cuda_float16 PASSED [0.0172s] [ 50%] 2025-12-04T12:38:28.6746676Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_2048_cuda_float32 PASSED [0.0626s] [ 50%] 2025-12-04T12:38:28.6747155Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_256_cuda_bfloat16 PASSED [0.0062s] [ 50%] 2025-12-04T12:38:28.6747606Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_256_cuda_float16 PASSED [0.0059s] [ 50%] 2025-12-04T12:38:28.6748071Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_256_cuda_float32 PASSED [0.0063s] [ 50%] 2025-12-04T12:38:28.6748519Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_32_cuda_bfloat16 PASSED [0.0055s] [ 50%] 2025-12-04T12:38:28.6748969Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_32_cuda_float16 PASSED [0.0055s] [ 51%] 2025-12-04T12:38:28.6749432Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_32_cuda_float32 PASSED [0.0054s] [ 51%] 2025-12-04T12:38:28.6749896Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_4096_cuda_bfloat16 PASSED [0.6054s] [ 51%] 2025-12-04T12:38:28.6750366Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_4096_cuda_float16 PASSED [0.0645s] [ 51%] 2025-12-04T12:38:28.6750822Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_4096_cuda_float32 PASSED [0.3877s] [ 51%] 2025-12-04T12:38:28.6751281Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_512_cuda_bfloat16 PASSED [0.0070s] [ 51%] 2025-12-04T12:38:28.6751748Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_512_cuda_float16 PASSED [0.0062s] [ 51%] 2025-12-04T12:38:28.6752233Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_512_cuda_float32 PASSED [0.0058s] [ 51%] 2025-12-04T12:38:28.6752702Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_64_cuda_bfloat16 PASSED [0.0053s] [ 51%] 2025-12-04T12:38:28.6753150Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_64_cuda_float16 PASSED [0.0054s] [ 51%] 2025-12-04T12:38:28.6753597Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_64_cuda_float32 PASSED [0.0057s] [ 51%] 2025-12-04T12:38:28.6754075Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_8192_cuda_bfloat16 PASSED [5.1921s] [ 51%] 2025-12-04T12:38:28.6754531Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_8192_cuda_float16 PASSED [0.5259s] [ 51%] 2025-12-04T12:38:28.6754985Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_cublas_deterministic_shape_8192_cuda_float32 PASSED [3.0181s] [ 52%] 2025-12-04T12:38:28.6755539Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_fp16_accum_and_fp32_out_failure_batch_size_1_backend_cublas_cuda PASSED [0.0048s] [ 52%] 2025-12-04T12:38:28.6756086Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_fp16_accum_and_fp32_out_failure_batch_size_1_backend_cublaslt_cuda PASSED [0.0043s] [ 52%] 2025-12-04T12:38:28.6756636Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_fp16_accum_and_fp32_out_failure_batch_size_32_backend_cublas_cuda PASSED [0.0043s] [ 52%] 2025-12-04T12:38:28.6757180Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_fp16_accum_and_fp32_out_failure_batch_size_32_backend_cublaslt_cuda PASSED [0.0038s] [ 52%] 2025-12-04T12:38:28.6757774Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_greencontext_carveout_cuda SKIPPED [0.0002s] (Green contexts are not supported) [ 52%] 2025-12-04T12:38:28.6758617Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_False_b_row_major_False_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 52%] 2025-12-04T12:38:28.6759447Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_False_b_row_major_False_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 52%] 2025-12-04T12:38:28.6760348Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_False_b_row_major_False_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 52%] 2025-12-04T12:38:28.6761168Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_False_b_row_major_True_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 52%] 2025-12-04T12:38:28.6761996Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_False_b_row_major_True_cuda_float16 SKIPPED [0.0005s] (Grouped gemm supported only on SM80 or greater) [ 52%] 2025-12-04T12:38:28.6762809Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_False_b_row_major_True_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 52%] 2025-12-04T12:38:28.6763644Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_True_b_row_major_False_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 52%] 2025-12-04T12:38:28.6764460Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_True_b_row_major_False_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 53%] 2025-12-04T12:38:28.6765278Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_True_b_row_major_False_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 53%] 2025-12-04T12:38:28.6766104Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_True_b_row_major_True_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 53%] 2025-12-04T12:38:28.6766962Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_True_b_row_major_True_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 53%] 2025-12-04T12:38:28.6767786Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_False_a_row_major_True_b_row_major_True_cuda_float32 SKIPPED [0.0004s] (Grouped gemm supported only on SM80 or greater) [ 53%] 2025-12-04T12:38:28.6768607Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_False_b_row_major_False_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 53%] 2025-12-04T12:38:28.6769434Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_False_b_row_major_False_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 53%] 2025-12-04T12:38:28.6770248Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_False_b_row_major_False_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 53%] 2025-12-04T12:38:28.6771080Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_False_b_row_major_True_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 53%] 2025-12-04T12:38:28.6771886Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_False_b_row_major_True_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 53%] 2025-12-04T12:38:28.6772728Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_False_b_row_major_True_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 53%] 2025-12-04T12:38:28.6773555Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_True_b_row_major_False_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 53%] 2025-12-04T12:38:28.6774372Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_True_b_row_major_False_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 53%] 2025-12-04T12:38:28.6775251Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_True_b_row_major_False_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 54%] 2025-12-04T12:38:28.6776056Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_True_b_row_major_True_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 54%] 2025-12-04T12:38:28.6776938Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_True_b_row_major_True_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 54%] 2025-12-04T12:38:28.6777746Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_2d_strided_True_a_row_major_True_b_row_major_True_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 54%] 2025-12-04T12:38:28.6778584Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_False_b_row_major_False_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 54%] 2025-12-04T12:38:28.6779397Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_False_b_row_major_False_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 54%] 2025-12-04T12:38:28.6780219Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_False_b_row_major_False_cuda_float32 SKIPPED [0.0004s] (Grouped gemm supported only on SM80 or greater) [ 54%] 2025-12-04T12:38:28.6781081Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_False_b_row_major_True_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 54%] 2025-12-04T12:38:28.6781889Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_False_b_row_major_True_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 54%] 2025-12-04T12:38:28.6782715Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_False_b_row_major_True_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 54%] 2025-12-04T12:38:28.6783534Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_True_b_row_major_False_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 54%] 2025-12-04T12:38:28.6784360Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_True_b_row_major_False_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 54%] 2025-12-04T12:38:28.6785171Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_True_b_row_major_False_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 54%] 2025-12-04T12:38:28.6785979Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_True_b_row_major_True_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 55%] 2025-12-04T12:38:28.6786798Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_True_b_row_major_True_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 55%] 2025-12-04T12:38:28.6787634Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_False_a_row_major_True_b_row_major_True_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 55%] 2025-12-04T12:38:28.6788465Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_False_b_row_major_False_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 55%] 2025-12-04T12:38:28.6789272Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_False_b_row_major_False_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 55%] 2025-12-04T12:38:28.6790158Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_False_b_row_major_False_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 55%] 2025-12-04T12:38:28.6790966Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_False_b_row_major_True_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 55%] 2025-12-04T12:38:28.6791787Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_False_b_row_major_True_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 55%] 2025-12-04T12:38:28.6792597Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_False_b_row_major_True_cuda_float32 SKIPPED [0.0004s] (Grouped gemm supported only on SM80 or greater) [ 55%] 2025-12-04T12:38:28.6793417Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_True_b_row_major_False_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 55%] 2025-12-04T12:38:28.6794236Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_True_b_row_major_False_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 55%] 2025-12-04T12:38:28.6795043Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_True_b_row_major_False_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 55%] 2025-12-04T12:38:28.6795894Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_True_b_row_major_True_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 55%] 2025-12-04T12:38:28.6796860Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_True_b_row_major_True_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 56%] 2025-12-04T12:38:28.6797675Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_2d_3d_strided_True_a_row_major_True_b_row_major_True_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 56%] 2025-12-04T12:38:28.6798505Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_False_b_row_major_False_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 56%] 2025-12-04T12:38:28.6799315Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_False_b_row_major_False_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 56%] 2025-12-04T12:38:28.6800151Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_False_b_row_major_False_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 56%] 2025-12-04T12:38:28.6800976Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_False_b_row_major_True_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 56%] 2025-12-04T12:38:28.6801804Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_False_b_row_major_True_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 56%] 2025-12-04T12:38:28.6802677Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_False_b_row_major_True_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 56%] 2025-12-04T12:38:28.6803507Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_True_b_row_major_False_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 56%] 2025-12-04T12:38:28.6804350Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_True_b_row_major_False_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 56%] 2025-12-04T12:38:28.6805211Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_True_b_row_major_False_cuda_float32 SKIPPED [0.0004s] (Grouped gemm supported only on SM80 or greater) [ 56%] 2025-12-04T12:38:28.6806023Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_True_b_row_major_True_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 56%] 2025-12-04T12:38:28.6806828Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_True_b_row_major_True_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 56%] 2025-12-04T12:38:28.6807650Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_False_a_row_major_True_b_row_major_True_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 57%] 2025-12-04T12:38:28.6808472Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_False_b_row_major_False_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 57%] 2025-12-04T12:38:28.6809298Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_False_b_row_major_False_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 57%] 2025-12-04T12:38:28.6810105Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_False_b_row_major_False_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 57%] 2025-12-04T12:38:28.6810968Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_False_b_row_major_True_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 57%] 2025-12-04T12:38:28.6811780Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_False_b_row_major_True_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 57%] 2025-12-04T12:38:28.6812602Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_False_b_row_major_True_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 57%] 2025-12-04T12:38:28.6813416Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_True_b_row_major_False_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 57%] 2025-12-04T12:38:28.6814219Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_True_b_row_major_False_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 57%] 2025-12-04T12:38:28.6815035Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_True_b_row_major_False_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 57%] 2025-12-04T12:38:28.6815843Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_True_b_row_major_True_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 57%] 2025-12-04T12:38:28.6816651Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_True_b_row_major_True_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 57%] 2025-12-04T12:38:28.6817549Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_2d_strided_True_a_row_major_True_b_row_major_True_cuda_float32 SKIPPED [0.0008s] (Grouped gemm supported only on SM80 or greater) [ 57%] 2025-12-04T12:38:28.6818389Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_False_b_row_major_False_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 58%] 2025-12-04T12:38:28.6819281Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_False_b_row_major_False_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 58%] 2025-12-04T12:38:28.6820096Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_False_b_row_major_False_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 58%] 2025-12-04T12:38:28.6820928Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_False_b_row_major_True_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 58%] 2025-12-04T12:38:28.6821747Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_False_b_row_major_True_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 58%] 2025-12-04T12:38:28.6822570Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_False_b_row_major_True_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 58%] 2025-12-04T12:38:28.6823395Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_True_b_row_major_False_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 58%] 2025-12-04T12:38:28.6824223Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_True_b_row_major_False_cuda_float16 SKIPPED [0.0004s] (Grouped gemm supported only on SM80 or greater) [ 58%] 2025-12-04T12:38:28.6825039Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_True_b_row_major_False_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 58%] 2025-12-04T12:38:28.6825957Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_True_b_row_major_True_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 58%] 2025-12-04T12:38:28.6826770Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_True_b_row_major_True_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 58%] 2025-12-04T12:38:28.6827574Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_False_a_row_major_True_b_row_major_True_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 58%] 2025-12-04T12:38:28.6828405Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_False_b_row_major_False_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 58%] 2025-12-04T12:38:28.6829215Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_False_b_row_major_False_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 59%] 2025-12-04T12:38:28.6830033Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_False_b_row_major_False_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 59%] 2025-12-04T12:38:28.6830850Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_False_b_row_major_True_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 59%] 2025-12-04T12:38:28.6831699Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_False_b_row_major_True_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 59%] 2025-12-04T12:38:28.6832506Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_False_b_row_major_True_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 59%] 2025-12-04T12:38:28.6833338Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_True_b_row_major_False_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 59%] 2025-12-04T12:38:28.6834206Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_True_b_row_major_False_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 59%] 2025-12-04T12:38:28.6835013Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_True_b_row_major_False_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 59%] 2025-12-04T12:38:28.6835836Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_True_b_row_major_True_cuda_bfloat16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 59%] 2025-12-04T12:38:28.6836643Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_True_b_row_major_True_cuda_float16 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 59%] 2025-12-04T12:38:28.6837458Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_3d_3d_strided_True_a_row_major_True_b_row_major_True_cuda_float32 SKIPPED [0.0002s] (Grouped gemm supported only on SM80 or greater) [ 59%] 2025-12-04T12:38:28.6838315Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/2d_a_row_major_False_b_row_major_False_max_autotune_False_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 59%] 2025-12-04T12:38:28.6839173Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/2d_a_row_major_False_b_row_major_False_max_autotune_True_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 59%] 2025-12-04T12:38:28.6840014Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/2d_a_row_major_False_b_row_major_True_max_autotune_False_cuda SKIPPED [0.0004s] (Grouped gemm with compile supported on SM90) [ 60%] 2025-12-04T12:38:28.6840874Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/2d_a_row_major_False_b_row_major_True_max_autotune_True_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 60%] 2025-12-04T12:38:28.6841729Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/2d_a_row_major_True_b_row_major_False_max_autotune_False_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 60%] 2025-12-04T12:38:28.6842559Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/2d_a_row_major_True_b_row_major_False_max_autotune_True_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 60%] 2025-12-04T12:38:28.6843403Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/2d_a_row_major_True_b_row_major_True_max_autotune_False_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 60%] 2025-12-04T12:38:28.6844232Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/2d_a_row_major_True_b_row_major_True_max_autotune_True_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 60%] 2025-12-04T12:38:28.6845121Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/3d_a_row_major_False_b_row_major_False_max_autotune_False_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 60%] 2025-12-04T12:38:28.6845966Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/3d_a_row_major_False_b_row_major_False_max_autotune_True_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 60%] 2025-12-04T12:38:28.6846855Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/3d_a_row_major_False_b_row_major_True_max_autotune_False_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 60%] 2025-12-04T12:38:28.6847690Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/3d_a_row_major_False_b_row_major_True_max_autotune_True_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 60%] 2025-12-04T12:38:28.6848529Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/3d_a_row_major_True_b_row_major_False_max_autotune_False_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 60%] 2025-12-04T12:38:28.6849461Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/3d_a_row_major_True_b_row_major_False_max_autotune_True_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 60%] 2025-12-04T12:38:28.6850297Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/3d_a_row_major_True_b_row_major_True_max_autotune_False_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 60%] 2025-12-04T12:38:28.6851139Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_2d/3d_a_row_major_True_b_row_major_True_max_autotune_True_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 61%] 2025-12-04T12:38:28.6851987Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/2d_a_row_major_False_b_row_major_False_max_autotune_False_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 61%] 2025-12-04T12:38:28.6852841Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/2d_a_row_major_False_b_row_major_False_max_autotune_True_cuda SKIPPED [0.0004s] (Grouped gemm with compile supported on SM90) [ 61%] 2025-12-04T12:38:28.6853778Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/2d_a_row_major_False_b_row_major_True_max_autotune_False_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 61%] 2025-12-04T12:38:28.6854633Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/2d_a_row_major_False_b_row_major_True_max_autotune_True_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 61%] 2025-12-04T12:38:28.6855513Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/2d_a_row_major_True_b_row_major_False_max_autotune_False_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 61%] 2025-12-04T12:38:28.6856348Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/2d_a_row_major_True_b_row_major_False_max_autotune_True_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 61%] 2025-12-04T12:38:28.6857254Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/2d_a_row_major_True_b_row_major_True_max_autotune_False_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 61%] 2025-12-04T12:38:28.6858090Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/2d_a_row_major_True_b_row_major_True_max_autotune_True_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 61%] 2025-12-04T12:38:28.6858948Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/3d_a_row_major_False_b_row_major_False_max_autotune_False_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 61%] 2025-12-04T12:38:28.6859787Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/3d_a_row_major_False_b_row_major_False_max_autotune_True_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 61%] 2025-12-04T12:38:28.6860643Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/3d_a_row_major_False_b_row_major_True_max_autotune_False_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 61%] 2025-12-04T12:38:28.6861508Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/3d_a_row_major_False_b_row_major_True_max_autotune_True_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 61%] 2025-12-04T12:38:28.6862361Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/3d_a_row_major_True_b_row_major_False_max_autotune_False_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 62%] 2025-12-04T12:38:28.6863199Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/3d_a_row_major_True_b_row_major_False_max_autotune_True_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 62%] 2025-12-04T12:38:28.6864100Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/3d_a_row_major_True_b_row_major_True_max_autotune_False_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 62%] 2025-12-04T12:38:28.6864944Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_grouped_gemm_compiled_op_3d/3d_a_row_major_True_b_row_major_True_max_autotune_True_cuda SKIPPED [0.0002s] (Grouped gemm with compile supported on SM90) [ 62%] 2025-12-04T12:38:28.6865409Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_input_dimension_checking_out_dtype_ops0_cuda PASSED [0.0046s] [ 62%] 2025-12-04T12:38:28.6865876Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_input_dimension_checking_out_dtype_ops1_cuda PASSED [0.0042s] [ 62%] 2025-12-04T12:38:28.6866332Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_input_dimension_checking_out_dtype_ops2_cuda PASSED [0.0036s] [ 62%] 2025-12-04T12:38:28.6866788Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_input_dimension_checking_out_dtype_ops3_cuda PASSED [0.0037s] [ 62%] 2025-12-04T12:38:28.6867373Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_backend_cublas_cuda PASSED [0.0038s] [ 62%] 2025-12-04T12:38:28.6867947Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0035s] [ 62%] 2025-12-04T12:38:28.6868529Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0035s] [ 62%] 2025-12-04T12:38:28.6869108Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0035s] [ 62%] 2025-12-04T12:38:28.6869725Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0037s] [ 62%] 2025-12-04T12:38:28.6870300Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0034s] [ 63%] 2025-12-04T12:38:28.6870865Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_backend_cublas_cuda PASSED [0.0035s] [ 63%] 2025-12-04T12:38:28.6871450Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0035s] [ 63%] 2025-12-04T12:38:28.6872019Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0039s] [ 63%] 2025-12-04T12:38:28.6872617Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0035s] [ 63%] 2025-12-04T12:38:28.6873183Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0035s] [ 63%] 2025-12-04T12:38:28.6873757Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0034s] [ 63%] 2025-12-04T12:38:28.6874330Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 63%] 2025-12-04T12:38:28.6874898Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0035s] [ 63%] 2025-12-04T12:38:28.6875511Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0034s] [ 63%] 2025-12-04T12:38:28.6876089Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0034s] [ 63%] 2025-12-04T12:38:28.6876665Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0034s] [ 63%] 2025-12-04T12:38:28.6877284Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_1_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0034s] [ 63%] 2025-12-04T12:38:28.6877880Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_1_batch_size0_backend_cublas_cuda PASSED [0.0038s] [ 64%] 2025-12-04T12:38:28.6878468Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0034s] [ 64%] 2025-12-04T12:38:28.6879033Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0034s] [ 64%] 2025-12-04T12:38:28.6879624Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0034s] [ 64%] 2025-12-04T12:38:28.6880188Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0034s] [ 64%] 2025-12-04T12:38:28.6880767Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0034s] [ 64%] 2025-12-04T12:38:28.6881346Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_32_batch_size0_backend_cublas_cuda PASSED [0.0038s] [ 64%] 2025-12-04T12:38:28.6881922Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0034s] [ 64%] 2025-12-04T12:38:28.6882502Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0034s] [ 64%] 2025-12-04T12:38:28.6883118Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0034s] [ 64%] 2025-12-04T12:38:28.6883695Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0034s] [ 64%] 2025-12-04T12:38:28.6884292Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0038s] [ 64%] 2025-12-04T12:38:28.6884860Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_64_batch_size0_backend_cublas_cuda PASSED [0.0040s] [ 64%] 2025-12-04T12:38:28.6885451Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0034s] [ 65%] 2025-12-04T12:38:28.6886021Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0035s] [ 65%] 2025-12-04T12:38:28.6886621Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0034s] [ 65%] 2025-12-04T12:38:28.6887189Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0034s] [ 65%] 2025-12-04T12:38:28.6887768Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_32_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0034s] [ 65%] 2025-12-04T12:38:28.6888349Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_backend_cublas_cuda PASSED [0.0038s] [ 65%] 2025-12-04T12:38:28.6888952Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0034s] [ 65%] 2025-12-04T12:38:28.6889535Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0034s] [ 65%] 2025-12-04T12:38:28.6890119Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0034s] [ 65%] 2025-12-04T12:38:28.6890718Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0034s] [ 65%] 2025-12-04T12:38:28.6891337Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0034s] [ 65%] 2025-12-04T12:38:28.6891908Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_backend_cublas_cuda PASSED [0.0037s] [ 65%] 2025-12-04T12:38:28.6892491Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0035s] [ 65%] 2025-12-04T12:38:28.6893061Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0035s] [ 66%] 2025-12-04T12:38:28.6893661Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0035s] [ 66%] 2025-12-04T12:38:28.6894231Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0035s] [ 66%] 2025-12-04T12:38:28.6894807Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0036s] [ 66%] 2025-12-04T12:38:28.6895392Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_backend_cublas_cuda PASSED [0.0039s] [ 66%] 2025-12-04T12:38:28.6896118Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0036s] [ 66%] 2025-12-04T12:38:28.6896769Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0035s] [ 66%] 2025-12-04T12:38:28.6897416Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0035s] [ 66%] 2025-12-04T12:38:28.6897997Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0034s] [ 66%] 2025-12-04T12:38:28.6898592Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_1_N_64_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0035s] [ 66%] 2025-12-04T12:38:28.6899161Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_backend_cublas_cuda PASSED [0.0038s] [ 66%] 2025-12-04T12:38:28.6899754Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0034s] [ 66%] 2025-12-04T12:38:28.6900324Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0034s] [ 66%] 2025-12-04T12:38:28.6900905Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0034s] [ 67%] 2025-12-04T12:38:28.6901486Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0035s] [ 67%] 2025-12-04T12:38:28.6902062Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0034s] [ 67%] 2025-12-04T12:38:28.6902698Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_backend_cublas_cuda PASSED [0.0039s] [ 67%] 2025-12-04T12:38:28.6903271Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0034s] [ 67%] 2025-12-04T12:38:28.6903860Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0035s] [ 67%] 2025-12-04T12:38:28.6904492Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0038s] [ 67%] 2025-12-04T12:38:28.6905096Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0035s] [ 67%] 2025-12-04T12:38:28.6905699Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0035s] [ 67%] 2025-12-04T12:38:28.6906264Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_backend_cublas_cuda PASSED [0.0039s] [ 67%] 2025-12-04T12:38:28.6906850Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0034s] [ 67%] 2025-12-04T12:38:28.6907420Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0034s] [ 67%] 2025-12-04T12:38:28.6908006Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0034s] [ 67%] 2025-12-04T12:38:28.6908581Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0035s] [ 68%] 2025-12-04T12:38:28.6909160Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_1_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0038s] [ 68%] 2025-12-04T12:38:28.6909739Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_1_batch_size0_backend_cublas_cuda PASSED [0.0034s] [ 68%] 2025-12-04T12:38:28.6910343Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0034s] [ 68%] 2025-12-04T12:38:28.6910926Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0034s] [ 68%] 2025-12-04T12:38:28.6911514Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0034s] [ 68%] 2025-12-04T12:38:28.6912083Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0034s] [ 68%] 2025-12-04T12:38:28.6912673Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0038s] [ 68%] 2025-12-04T12:38:28.6913246Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_32_batch_size0_backend_cublas_cuda PASSED [0.0034s] [ 68%] 2025-12-04T12:38:28.6913834Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0034s] [ 68%] 2025-12-04T12:38:28.6914405Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0034s] [ 68%] 2025-12-04T12:38:28.6914992Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0034s] [ 68%] 2025-12-04T12:38:28.6915571Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0035s] [ 68%] 2025-12-04T12:38:28.6916188Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0044s] [ 69%] 2025-12-04T12:38:28.6916769Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_64_batch_size0_backend_cublas_cuda PASSED [0.0035s] [ 69%] 2025-12-04T12:38:28.6917349Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0035s] [ 69%] 2025-12-04T12:38:28.6917970Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0034s] [ 69%] 2025-12-04T12:38:28.6918588Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0035s] [ 69%] 2025-12-04T12:38:28.6919163Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0034s] [ 69%] 2025-12-04T12:38:28.6919756Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_32_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0039s] [ 69%] 2025-12-04T12:38:28.6920323Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_backend_cublas_cuda PASSED [0.0034s] [ 69%] 2025-12-04T12:38:28.6920915Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0035s] [ 69%] 2025-12-04T12:38:28.6921489Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0034s] [ 69%] 2025-12-04T12:38:28.6922070Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0034s] [ 69%] 2025-12-04T12:38:28.6922650Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0034s] [ 69%] 2025-12-04T12:38:28.6923227Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0038s] [ 69%] 2025-12-04T12:38:28.6923841Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_backend_cublas_cuda PASSED [0.0034s] [ 70%] 2025-12-04T12:38:28.6924419Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0037s] [ 70%] 2025-12-04T12:38:28.6924995Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0034s] [ 70%] 2025-12-04T12:38:28.6925598Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0035s] [ 70%] 2025-12-04T12:38:28.6926172Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0034s] [ 70%] 2025-12-04T12:38:28.6926760Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0038s] [ 70%] 2025-12-04T12:38:28.6927328Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_backend_cublas_cuda PASSED [0.0034s] [ 70%] 2025-12-04T12:38:28.6927920Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0034s] [ 70%] 2025-12-04T12:38:28.6928497Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0034s] [ 70%] 2025-12-04T12:38:28.6929083Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0034s] [ 70%] 2025-12-04T12:38:28.6929715Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0035s] [ 70%] 2025-12-04T12:38:28.6930297Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0038s] [ 70%] 2025-12-04T12:38:28.6930876Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_backend_cublas_cuda PASSED [0.0034s] [ 70%] 2025-12-04T12:38:28.6931481Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0034s] [ 71%] 2025-12-04T12:38:28.6932083Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0034s] [ 71%] 2025-12-04T12:38:28.6932686Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0034s] [ 71%] 2025-12-04T12:38:28.6933252Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0035s] [ 71%] 2025-12-04T12:38:28.6933839Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0038s] [ 71%] 2025-12-04T12:38:28.6934409Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_backend_cublas_cuda PASSED [0.0034s] [ 71%] 2025-12-04T12:38:28.6935004Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0034s] [ 71%] 2025-12-04T12:38:28.6935575Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0034s] [ 71%] 2025-12-04T12:38:28.6936157Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0034s] [ 71%] 2025-12-04T12:38:28.6936736Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0034s] [ 71%] 2025-12-04T12:38:28.6937420Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0038s] [ 71%] 2025-12-04T12:38:28.6938010Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_backend_cublas_cuda PASSED [0.0034s] [ 71%] 2025-12-04T12:38:28.6938584Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0034s] [ 71%] 2025-12-04T12:38:28.6939153Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0035s] [ 72%] 2025-12-04T12:38:28.6939757Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0035s] [ 72%] 2025-12-04T12:38:28.6940324Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0035s] [ 72%] 2025-12-04T12:38:28.6940922Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_1_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0039s] [ 72%] 2025-12-04T12:38:28.6941489Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_1_batch_size0_backend_cublas_cuda PASSED [0.0034s] [ 72%] 2025-12-04T12:38:28.6942083Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0034s] [ 72%] 2025-12-04T12:38:28.6942653Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0034s] [ 72%] 2025-12-04T12:38:28.6943272Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0035s] [ 72%] 2025-12-04T12:38:28.6943853Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0034s] [ 72%] 2025-12-04T12:38:28.6944433Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0042s] [ 72%] 2025-12-04T12:38:28.6945049Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_32_batch_size0_backend_cublas_cuda PASSED [0.0034s] [ 72%] 2025-12-04T12:38:28.6945658Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0035s] [ 72%] 2025-12-04T12:38:28.6946239Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0035s] [ 72%] 2025-12-04T12:38:28.6946837Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0035s] [ 73%] 2025-12-04T12:38:28.6947405Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0035s] [ 73%] 2025-12-04T12:38:28.6948001Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0040s] [ 73%] 2025-12-04T12:38:28.6948571Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_64_batch_size0_backend_cublas_cuda PASSED [0.0035s] [ 73%] 2025-12-04T12:38:28.6949148Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0035s] [ 73%] 2025-12-04T12:38:28.6949737Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0035s] [ 73%] 2025-12-04T12:38:28.6950321Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0035s] [ 73%] 2025-12-04T12:38:28.6950936Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0035s] [ 73%] 2025-12-04T12:38:28.6951516Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_32_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0039s] [ 73%] 2025-12-04T12:38:28.6952098Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_backend_cublas_cuda PASSED [0.0034s] [ 73%] 2025-12-04T12:38:28.6952672Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0034s] [ 73%] 2025-12-04T12:38:28.6953242Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0034s] [ 73%] 2025-12-04T12:38:28.6953833Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0035s] [ 73%] 2025-12-04T12:38:28.6954400Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0040s] [ 74%] 2025-12-04T12:38:28.6954993Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0035s] [ 74%] 2025-12-04T12:38:28.6955564Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_backend_cublas_cuda PASSED [0.0034s] [ 74%] 2025-12-04T12:38:28.6956141Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0034s] [ 74%] 2025-12-04T12:38:28.6956776Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0035s] [ 74%] 2025-12-04T12:38:28.6957360Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0034s] [ 74%] 2025-12-04T12:38:28.6957946Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0044s] [ 74%] 2025-12-04T12:38:28.6958554Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0034s] [ 74%] 2025-12-04T12:38:28.6959235Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_backend_cublas_cuda PASSED [0.0034s] [ 74%] 2025-12-04T12:38:28.6959814Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0034s] [ 74%] 2025-12-04T12:38:28.6960383Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0035s] [ 74%] 2025-12-04T12:38:28.6960983Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0035s] [ 74%] 2025-12-04T12:38:28.6961549Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0039s] [ 74%] 2025-12-04T12:38:28.6962141Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0034s] [ 75%] 2025-12-04T12:38:28.6962703Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_backend_cublas_cuda PASSED [0.0044s] [ 75%] 2025-12-04T12:38:28.6963268Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0042s] [ 75%] 2025-12-04T12:38:28.6963841Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0043s] [ 75%] 2025-12-04T12:38:28.6964445Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0045s] [ 75%] 2025-12-04T12:38:28.6965011Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0047s] [ 75%] 2025-12-04T12:38:28.6965583Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0042s] [ 75%] 2025-12-04T12:38:28.6966140Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_backend_cublas_cuda PASSED [0.0042s] [ 75%] 2025-12-04T12:38:28.6966717Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0042s] [ 75%] 2025-12-04T12:38:28.6967278Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0042s] [ 75%] 2025-12-04T12:38:28.6967866Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0043s] [ 75%] 2025-12-04T12:38:28.6968426Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0047s] [ 75%] 2025-12-04T12:38:28.6969010Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0042s] [ 75%] 2025-12-04T12:38:28.6969567Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_backend_cublas_cuda PASSED [0.0042s] [ 75%] 2025-12-04T12:38:28.6970165Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0042s] [ 76%] 2025-12-04T12:38:28.6970742Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0042s] [ 76%] 2025-12-04T12:38:28.6971316Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0042s] [ 76%] 2025-12-04T12:38:28.6971923Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0045s] [ 76%] 2025-12-04T12:38:28.6972522Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_1_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0041s] [ 76%] 2025-12-04T12:38:28.6973083Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_1_batch_size0_backend_cublas_cuda PASSED [0.0042s] [ 76%] 2025-12-04T12:38:28.6973666Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0041s] [ 76%] 2025-12-04T12:38:28.6974226Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0042s] [ 76%] 2025-12-04T12:38:28.6974814Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0042s] [ 76%] 2025-12-04T12:38:28.6975373Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0046s] [ 76%] 2025-12-04T12:38:28.6975958Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0041s] [ 76%] 2025-12-04T12:38:28.6976516Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_32_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 76%] 2025-12-04T12:38:28.6977170Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0042s] [ 76%] 2025-12-04T12:38:28.6977752Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0042s] [ 77%] 2025-12-04T12:38:28.6978383Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0042s] [ 77%] 2025-12-04T12:38:28.6978967Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0045s] [ 77%] 2025-12-04T12:38:28.6979544Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0041s] [ 77%] 2025-12-04T12:38:28.6980103Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_64_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 77%] 2025-12-04T12:38:28.6980681Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0042s] [ 77%] 2025-12-04T12:38:28.6981245Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0041s] [ 77%] 2025-12-04T12:38:28.6981834Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0042s] [ 77%] 2025-12-04T12:38:28.6982399Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0045s] [ 77%] 2025-12-04T12:38:28.6982969Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_32_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0041s] [ 77%] 2025-12-04T12:38:28.6983575Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_backend_cublas_cuda PASSED [0.0040s] [ 77%] 2025-12-04T12:38:28.6984141Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0044s] [ 77%] 2025-12-04T12:38:28.6984714Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0042s] [ 77%] 2025-12-04T12:38:28.6985318Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0041s] [ 78%] 2025-12-04T12:38:28.6985923Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0045s] [ 78%] 2025-12-04T12:38:28.6986493Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0041s] [ 78%] 2025-12-04T12:38:28.6987067Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 78%] 2025-12-04T12:38:28.6987652Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0041s] [ 78%] 2025-12-04T12:38:28.6988220Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0042s] [ 78%] 2025-12-04T12:38:28.6988813Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0042s] [ 78%] 2025-12-04T12:38:28.6989377Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0045s] [ 78%] 2025-12-04T12:38:28.6989951Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0041s] [ 78%] 2025-12-04T12:38:28.6990532Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 78%] 2025-12-04T12:38:28.6991103Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0041s] [ 78%] 2025-12-04T12:38:28.6991716Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0041s] [ 78%] 2025-12-04T12:38:28.6992295Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0042s] [ 78%] 2025-12-04T12:38:28.6992870Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0045s] [ 79%] 2025-12-04T12:38:28.6993450Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_1_N_64_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0041s] [ 79%] 2025-12-04T12:38:28.6994010Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_backend_cublas_cuda PASSED [0.0040s] [ 79%] 2025-12-04T12:38:28.6994596Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0041s] [ 79%] 2025-12-04T12:38:28.6995158Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0041s] [ 79%] 2025-12-04T12:38:28.6995749Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0042s] [ 79%] 2025-12-04T12:38:28.6996465Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0044s] [ 79%] 2025-12-04T12:38:28.6997095Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0041s] [ 79%] 2025-12-04T12:38:28.6997671Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_backend_cublas_cuda PASSED [0.0040s] [ 79%] 2025-12-04T12:38:28.6998244Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0041s] [ 79%] 2025-12-04T12:38:28.6998825Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0041s] [ 79%] 2025-12-04T12:38:28.6999490Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0051s] [ 79%] 2025-12-04T12:38:28.7000053Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0042s] [ 79%] 2025-12-04T12:38:28.7000639Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0041s] [ 80%] 2025-12-04T12:38:28.7001199Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 80%] 2025-12-04T12:38:28.7001788Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0041s] [ 80%] 2025-12-04T12:38:28.7002357Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0042s] [ 80%] 2025-12-04T12:38:28.7002955Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0045s] [ 80%] 2025-12-04T12:38:28.7003517Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0041s] [ 80%] 2025-12-04T12:38:28.7004094Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_1_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0041s] [ 80%] 2025-12-04T12:38:28.7004671Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_1_batch_size0_backend_cublas_cuda PASSED [0.0043s] [ 80%] 2025-12-04T12:38:28.7005288Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0042s] [ 80%] 2025-12-04T12:38:28.7005869Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0042s] [ 80%] 2025-12-04T12:38:28.7006443Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0047s] [ 80%] 2025-12-04T12:38:28.7007006Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0041s] [ 80%] 2025-12-04T12:38:28.7007592Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0042s] [ 80%] 2025-12-04T12:38:28.7008154Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_32_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 81%] 2025-12-04T12:38:28.7008739Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0042s] [ 81%] 2025-12-04T12:38:28.7009310Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0043s] [ 81%] 2025-12-04T12:38:28.7009906Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0047s] [ 81%] 2025-12-04T12:38:28.7010506Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0042s] [ 81%] 2025-12-04T12:38:28.7011080Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0041s] [ 81%] 2025-12-04T12:38:28.7011662Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_64_batch_size0_backend_cublas_cuda PASSED [0.0042s] [ 81%] 2025-12-04T12:38:28.7012236Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0042s] [ 81%] 2025-12-04T12:38:28.7012874Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0043s] [ 81%] 2025-12-04T12:38:28.7013454Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0047s] [ 81%] 2025-12-04T12:38:28.7014020Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0041s] [ 81%] 2025-12-04T12:38:28.7014605Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_32_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0042s] [ 81%] 2025-12-04T12:38:28.7015170Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 81%] 2025-12-04T12:38:28.7015755Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0041s] [ 82%] 2025-12-04T12:38:28.7016323Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0042s] [ 82%] 2025-12-04T12:38:28.7016968Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0046s] [ 82%] 2025-12-04T12:38:28.7017534Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0041s] [ 82%] 2025-12-04T12:38:28.7018105Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0041s] [ 82%] 2025-12-04T12:38:28.7018718Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 82%] 2025-12-04T12:38:28.7019293Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0042s] [ 82%] 2025-12-04T12:38:28.7019874Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0043s] [ 82%] 2025-12-04T12:38:28.7020452Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0050s] [ 82%] 2025-12-04T12:38:28.7021017Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0041s] [ 82%] 2025-12-04T12:38:28.7021606Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0041s] [ 82%] 2025-12-04T12:38:28.7022168Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 82%] 2025-12-04T12:38:28.7022752Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0041s] [ 82%] 2025-12-04T12:38:28.7023315Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0043s] [ 83%] 2025-12-04T12:38:28.7023926Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0053s] [ 83%] 2025-12-04T12:38:28.7024499Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0041s] [ 83%] 2025-12-04T12:38:28.7025079Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_32_N_64_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0041s] [ 83%] 2025-12-04T12:38:28.7025650Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 83%] 2025-12-04T12:38:28.7026300Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0041s] [ 83%] 2025-12-04T12:38:28.7026877Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0042s] [ 83%] 2025-12-04T12:38:28.7027454Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0046s] [ 83%] 2025-12-04T12:38:28.7028016Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0040s] [ 83%] 2025-12-04T12:38:28.7028596Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0041s] [ 83%] 2025-12-04T12:38:28.7029159Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 83%] 2025-12-04T12:38:28.7029744Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0042s] [ 83%] 2025-12-04T12:38:28.7030309Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0043s] [ 83%] 2025-12-04T12:38:28.7030887Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0049s] [ 84%] 2025-12-04T12:38:28.7031461Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0043s] [ 84%] 2025-12-04T12:38:28.7032069Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0043s] [ 84%] 2025-12-04T12:38:28.7032644Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_backend_cublas_cuda PASSED [0.0045s] [ 84%] 2025-12-04T12:38:28.7033211Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0043s] [ 84%] 2025-12-04T12:38:28.7033789Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0044s] [ 84%] 2025-12-04T12:38:28.7034367Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0048s] [ 84%] 2025-12-04T12:38:28.7034925Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0042s] [ 84%] 2025-12-04T12:38:28.7035513Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_1_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0042s] [ 84%] 2025-12-04T12:38:28.7036077Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_1_batch_size0_backend_cublas_cuda PASSED [0.0042s] [ 84%] 2025-12-04T12:38:28.7036659Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0043s] [ 84%] 2025-12-04T12:38:28.7037252Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0042s] [ 84%] 2025-12-04T12:38:28.7037832Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0047s] [ 84%] 2025-12-04T12:38:28.7038403Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0041s] [ 85%] 2025-12-04T12:38:28.7038975Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0041s] [ 85%] 2025-12-04T12:38:28.7039578Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_32_batch_size0_backend_cublas_cuda PASSED [0.0042s] [ 85%] 2025-12-04T12:38:28.7040178Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0042s] [ 85%] 2025-12-04T12:38:28.7040752Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0052s] [ 85%] 2025-12-04T12:38:28.7041344Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0046s] [ 85%] 2025-12-04T12:38:28.7041911Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0041s] [ 85%] 2025-12-04T12:38:28.7042499Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0041s] [ 85%] 2025-12-04T12:38:28.7043062Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_64_batch_size0_backend_cublas_cuda PASSED [0.0042s] [ 85%] 2025-12-04T12:38:28.7043646Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0045s] [ 85%] 2025-12-04T12:38:28.7044215Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0047s] [ 85%] 2025-12-04T12:38:28.7044794Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0042s] [ 85%] 2025-12-04T12:38:28.7045399Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0041s] [ 85%] 2025-12-04T12:38:28.7045976Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_32_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0041s] [ 86%] 2025-12-04T12:38:28.7046550Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 86%] 2025-12-04T12:38:28.7047120Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0042s] [ 86%] 2025-12-04T12:38:28.7047686Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0047s] [ 86%] 2025-12-04T12:38:28.7048274Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0046s] [ 86%] 2025-12-04T12:38:28.7048837Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0042s] [ 86%] 2025-12-04T12:38:28.7049420Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0042s] [ 86%] 2025-12-04T12:38:28.7049986Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 86%] 2025-12-04T12:38:28.7050571Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0042s] [ 86%] 2025-12-04T12:38:28.7051171Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0047s] [ 86%] 2025-12-04T12:38:28.7051747Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0046s] [ 86%] 2025-12-04T12:38:28.7052330Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0042s] [ 86%] 2025-12-04T12:38:28.7052940Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0042s] [ 86%] 2025-12-04T12:38:28.7053545Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_backend_cublas_cuda PASSED [0.0042s] [ 87%] 2025-12-04T12:38:28.7054121Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0042s] [ 87%] 2025-12-04T12:38:28.7054685Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0051s] [ 87%] 2025-12-04T12:38:28.7055278Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0045s] [ 87%] 2025-12-04T12:38:28.7055838Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0041s] [ 87%] 2025-12-04T12:38:28.7056430Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float16_M_64_N_64_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0042s] [ 87%] 2025-12-04T12:38:28.7057041Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_backend_cublas_cuda PASSED [0.0037s] [ 87%] 2025-12-04T12:38:28.7057611Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0037s] [ 87%] 2025-12-04T12:38:28.7058183Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0041s] [ 87%] 2025-12-04T12:38:28.7058796Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0037s] [ 87%] 2025-12-04T12:38:28.7059367Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0036s] [ 87%] 2025-12-04T12:38:28.7059937Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 87%] 2025-12-04T12:38:28.7060509Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_backend_cublas_cuda PASSED [0.0037s] [ 87%] 2025-12-04T12:38:28.7061071Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0037s] [ 88%] 2025-12-04T12:38:28.7061631Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0041s] [ 88%] 2025-12-04T12:38:28.7062222Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0036s] [ 88%] 2025-12-04T12:38:28.7062786Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0037s] [ 88%] 2025-12-04T12:38:28.7063372Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0040s] [ 88%] 2025-12-04T12:38:28.7063929Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_backend_cublas_cuda PASSED [0.0037s] [ 88%] 2025-12-04T12:38:28.7064528Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0037s] [ 88%] 2025-12-04T12:38:28.7065100Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0041s] [ 88%] 2025-12-04T12:38:28.7065671Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0036s] [ 88%] 2025-12-04T12:38:28.7066278Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0036s] [ 88%] 2025-12-04T12:38:28.7066883Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_1_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 88%] 2025-12-04T12:38:28.7067457Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_1_batch_size0_backend_cublas_cuda PASSED [0.0036s] [ 88%] 2025-12-04T12:38:28.7068023Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0037s] [ 88%] 2025-12-04T12:38:28.7068584Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0040s] [ 89%] 2025-12-04T12:38:28.7069171Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0036s] [ 89%] 2025-12-04T12:38:28.7069731Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0036s] [ 89%] 2025-12-04T12:38:28.7070315Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0036s] [ 89%] 2025-12-04T12:38:28.7070877Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_32_batch_size0_backend_cublas_cuda PASSED [0.0037s] [ 89%] 2025-12-04T12:38:28.7071443Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0037s] [ 89%] 2025-12-04T12:38:28.7072017Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0041s] [ 89%] 2025-12-04T12:38:28.7072624Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0037s] [ 89%] 2025-12-04T12:38:28.7073201Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0037s] [ 89%] 2025-12-04T12:38:28.7073776Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 89%] 2025-12-04T12:38:28.7074344Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_64_batch_size0_backend_cublas_cuda PASSED [0.0037s] [ 89%] 2025-12-04T12:38:28.7074923Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0037s] [ 89%] 2025-12-04T12:38:28.7075493Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0041s] [ 89%] 2025-12-04T12:38:28.7076083Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0037s] [ 90%] 2025-12-04T12:38:28.7076644Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0037s] [ 90%] 2025-12-04T12:38:28.7077233Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_32_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 90%] 2025-12-04T12:38:28.7077829Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_backend_cublas_cuda PASSED [0.0037s] [ 90%] 2025-12-04T12:38:28.7078398Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0037s] [ 90%] 2025-12-04T12:38:28.7078976Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0040s] [ 90%] 2025-12-04T12:38:28.7079600Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0036s] [ 90%] 2025-12-04T12:38:28.7080205Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0036s] [ 90%] 2025-12-04T12:38:28.7080775Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 90%] 2025-12-04T12:38:28.7081339Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_backend_cublas_cuda PASSED [0.0037s] [ 90%] 2025-12-04T12:38:28.7081924Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0045s] [ 90%] 2025-12-04T12:38:28.7082493Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0036s] [ 90%] 2025-12-04T12:38:28.7083086Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0040s] [ 90%] 2025-12-04T12:38:28.7083647Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0036s] [ 91%] 2025-12-04T12:38:28.7084227Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 91%] 2025-12-04T12:38:28.7084787Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_backend_cublas_cuda PASSED [0.0036s] [ 91%] 2025-12-04T12:38:28.7085355Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0041s] [ 91%] 2025-12-04T12:38:28.7085973Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0036s] [ 91%] 2025-12-04T12:38:28.7086551Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0037s] [ 91%] 2025-12-04T12:38:28.7087192Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0036s] [ 91%] 2025-12-04T12:38:28.7087804Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_1_N_64_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0036s] [ 91%] 2025-12-04T12:38:28.7095186Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_backend_cublas_cuda PASSED [0.0036s] [ 91%] 2025-12-04T12:38:28.7095804Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0040s] [ 91%] 2025-12-04T12:38:28.7096633Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0036s] [ 91%] 2025-12-04T12:38:28.7097290Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0036s] [ 91%] 2025-12-04T12:38:28.7097861Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0036s] [ 91%] 2025-12-04T12:38:28.7098588Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 92%] 2025-12-04T12:38:28.7099154Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_backend_cublas_cuda PASSED [0.0037s] [ 92%] 2025-12-04T12:38:28.7099744Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0042s] [ 92%] 2025-12-04T12:38:28.7100315Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0037s] [ 92%] 2025-12-04T12:38:28.7101017Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0037s] [ 92%] 2025-12-04T12:38:28.7101580Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0036s] [ 92%] 2025-12-04T12:38:28.7102154Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 92%] 2025-12-04T12:38:28.7102724Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_backend_cublas_cuda PASSED [0.0037s] [ 92%] 2025-12-04T12:38:28.7103292Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0041s] [ 92%] 2025-12-04T12:38:28.7103873Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0036s] [ 92%] 2025-12-04T12:38:28.7104450Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0036s] [ 92%] 2025-12-04T12:38:28.7105008Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0036s] [ 92%] 2025-12-04T12:38:28.7105593Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_1_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 92%] 2025-12-04T12:38:28.7106152Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_1_batch_size0_backend_cublas_cuda PASSED [0.0037s] [ 93%] 2025-12-04T12:38:28.7106776Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0041s] [ 93%] 2025-12-04T12:38:28.7107345Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0037s] [ 93%] 2025-12-04T12:38:28.7107929Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0037s] [ 93%] 2025-12-04T12:38:28.7108502Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0036s] [ 93%] 2025-12-04T12:38:28.7109073Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0036s] [ 93%] 2025-12-04T12:38:28.7109644Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_32_batch_size0_backend_cublas_cuda PASSED [0.0037s] [ 93%] 2025-12-04T12:38:28.7110214Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0044s] [ 93%] 2025-12-04T12:38:28.7110796Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0036s] [ 93%] 2025-12-04T12:38:28.7111375Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0036s] [ 93%] 2025-12-04T12:38:28.7111973Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0036s] [ 93%] 2025-12-04T12:38:28.7112561Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 93%] 2025-12-04T12:38:28.7113125Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_64_batch_size0_backend_cublas_cuda PASSED [0.0037s] [ 93%] 2025-12-04T12:38:28.7113706Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0040s] [ 94%] 2025-12-04T12:38:28.7114337Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0036s] [ 94%] 2025-12-04T12:38:28.7114915Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0036s] [ 94%] 2025-12-04T12:38:28.7115497Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0036s] [ 94%] 2025-12-04T12:38:28.7116068Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_32_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 94%] 2025-12-04T12:38:28.7116640Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_backend_cublas_cuda PASSED [0.0037s] [ 94%] 2025-12-04T12:38:28.7117213Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0041s] [ 94%] 2025-12-04T12:38:28.7117790Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0037s] [ 94%] 2025-12-04T12:38:28.7118368Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0036s] [ 94%] 2025-12-04T12:38:28.7118928Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0037s] [ 94%] 2025-12-04T12:38:28.7119511Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 94%] 2025-12-04T12:38:28.7120104Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_backend_cublas_cuda PASSED [0.0037s] [ 94%] 2025-12-04T12:38:28.7120688Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0040s] [ 94%] 2025-12-04T12:38:28.7121254Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0037s] [ 95%] 2025-12-04T12:38:28.7121830Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0036s] [ 95%] 2025-12-04T12:38:28.7122412Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0036s] [ 95%] 2025-12-04T12:38:28.7122987Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0036s] [ 95%] 2025-12-04T12:38:28.7123561Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_backend_cublas_cuda PASSED [0.0037s] [ 95%] 2025-12-04T12:38:28.7124137Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0040s] [ 95%] 2025-12-04T12:38:28.7124701Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0037s] [ 95%] 2025-12-04T12:38:28.7125331Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0040s] [ 95%] 2025-12-04T12:38:28.7125898Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0037s] [ 95%] 2025-12-04T12:38:28.7126487Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_32_N_64_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0036s] [ 95%] 2025-12-04T12:38:28.7127042Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_backend_cublas_cuda PASSED [0.0037s] [ 95%] 2025-12-04T12:38:28.7127701Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0040s] [ 95%] 2025-12-04T12:38:28.7128261Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0037s] [ 95%] 2025-12-04T12:38:28.7128836Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0036s] [ 96%] 2025-12-04T12:38:28.7129410Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0037s] [ 96%] 2025-12-04T12:38:28.7129979Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0040s] [ 96%] 2025-12-04T12:38:28.7130550Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_backend_cublas_cuda PASSED [0.0046s] [ 96%] 2025-12-04T12:38:28.7131123Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0036s] [ 96%] 2025-12-04T12:38:28.7131690Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0036s] [ 96%] 2025-12-04T12:38:28.7132283Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0037s] [ 96%] 2025-12-04T12:38:28.7132850Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0037s] [ 96%] 2025-12-04T12:38:28.7133464Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 96%] 2025-12-04T12:38:28.7134028Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 96%] 2025-12-04T12:38:28.7134610Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0036s] [ 96%] 2025-12-04T12:38:28.7135171Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0037s] [ 96%] 2025-12-04T12:38:28.7135745Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0037s] [ 96%] 2025-12-04T12:38:28.7136318Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0037s] [ 97%] 2025-12-04T12:38:28.7136955Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_1_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 97%] 2025-12-04T12:38:28.7137547Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_1_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 97%] 2025-12-04T12:38:28.7138124Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0036s] [ 97%] 2025-12-04T12:38:28.7138731Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0037s] [ 97%] 2025-12-04T12:38:28.7139321Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0037s] [ 97%] 2025-12-04T12:38:28.7139881Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0037s] [ 97%] 2025-12-04T12:38:28.7140471Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 97%] 2025-12-04T12:38:28.7141095Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_32_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 97%] 2025-12-04T12:38:28.7141779Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0036s] [ 97%] 2025-12-04T12:38:28.7142355Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0037s] [ 97%] 2025-12-04T12:38:28.7142932Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0036s] [ 97%] 2025-12-04T12:38:28.7143515Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0037s] [ 97%] 2025-12-04T12:38:28.7144087Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 98%] 2025-12-04T12:38:28.7144665Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_64_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 98%] 2025-12-04T12:38:28.7145235Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0036s] [ 98%] 2025-12-04T12:38:28.7145805Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0037s] [ 98%] 2025-12-04T12:38:28.7146403Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0037s] [ 98%] 2025-12-04T12:38:28.7147006Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0037s] [ 98%] 2025-12-04T12:38:28.7147597Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_32_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 98%] 2025-12-04T12:38:28.7148160Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 98%] 2025-12-04T12:38:28.7148729Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_1_batch_size0_backend_cublaslt_cuda PASSED [0.0036s] [ 98%] 2025-12-04T12:38:28.7149312Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_16_backend_cublas_cuda PASSED [0.0037s] [ 98%] 2025-12-04T12:38:28.7149887Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_16_backend_cublaslt_cuda PASSED [0.0043s] [ 98%] 2025-12-04T12:38:28.7150463Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_backend_cublas_cuda PASSED [0.0037s] [ 98%] 2025-12-04T12:38:28.7151036Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_1_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 98%] 2025-12-04T12:38:28.7151617Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 99%] 2025-12-04T12:38:28.7152185Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_32_batch_size0_backend_cublaslt_cuda PASSED [0.0036s] [ 99%] 2025-12-04T12:38:28.7152786Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_16_backend_cublas_cuda PASSED [0.0037s] [ 99%] 2025-12-04T12:38:28.7153377Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_16_backend_cublaslt_cuda PASSED [0.0037s] [ 99%] 2025-12-04T12:38:28.7153944Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_backend_cublas_cuda PASSED [0.0037s] [ 99%] 2025-12-04T12:38:28.7154561Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_32_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 99%] 2025-12-04T12:38:28.7155154Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_backend_cublas_cuda PASSED [0.0041s] [ 99%] 2025-12-04T12:38:28.7155730Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_64_batch_size0_backend_cublaslt_cuda PASSED [0.0036s] [ 99%] 2025-12-04T12:38:28.7156306Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_16_backend_cublas_cuda PASSED [0.0037s] [ 99%] 2025-12-04T12:38:28.7156882Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_16_backend_cublaslt_cuda PASSED [0.0040s] [ 99%] 2025-12-04T12:38:28.7157456Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_backend_cublas_cuda PASSED [0.0037s] [ 99%] 2025-12-04T12:38:28.7158040Z test_matmul_cuda.py::TestMatmulCudaCUDA::test_mm_bmm_dtype_overload_float32_M_64_N_64_K_64_batch_size_1_backend_cublaslt_cuda PASSED [0.0037s] [ 99%] 2025-12-04T12:38:28.7158746Z test_matmul_cuda.py::TestMixedDtypesLinearCudaCUDA::test_mixed_dtypes_linear_cuda_bfloat16 SKIPPED [0.0003s] (mixed dtypes linear only supported on SM 8.x) [ 99%] 2025-12-04T12:38:28.7159441Z test_matmul_cuda.py::TestMixedDtypesLinearCudaCUDA::test_mixed_dtypes_linear_cuda_float16 SKIPPED [0.0002s] (mixed dtypes linear only supported on SM 8.x) [100%] 2025-12-04T12:38:28.7159449Z 2025-12-04T12:38:28.7160084Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-829800aaf6d2511a.xml - 2025-12-04T12:38:28.7160372Z ======== 1171 passed, 133 skipped, 280 deselected in 133.12s (0:02:13) ========= 2025-12-04T12:38:28.7173748Z The following tests failed consistently: ['test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda', 'test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda', 'test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda', 'test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda', 'test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda', 'test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda', 'test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda', 'test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_32_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda', 'test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda', 'test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda', 'test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda', 'test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_32_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda', 'test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublas_cuda', 'test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_False_backend_cublaslt_cuda', 'test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublas_cuda', 'test/test_matmul_cuda.py::TestMatmulCudaCUDA::test_addmm_baddmm_dtype_overload_bfloat16_M_64_N_64_K_64_batch_size0_broadcast_self_True_high_precision_self_True_backend_cublaslt_cuda'] 2025-12-04T12:38:28.7173914Z 2025-12-04T12:38:28.7174352Z FINISHED PRINTING LOG FILE of test_matmul_cuda 1/1 (test/test-reports/test_matmul_cuda_1.1_e0f80f5015105a39_.log) 2025-12-04T12:38:28.7174360Z 2025-12-04T12:38:28.7174658Z Finished test_matmul_cuda 1/1 ... [2025-12-04 12:38:28.201329][12291.891710108], took 13.08min 2025-12-04T12:38:28.7175371Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-d7ddbde7cf40ab13.xml 2025-12-04T12:38:28.7176062Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-63a23e8fc1ef8137.xml 2025-12-04T12:38:28.7176747Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-11f9c4926783dfef.xml 2025-12-04T12:38:28.7177485Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-737659f69f9145f2.xml 2025-12-04T12:38:28.7178211Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2099d82b2931fb99.xml 2025-12-04T12:38:28.7178943Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-5f0933bf702563bb.xml 2025-12-04T12:38:28.7179621Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-ed1b6e1c95269b63.xml 2025-12-04T12:38:28.7180312Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-ccb90d6b004c5139.xml 2025-12-04T12:38:28.7180990Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a3b2bb4f2a94fb32.xml 2025-12-04T12:38:28.7181674Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-581c94c41301ad68.xml 2025-12-04T12:38:28.7182351Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-809b8dfa72183a04.xml 2025-12-04T12:38:28.7183027Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-59f3dc780871f4a4.xml 2025-12-04T12:38:28.7183716Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2648ef692c69d7ff.xml 2025-12-04T12:38:28.7184445Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-1e4f023d2e76e67f.xml 2025-12-04T12:38:28.7185130Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-88a75603bd167dcb.xml 2025-12-04T12:38:28.7467101Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-58d14c6ca3ad4514.xml 2025-12-04T12:38:28.7877507Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-4358034553dc4d55.xml 2025-12-04T12:38:28.8186191Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-e604dacd305a4e36.xml 2025-12-04T12:38:28.8457902Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-29c18a70d3fdf25d.xml 2025-12-04T12:38:28.8770362Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2a489f89b00449c3.xml 2025-12-04T12:38:28.9056704Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-1e3f7f3e1ddc719e.xml 2025-12-04T12:38:28.9323450Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-0a2928546dd5ddd2.xml 2025-12-04T12:38:28.9663709Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-36be7a929b5f8cea.xml 2025-12-04T12:38:29.0044631Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-db4ab888748348ef.xml 2025-12-04T12:38:29.0358344Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a61e284ca72dd15d.xml 2025-12-04T12:38:29.0665448Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-8da3a511cea25a62.xml 2025-12-04T12:38:29.0937554Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-aaebada1c5676813.xml 2025-12-04T12:38:29.1203903Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-c9e9611bbdf13a2b.xml 2025-12-04T12:38:29.1481718Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-963810090dd7c38b.xml 2025-12-04T12:38:29.1787956Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-202f1cfa5f6f2e59.xml 2025-12-04T12:38:29.2460375Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a5f1ba0194e0f361.xml 2025-12-04T12:38:29.2792543Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-784ae621331a35c7.xml 2025-12-04T12:38:29.3090689Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2c9da68c717b465c.xml 2025-12-04T12:38:29.3361016Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a06aff42aa9d7210.xml 2025-12-04T12:38:29.3648302Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-66f0187c8b5c56f0.xml 2025-12-04T12:38:29.3924776Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a93a155b666a377d.xml 2025-12-04T12:38:29.4193759Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-f6900cb9a4d4e17a.xml 2025-12-04T12:38:29.4488445Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-b4de48ce2713ebfd.xml 2025-12-04T12:38:29.4797305Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-9d9a1cc8efaa0b76.xml 2025-12-04T12:38:29.5089619Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-f2c6d1655dbd05d5.xml 2025-12-04T12:38:29.5380347Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-30cf0cd6fabb5398.xml 2025-12-04T12:38:29.5675776Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-29e9b8d6f2f6950a.xml 2025-12-04T12:38:29.5913496Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-0ebf80986aa75a9d.xml 2025-12-04T12:38:29.6352987Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-ed818918883fb5a6.xml 2025-12-04T12:38:29.6830396Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-688ec0aa50c9a8c6.xml 2025-12-04T12:38:29.7117934Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-6fda2bf2ae878b46.xml 2025-12-04T12:38:29.7413807Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-0857a7a11662fc06.xml 2025-12-04T12:38:29.7686232Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a3e03d3f7a7bbb81.xml 2025-12-04T12:38:29.7954532Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-829800aaf6d2511a.xml 2025-12-04T12:38:30.9837861Z Uploading logs for 57119749282 to S3 2025-12-04T12:38:31.3430619Z Uploading artifacts took 1.49 seconds 2025-12-04T12:38:31.3431468Z test_matmul_cuda 1/1 failed! 2025-12-04T12:38:31.3435904Z Running test_dataloader 2/2 ... [2025-12-04 12:38:31.343382][12295.033772149] 2025-12-04T12:38:31.3436709Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T12:38:31.3441235Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_dataloader.py', '--shard-id=2', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 12:38:31.343840] 2025-12-04T12:48:37.6875082Z 2025-12-04T12:48:37.6876101Z test_dataloader 2/2 was successful, full logs can be found in artifacts with path test/test-reports/test_dataloader_2.2_8685276eb73ca04c_.log 2025-12-04T12:48:37.6912910Z Running 92 items in this shard: test/test_dataloader.py::TestDatasetRandomSplit::test_lengths_must_equal_dataset_size, test/test_dataloader.py::TestDatasetRandomSplit::test_slicing_of_subset_of_dataset, test/test_dataloader.py::TestDatasetRandomSplit::test_splits_generator, test/test_dataloader.py::TestDatasetRandomSplit::test_splits_have_correct_size, test/test_dataloader.py::TestDatasetRandomSplit::test_splits_reproducibility, test/test_dataloader.py::TestTensorDataset::test_getitem_1d, test/test_dataloader.py::TestTensorDataset::test_single_tensor, test/test_dataloader.py::TestStackDataset::test_empty, test/test_dataloader.py::TestStackDataset::test_len, test/test_dataloader.py::TestConcatDataset::test_add_dataset, test/test_dataloader.py::TestConcatDataset::test_concat_raises_index_error, test/test_dataloader.py::TestConcatDataset::test_concat_two_singletons, test/test_dataloader.py::TestDataLoader::test_batch_sampler, test/test_dataloader.py::TestDataLoader::test_bulk_loading_nobatch, test/test_dataloader.py::TestDataLoader::test_chain_iterable_style_dataset, test/test_dataloader.py::TestDataLoader::test_default_collate_bad_sequence_type, test/test_dataloader.py::TestDataLoader::test_default_collate_dtype, test/test_dataloader.py::TestDataLoader::test_default_collate_numpy_memmap, test/test_dataloader.py::TestDataLoader::test_default_convert_mapping_keep_type, test/test_dataloader.py::TestDataLoader::test_default_convert_sequence_keep_type, test/test_dataloader.py::TestDataLoader::test_distributed_sampler_invalid_rank, test/test_dataloader.py::TestDataLoader::test_error_workers, test/test_dataloader.py::TestDataLoader::test_growing_dataset, test/test_dataloader.py::TestDataLoader::test_invalid_ctor_args_combinations, test/test_dataloader.py::TestDataLoader::test_large_sampler_indices, test/test_dataloader.py::TestDataLoader::test_multiple_dataloaders, test/test_dataloader.py::TestDataLoader::test_multiprocessing_iterdatapipe, test/test_dataloader.py::TestDataLoader::test_numpy_gen_state, test/test_dataloader.py::TestDataLoader::test_numpy_scalars, test/test_dataloader.py::TestDataLoader::test_proper_exit, test/test_dataloader.py::TestDataLoader::test_segfault, test/test_dataloader.py::TestDataLoader::test_sequential_batch, test/test_dataloader.py::TestDataLoader::test_sequential_nonbatch, test/test_dataloader.py::TestDataLoader::test_sequential_pin_memory, test/test_dataloader.py::TestDataLoader::test_sequential_workers, test/test_dataloader.py::TestDataLoader::test_shuffle_batch_none, test/test_dataloader.py::TestDataLoader::test_shuffle_batch_workers, test/test_dataloader.py::TestDataLoader::test_shuffle_batch_workers_prefetch, test/test_dataloader.py::TestDataLoader::test_shuffle_pin_memory, test/test_dataloader.py::TestDataLoader::test_shuffle_reproducibility, test/test_dataloader.py::TestDataLoader::test_shuffle_workers, test/test_dataloader.py::TestDataLoader::test_timeout, test/test_dataloader.py::TestDataLoader::test_typing, test/test_dataloader.py::TestDataLoader::test_worker_init_fn_forkserver, test/test_dataloader.py::TestDataLoader::test_worker_seed, test/test_dataloader.py::TestDataLoader::test_worker_seed_reproducibility, test/test_dataloader.py::IntegrationTestDataLoaderDataPipe::test_shuffler_iterdatapipe, test/test_dataloader.py::TestStringDataLoader::test_shuffle_pin_memory, test/test_dataloader.py::TestDictDataLoader::test_pin_memory, test/test_dataloader.py::TestDictDataLoader::test_pin_memory_device, test/test_dataloader.py::TestDictDataLoader::test_sequential_batch, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_builtin_collection_conversion, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_bulk_loading_nobatch, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_chain_iterable_style_dataset, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_default_collate_bad_sequence_type, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_default_collate_shared_tensor, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_default_convert_mapping_keep_type, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_distributed_sampler_invalid_rank, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_early_exit, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_error_in_init, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_error_workers, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_excessive_thread_creation_warning, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_fd_limit_exceeded, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_growing_dataset, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_invalid_assign_after_init, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_len, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_multi_epochs_reproducibility, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_multiple_dataloaders, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_multiprocessing_contexts, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_multiprocessing_iterdatapipe_with_dill, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_no_segfault, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_numpy, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_numpy_gen_state, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_proper_exit, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_random_sampler_len_with_replacement, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_random_sampler_len_without_replacement, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_sampler, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_sampler_reproducibility, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_sequential_batch, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_sequential_nonbatch, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_sequential_pin_memory, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_shuffle_batch, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_shuffle_reproducibility, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_typing, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_worker_init_fn_forkserver, test/test_dataloader.py::TestDataLoaderPersistentWorkers::test_worker_seed, test/test_dataloader.py::TestCustomPinFn::test_custom_batch_pin_worker, test/test_dataloader.py::TestIndividualWorkerQueue::test_ind_worker_queue, test/test_dataloader.py::TestSetAffinity::test_set_affinity_in_worker_init, test/test_dataloader.py::TestOutOfOrderDataLoader::test_in_order_iterable_ds, test/test_dataloader.py::TestDataLoaderDeviceTypeCUDA::test_nested_tensor_multiprocessing_context_forkserver_cuda, test/test_dataloader.py::TestDataLoaderDeviceTypeCUDA::test_sparse_tensor_multiprocessing_context_forkserver_cuda 2025-12-04T12:48:37.6949074Z 2025-12-04T12:48:37.6949383Z Finished test_dataloader 2/2 ... [2025-12-04 12:48:37.686889][12901.377278624], took 10.11min 2025-12-04T12:48:37.7257776Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_dataloader/test_dataloader-f02fdb9181a8d8ee.xml 2025-12-04T12:48:37.8134044Z Running test_decomp 5/17 ... [2025-12-04 12:48:37.813131][12901.503522459] 2025-12-04T12:48:37.8134565Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T12:48:37.8138240Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_decomp.py', '--shard-id=5', '--num-shards=17', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 12:48:37.813571] 2025-12-04T13:01:25.2949663Z 2025-12-04T13:01:25.2950628Z test_decomp 5/17 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_5.17_3ed363c2fcab4a0f_.log 2025-12-04T13:01:25.3155723Z Running 538 items in this shard: test/test_decomp.py::TestDecompCUDA::test_comprehensive_H_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_H_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_T_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive___getitem___cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive___getitem___cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rmatmul___cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rmod___cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rmul___cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rpow___cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive__chunk_cat_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive__softmax_backward_data_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_abs_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_acos_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_acosh_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_add_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addbmm_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addcdiv_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addcdiv_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addmm_decomposed_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addmv_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addr_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_alias_copy_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_allclose_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_any_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_argsort_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_argsort_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_argsort_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_argsort_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_argsort_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_copy_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_scatter_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_scatter_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_asin_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atan_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atanh_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atanh_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atleast_1d_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atleast_2d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atleast_3d_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atleast_3d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_baddbmm_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bernoulli_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_block_diag_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_block_diag_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_broadcast_tensors_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_broadcast_tensors_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bucketize_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_byte_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_byte_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_byte_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cdouble_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ceil_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cfloat_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_char_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cholesky_solve_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_clamp_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_clamp_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_clamp_min_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_clamp_min_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_combinations_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_constant_pad_nd_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_contiguous_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cummin_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_deg2rad_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_deg2rad_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diag_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diagflat_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diagonal_scatter_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_div_no_rounding_mode_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_double_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_double_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_eq_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_eq_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_eq_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_erfinv_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_exp_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expand_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expm1_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expm1_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_exponential_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_fftshift_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_hfft2_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifft2_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifft_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ihfft2_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ihfft_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ihfftn_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_irfft2_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_irfft2_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_irfft_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_irfft_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_irfftn_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_rfftn_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fill_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_flatten_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fliplr_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_float_power_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_float_power_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_floor_divide_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fmax_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fmin_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_full_like_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gcd_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gcd_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_geometric_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_geqrf_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gradient_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gradient_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_half_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_hash_tensor_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_hsplit_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_hypot_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_i0_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_fill_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_reduce_mean_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_int_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isfinite_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isfinite_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isnan_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isposinf_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isreal_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isreal_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_item_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_item_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_2inputs_2outputs_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_4inputs_with_extra_args_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_binary_return_by_ref_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_binary_return_by_ref_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_kron_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ldexp_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_le_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_cross_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_eigh_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_eigvalsh_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_inv_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_inv_ex_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_lu_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_matrix_rank_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_norm_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_pinv_singular_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_pinv_singular_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_vander_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_vecdot_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linspace_tensor_overload_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linspace_tensor_overload_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log10_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log1p_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log_softmax_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logaddexp2_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_not_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_not_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_not_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_or_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_xor_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logit_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logspace_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logspace_tensor_overload_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logsumexp_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_lt_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_lu_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mH_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mH_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mT_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mT_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_cumprod_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_cumprod_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_fill_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_median_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_normalize_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_select_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_select_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_std_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_sum_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_sum_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_var_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_meshgrid_variadic_tensors_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_min_binary_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_min_reduction_with_dim_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_min_reduction_with_dim_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_minimum_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mode_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_multinomial_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mv_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mvlgamma_mvlgamma_p_1_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nansum_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nansum_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nansum_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_narrow_copy_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_narrow_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_native_batch_norm_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_full_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_ones_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_ones_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_adaptive_max_pool1d_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_batch_norm_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_celu_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_conv1d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_conv_transpose1d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_cross_entropy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_gelu_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_hardtanh_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_interpolate_area_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_interpolate_trilinear_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_layer_norm_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_leaky_relu_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_margin_ranking_loss_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_pool1d_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_pool3d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_unpool1d_grad_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_unpool2d_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_unpool3d_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_multilabel_margin_loss_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_multilabel_margin_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_multilabel_soft_margin_loss_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_circular_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_replicate_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_replicate_negative_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pixel_shuffle_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pixel_unshuffle_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_relu6_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_softshrink_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_softsign_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_threshold_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_triplet_margin_loss_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_unfold_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_unfold_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_upsample_bilinear_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nonzero_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nonzero_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_normal_number_mean_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ones_like_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ones_like_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ones_like_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ormqr_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_outer_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_permute_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_permute_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_pinverse_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polar_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_0_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_2_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_3_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_positive_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_pow_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_pow_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_pow_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_prod_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_prod_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_randint_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_repeat_interleave_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_reshape_as_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resize_as__cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resize_as__cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resolve_conj_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resolve_conj_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resolve_conj_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resolve_conj_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resolve_neg_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resolve_neg_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rot90_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_add_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_add_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_amin_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_amin_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_prod_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_searchsorted_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_select_scatter_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_select_scatter_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sgn_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sgn_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sigmoid_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sigmoid_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_signal_windows_bartlett_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_signal_windows_cosine_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_signal_windows_hamming_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sin_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sin_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sin_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sinc_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sinh_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_slice_scatter_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_softmax_with_dtype_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sort_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_bessel_j0_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_chebyshev_polynomial_t_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_chebyshev_polynomial_t_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_i0e_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_i0e_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_i1_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_i1_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_i1_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_i1e_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_modified_bessel_i1_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_modified_bessel_k0_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_ndtr_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_ndtri_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_spherical_bessel_j0_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_list_args_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_with_sizes_copy_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_squeeze_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_squeeze_copy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_squeeze_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_squeeze_multiple_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_squeeze_multiple_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_stack_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_std_unbiased_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_std_unbiased_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sub_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sum_to_size_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_t_copy_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_t_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_t_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_take_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_take_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tan_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tile_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tile_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_to_sparse_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_topk_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_topk_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_torch_ops_aten__safe_softmax_default_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trapezoid_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trapezoid_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trapz_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trapz_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_triu_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_triu_indices_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_true_divide_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unbind_copy_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unbind_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unbind_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unique_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsafe_chunk_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsqueeze_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_var_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_var_mean_unbiased_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_var_unbiased_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_vdot_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_view_as_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_vsplit_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_where_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_zero__cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_zeros_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick__batch_norm_with_update_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick__chunk_cat_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_abs_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_acos_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_acosh_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_acosh_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_add_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_add_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_add_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_addmv_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_addr_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_addr_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_all_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_all_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_amin_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_amin_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_any_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_as_strided_scatter_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_as_strided_scatter_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_asin_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_atan_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_atanh_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_atanh_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_bitwise_not_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_bitwise_right_shift_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_cat_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_cat_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_cauchy_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_ceil_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_ceil_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_clamp_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_clamp_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_clamp_max_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_clone_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_conj_physical_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_constant_pad_nd_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_constant_pad_nd_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_copysign_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_logsumexp_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_norm_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_renorm_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_special_entr_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_t_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_cos_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_cos_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_cos_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_count_nonzero_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_cumprod_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_cumsum_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_cumsum_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_deg2rad_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_diag_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_diagonal_copy_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_diagonal_copy_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_diagonal_scatter_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_diagonal_scatter_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_digamma_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_div_floor_rounding_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_div_floor_rounding_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_empty_like_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_erfc_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_erfinv_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_exp2_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_expand_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_expand_copy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_expand_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_expand_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_expand_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_expm1_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_eye_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_eye_cuda_float8_e4m3fnuz, test/test_decomp.py::TestDecompCUDA::test_quick_fft_fft2_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_fft_fft2_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_fft_fft_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_fft_fftn_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_fftn_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_fft_hfftn_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ifft2_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ifft_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ihfft2_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ihfft2_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ihfft2_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ihfft_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_irfft2_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_floor_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_fmax_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_fmin_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_fmod_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_i0_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_i0_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_index_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_index_copy_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_index_fill_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_index_select_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_isinf_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_item_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_item_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_item_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_le_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_linalg_diagonal_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_linspace_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_log1p_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_log2_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_log_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_log_softmax_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_logaddexp_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_logical_and_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_logical_not_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_logical_or_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_logspace_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_logsumexp_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_logsumexp_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_logsumexp_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_lt_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_native_dropout_backward_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_ne_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_neg_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_new_empty_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_new_empty_strided_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_new_empty_strided_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_new_empty_strided_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_new_ones_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_new_ones_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_new_zeros_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_new_zeros_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_binary_cross_entropy_with_logits_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_hardtanh_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_hardtanh_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_huber_loss_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_logsigmoid_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_max_unpool3d_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_mse_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_pad_constant_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_relu_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_softshrink_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_norm_nuc_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_ones_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_ones_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_permute_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_polar_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_rad2deg_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_renorm_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_repeat_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_repeat_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_repeat_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_repeat_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_roll_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_round_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_rsqrt_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_rsub_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_rsub_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_select_scatter_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_sign_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_signbit_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_sinc_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_slice_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_softmax_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_special_erfcx_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_special_i0e_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_special_log_ndtr_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_special_zeta_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_special_zeta_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_split_with_sizes_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_sqrt_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_copy_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_copy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_multiple_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_multiple_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_sub_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_t_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_t_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_tanh_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_tanh_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_trace_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_tril_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_unbind_copy_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_unfold_copy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_unfold_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_unfold_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_unfold_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_uniform_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_unsafe_split_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_unsqueeze_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_var_unbiased_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_view_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_rnn_decomp_module_nn_GRU_train_mode_cuda_float64, test/test_decomp.py::DecompOneOffTestsCUDA::test_rms_norm_decomp_cuda_cuda 2025-12-04T13:01:25.3356402Z 2025-12-04T13:01:25.3356733Z Finished test_decomp 5/17 ... [2025-12-04 13:01:25.295584][13668.985973013], took 12.79min 2025-12-04T13:01:25.3357805Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_decomp/test_decomp-2c8b2ff0efe02fa9.xml 2025-12-04T13:01:26.5887332Z Uploading artifacts took 1.16 seconds 2025-12-04T13:01:26.5891504Z Running test_decomp 10/17 ... [2025-12-04 13:01:26.588944][13670.279334986] 2025-12-04T13:01:26.5892029Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T13:01:26.5896526Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_decomp.py', '--shard-id=10', '--num-shards=17', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 13:01:26.589391] 2025-12-04T13:10:57.2587201Z 2025-12-04T13:10:57.2588143Z test_decomp 10/17 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_10.17_cae88f2962bfc8d3_.log 2025-12-04T13:10:57.2781107Z Running 502 items in this shard: test/test_decomp.py::TestDecompCUDA::test_batch_norm_unflatten_weight_bias_cuda, test/test_decomp.py::TestDecompCUDA::test_cat_single_input_cuda, test/test_decomp.py::TestDecompCUDA::test_comprehensive___radd___cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rmatmul___cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rmod___cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive__unsafe_masked_index_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive__unsafe_masked_index_put_accumulate_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive__unsafe_masked_index_put_accumulate_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_acos_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_acosh_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_acosh_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_acosh_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_add_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addcmul_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_amax_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_any_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_argmax_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_copy_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_scatter_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_asin_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atleast_1d_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atleast_2d_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atleast_2d_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atleast_3d_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atleast_3d_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bfloat16_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bfloat16_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bfloat16_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bitwise_left_shift_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bitwise_or_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bitwise_xor_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_broadcast_to_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cartesian_prod_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cfloat_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cfloat_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_chalf_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_chalf_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_chalf_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cholesky_inverse_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_chunk_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_clamp_max_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_column_stack_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_combinations_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_combinations_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_complex_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_constant_pad_nd_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_constant_pad_nd_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_contiguous_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_copysign_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_copysign_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_copysign_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_copysign_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_corrcoef_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cosh_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_count_nonzero_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cummax_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cummax_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cummin_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cumprod_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cumulative_trapezoid_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diag_embed_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diagflat_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_dist_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_div_trunc_rounding_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_dstack_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_permuted_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_eq_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_equal_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_erf_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_erfc_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_erfc_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_erfinv_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_exp2_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_exp_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expand_copy_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expand_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_eye_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_eye_cuda_float8_e4m3fnuz, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_fft2_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_hfft_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifft2_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifft_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifftn_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifftshift_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ihfft_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_irfft2_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_irfft_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_irfftn_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_rfft2_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_rfft_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_rfftn_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_rfftn_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_rfftn_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fill_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_flip_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fliplr_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_flipud_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_floor_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_floor_divide_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fmax_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fmin_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fmod_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_frac_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_frexp_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_frexp_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_full_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_full_like_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gather_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gather_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gcd_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ge_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_grid_sampler_2d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_grid_sampler_3d_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_heaviside_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_histc_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_hstack_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_i0_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_put_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_reduce_amin_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_select_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_inner_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_int_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isclose_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_item_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_item_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_4inputs_with_extra_args_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_binary_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_binary_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_binary_return_by_ref_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_kron_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_cholesky_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_cholesky_ex_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_det_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_diagonal_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_diagonal_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_eigvals_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_ldl_factor_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_lstsq_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_matrix_power_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_vander_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linspace_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log10_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log_softmax_with_dtype_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logcumsumexp_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_and_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_or_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logit_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logspace_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logspace_tensor_overload_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logspace_tensor_overload_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_argmax_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_cumprod_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_cumsum_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_fill_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_logsumexp_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_mean_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_prod_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_scatter_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_std_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_sum_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_var_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_max_binary_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_max_binary_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_max_pool2d_with_indices_backward_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_max_pool2d_with_indices_backward_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_max_pool2d_with_indices_backward_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_max_reduction_with_dim_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_maximum_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_min_binary_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_min_binary_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_min_reduction_with_dim_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_minimum_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mm_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mode_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_movedim_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mul_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mvlgamma_mvlgamma_p_3_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mvlgamma_mvlgamma_p_3_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nan_to_num_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nansum_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_narrow_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ne_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_empty_strided_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_ones_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_zeros_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_adaptive_max_pool1d_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_adaptive_max_pool1d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_avg_pool3d_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_batch_norm_without_cudnn_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_cosine_similarity_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_cosine_similarity_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_hardtanh_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_huber_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_interpolate_nearest-exact_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_margin_ranking_loss_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_pool3d_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_mse_loss_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_multi_head_attention_forward_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_constant_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_reflect_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_reflect_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_replicate_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_replicate_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pixel_shuffle_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pixel_shuffle_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_relu_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_selu_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_selu_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_softmin_with_dtype_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_softmin_with_dtype_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_softplus_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_threshold_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_triplet_margin_loss_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_unfold_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nonzero_static_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nonzero_static_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_norm_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_norm_nuc_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_norm_nuc_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_normal_in_place_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_normal_in_place_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_prod_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_put_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_qr_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_qr_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rand_like_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_randint_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_randint_like_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_randn_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ravel_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_real_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_repeat_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_repeat_interleave_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_reshape_as_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_reshape_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resize__cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resolve_conj_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resolve_neg_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_roll_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_round_decimals_neg_3_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rsqrt_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rsub_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rsub_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_add_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_searchsorted_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sgn_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_signal_windows_general_hamming_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_signal_windows_hamming_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sin_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sinc_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sinh_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sinh_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_softmax_with_dtype_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sort_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sparse_sampled_addmm_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_bessel_j1_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_bessel_j1_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_chebyshev_polynomial_u_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_chebyshev_polynomial_w_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_entr_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_erfcx_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_i1e_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_legendre_polynomial_p_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_log_ndtr_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_modified_bessel_k0_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_modified_bessel_k1_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_polygamma_special_polygamma_n_0_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_scaled_modified_bessel_k1_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_scaled_modified_bessel_k1_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_t_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_u_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_w_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_spherical_bessel_j0_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_xlog1py_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_zeta_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_list_args_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_with_sizes_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_with_sizes_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sqrt_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_square_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_square_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_square_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_square_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_squeeze_copy_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_squeeze_multiple_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sub_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sum_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sum_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sum_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sum_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_svd_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_take_along_dim_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_take_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_take_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tan_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tanh_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tile_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trace_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_transpose_copy_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_transpose_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_transpose_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trapezoid_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trapz_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trapz_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tril_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_triu_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_true_divide_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_true_divide_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unflatten_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unflatten_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unfold_copy_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unfold_copy_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_uniform_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unique_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unravel_index_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unravel_index_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsafe_chunk_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_var_mean_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_var_mean_unbiased_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_view_as_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_view_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_view_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_view_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_vsplit_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_vsplit_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_vstack_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_xlogy_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_xlogy_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_zero__cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_zero__cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_zeros_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick__batch_norm_with_update_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick__native_batch_norm_legit_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick__softmax_backward_data_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick__softmax_backward_data_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick__unsafe_masked_index_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick__unsafe_masked_index_put_accumulate_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick__unsafe_masked_index_put_accumulate_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick__upsample_bilinear2d_aa_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_abs_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_abs_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_abs_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_addcmul_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_addmm_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_addmv_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_alias_copy_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_amin_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_aminmax_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_any_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_as_strided_copy_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_as_strided_scatter_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_asin_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_asin_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_asinh_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_atan2_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_atan_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_bitwise_not_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_bitwise_xor_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_block_diag_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_bucketize_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_ceil_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_clamp_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_clamp_max_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_clamp_max_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_clone_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_conj_physical_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_constant_pad_nd_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_constant_pad_nd_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_copysign_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward__unsafe_masked_index_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_bernoulli_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_logaddexp_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_native_dropout_backward_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_nn_functional_max_unpool2d_grad_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_nn_functional_max_unpool3d_grad_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_cosh_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_cosh_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_cumsum_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_diag_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_diag_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_diagonal_scatter_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_diagonal_scatter_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_div_no_rounding_mode_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_dot_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_empty_like_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_empty_like_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_eq_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_erfinv_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_exp2_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_expand_copy_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_eye_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_eye_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_fft_fft_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_fft_fftn_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_fft_fftn_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_fft_hfft2_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_fft_hfft2_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_fft_hfft2_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_fft_hfft_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_fft_hfftn_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_fft_hfftn_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ifft2_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_irfft2_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_fft_rfftn_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_full_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_ge_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_index_copy_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_index_fill_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_index_select_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_isinf_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_isnan_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_isnan_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_isneginf_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_isneginf_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_isneginf_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_item_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_item_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_lerp_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_lgamma_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_linalg_cross_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_linalg_diagonal_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_log_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_logical_and_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_logical_and_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_logical_not_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_logical_or_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_logspace_tensor_overload_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_lt_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_lt_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_masked_fill_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_maximum_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_meshgrid_list_of_tensors_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_meshgrid_list_of_tensors_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_minimum_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_mul_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_mul_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_mv_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_mvlgamma_mvlgamma_p_1_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_mvlgamma_mvlgamma_p_5_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_nan_to_num_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_ne_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_new_empty_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_new_full_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_hardshrink_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_mish_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_mse_loss_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_pad_constant_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_prelu_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_relu6_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_relu_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_ones_like_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_permute_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_permute_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_pow_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_randn_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_renorm_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_repeat_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_rot90_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_rot90_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_round_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_round_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_round_decimals_0_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_round_decimals_neg_3_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_select_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_sin_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_sinh_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_slice_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_slice_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_slice_scatter_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_special_entr_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_special_i0e_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_special_log_ndtr_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_split_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_split_list_args_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_split_list_args_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_sqrt_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_sqrt_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_stack_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_std_mean_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_sum_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_t_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_take_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_trace_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_tril_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_tril_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_tril_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_triu_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_unbind_copy_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_unsqueeze_copy_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_var_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_view_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_view_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_view_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_view_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_xlogy_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_xlogy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_zeros_like_cuda_int8, test/test_decomp.py::DecompOneOffTestsCUDA::test_contiguous_softmax_cuda 2025-12-04T13:10:57.2971567Z 2025-12-04T13:10:57.2971896Z Finished test_decomp 10/17 ... [2025-12-04 13:10:57.259350][14240.949740899], took 9.51min 2025-12-04T13:10:57.2975224Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_decomp/test_decomp-0c1ee1f6151b681e.xml 2025-12-04T13:10:57.3911522Z Running test_decomp 15/17 ... [2025-12-04 13:10:57.390831][14241.081220799] 2025-12-04T13:10:57.3912054Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T13:10:57.3915570Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_decomp.py', '--shard-id=15', '--num-shards=17', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 13:10:57.391289] 2025-12-04T13:22:40.3875208Z 2025-12-04T13:22:40.3876470Z test_decomp 15/17 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_15.17_7513624a35207a1d_.log 2025-12-04T13:22:40.4079558Z Running 525 items in this shard: test/test_decomp.py::TestDecompCUDA::test_comprehensive_H_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rand___cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rdiv___cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rdiv___cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rpow___cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rpow___cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rsub___cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive__native_batch_norm_legit_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive__segment_reduce_lengths_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive__unsafe_masked_index_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive__unsafe_masked_index_put_accumulate_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_abs_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_acos_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_add_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addcmul_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addr_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_alias_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_aminmax_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_any_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_argwhere_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_argwhere_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_copy_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_scatter_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_asinh_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atan_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atanh_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atleast_1d_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atleast_1d_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atleast_3d_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bfloat16_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bfloat16_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bitwise_and_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bitwise_and_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_block_diag_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bool_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_broadcast_tensors_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bucketize_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cartesian_prod_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cartesian_prod_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cat_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cat_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cat_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cauchy_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ceil_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cfloat_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cfloat_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_chalf_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_char_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cholesky_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_chunk_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_chunk_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_chunk_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_clamp_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_clamp_min_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_clamp_min_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_combinations_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_conj_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_constant_pad_nd_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_constant_pad_nd_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_count_nonzero_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cov_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cross_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cummin_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cumprod_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_deg2rad_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diag_embed_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diagflat_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diagonal_copy_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diagonal_scatter_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diff_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diff_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_div_trunc_rounding_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_div_trunc_rounding_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_dsplit_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_dstack_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_permuted_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_strided_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_strided_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_equal_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expand_as_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expm1_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_fft_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_fft_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_fftn_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_fftshift_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_hfftn_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifft2_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifftn_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifftn_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ihfft_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ihfftn_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_irfft_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_rfft_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fill_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_flatten_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_float_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_float_power_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fmin_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_full_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_full_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_full_like_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gather_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gather_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gcd_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gcd_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ge_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ge_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_geqrf_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gt_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_hash_tensor_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_heaviside_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_hsplit_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_hstack_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_hstack_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_igamma_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_copy_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_fill_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_put_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_put_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_reduce_amax_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_select_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_int_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isclose_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isfinite_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isinf_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isinf_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isinf_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isreal_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_kron_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_cond_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_cond_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_cross_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_diagonal_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_eigh_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_inv_ex_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_lstsq_grad_oriented_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_matrix_norm_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_matrix_norm_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_matrix_power_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_matrix_power_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_norm_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_solve_ex_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linspace_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linspace_tensor_overload_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log1p_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log2_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logcumsumexp_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_and_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_or_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logspace_tensor_overload_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logsumexp_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_long_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_lu_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_lu_solve_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_lu_unpack_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mT_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_amax_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_amin_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_cumsum_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_fill_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_mean_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_normalize_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_scatter_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_scatter_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_select_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_softmin_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_softmin_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_sum_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_max_binary_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_median_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_min_reduction_no_dim_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_min_reduction_with_dim_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mode_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_movedim_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_movedim_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mvlgamma_mvlgamma_p_1_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mvlgamma_mvlgamma_p_5_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mvlgamma_mvlgamma_p_5_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nan_to_num_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nanmedian_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nansum_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_empty_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_empty_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_full_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_full_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_ones_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_zeros_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_adaptive_avg_pool1d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_channel_shuffle_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_conv3d_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_conv_transpose2d_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_cosine_embedding_loss_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_cosine_similarity_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_dropout3d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_dropout_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_embedding_bag_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_fractional_max_pool3d_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_hardshrink_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_hardsigmoid_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_hardswish_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_hinge_embedding_loss_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_interpolate_bicubic_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_interpolate_nearest_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_kl_div_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_linear_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_pool2d_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_pool3d_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_pool3d_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_unpool3d_grad_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_mish_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_multi_margin_loss_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_multi_margin_loss_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_normalize_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_circular_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_constant_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_reflect_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pairwise_distance_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pdist_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pixel_shuffle_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pixel_shuffle_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pixel_unshuffle_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pixel_unshuffle_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pixel_unshuffle_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_smooth_l1_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_softmin_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_softmin_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_softsign_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_triplet_margin_loss_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_unfold_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nonzero_static_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nonzero_static_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_norm_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_norm_fro_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_norm_inf_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ones_like_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_permute_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_pinverse_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_1_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_1_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_4_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_4_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_pow_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_prod_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_prod_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rad2deg_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rad2deg_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rand_like_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_randint_like_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_remainder_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_repeat_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_repeat_interleave_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_reshape_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resolve_neg_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_round_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rsub_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rsub_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_add_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_amax_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_amax_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_amin_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_mean_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_prod_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_sum_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_sum_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_select_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_short_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_short_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_short_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sigmoid_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sigmoid_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sign_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_signal_windows_exponential_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sinc_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sinh_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_slice_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_softmax_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_softmax_with_dtype_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sort_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_bessel_j1_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_chebyshev_polynomial_t_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_chebyshev_polynomial_v_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_chebyshev_polynomial_v_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_entr_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_hermite_polynomial_h_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_i1e_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_legendre_polynomial_p_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_log_ndtr_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_modified_bessel_k0_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_ndtri_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_ndtri_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_polygamma_special_polygamma_n_0_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_scaled_modified_bessel_k1_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_zeta_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_list_args_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_with_sizes_copy_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sqrt_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_squeeze_copy_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_squeeze_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_squeeze_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_stack_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_std_mean_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_stft_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_stft_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sub_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sum_to_size_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sum_to_size_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_t_copy_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_t_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_take_along_dim_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_take_along_dim_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_take_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_take_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tensordot_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tile_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_topk_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_topk_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_torch_ops_aten__safe_softmax_default_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trace_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trace_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_transpose_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trapezoid_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trapz_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tril_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tril_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_triu_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unbind_copy_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unfold_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_uniform_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unique_consecutive_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsafe_chunk_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsqueeze_copy_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsqueeze_copy_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsqueeze_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsqueeze_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsqueeze_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_var_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_var_mean_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_vdot_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_view_as_complex_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_view_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_vsplit_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_vsplit_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_vstack_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_where_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_zeros_like_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick__native_batch_norm_legit_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick__unsafe_masked_index_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_acos_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_acos_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_acosh_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_addcmul_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_addmm_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_addmm_decomposed_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_addmv_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_alias_copy_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_alias_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_amax_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_amin_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_as_strided_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_as_strided_scatter_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_asin_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_asinh_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_atan2_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_atan_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_atan_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_atanh_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_bitwise_and_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_bitwise_left_shift_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_bitwise_or_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_bitwise_right_shift_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_bitwise_xor_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_block_diag_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_block_diag_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_bucketize_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_cat_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_cauchy_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_clamp_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_clamp_min_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_clone_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_baddbmm_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_nn_functional_max_unpool2d_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_nn_functional_max_unpool3d_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_nn_functional_softshrink_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_split_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_squeeze_copy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_unbind_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_unsqueeze_copy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_cos_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_cosh_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_cumprod_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_cumsum_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_diag_embed_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_diagonal_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_diagonal_scatter_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_digamma_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_div_no_rounding_mode_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_div_trunc_rounding_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_div_trunc_rounding_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_empty_strided_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_erfinv_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_exp2_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_expand_copy_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_expand_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_expand_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_expm1_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_fft_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_hfft_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ifft_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ihfft_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_fft_irfftn_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_fft_rfft2_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_fill_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_fill_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_full_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_gcd_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_geometric_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_gt_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_hypot_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_index_fill_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_index_select_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_isin_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_isinf_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_isinf_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_isnan_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_isnan_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_le_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_linalg_cross_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_linspace_tensor_overload_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_linspace_tensor_overload_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_linspace_tensor_overload_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_log1p_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_log_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_logaddexp2_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_logical_or_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_logical_or_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_logical_xor_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_logspace_tensor_overload_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_logsumexp_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_logsumexp_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_maximum_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_maximum_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_minimum_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_mul_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_mvlgamma_mvlgamma_p_1_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_mvlgamma_mvlgamma_p_5_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_mvlgamma_mvlgamma_p_5_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_ne_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_new_empty_strided_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_new_empty_strided_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_new_full_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_new_full_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_new_full_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_nextafter_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_nextafter_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_binary_cross_entropy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_elu_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_hardtanh_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_max_unpool2d_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_max_unpool2d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_mish_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_pad_constant_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_relu_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_softplus_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_softshrink_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_norm_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_norm_inf_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_norm_nuc_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_norm_nuc_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_normal_in_place_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_permute_copy_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_pow_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_pow_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_rad2deg_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_randn_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_reciprocal_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_remainder_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_rot90_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_rsqrt_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_sgn_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_sigmoid_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_signbit_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_sinh_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_sinh_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_sinh_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_slice_scatter_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_special_erfcx_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_special_i1e_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_special_log_ndtr_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_special_ndtr_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_special_ndtri_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_split_list_args_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_split_with_sizes_copy_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_std_unbiased_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_sub_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_t_copy_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_t_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_tan_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_trace_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_transpose_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_tril_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_tril_indices_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_triu_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_unbind_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_unbind_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_unfold_copy_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_uniform_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_unsafe_split_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_unsafe_split_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_unsqueeze_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_unsqueeze_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_var_mean_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_var_mean_unbiased_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_where_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_where_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_xlogy_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_zero__cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_zero__cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_zeros_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_rnn_decomp_module_nn_LSTM_train_mode_cuda_float32, test/test_decomp.py::DecompOneOffTestsCUDA::test_sdpa_nn_functional_scaled_dot_product_attention_cuda_float16 2025-12-04T13:22:40.4278680Z 2025-12-04T13:22:40.4278982Z Finished test_decomp 15/17 ... [2025-12-04 13:22:40.388254][14944.07864089], took 11.72min 2025-12-04T13:22:40.4280059Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_decomp/test_decomp-68b4836d5d5fa3ca.xml 2025-12-04T13:22:41.7240404Z Uploading artifacts took 1.20 seconds 2025-12-04T13:22:41.7244463Z Running test_ci_sanity_check_fail 1/1 ... [2025-12-04 13:22:41.724253][14945.414645474] 2025-12-04T13:22:41.7246050Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T13:22:41.7249186Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ci_sanity_check_fail.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 13:22:41.724681] 2025-12-04T13:23:03.4112854Z Finished test_ci_sanity_check_fail 1/1 ... [2025-12-04 13:23:03.410887][14967.101276735], took 0.36min 2025-12-04T13:23:03.4495433Z Running test_meta 1/5 ... [2025-12-04 13:23:03.449278][14967.139669398] 2025-12-04T13:23:03.4497629Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T13:23:03.4499300Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_meta.py', '--shard-id=1', '--num-shards=5', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 13:23:03.449696] 2025-12-04T13:47:40.4097332Z 2025-12-04T13:47:40.4098225Z test_meta 1/5 was successful, full logs can be found in artifacts with path test/test-reports/test_meta_1.5_23737c7a1f6b460a_.log 2025-12-04T13:47:40.7470422Z Running 7984 items in this shard: test/test_meta.py::TestMetaConverter::test_channels_last_leaf, test/test_meta.py::TestMetaConverter::test_view_of_view_of_leaf, test/test_meta.py::TestMetaCUDA::test_batch_norm_backward_output_mask1_cuda, test/test_meta.py::TestMetaCUDA::test_binary_ufuncs_mixed_dtype___rpow___cuda_float32, test/test_meta.py::TestMetaCUDA::test_binary_ufuncs_mixed_dtype__refs__conversions_complex_cuda_float32, test/test_meta.py::TestMetaCUDA::test_binary_ufuncs_mixed_dtype__refs__conversions_polar_cuda_float32, test/test_meta.py::TestMetaCUDA::test_binary_ufuncs_mixed_dtype__refs_clamp_min_cuda_float32, test/test_meta.py::TestMetaCUDA::test_binary_ufuncs_mixed_dtype__refs_copysign_cuda_float32, test/test_meta.py::TestMetaCUDA::test_binary_ufuncs_mixed_dtype__refs_floor_divide_cuda_float32, test/test_meta.py::TestMetaCUDA::test_binary_ufuncs_mixed_dtype__refs_fmax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_binary_ufuncs_mixed_dtype__refs_fmod_cuda_float32, test/test_meta.py::TestMetaCUDA::test_binary_ufuncs_mixed_dtype__refs_gt_cuda_float32, test/test_meta.py::TestMetaCUDA::test_binary_ufuncs_mixed_dtype__refs_igammac_cuda_float32, test/test_meta.py::TestMetaCUDA::test_binary_ufuncs_mixed_dtype__refs_le_cuda_float32, test/test_meta.py::TestMetaCUDA::test_binary_ufuncs_mixed_dtype__refs_logaddexp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_binary_ufuncs_mixed_dtype__refs_true_divide_cuda_float32, test/test_meta.py::TestMetaCUDA::test_binary_ufuncs_mixed_dtype_eq_cuda_float32, test/test_meta.py::TestMetaCUDA::test_binary_ufuncs_mixed_dtype_heaviside_cuda_float32, test/test_meta.py::TestMetaCUDA::test_binary_ufuncs_mixed_dtype_jiterator_binary_return_by_ref_cuda_float32, test/test_meta.py::TestMetaCUDA::test_binary_ufuncs_mixed_dtype_le_cuda_float32, test/test_meta.py::TestMetaCUDA::test_binary_ufuncs_mixed_dtype_lt_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_H_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_H_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_T_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_T_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace___getitem___cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace___getitem___cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace___radd___cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace___radd___cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace___rand___cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace___rand___cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace___rdiv___cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace___rdiv___cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace___rmatmul___cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace___rmod___cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace___rmod___cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace___rmul___cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace___rmul___cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace___rmul___cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace___rpow___cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace___rpow___cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace___rpow___cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace___rsub___cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace___rsub___cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__batch_norm_with_update_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__chunk_cat_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_abs_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_abs_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_abs_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_acos_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_acos_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_acos_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_acos_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_add_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_add_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_add_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_addcdiv_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_addcdiv_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_addcdiv_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_addcdiv_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_addcdiv_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_addcdiv_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_addcmul_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_addcmul_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_addcmul_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_asin_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_asin_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_asin_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_atan_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_atan_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_ceil_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_ceil_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_clamp_max_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_clamp_min_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_copy_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_copy_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_cos_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_cos_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_cos_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_cos_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_cos_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_cos_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_cosh_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_cosh_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_div_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_erf_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_erf_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_erf_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_erfc_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_erfc_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_exp_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_exp_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_exp_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_exp_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_expm1_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_floor_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_frac_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_frac_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_lerp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_lerp_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_lgamma_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_log10_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_log10_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_log10_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_log10_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_log1p_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_log1p_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_log2_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_log2_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_log_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_max_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_max_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_max_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_max_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_maximum_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_maximum_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_maximum_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_minimum_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_minimum_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_mul_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_mul_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_mul_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_mul_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_neg_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_neg_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_neg_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_norm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_norm_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_pow_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_pow_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_pow_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_reciprocal_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_reciprocal_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_round_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_rsqrt_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_rsqrt_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_rsqrt_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_sigmoid_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_sigmoid_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_sigmoid_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_sign_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_sign_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_sign_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_sign_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_sin_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_sin_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_sin_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_sin_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_sinh_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_sinh_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_sqrt_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_sub_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_sub_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_sub_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_sub_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_tan_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_tanh_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_tanh_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_tanh_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_trunc_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_zero_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_zero_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__foreach_zero_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__segment_reduce_offsets_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__segment_reduce_offsets_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__softmax_backward_data_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__unsafe_masked_index_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__unsafe_masked_index_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__unsafe_masked_index_put_accumulate_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__unsafe_masked_index_put_accumulate_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__unsafe_masked_index_put_accumulate_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace__upsample_bilinear2d_aa_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_abs_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_abs_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_abs_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_acos_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_acos_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_add_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_add_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_add_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_add_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_addbmm_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_addbmm_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_addbmm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_addcdiv_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_addcdiv_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_addcdiv_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_addcmul_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_addcmul_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_addmm_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_addmv_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_addr_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_addr_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_addr_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_addr_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_alias_copy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_allclose_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_amax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_amax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_amax_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_amin_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_amin_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_amin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_aminmax_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_aminmax_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_angle_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_angle_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_angle_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_any_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_arange_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_arange_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_arange_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_argmin_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_argmin_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_argmin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_argwhere_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_as_strided_copy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_as_strided_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_as_strided_copy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_as_strided_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_as_strided_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_as_strided_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_as_strided_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_as_strided_partial_views_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_as_strided_partial_views_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_as_strided_partial_views_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_as_strided_partial_views_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_as_strided_partial_views_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_as_strided_scatter_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_as_strided_scatter_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_as_strided_scatter_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_asin_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_asin_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_asin_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_asin_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_asinh_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_asinh_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_asinh_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_atan2_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_atan2_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_atan2_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_atan2_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_atan_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_atanh_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_atanh_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_atanh_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_atanh_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_atleast_1d_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_atleast_1d_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_atleast_1d_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_atleast_2d_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_atleast_2d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_atleast_3d_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_atleast_3d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_bfloat16_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_bfloat16_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_bfloat16_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_bfloat16_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_bitwise_left_shift_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_bitwise_left_shift_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_bitwise_not_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_bitwise_not_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_bitwise_xor_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_bitwise_xor_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_block_diag_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_block_diag_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_bmm_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_bool_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_bool_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_bool_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_broadcast_tensors_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_broadcast_tensors_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_broadcast_tensors_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_broadcast_tensors_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_broadcast_to_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_broadcast_to_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_bucketize_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_bucketize_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_byte_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_byte_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cartesian_prod_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cartesian_prod_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cartesian_prod_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cat_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cat_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cat_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cat_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cauchy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cauchy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cdouble_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cdouble_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cdouble_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cdouble_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_ceil_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_ceil_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_ceil_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cfloat_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_chalf_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_chalf_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_chalf_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_char_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_char_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_char_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_char_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cholesky_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cholesky_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cholesky_inverse_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_chunk_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_chunk_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_chunk_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_clamp_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_clamp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_clamp_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_clamp_max_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_clamp_max_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_clamp_min_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_clone_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_clone_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_clone_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_column_stack_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_combinations_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_combinations_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_conj_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_conj_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_conj_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_conj_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_conj_physical_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_constant_pad_nd_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_constant_pad_nd_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_contiguous_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_corrcoef_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_corrcoef_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_corrcoef_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_corrcoef_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cos_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cos_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cosh_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cosh_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_count_nonzero_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_count_nonzero_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_count_nonzero_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_count_nonzero_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cov_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cross_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cross_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cross_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cummax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cummin_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cummin_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cummin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cumprod_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cumsum_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cumulative_trapezoid_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cumulative_trapezoid_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cumulative_trapezoid_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cumulative_trapezoid_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cumulative_trapezoid_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_cumulative_trapezoid_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_deg2rad_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diag_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diag_embed_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diag_embed_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diag_embed_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diag_embed_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diag_embed_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diagflat_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diagflat_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diagflat_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diagflat_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diagflat_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diagonal_copy_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diagonal_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diagonal_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diagonal_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diagonal_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diagonal_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diagonal_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diagonal_scatter_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diagonal_scatter_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diagonal_scatter_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diagonal_scatter_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diff_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_diff_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_digamma_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_digamma_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_digamma_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_dist_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_div_floor_rounding_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_div_no_rounding_mode_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_div_no_rounding_mode_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_div_trunc_rounding_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_div_trunc_rounding_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_dot_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_double_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_double_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_double_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_dsplit_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_dsplit_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_dsplit_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_dsplit_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_dstack_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_dstack_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_einsum_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_empty_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_empty_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_empty_like_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_empty_like_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_empty_permuted_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_empty_permuted_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_empty_permuted_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_empty_permuted_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_empty_strided_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_empty_strided_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_eq_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_eq_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_equal_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_equal_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_erf_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_erfc_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_erfc_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_erfc_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_erfinv_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_erfinv_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_erfinv_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_exp2_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_exp2_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_exp2_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_exp_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_exp_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_expand_as_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_expand_as_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_expand_copy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_expand_copy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_expand_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_expand_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_expm1_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_expm1_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_expm1_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_exponential_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_eye_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_eye_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_fft2_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_fft2_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_fft_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_fft_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_fftn_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_fftshift_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_fftshift_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_fftshift_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_hfft2_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_hfft2_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_hfft_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_hfft_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_hfft_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_hfftn_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_hfftn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_ifft2_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_ifft_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_ifftn_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_ifftn_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_ifftshift_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_ihfft2_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_ihfft_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_ihfft_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_ihfft_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_ihfftn_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_ihfftn_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_irfft2_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_irfft2_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_irfft_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_irfft_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_irfft_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_irfftn_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_irfftn_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_irfftn_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_rfft2_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_rfft2_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fft_rfftn_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fill_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fill_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_flatten_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_flip_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_flip_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fliplr_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fliplr_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fliplr_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_flipud_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_float_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_float_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_float_power_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_float_power_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_floor_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_floor_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_floor_divide_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_floor_divide_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fmax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fmax_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fmax_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fmin_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fmin_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fmin_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fmod_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fmod_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fmod_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_fmod_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_frac_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_frexp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_gather_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_gather_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_gather_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_ge_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_ge_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_geometric_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_geometric_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_geqrf_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_geqrf_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_gradient_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_gradient_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_gradient_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_grid_sampler_2d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_grid_sampler_2d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_gt_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_gt_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_half_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_half_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_half_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_half_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_hash_tensor_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_hash_tensor_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_hash_tensor_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_hash_tensor_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_heaviside_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_heaviside_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_heaviside_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_heaviside_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_histc_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_hsplit_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_hsplit_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_hstack_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_hypot_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_i0_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_i0_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_i0_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_add_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_add_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_add_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_add_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_add_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_copy_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_copy_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_fill_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_fill_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_fill_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_reduce_amax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_reduce_amin_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_reduce_amin_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_reduce_amin_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_reduce_mean_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_reduce_mean_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_reduce_prod_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_reduce_prod_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_reduce_prod_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_select_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_select_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_select_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_index_select_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_inner_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_int_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_int_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_int_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_isclose_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_isclose_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_isclose_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_isclose_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_isfinite_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_isfinite_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_isfinite_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_isinf_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_isinf_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_isneginf_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_isneginf_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_isneginf_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_isposinf_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_isposinf_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_isreal_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_isreal_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_item_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_item_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_item_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_jiterator_2inputs_2outputs_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_jiterator_2inputs_2outputs_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_jiterator_2inputs_2outputs_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_jiterator_4inputs_with_extra_args_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_jiterator_4inputs_with_extra_args_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_jiterator_binary_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_jiterator_binary_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_jiterator_binary_return_by_ref_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_jiterator_binary_return_by_ref_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_jiterator_binary_return_by_ref_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_jiterator_unary_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_jiterator_unary_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_jiterator_unary_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_jiterator_unary_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_jiterator_unary_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_kron_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_kron_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_kron_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_kron_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_kthvalue_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_kthvalue_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_kthvalue_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_kthvalue_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_lcm_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_ldexp_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_le_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_le_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_le_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_cholesky_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_cholesky_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_cholesky_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_cond_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_cross_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_cross_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_diagonal_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_diagonal_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_eigvals_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_eigvals_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_householder_product_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_inv_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_inv_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_inv_ex_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_inv_ex_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_ldl_solve_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_lu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_matrix_power_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_multi_dot_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_multi_dot_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_norm_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_norm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_norm_subgradients_at_zero_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_norm_subgradients_at_zero_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_pinv_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_pinv_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_pinv_hermitian_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_pinv_hermitian_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_pinv_hermitian_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_qr_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_slogdet_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_solve_ex_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_solve_triangular_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_solve_triangular_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_svd_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_svd_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_tensorinv_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_tensorsolve_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_vander_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_vander_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_vander_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linalg_vector_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linspace_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_linspace_tensor_overload_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_log10_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_log1p_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_log1p_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_log1p_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_log2_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_log_normal_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_log_softmax_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_log_softmax_with_dtype_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_log_softmax_with_dtype_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_log_softmax_with_dtype_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_logaddexp_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_logdet_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_logical_and_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_logical_not_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_logical_not_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_logical_not_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_logical_not_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_logical_or_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_logical_or_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_logical_or_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_logical_xor_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_logit_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_logspace_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_logspace_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_logspace_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_logspace_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_logsumexp_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_logsumexp_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_long_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_long_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_long_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_lt_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_mH_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_mH_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_mH_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_mT_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_mT_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_mT_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_amax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_amax_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_amin_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_amin_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_amin_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_argmax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_argmin_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_argmin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_cumsum_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_cumsum_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_cumsum_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_fill_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_fill_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_fill_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_log_softmax_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_log_softmax_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_logsumexp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_mean_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_mean_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_mean_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_median_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_median_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_normalize_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_prod_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_prod_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_scatter_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_scatter_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_scatter_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_scatter_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_scatter_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_select_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_softmin_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_std_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_std_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_std_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_sum_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_sum_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_var_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_masked_var_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_matmul_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_max_binary_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_max_pool2d_with_indices_backward_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_max_reduction_no_dim_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_max_reduction_no_dim_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_max_reduction_with_dim_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_max_reduction_with_dim_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_maximum_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_maximum_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_maximum_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_maximum_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_maximum_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_median_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_median_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_median_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_meshgrid_list_of_tensors_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_meshgrid_list_of_tensors_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_meshgrid_list_of_tensors_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_meshgrid_list_of_tensors_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_meshgrid_list_of_tensors_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_meshgrid_variadic_tensors_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_meshgrid_variadic_tensors_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_meshgrid_variadic_tensors_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_meshgrid_variadic_tensors_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_min_binary_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_min_binary_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_min_binary_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_min_binary_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_min_reduction_no_dim_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_min_reduction_with_dim_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_min_reduction_with_dim_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_minimum_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_movedim_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_movedim_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_movedim_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_msort_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_mul_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_mul_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_mul_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_multinomial_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_multinomial_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_mv_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_mv_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_mvlgamma_mvlgamma_p_1_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_mvlgamma_mvlgamma_p_1_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_mvlgamma_mvlgamma_p_1_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_mvlgamma_mvlgamma_p_3_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_mvlgamma_mvlgamma_p_3_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_mvlgamma_mvlgamma_p_5_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_mvlgamma_mvlgamma_p_5_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_mvlgamma_mvlgamma_p_5_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nan_to_num_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nan_to_num_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nanmean_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nanmedian_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nanmedian_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nanmedian_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nansum_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nansum_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_narrow_copy_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_narrow_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_narrow_copy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_narrow_copy_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_native_dropout_backward_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_native_dropout_backward_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_ne_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_new_empty_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_new_empty_strided_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_new_empty_strided_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_new_empty_strided_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_new_empty_strided_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_new_full_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_new_full_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_new_full_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_new_full_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_new_full_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_new_ones_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_new_ones_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_new_zeros_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_new_zeros_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nextafter_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_adaptive_avg_pool3d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_adaptive_max_pool1d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_alpha_dropout_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_batch_norm_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_binary_cross_entropy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_celu_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_channel_shuffle_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_channel_shuffle_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_channel_shuffle_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_conv1d_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_conv1d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_conv2d_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_conv_transpose1d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_conv_transpose1d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_conv_transpose3d_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_conv_transpose3d_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_cosine_embedding_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_cosine_similarity_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_cross_entropy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_ctc_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_elu_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_elu_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_embedding_bag_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_embedding_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_feature_alpha_dropout_with_train_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_feature_alpha_dropout_without_train_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_fractional_max_pool2d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_gaussian_nll_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_gelu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_group_norm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_hardshrink_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_hardsigmoid_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_hardtanh_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_hardtanh_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_hinge_embedding_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_huber_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_interpolate_area_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_interpolate_bicubic_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_interpolate_bicubic_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_interpolate_bilinear_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_interpolate_linear_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_interpolate_nearest-exact_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_interpolate_trilinear_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_kl_div_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_leaky_relu_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_leaky_relu_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_linear_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_linear_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_local_response_norm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_local_response_norm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_local_response_norm_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_logsigmoid_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_margin_ranking_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_margin_ranking_loss_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_max_pool1d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_max_pool2d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_max_unpool2d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_max_unpool2d_grad_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_max_unpool2d_grad_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_max_unpool3d_grad_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_max_unpool3d_grad_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_mish_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_multi_head_attention_forward_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_multi_margin_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_multi_margin_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_multilabel_margin_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_multilabel_soft_margin_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_multilabel_soft_margin_loss_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_multilabel_soft_margin_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_nll_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_normalize_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_normalize_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_pad_circular_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_pad_circular_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_pad_circular_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_pad_circular_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_pad_constant_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_pad_reflect_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_pad_reflect_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_pad_replicate_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_pad_replicate_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_pad_replicate_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_pad_replicate_negative_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_pad_replicate_negative_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_pairwise_distance_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_pixel_shuffle_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_pixel_unshuffle_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_pixel_unshuffle_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_poisson_nll_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_prelu_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_relu6_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_relu_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_rms_norm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_silu_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_smooth_l1_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_soft_margin_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_softmin_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_softmin_with_dtype_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_softmin_with_dtype_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_softplus_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_softplus_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_softsign_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_softsign_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_tanhshrink_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_threshold_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_threshold_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_threshold_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_triplet_margin_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_triplet_margin_loss_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_triplet_margin_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_triplet_margin_loss_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_triplet_margin_with_distance_loss_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_unfold_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_unfold_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nn_functional_upsample_bilinear_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nonzero_static_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_nonzero_static_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_norm_fro_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_norm_inf_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_norm_inf_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_norm_nuc_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_normal_in_place_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_normal_in_place_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_normal_number_mean_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_ones_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_ones_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_ones_like_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_ones_like_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_ones_like_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_outer_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_outer_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_outer_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_polygamma_polygamma_n_0_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_polygamma_polygamma_n_1_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_polygamma_polygamma_n_1_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_polygamma_polygamma_n_1_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_polygamma_polygamma_n_2_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_polygamma_polygamma_n_2_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_polygamma_polygamma_n_3_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_polygamma_polygamma_n_3_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_polygamma_polygamma_n_4_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_polygamma_polygamma_n_4_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_positive_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_positive_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_positive_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_pow_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_prod_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_prod_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_prod_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_prod_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_put_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_put_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_rad2deg_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_rad2deg_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_rad2deg_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_randint_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_randint_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_randint_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_randint_like_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_randint_like_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_randint_like_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_randint_like_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_randn_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_randn_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_ravel_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_real_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_real_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_reciprocal_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_reciprocal_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_reciprocal_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_remainder_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_repeat_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_repeat_interleave_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_repeat_interleave_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_repeat_interleave_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_repeat_interleave_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_repeat_interleave_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_reshape_as_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_reshape_as_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_reshape_as_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_reshape_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_reshape_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_reshape_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_reshape_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_reshape_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_resize__cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_resize__cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_resolve_conj_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_resolve_conj_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_resolve_conj_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_resolve_neg_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_resolve_neg_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_resolve_neg_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_resolve_neg_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_roll_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_roll_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_roll_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_roll_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_rot90_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_rot90_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_round_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_round_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_round_decimals_neg_3_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_rsqrt_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_rsqrt_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_rsqrt_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_rsub_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_rsub_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scalar_tensor_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scalar_tensor_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scalar_tensor_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scatter_add_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scatter_add_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scatter_add_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scatter_add_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scatter_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scatter_reduce_amax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scatter_reduce_amax_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scatter_reduce_amin_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scatter_reduce_amin_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scatter_reduce_amin_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scatter_reduce_amin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scatter_reduce_mean_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scatter_reduce_mean_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scatter_reduce_mean_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scatter_reduce_mean_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scatter_reduce_prod_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scatter_reduce_prod_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scatter_reduce_sum_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_scatter_reduce_sum_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_searchsorted_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_select_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sgn_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sgn_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sgn_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sgn_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sgn_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_short_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_short_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sigmoid_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sigmoid_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sign_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sign_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sign_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sign_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_signal_windows_bartlett_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_signal_windows_exponential_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_signal_windows_exponential_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_signal_windows_general_hamming_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_signal_windows_hamming_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_signal_windows_kaiser_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_signal_windows_nuttall_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_signal_windows_nuttall_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_signbit_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_signbit_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sin_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sinc_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sinc_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sinc_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sinh_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sinh_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_slice_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_slice_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_slice_scatter_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_slice_scatter_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_softmax_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_softmax_with_dtype_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_softmax_with_dtype_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_softmax_with_dtype_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_softmax_with_dtype_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_softmax_with_dtype_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_softmax_with_dtype_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sort_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sparse_mm_reduce_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sparse_sampled_addmm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_airy_ai_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_airy_ai_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_airy_ai_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_airy_ai_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_bessel_j0_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_bessel_j0_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_bessel_j0_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_bessel_j1_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_bessel_j1_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_bessel_j1_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_bessel_j1_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_bessel_j1_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_bessel_y0_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_bessel_y1_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_bessel_y1_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_chebyshev_polynomial_t_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_chebyshev_polynomial_u_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_chebyshev_polynomial_u_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_chebyshev_polynomial_w_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_chebyshev_polynomial_w_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_chebyshev_polynomial_w_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_chebyshev_polynomial_w_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_chebyshev_polynomial_w_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_entr_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_entr_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_entr_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_entr_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_erfcx_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_erfcx_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_hermite_polynomial_h_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_hermite_polynomial_h_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_hermite_polynomial_he_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_hermite_polynomial_he_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_i0e_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_i0e_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_i1_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_i1e_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_i1e_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_laguerre_polynomial_l_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_laguerre_polynomial_l_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_legendre_polynomial_p_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_legendre_polynomial_p_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_log_ndtr_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_log_ndtr_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_log_ndtr_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_modified_bessel_i0_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_modified_bessel_i1_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_modified_bessel_i1_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_modified_bessel_k0_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_modified_bessel_k0_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_modified_bessel_k0_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_modified_bessel_k1_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_modified_bessel_k1_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_polygamma_special_polygamma_n_0_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_polygamma_special_polygamma_n_0_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_polygamma_special_polygamma_n_0_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_scaled_modified_bessel_k0_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_scaled_modified_bessel_k0_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_scaled_modified_bessel_k1_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_shifted_chebyshev_polynomial_t_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_shifted_chebyshev_polynomial_t_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_shifted_chebyshev_polynomial_u_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_shifted_chebyshev_polynomial_u_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_shifted_chebyshev_polynomial_u_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_shifted_chebyshev_polynomial_w_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_spherical_bessel_j0_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_spherical_bessel_j0_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_xlog1py_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_special_zeta_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_split_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_split_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_split_list_args_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_split_list_args_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_split_list_args_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_split_with_sizes_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_split_with_sizes_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_split_with_sizes_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_split_with_sizes_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_square_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_squeeze_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_squeeze_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_squeeze_multiple_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_squeeze_multiple_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_stack_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_stack_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_stack_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_std_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sum_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sum_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_sum_to_size_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_t_copy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_t_copy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_t_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_t_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_t_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_t_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_t_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_t_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_take_along_dim_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_take_along_dim_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_tan_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_tan_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_tan_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_tanh_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_tanh_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_tanh_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_tensor_split_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_tensor_split_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_tensordot_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_tensordot_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_tile_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_tile_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_tile_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_tile_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_to_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_to_sparse_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_topk_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_torch__scaled_mm_v2_cuda_float8_e4m3fn, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_torch_ops_aten__safe_softmax_default_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_trace_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_transpose_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_transpose_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_transpose_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_transpose_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_transpose_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_trapezoid_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_trapz_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_trapz_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_tril_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_tril_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_tril_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_triu_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_triu_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_triu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_triu_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_triu_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_triu_indices_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_triu_indices_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_true_divide_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_true_divide_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_true_divide_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_true_divide_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_trunc_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unbind_copy_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unbind_copy_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unbind_copy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unbind_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unbind_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unbind_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unbind_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unflatten_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unflatten_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unflatten_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unfold_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unfold_copy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unfold_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unfold_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unfold_copy_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unique_consecutive_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unique_consecutive_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unique_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unique_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unravel_index_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unsafe_chunk_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unsafe_chunk_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unsafe_chunk_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unsafe_split_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unsqueeze_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_unsqueeze_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_var_mean_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_var_mean_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_var_mean_unbiased_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_var_mean_unbiased_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_var_unbiased_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_view_as_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_view_as_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_view_as_real_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_view_as_real_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_view_copy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_view_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_view_copy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_view_copy_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_view_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_view_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_vsplit_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_vsplit_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_vsplit_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_vstack_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_vstack_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_vstack_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_where_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_where_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_where_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_where_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_xlogy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_xlogy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_zero__cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_zero__cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_zero__cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_zeros_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_zeros_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_zeros_like_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_zeros_like_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_inplace_zeros_like_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_T_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_T_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_T_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_T_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___getitem___cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___getitem___cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___radd___cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___rand___cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___rand___cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___rdiv___cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___rmatmul___cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___rmatmul___cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___rmatmul___cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___rmod___cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___rmul___cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___ror___cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___ror___cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___rpow___cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___rpow___cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___rpow___cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___rpow___cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___rpow___cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___rpow___cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___rsub___cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace___rxor___cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__batch_norm_with_update_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__chunk_cat_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__chunk_cat_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__chunk_cat_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_abs_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_abs_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_acos_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_acos_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_acos_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_add_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_add_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_addcmul_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_asin_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_asin_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_asin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_atan_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_atan_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_atan_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_ceil_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_ceil_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_ceil_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_ceil_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_clamp_max_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_clamp_max_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_clamp_max_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_cos_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_cosh_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_cosh_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_div_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_div_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_erf_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_erf_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_erf_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_erf_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_erfc_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_exp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_exp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_exp_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_expm1_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_frac_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_lerp_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_lerp_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_lerp_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_lerp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_lerp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_lerp_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_lerp_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_lgamma_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_lgamma_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_log10_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_log10_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_log1p_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_log1p_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_log2_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_log2_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_log_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_log_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_max_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_max_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_max_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_max_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_max_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_maximum_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_minimum_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_minimum_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_minimum_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_minimum_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_minimum_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_minimum_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_minimum_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_mul_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_mul_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_neg_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_neg_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_norm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_pow_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_pow_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_reciprocal_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_reciprocal_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_rsqrt_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_rsqrt_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_rsqrt_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_rsqrt_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_rsqrt_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_rsqrt_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_sigmoid_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_sigmoid_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_sigmoid_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_sign_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_sign_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_sign_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_sin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_sinh_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_sinh_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_sinh_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_sinh_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_sinh_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_sqrt_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_sqrt_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_sqrt_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_sub_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_tan_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_trunc_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_zero_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_zero_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_zero_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__foreach_zero_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__segment_reduce_lengths_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__softmax_backward_data_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__unsafe_masked_index_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__unsafe_masked_index_put_accumulate_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__unsafe_masked_index_put_accumulate_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__unsafe_masked_index_put_accumulate_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__unsafe_masked_index_put_accumulate_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__upsample_bilinear2d_aa_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace__upsample_bilinear2d_aa_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_abs_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_abs_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_acos_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_acos_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_acos_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_acosh_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_acosh_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_add_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_add_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_add_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_add_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_addbmm_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_addbmm_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_addcdiv_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_addcdiv_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_addcmul_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_addmm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_addmm_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_addr_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_addr_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_alias_copy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_all_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_all_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_all_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_allclose_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_amax_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_amax_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_amax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_amax_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_amax_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_amin_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_amin_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_amin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_aminmax_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_aminmax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_angle_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_angle_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_angle_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_angle_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_any_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_arange_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_argmax_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_argmin_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_argmin_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_argsort_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_argsort_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_as_strided_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_as_strided_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_as_strided_copy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_as_strided_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_as_strided_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_as_strided_partial_views_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_as_strided_partial_views_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_as_strided_partial_views_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_as_strided_scatter_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_as_strided_scatter_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_asin_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_asinh_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_atan2_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_atan2_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_atan_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_atan_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_atan_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_atanh_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_atleast_1d_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_atleast_1d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_atleast_1d_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_atleast_3d_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_bfloat16_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_bfloat16_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_bitwise_and_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_bitwise_and_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_bitwise_left_shift_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_bitwise_not_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_bitwise_or_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_bitwise_or_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_bitwise_right_shift_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_bitwise_xor_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_block_diag_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_block_diag_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_block_diag_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_bmm_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_bool_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_bool_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_broadcast_to_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_broadcast_to_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_bucketize_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_bucketize_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_byte_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_byte_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_byte_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cartesian_prod_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cat_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cdist_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cdouble_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_ceil_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_ceil_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_ceil_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cfloat_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cfloat_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cfloat_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cfloat_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_chalf_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_chalf_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_char_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_char_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_char_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cholesky_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cholesky_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cholesky_inverse_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cholesky_solve_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_chunk_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_clamp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_clamp_max_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_clamp_min_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_clamp_min_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_clamp_min_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_clone_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_clone_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_clone_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_column_stack_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_column_stack_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_column_stack_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_column_stack_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_combinations_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_conj_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_conj_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_conj_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_conj_physical_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_conj_physical_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_conj_physical_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_conj_physical_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_conj_physical_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_constant_pad_nd_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_constant_pad_nd_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_constant_pad_nd_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_contiguous_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_contiguous_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_contiguous_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_copysign_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_copysign_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_copysign_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_copysign_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_corrcoef_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_corrcoef_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cos_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_count_nonzero_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_count_nonzero_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cov_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cross_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cross_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cross_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cummax_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cummax_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cummax_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cummax_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cummin_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cumsum_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cumsum_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cumulative_trapezoid_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_cumulative_trapezoid_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_deg2rad_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_deg2rad_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_deg2rad_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_deg2rad_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diag_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diag_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diag_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diag_embed_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diag_embed_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diagflat_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diagflat_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diagflat_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diagonal_copy_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diagonal_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diagonal_copy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diagonal_copy_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diagonal_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diagonal_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diagonal_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diagonal_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diagonal_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diagonal_scatter_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diagonal_scatter_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_diff_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_digamma_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_div_floor_rounding_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_div_floor_rounding_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_div_floor_rounding_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_div_trunc_rounding_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_dot_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_dot_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_double_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_double_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_dstack_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_dstack_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_dstack_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_empty_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_empty_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_empty_like_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_empty_like_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_empty_like_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_empty_like_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_empty_permuted_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_empty_permuted_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_empty_strided_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_empty_strided_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_empty_strided_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_eq_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_equal_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_equal_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_equal_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_erf_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_erf_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_erfc_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_erfinv_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_erfinv_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_exp2_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_exp2_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_exp2_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_exp_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_expand_as_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_expand_copy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_expand_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_expand_copy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_expand_copy_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_expand_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_expand_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_expm1_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_expm1_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_expm1_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_eye_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_eye_cuda_float8_e4m3fn, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_eye_cuda_float8_e4m3fnuz, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_eye_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_fft2_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_fft2_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_fft2_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_fft2_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_fft_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_fftn_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_fftn_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_fftshift_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_hfft2_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_hfft2_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_hfft_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_hfft_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_hfft_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_hfft_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_hfftn_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_hfftn_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_hfftn_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_ifft2_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_ifft2_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_ifft2_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_ifft_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_ifft_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_ifft_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_ifftn_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_ifftn_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_ifftn_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_ifftshift_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_ifftshift_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_ifftshift_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_ihfft2_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_ihfft2_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_ihfft2_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_ihfft_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_ihfftn_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_irfft2_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_irfft2_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_irfft2_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_irfft2_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_irfft2_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_irfft_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_irfft_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_irfft_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_irfftn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_irfftn_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_irfftn_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_rfft2_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_rfft2_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_rfft_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_rfft_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fft_rfftn_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fill_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_flatten_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_flatten_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_flip_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_flip_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_flip_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_flip_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_flip_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fliplr_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_flipud_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_flipud_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_flipud_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_flipud_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_float_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_float_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_float_power_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_float_power_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_floor_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_floor_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_floor_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_floor_divide_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_floor_divide_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fmax_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fmax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fmin_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fmin_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fmod_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_fmod_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_frexp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_full_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_full_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_gather_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_gcd_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_ge_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_geometric_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_geometric_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_geqrf_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_geqrf_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_gradient_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_gradient_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_gradient_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_grid_sampler_2d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_grid_sampler_2d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_gt_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_gt_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_half_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_half_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_half_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_half_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_hash_tensor_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_hash_tensor_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_heaviside_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_heaviside_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_histc_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_hsplit_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_hstack_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_hstack_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_hstack_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_i0_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_i0_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_i0_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_igammac_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_index_add_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_index_add_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_index_add_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_index_copy_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_index_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_index_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_index_fill_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_index_reduce_amax_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_index_reduce_amax_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_index_reduce_amin_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_index_reduce_mean_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_index_reduce_mean_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_index_reduce_mean_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_index_reduce_prod_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_index_reduce_prod_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_index_select_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_index_select_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_index_select_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_index_select_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_inner_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_inner_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_int_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_int_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_int_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_isclose_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_isclose_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_isfinite_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_isfinite_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_isfinite_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_isin_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_isin_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_isinf_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_isinf_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_isinf_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_isneginf_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_isposinf_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_isposinf_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_isreal_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_isreal_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_item_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_item_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_jiterator_2inputs_2outputs_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_jiterator_2inputs_2outputs_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_jiterator_4inputs_with_extra_args_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_jiterator_4inputs_with_extra_args_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_jiterator_4inputs_with_extra_args_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_jiterator_binary_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_jiterator_binary_return_by_ref_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_jiterator_binary_return_by_ref_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_jiterator_binary_return_by_ref_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_jiterator_binary_return_by_ref_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_jiterator_binary_return_by_ref_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_jiterator_unary_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_jiterator_unary_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_kron_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_kron_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_kthvalue_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_kthvalue_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_le_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_le_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_le_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_le_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_lgamma_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_lgamma_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_lgamma_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_cholesky_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_cholesky_ex_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_cross_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_cross_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_cross_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_det_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_diagonal_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_diagonal_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_diagonal_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_diagonal_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_diagonal_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_diagonal_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_eig_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_eigh_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_eigh_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_eigvals_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_eigvals_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_eigvalsh_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_householder_product_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_inv_ex_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_ldl_factor_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_lstsq_grad_oriented_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_lu_factor_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_lu_solve_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_lu_solve_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_matrix_norm_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_matrix_power_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_matrix_rank_hermitian_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_matrix_rank_hermitian_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_multi_dot_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_multi_dot_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_norm_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_norm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_norm_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_norm_subgradients_at_zero_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_pinv_hermitian_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_pinv_hermitian_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_pinv_singular_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_qr_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_qr_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_svdvals_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_tensorinv_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_tensorsolve_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_tensorsolve_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_vander_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_vander_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_vander_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_vecdot_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linalg_vector_norm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linspace_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linspace_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linspace_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linspace_tensor_overload_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linspace_tensor_overload_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_linspace_tensor_overload_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_log10_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_log1p_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_log1p_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_log2_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_log_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_log_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_log_normal_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_log_softmax_with_dtype_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_log_softmax_with_dtype_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_log_softmax_with_dtype_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logaddexp_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logcumsumexp_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logcumsumexp_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logdet_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logdet_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logical_not_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logical_not_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logical_not_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logical_or_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logical_or_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logical_or_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logical_xor_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logit_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logit_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logspace_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logspace_tensor_overload_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logspace_tensor_overload_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logspace_tensor_overload_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logspace_tensor_overload_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logspace_tensor_overload_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logsumexp_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_logsumexp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_long_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_long_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_long_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_long_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_long_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_lt_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_lt_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_lt_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_lt_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_lt_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_lu_solve_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_lu_unpack_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_lu_unpack_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mH_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mH_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mH_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mH_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mT_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mT_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mT_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mT_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_amax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_amax_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_amin_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_amin_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_argmin_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_argmin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_cumprod_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_cumprod_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_cumprod_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_cumprod_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_cumprod_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_cumsum_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_fill_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_fill_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_log_softmax_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_logsumexp_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_logsumexp_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_mean_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_normalize_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_normalize_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_normalize_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_prod_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_prod_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_scatter_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_scatter_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_select_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_softmin_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_sum_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_sum_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_var_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_var_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_masked_var_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_matmul_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_matmul_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_matrix_exp_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_max_binary_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_max_pool2d_with_indices_backward_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_max_reduction_no_dim_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_max_reduction_no_dim_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_max_reduction_no_dim_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_max_reduction_with_dim_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_maximum_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_maximum_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_maximum_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mean_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_median_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_meshgrid_list_of_tensors_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_meshgrid_list_of_tensors_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_meshgrid_list_of_tensors_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_meshgrid_list_of_tensors_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_meshgrid_variadic_tensors_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_min_binary_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_min_reduction_no_dim_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_min_reduction_with_dim_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_minimum_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_minimum_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_minimum_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mm_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mode_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_movedim_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_msort_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mul_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mul_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mul_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mul_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mv_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mvlgamma_mvlgamma_p_1_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mvlgamma_mvlgamma_p_1_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mvlgamma_mvlgamma_p_1_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mvlgamma_mvlgamma_p_3_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mvlgamma_mvlgamma_p_5_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mvlgamma_mvlgamma_p_5_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mvlgamma_mvlgamma_p_5_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mvlgamma_mvlgamma_p_5_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_mvlgamma_mvlgamma_p_5_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nan_to_num_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nanmean_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nanmedian_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nansum_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_narrow_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_native_dropout_backward_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_native_dropout_backward_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_native_dropout_backward_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_ne_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_neg_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_neg_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_neg_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_neg_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_new_empty_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_new_empty_strided_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_new_empty_strided_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_new_empty_strided_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_new_empty_strided_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_new_empty_strided_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_new_full_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_new_full_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_new_full_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_new_full_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_new_full_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_new_full_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_new_full_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_new_ones_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_new_zeros_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_new_zeros_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nextafter_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nextafter_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_adaptive_avg_pool3d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_adaptive_avg_pool3d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_adaptive_max_pool1d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_adaptive_max_pool3d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_alpha_dropout_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_avg_pool1d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_avg_pool1d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_avg_pool2d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_avg_pool3d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_avg_pool3d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_batch_norm_without_cudnn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_bilinear_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_binary_cross_entropy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_binary_cross_entropy_with_logits_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_celu_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_celu_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_channel_shuffle_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_channel_shuffle_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_channel_shuffle_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_channel_shuffle_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_conv1d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_conv1d_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_conv2d_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_conv3d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_conv3d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_conv_transpose1d_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_conv_transpose1d_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_conv_transpose1d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_conv_transpose1d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_conv_transpose2d_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_conv_transpose3d_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_conv_transpose3d_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_cosine_embedding_loss_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_cosine_similarity_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_ctc_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_dropout3d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_dropout_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_embedding_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_embedding_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_feature_alpha_dropout_with_train_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_feature_alpha_dropout_with_train_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_feature_alpha_dropout_without_train_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_feature_alpha_dropout_without_train_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_feature_alpha_dropout_without_train_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_fractional_max_pool3d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_fractional_max_pool3d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_gaussian_nll_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_gaussian_nll_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_gelu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_glu_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_grid_sample_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_group_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_hardshrink_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_hardsigmoid_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_hardswish_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_hardtanh_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_hinge_embedding_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_huber_loss_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_huber_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_interpolate_bicubic_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_interpolate_bicubic_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_interpolate_bilinear_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_interpolate_bilinear_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_interpolate_bilinear_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_interpolate_nearest-exact_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_interpolate_trilinear_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_interpolate_trilinear_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_l1_loss_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_leaky_relu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_linear_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_local_response_norm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_logsigmoid_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_margin_ranking_loss_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_max_pool1d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_max_pool1d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_max_unpool1d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_max_unpool2d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_max_unpool3d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_max_unpool3d_grad_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_mish_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_multi_margin_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_multilabel_margin_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_multilabel_soft_margin_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_multilabel_soft_margin_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_nll_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_normalize_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pad_circular_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pad_circular_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pad_circular_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pad_circular_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pad_constant_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pad_constant_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pad_constant_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pad_constant_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pad_reflect_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pad_reflect_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pad_reflect_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pad_replicate_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pad_replicate_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pad_replicate_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pad_replicate_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pad_replicate_negative_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pairwise_distance_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pairwise_distance_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pairwise_distance_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pdist_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pixel_shuffle_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_pixel_unshuffle_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_poisson_nll_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_poisson_nll_loss_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_prelu_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_relu6_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_relu6_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_relu_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_relu_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_relu_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_rms_norm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_rms_norm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_rms_norm_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_rrelu_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_selu_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_soft_margin_loss_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_softshrink_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_softsign_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_softsign_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_tanhshrink_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_tanhshrink_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_threshold_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_threshold_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_triplet_margin_loss_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_triplet_margin_loss_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_triplet_margin_with_distance_loss_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_triplet_margin_with_distance_loss_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_unfold_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_upsample_nearest_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nn_functional_upsample_nearest_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nonzero_static_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nonzero_static_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nonzero_static_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_nonzero_static_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_norm_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_norm_fro_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_norm_inf_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_norm_inf_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_norm_nuc_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_normal_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_normal_in_place_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_normal_in_place_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_ones_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_ones_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_ones_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_ones_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_ones_like_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_ones_like_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_ormqr_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_outer_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_outer_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_outer_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_pca_lowrank_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_permute_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_permute_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_permute_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_permute_copy_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_pinverse_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_pinverse_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_pinverse_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_polar_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_polygamma_polygamma_n_0_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_polygamma_polygamma_n_0_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_polygamma_polygamma_n_0_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_polygamma_polygamma_n_0_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_polygamma_polygamma_n_1_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_polygamma_polygamma_n_1_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_polygamma_polygamma_n_2_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_polygamma_polygamma_n_2_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_polygamma_polygamma_n_3_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_polygamma_polygamma_n_3_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_polygamma_polygamma_n_3_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_polygamma_polygamma_n_4_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_positive_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_positive_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_pow_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_prod_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_prod_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_prod_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_prod_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_prod_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_rad2deg_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_rad2deg_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_rad2deg_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_rand_like_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_rand_like_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_randint_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_randint_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_randint_like_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_randint_like_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_randint_like_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_randn_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_randn_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_randn_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_randn_like_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_randn_like_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_real_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_real_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_real_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_reciprocal_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_reciprocal_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_remainder_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_renorm_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_repeat_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_repeat_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_repeat_interleave_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_reshape_as_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_reshape_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_reshape_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_reshape_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_reshape_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_resize_as__cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_resize_as__cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_resolve_neg_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_resolve_neg_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_resolve_neg_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_roll_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_roll_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_roll_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_rot90_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_rot90_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_rot90_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_rot90_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_round_decimals_0_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_rsqrt_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_rsqrt_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_rsqrt_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_rsqrt_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_rsqrt_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_rsub_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_rsub_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_scalar_tensor_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_scalar_tensor_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_scalar_tensor_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_scatter_add_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_scatter_add_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_scatter_add_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_scatter_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_scatter_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_scatter_reduce_amax_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_scatter_reduce_prod_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_searchsorted_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_searchsorted_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_select_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_select_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_select_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_select_scatter_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_select_scatter_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sgn_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_short_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sigmoid_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sigmoid_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sigmoid_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sigmoid_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sign_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sign_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sign_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sign_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_signal_windows_blackman_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_signal_windows_cosine_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_signal_windows_hann_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sin_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sin_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sin_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sin_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sin_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sinc_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sinc_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sinh_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sinh_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sinh_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sinh_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_slice_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_slice_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_slice_scatter_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_softmax_with_dtype_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_softmax_with_dtype_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_softmax_with_dtype_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_softmax_with_dtype_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sort_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sort_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sort_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sort_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sparse_mm_reduce_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sparse_sampled_addmm_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sparse_sampled_addmm_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_bessel_j0_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_bessel_j0_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_bessel_j0_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_bessel_j1_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_bessel_y0_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_bessel_y0_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_bessel_y0_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_bessel_y1_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_chebyshev_polynomial_u_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_chebyshev_polynomial_w_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_entr_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_erfcx_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_hermite_polynomial_h_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_hermite_polynomial_h_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_i0e_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_i0e_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_i0e_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_i1_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_i1e_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_i1e_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_i1e_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_laguerre_polynomial_l_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_laguerre_polynomial_l_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_laguerre_polynomial_l_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_legendre_polynomial_p_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_log_ndtr_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_modified_bessel_i0_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_modified_bessel_i0_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_modified_bessel_i0_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_modified_bessel_i0_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_modified_bessel_i1_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_modified_bessel_k0_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_modified_bessel_k0_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_modified_bessel_k0_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_modified_bessel_k1_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_ndtr_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_ndtr_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_ndtr_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_polygamma_special_polygamma_n_0_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_scaled_modified_bessel_k0_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_scaled_modified_bessel_k1_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_scaled_modified_bessel_k1_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_scaled_modified_bessel_k1_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_shifted_chebyshev_polynomial_t_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_shifted_chebyshev_polynomial_t_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_shifted_chebyshev_polynomial_t_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_shifted_chebyshev_polynomial_u_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_shifted_chebyshev_polynomial_u_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_shifted_chebyshev_polynomial_v_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_shifted_chebyshev_polynomial_v_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_shifted_chebyshev_polynomial_w_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_spherical_bessel_j0_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_xlog1py_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_special_zeta_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_split_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_split_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_split_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_split_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_split_list_args_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_split_list_args_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_split_list_args_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_split_list_args_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_split_with_sizes_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_split_with_sizes_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_split_with_sizes_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sqrt_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_squeeze_copy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_squeeze_copy_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_squeeze_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_squeeze_copy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_squeeze_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_squeeze_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_squeeze_multiple_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_stack_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_stack_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_stack_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_stack_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_std_mean_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_std_mean_unbiased_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sub_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sub_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sum_to_size_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sum_to_size_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_sum_to_size_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_svd_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_svd_lowrank_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_svd_lowrank_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_t_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_t_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_t_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_take_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_take_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_take_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_tan_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_tan_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_tanh_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_tanh_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_tanh_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_tensor_split_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_tensor_split_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_tensor_split_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_tensor_split_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_tile_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_tile_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_tile_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_tile_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_to_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_to_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_to_sparse_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_to_sparse_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_to_sparse_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_torch__scaled_mm_cuda_float8_e4m3fn, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_torch_ops_aten__efficient_attention_forward_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_torch_ops_aten__flash_attention_forward_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_torch_ops_aten__safe_softmax_default_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_trace_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_trace_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_trace_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_trace_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_transpose_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_transpose_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_transpose_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_transpose_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_trapezoid_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_trapezoid_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_trapezoid_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_trapz_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_triangular_solve_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_tril_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_tril_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_tril_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_tril_indices_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_triu_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_triu_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_triu_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_true_divide_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_true_divide_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_true_divide_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_true_divide_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_true_divide_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_true_divide_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_trunc_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_trunc_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_trunc_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_trunc_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unbind_copy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unbind_copy_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unbind_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unbind_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unbind_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unbind_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unflatten_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unflatten_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unflatten_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unfold_copy_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unfold_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_uniform_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unique_consecutive_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unique_consecutive_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unique_consecutive_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unique_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unique_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unsafe_chunk_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unsafe_chunk_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unsafe_chunk_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unsafe_split_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unsqueeze_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unsqueeze_copy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unsqueeze_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unsqueeze_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unsqueeze_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unsqueeze_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_unsqueeze_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_var_mean_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_var_mean_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_var_mean_unbiased_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_var_unbiased_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_var_unbiased_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_var_unbiased_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_var_unbiased_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_vdot_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_view_as_complex_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_view_as_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_view_as_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_view_as_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_view_copy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_view_copy_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_view_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_view_copy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_view_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_view_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_vsplit_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_vsplit_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_vstack_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_vstack_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_where_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_where_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_where_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_xlogy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_xlogy_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_zero__cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_zero__cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_zero__cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_zero__cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_zeros_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_zeros_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_zeros_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_meta_outplace_zeros_like_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_H_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_H_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_T_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_T_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_T_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___getitem___cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___getitem___cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___getitem___cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___radd___cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___radd___cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___radd___cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___rand___cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___rdiv___cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___rdiv___cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___rdiv___cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___rdiv___cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___rdiv___cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___rmod___cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___rmod___cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___rmod___cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___rmul___cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___rmul___cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___rpow___cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___rpow___cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___rpow___cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___rpow___cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___rsub___cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace___rsub___cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__chunk_cat_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__chunk_cat_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__chunk_cat_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__chunk_cat_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__chunk_cat_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_abs_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_acos_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_acos_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_acos_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_acos_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_acos_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_acos_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_add_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_add_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_addcdiv_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_addcdiv_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_addcdiv_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_addcdiv_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_addcmul_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_atan_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_atan_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_atan_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_ceil_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_ceil_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_ceil_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_ceil_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_clamp_max_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_clamp_max_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_clamp_min_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_clamp_min_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_clamp_min_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_clamp_min_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_copy_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_copy_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_cos_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_cos_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_cos_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_cosh_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_cosh_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_cosh_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_cosh_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_div_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_div_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_erf_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_erfc_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_erfc_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_exp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_expm1_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_expm1_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_floor_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_floor_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_floor_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_frac_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_lerp_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_lerp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_lerp_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_lerp_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_log10_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_log1p_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_log1p_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_log1p_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_log1p_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_log1p_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_log2_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_log2_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_log_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_log_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_log_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_max_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_max_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_max_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_max_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_maximum_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_maximum_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_maximum_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_maximum_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_minimum_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_minimum_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_minimum_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_mul_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_mul_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_neg_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_neg_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_neg_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_norm_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_norm_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_pow_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_pow_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_pow_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_reciprocal_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_reciprocal_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_reciprocal_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_reciprocal_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_round_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_round_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_round_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_rsqrt_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_sigmoid_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_sigmoid_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_sigmoid_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_sign_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_sign_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_sin_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_sin_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_sinh_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_sqrt_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_sqrt_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_sqrt_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_sqrt_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_sqrt_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_sub_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_sub_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_sub_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_sub_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_sub_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_tan_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_tan_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_tanh_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_zero_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__foreach_zero_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__native_batch_norm_legit_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__native_batch_norm_legit_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__segment_reduce_lengths_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__unsafe_masked_index_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__unsafe_masked_index_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__unsafe_masked_index_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace__unsafe_masked_index_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_abs_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_abs_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_abs_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_acosh_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_add_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_add_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_add_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_add_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_addcdiv_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_addcmul_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_addcmul_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_addmm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_addmm_decomposed_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_addmv_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_addmv_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_addr_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_addr_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_addr_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_alias_copy_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_alias_copy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_alias_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides___rand___cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides__chunk_cat_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides__foreach_abs_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides__foreach_acos_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides__foreach_add_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides__foreach_addcdiv_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides__foreach_ceil_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides__foreach_copy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides__foreach_cos_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides__foreach_div_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides__foreach_log2_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides__foreach_log_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides__foreach_pow_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides__foreach_sinh_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides__segment_reduce_offsets_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides__unsafe_masked_index_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_abs_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_addcdiv_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_all_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_allclose_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_amin_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_arange_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_as_strided_copy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_as_strided_partial_views_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_asin_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_atleast_1d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_char_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_column_stack_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_combinations_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_conj_physical_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_corrcoef_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_cosh_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_count_nonzero_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_cumulative_trapezoid_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_diagonal_copy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_diagonal_scatter_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_einsum_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_erf_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_exp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_expand_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_fft_hfft2_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_fft_rfftn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_fill_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_fmax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_frac_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_geqrf_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_gradient_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_heaviside_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_index_put_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_isfinite_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_isinf_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_istft_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_kron_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_ldexp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_le_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_linalg_lu_factor_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_linalg_lu_solve_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_linalg_multi_dot_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_linalg_pinv_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_linalg_pinv_hermitian_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_linalg_tensorsolve_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_linalg_vecdot_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_log_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_log_softmax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_masked_amax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_masked_cumprod_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_masked_normalize_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_max_binary_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_multinomial_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_mv_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_mvlgamma_mvlgamma_p_1_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_mvlgamma_mvlgamma_p_3_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_nanquantile_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_native_layer_norm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_new_zeros_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_nn_functional_adaptive_max_pool2d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_nn_functional_avg_pool1d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_nn_functional_binary_cross_entropy_with_logits_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_nn_functional_cross_entropy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_nn_functional_elu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_nn_functional_huber_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_nn_functional_leaky_relu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_nn_functional_multi_margin_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_nn_functional_nll_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_nn_functional_one_hot_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_nn_functional_pad_circular_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_nn_functional_pad_replicate_negative_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_nn_functional_rrelu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_nn_functional_soft_margin_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_nn_functional_softmin_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_nn_functional_triplet_margin_with_distance_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_nn_functional_upsample_bilinear_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_norm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_pca_lowrank_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_permute_copy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_polygamma_polygamma_n_4_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_ravel_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_reciprocal_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_reshape_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_resolve_neg_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_roll_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_rsub_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_scatter_add_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_signal_windows_blackman_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_signal_windows_general_cosine_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_signal_windows_general_hamming_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_signal_windows_hann_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_signal_windows_kaiser_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_sin_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_softmax_with_dtype_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_sort_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_special_hermite_polynomial_he_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_special_polygamma_special_polygamma_n_0_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_special_scaled_modified_bessel_k1_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_special_shifted_chebyshev_polynomial_v_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_special_xlog1py_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_square_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_sum_to_size_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_t_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_tan_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_torch__scaled_mm_cuda_float8_e4m3fn, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_transpose_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_trapezoid_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_trapz_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_uniform_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_unsafe_chunk_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_view_as_complex_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_view_as_real_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_vsplit_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_vstack_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_all_strides_zeros_like_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_allclose_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_amax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_amin_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_amin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_aminmax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_aminmax_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_aminmax_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_angle_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_any_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_any_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_arange_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_arange_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_argmax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_argmin_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_argmin_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_argmin_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_argmin_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_argsort_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_argsort_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_argsort_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_argwhere_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_argwhere_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_as_strided_copy_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_as_strided_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_as_strided_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_as_strided_partial_views_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_as_strided_partial_views_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_as_strided_partial_views_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_as_strided_scatter_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_as_strided_scatter_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_as_strided_scatter_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_asin_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_asinh_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_asinh_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_asinh_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_asinh_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_atan2_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_atan_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_atan_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_atanh_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_atleast_1d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_atleast_1d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_atleast_1d_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_atleast_2d_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_atleast_2d_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_atleast_3d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_bfloat16_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_bfloat16_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_bfloat16_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_bfloat16_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_bitwise_not_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_bitwise_or_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_bitwise_or_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_bitwise_xor_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_block_diag_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_block_diag_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_block_diag_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_block_diag_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_block_diag_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_block_diag_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_bmm_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_bool_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_bool_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_bool_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_broadcast_tensors_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_broadcast_tensors_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_broadcast_tensors_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_broadcast_tensors_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_bucketize_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_byte_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cartesian_prod_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cartesian_prod_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cartesian_prod_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cartesian_prod_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cartesian_prod_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cat_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cauchy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cdouble_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cdouble_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_ceil_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_ceil_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_ceil_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cfloat_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cfloat_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cfloat_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cfloat_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_chalf_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_char_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_char_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cholesky_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cholesky_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cholesky_inverse_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cholesky_solve_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_chunk_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_chunk_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_chunk_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_clamp_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_clamp_max_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_clamp_min_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_clamp_min_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_clone_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_clone_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_clone_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_column_stack_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_column_stack_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_column_stack_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_column_stack_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_column_stack_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_combinations_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_combinations_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_complex_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_complex_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_conj_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_conj_physical_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_conj_physical_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_conj_physical_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_constant_pad_nd_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_constant_pad_nd_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_constant_pad_nd_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_contiguous_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_contiguous_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_contiguous_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_copysign_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_copysign_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_corrcoef_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_corrcoef_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cos_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cos_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cos_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_count_nonzero_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_count_nonzero_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_count_nonzero_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cov_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cross_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cross_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cummax_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cummax_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cumprod_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cumprod_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cumsum_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cumsum_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cumsum_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_cumulative_trapezoid_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_deg2rad_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_deg2rad_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_deg2rad_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_deg2rad_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_diag_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_diag_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_diag_embed_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_diag_embed_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_diag_embed_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_diagflat_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_diagonal_copy_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_diagonal_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_diagonal_copy_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_diagonal_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_diagonal_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_diagonal_scatter_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_diagonal_scatter_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_diff_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_diff_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_digamma_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_div_no_rounding_mode_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_div_trunc_rounding_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_div_trunc_rounding_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_dot_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_dot_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_double_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_double_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_double_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_dsplit_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_dsplit_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_dsplit_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_dstack_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_empty_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_empty_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_empty_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_empty_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_empty_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_empty_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_empty_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_empty_permuted_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_empty_permuted_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_empty_permuted_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_empty_permuted_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_empty_strided_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_empty_strided_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_equal_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_equal_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_erfc_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_erfc_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_erfinv_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_erfinv_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_erfinv_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_exp2_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_exp_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_exp_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_exp_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_exp_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_expand_as_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_expand_as_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_expand_as_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_expand_as_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_expand_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_expm1_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_eye_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_eye_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_eye_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_fft2_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_fft2_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_fft2_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_fft2_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_fft_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_fft_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_fftshift_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_fftshift_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_fftshift_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_fftshift_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_fftshift_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_hfft2_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_hfft2_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_hfft_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_hfft_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_hfft_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_ifft2_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_ifft_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_ifft_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_ifft_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_ifftn_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_ifftn_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_ifftshift_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_ifftshift_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_ifftshift_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_ifftshift_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_ihfft2_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_ihfft_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_ihfftn_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_irfft2_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_irfft2_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_irfft2_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_irfft2_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_irfft_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_irfft_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_irfft_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_irfftn_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_irfftn_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_rfft2_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_rfft2_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_rfft2_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_rfft_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_rfft_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_rfft_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_rfftn_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_rfftn_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_rfftn_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_rfftn_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fft_rfftn_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fill_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_flatten_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_flatten_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_flip_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_flip_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fliplr_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fliplr_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fliplr_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_float_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_float_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_float_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_float_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_float_power_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_float_power_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_floor_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_floor_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_floor_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_floor_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_floor_divide_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_floor_divide_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_floor_divide_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_floor_divide_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fmax_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fmin_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fmin_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fmin_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fmin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fmod_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_fmod_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_full_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_full_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_full_like_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_full_like_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_full_like_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_gather_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_geometric_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_gradient_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_gradient_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_gt_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_half_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_hash_tensor_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_hash_tensor_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_hash_tensor_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_heaviside_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_hsplit_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_hstack_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_hstack_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_hstack_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_hstack_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_hypot_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_i0_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_i0_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_igammac_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_imag_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_add_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_add_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_add_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_add_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_copy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_copy_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_copy_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_copy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_copy_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_fill_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_fill_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_fill_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_fill_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_put_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_reduce_amax_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_reduce_amax_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_reduce_amin_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_reduce_amin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_reduce_mean_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_reduce_prod_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_reduce_prod_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_select_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_select_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_select_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_select_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_index_select_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_inner_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_inner_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_int_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_int_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_int_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_isclose_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_isclose_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_isclose_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_isclose_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_isfinite_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_isfinite_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_isinf_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_isnan_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_isneginf_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_isneginf_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_isposinf_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_isposinf_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_isposinf_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_isreal_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_isreal_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_isreal_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_isreal_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_item_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_item_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_item_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_item_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_item_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_item_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_jiterator_2inputs_2outputs_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_jiterator_4inputs_with_extra_args_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_jiterator_binary_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_jiterator_binary_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_jiterator_binary_return_by_ref_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_jiterator_binary_return_by_ref_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_jiterator_unary_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_jiterator_unary_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_kron_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_kron_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_kthvalue_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_lcm_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_ldexp_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_ldexp_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_le_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_le_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_lgamma_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_lgamma_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_lgamma_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_cholesky_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_cross_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_cross_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_cross_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_det_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_diagonal_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_diagonal_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_eig_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_eigh_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_eigh_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_eigvals_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_eigvals_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_inv_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_inv_ex_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_ldl_factor_ex_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_ldl_factor_ex_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_lstsq_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_lstsq_grad_oriented_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_lstsq_grad_oriented_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_lu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_lu_factor_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_lu_factor_ex_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_lu_solve_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_matrix_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_matrix_norm_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_matrix_power_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_matrix_rank_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_matrix_rank_hermitian_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_multi_dot_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_norm_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_norm_subgradients_at_zero_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_pinv_hermitian_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_pinv_singular_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_qr_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_qr_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_solve_triangular_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_svd_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_tensorsolve_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_vander_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linalg_vecdot_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linspace_tensor_overload_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linspace_tensor_overload_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_linspace_tensor_overload_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_log10_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_log10_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_log1p_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_log1p_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_log2_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_log_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_log_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_log_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_log_softmax_with_dtype_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logaddexp_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logcumsumexp_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logcumsumexp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logical_and_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logical_not_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logical_not_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logical_not_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logical_or_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logical_or_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logical_or_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logical_xor_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logical_xor_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logical_xor_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logit_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logit_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logspace_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logspace_tensor_overload_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logspace_tensor_overload_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logspace_tensor_overload_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_logsumexp_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_long_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_long_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_long_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_long_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_lt_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_lt_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_lt_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_lu_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_lu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_lu_solve_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_lu_solve_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_lu_solve_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_lu_unpack_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_mH_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_mT_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_mT_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_mT_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_amax_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_amax_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_argmax_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_argmin_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_argmin_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_argmin_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_cumprod_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_cumsum_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_cumsum_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_cumsum_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_log_softmax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_logaddexp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_logsumexp_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_logsumexp_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_logsumexp_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_logsumexp_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_median_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_normalize_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_prod_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_scatter_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_scatter_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_select_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_softmax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_softmin_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_std_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_std_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_sum_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_var_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_var_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_var_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_masked_var_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_matmul_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_matmul_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_matrix_exp_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_max_binary_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_max_pool2d_with_indices_backward_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_max_reduction_no_dim_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_max_reduction_no_dim_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_max_reduction_no_dim_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_max_reduction_with_dim_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_maximum_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_maximum_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_maximum_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_median_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_median_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_meshgrid_list_of_tensors_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_meshgrid_variadic_tensors_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_meshgrid_variadic_tensors_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_meshgrid_variadic_tensors_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_meshgrid_variadic_tensors_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_min_reduction_no_dim_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_min_reduction_with_dim_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_min_reduction_with_dim_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_minimum_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_minimum_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_mm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_mode_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_mode_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_movedim_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_msort_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_msort_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_msort_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_mul_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_mul_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_multinomial_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_mv_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_mvlgamma_mvlgamma_p_1_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_mvlgamma_mvlgamma_p_1_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_mvlgamma_mvlgamma_p_3_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_mvlgamma_mvlgamma_p_3_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_mvlgamma_mvlgamma_p_3_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_mvlgamma_mvlgamma_p_3_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_mvlgamma_mvlgamma_p_5_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_mvlgamma_mvlgamma_p_5_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_mvlgamma_mvlgamma_p_5_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nan_to_num_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nan_to_num_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nanmean_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nanmean_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nanmean_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nanmean_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nanmedian_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nanmedian_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nanmedian_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nanmedian_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nanquantile_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nansum_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nansum_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nansum_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_narrow_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_narrow_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_narrow_copy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_narrow_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_narrow_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_narrow_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_narrow_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_narrow_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_native_batch_norm_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_native_layer_norm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_ne_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_ne_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_ne_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_neg_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_neg_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_new_empty_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_new_empty_strided_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_new_full_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_new_ones_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_new_ones_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_new_ones_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_new_zeros_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_new_zeros_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nextafter_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nextafter_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_adaptive_avg_pool1d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_adaptive_avg_pool2d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_adaptive_avg_pool2d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_adaptive_avg_pool2d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_adaptive_avg_pool3d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_adaptive_max_pool1d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_adaptive_max_pool2d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_adaptive_max_pool3d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_avg_pool1d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_avg_pool1d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_batch_norm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_bilinear_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_bilinear_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_binary_cross_entropy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_binary_cross_entropy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_celu_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_channel_shuffle_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_channel_shuffle_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_channel_shuffle_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_channel_shuffle_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_channel_shuffle_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_conv1d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_conv1d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_conv3d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_conv_transpose1d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_conv_transpose2d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_conv_transpose2d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_conv_transpose3d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_conv_transpose3d_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_conv_transpose3d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_cosine_embedding_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_cosine_embedding_loss_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_cosine_embedding_loss_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_cosine_embedding_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_cosine_embedding_loss_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_cosine_similarity_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_ctc_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_dropout2d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_dropout2d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_dropout3d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_dropout_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_feature_alpha_dropout_with_train_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_feature_alpha_dropout_without_train_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_feature_alpha_dropout_without_train_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_feature_alpha_dropout_without_train_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_feature_alpha_dropout_without_train_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_fractional_max_pool3d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_fractional_max_pool3d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_gelu_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_gelu_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_grid_sample_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_group_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_hardshrink_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_hardswish_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_hardswish_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_hardtanh_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_hardtanh_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_hinge_embedding_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_instance_norm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_interpolate_bilinear_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_interpolate_nearest_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_interpolate_trilinear_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_interpolate_trilinear_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_linear_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_linear_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_local_response_norm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_logsigmoid_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_logsigmoid_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_margin_ranking_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_margin_ranking_loss_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_max_pool3d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_max_unpool1d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_max_unpool1d_grad_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_max_unpool2d_grad_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_max_unpool3d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_mish_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_mse_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_mse_loss_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_multi_head_attention_forward_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_multi_head_attention_forward_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_multi_head_attention_forward_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_multi_margin_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_multi_margin_loss_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_multi_margin_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_nll_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_pad_circular_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_pad_constant_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_pad_constant_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_pad_reflect_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_pad_reflect_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_pad_replicate_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_pad_replicate_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_pad_replicate_negative_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_pad_replicate_negative_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_pad_replicate_negative_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_pad_replicate_negative_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_pairwise_distance_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_pairwise_distance_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_pairwise_distance_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_pairwise_distance_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_pdist_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_pixel_shuffle_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_pixel_unshuffle_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_prelu_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_prelu_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_prelu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_prelu_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_relu_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_relu_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_rms_norm_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_rms_norm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_rrelu_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_scaled_dot_product_attention_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_silu_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_silu_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_smooth_l1_loss_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_soft_margin_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_softmin_with_dtype_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_softshrink_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_softsign_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_softsign_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_softsign_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_softsign_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_softsign_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_tanhshrink_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_tanhshrink_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_threshold_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_threshold_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_triplet_margin_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_triplet_margin_loss_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_triplet_margin_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_triplet_margin_loss_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_triplet_margin_loss_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_triplet_margin_with_distance_loss_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_triplet_margin_with_distance_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_triplet_margin_with_distance_loss_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_unfold_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_unfold_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_unfold_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_upsample_nearest_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nn_functional_upsample_nearest_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nonzero_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nonzero_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nonzero_static_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_nonzero_static_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_norm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_norm_fro_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_norm_fro_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_norm_inf_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_norm_inf_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_norm_inf_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_norm_nuc_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_normal_in_place_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_normal_in_place_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_normal_number_mean_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_ones_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_ones_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_ones_like_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_ones_like_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_ones_like_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_ones_like_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_ormqr_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_outer_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_outer_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_outer_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_outer_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_pca_lowrank_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_permute_copy_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_permute_copy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_permute_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_permute_copy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_permute_copy_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_permute_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_permute_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_permute_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_polygamma_polygamma_n_0_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_polygamma_polygamma_n_0_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_polygamma_polygamma_n_1_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_polygamma_polygamma_n_1_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_polygamma_polygamma_n_2_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_polygamma_polygamma_n_3_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_polygamma_polygamma_n_3_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_polygamma_polygamma_n_4_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_polygamma_polygamma_n_4_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_positive_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_prod_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_prod_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_prod_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_prod_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_prod_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_put_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_put_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_put_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_quantile_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_rad2deg_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_rad2deg_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_rad2deg_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_randint_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_randint_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_randint_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_randint_like_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_randn_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_randn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_randn_like_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_randn_like_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_randn_like_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_randn_like_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_ravel_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_ravel_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_ravel_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_ravel_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_real_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_real_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_reciprocal_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_reciprocal_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_remainder_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_remainder_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_remainder_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_remainder_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_repeat_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_repeat_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_reshape_as_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_reshape_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_reshape_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_reshape_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_resize__cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_resize__cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_resize__cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_resize__cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_resize_as__cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_resolve_conj_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_resolve_neg_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_roll_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_roll_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_rot90_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_rot90_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_rot90_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_round_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_round_decimals_3_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_rsqrt_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_rsqrt_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_rsqrt_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_rsub_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scalar_tensor_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scalar_tensor_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scalar_tensor_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scalar_tensor_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scatter_add_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scatter_add_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scatter_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scatter_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scatter_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scatter_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scatter_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scatter_reduce_amax_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scatter_reduce_amax_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scatter_reduce_amin_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scatter_reduce_amin_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scatter_reduce_mean_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scatter_reduce_mean_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scatter_reduce_prod_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scatter_reduce_prod_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_scatter_reduce_sum_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_searchsorted_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_searchsorted_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_searchsorted_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_searchsorted_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_select_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_select_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_select_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_select_scatter_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_select_scatter_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sgn_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sgn_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_short_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sigmoid_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sigmoid_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sigmoid_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sigmoid_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sign_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sign_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_signal_windows_bartlett_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_signal_windows_gaussian_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_signal_windows_hann_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_signbit_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_signbit_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_signbit_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_signbit_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sin_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sin_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sin_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sinc_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sinc_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sinc_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sinc_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sinc_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sinc_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sinh_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sinh_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sinh_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_slice_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_slice_scatter_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_softmax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sort_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sort_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sparse_mm_reduce_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sparse_mm_reduce_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sparse_sampled_addmm_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sparse_sampled_addmm_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_airy_ai_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_airy_ai_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_airy_ai_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_bessel_j0_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_bessel_j0_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_bessel_j0_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_bessel_j1_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_bessel_y1_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_bessel_y1_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_chebyshev_polynomial_t_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_chebyshev_polynomial_t_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_chebyshev_polynomial_t_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_chebyshev_polynomial_u_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_chebyshev_polynomial_u_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_chebyshev_polynomial_u_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_chebyshev_polynomial_v_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_chebyshev_polynomial_v_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_chebyshev_polynomial_v_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_chebyshev_polynomial_v_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_chebyshev_polynomial_v_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_chebyshev_polynomial_w_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_erfcx_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_erfcx_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_hermite_polynomial_h_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_hermite_polynomial_h_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_hermite_polynomial_h_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_hermite_polynomial_he_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_hermite_polynomial_he_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_laguerre_polynomial_l_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_legendre_polynomial_p_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_log_ndtr_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_log_ndtr_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_modified_bessel_i1_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_modified_bessel_k0_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_modified_bessel_k0_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_modified_bessel_k1_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_modified_bessel_k1_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_modified_bessel_k1_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_modified_bessel_k1_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_ndtri_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_ndtri_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_ndtri_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_polygamma_special_polygamma_n_0_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_polygamma_special_polygamma_n_0_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_polygamma_special_polygamma_n_0_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_polygamma_special_polygamma_n_0_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_scaled_modified_bessel_k1_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_shifted_chebyshev_polynomial_u_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_shifted_chebyshev_polynomial_v_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_shifted_chebyshev_polynomial_v_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_shifted_chebyshev_polynomial_w_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_shifted_chebyshev_polynomial_w_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_shifted_chebyshev_polynomial_w_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_spherical_bessel_j0_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_spherical_bessel_j0_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_special_spherical_bessel_j0_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_split_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_split_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_split_list_args_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_split_list_args_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_split_list_args_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_split_list_args_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_split_with_sizes_copy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_split_with_sizes_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_split_with_sizes_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_split_with_sizes_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sqrt_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_square_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_squeeze_copy_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_squeeze_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_squeeze_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_squeeze_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_squeeze_multiple_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_squeeze_multiple_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_stack_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_stack_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_stack_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_std_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_std_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_std_mean_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_std_mean_unbiased_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_std_unbiased_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_std_unbiased_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_std_unbiased_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sub_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sub_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sub_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sub_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sub_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sub_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sub_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sum_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sum_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sum_to_size_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sum_to_size_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sum_to_size_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sum_to_size_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_sum_to_size_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_t_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_t_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_t_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_t_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_take_along_dim_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_take_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_take_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_take_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_take_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_take_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_tanh_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_tanh_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_tanh_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_tanh_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_tensor_split_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_tensor_split_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_tensor_split_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_tensordot_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_tensordot_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_tensordot_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_tile_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_tile_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_to_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_to_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_to_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_to_sparse_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_to_sparse_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_topk_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_torch__scaled_mm_cuda_float8_e4m3fn, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_torch_ops_aten__safe_softmax_default_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_torch_ops_aten__safe_softmax_default_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_transpose_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_transpose_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_trapezoid_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_trapezoid_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_trapz_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_trapz_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_triangular_solve_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_tril_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_tril_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_tril_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_tril_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_triu_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_triu_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_triu_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_triu_indices_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_true_divide_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_true_divide_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unbind_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unbind_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unbind_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unbind_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unbind_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unbind_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unflatten_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unflatten_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unflatten_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unflatten_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unfold_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unfold_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unfold_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unfold_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unfold_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unfold_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unfold_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unique_consecutive_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unique_consecutive_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unique_consecutive_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unique_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unique_cuda_uint32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unique_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unravel_index_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unsafe_chunk_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unsafe_chunk_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unsafe_split_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unsafe_split_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unsafe_split_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unsqueeze_copy_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unsqueeze_copy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unsqueeze_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unsqueeze_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_unsqueeze_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_var_mean_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_var_mean_unbiased_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_var_mean_unbiased_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_var_mean_unbiased_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_vdot_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_view_as_complex_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_view_as_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_view_as_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_view_as_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_view_as_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_view_copy_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_view_copy_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_view_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_view_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_view_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_view_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_view_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_view_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_vsplit_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_vsplit_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_vsplit_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_vstack_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_vstack_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_vstack_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_vstack_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_xlogy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_xlogy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_xlogy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_zero__cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_zero__cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_zero__cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_zero__cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_zeros_like_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_zeros_like_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_inplace_zeros_like_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_H_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_T_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___getitem___cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___getitem___cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___getitem___cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___radd___cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___radd___cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___radd___cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___radd___cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___radd___cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___rand___cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___rand___cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___rand___cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___rmatmul___cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___rmod___cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___rmod___cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___rmul___cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___rmul___cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___ror___cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___ror___cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___rpow___cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___rpow___cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___rsub___cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___rxor___cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace___rxor___cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__batch_norm_with_update_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__chunk_cat_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__chunk_cat_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_abs_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_abs_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_abs_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_acos_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_acos_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_acos_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_addcdiv_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_addcdiv_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_addcmul_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_addcmul_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_addcmul_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_addcmul_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_asin_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_asin_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_asin_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_asin_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_atan_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_atan_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_ceil_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_clamp_max_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_clamp_min_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_clamp_min_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_clamp_min_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_cos_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_cos_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_cos_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_cosh_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_cosh_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_cosh_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_div_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_div_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_erf_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_erfc_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_erfc_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_erfc_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_exp_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_expm1_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_expm1_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_expm1_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_floor_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_floor_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_floor_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_lerp_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_lerp_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_lerp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_lerp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_lerp_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_lgamma_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_lgamma_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_lgamma_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_lgamma_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_log10_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_log1p_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_log1p_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_log1p_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_log2_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_log2_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_log2_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_log_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_max_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_max_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_max_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_maximum_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_maximum_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_minimum_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_minimum_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_minimum_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_minimum_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_mul_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_mul_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_mul_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_neg_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_norm_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_pow_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_pow_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_reciprocal_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_reciprocal_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_reciprocal_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_reciprocal_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_reciprocal_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_round_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_round_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_round_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_round_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_rsqrt_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_rsqrt_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_rsqrt_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_rsqrt_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_sigmoid_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_sigmoid_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_sigmoid_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_sign_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_sign_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_sign_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_sin_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_sinh_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_sinh_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_sinh_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_sinh_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_sinh_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_sinh_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_sqrt_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_sqrt_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_sqrt_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_sqrt_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_sub_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_sub_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_tan_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_tan_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_tanh_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_tanh_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_trunc_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_trunc_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_trunc_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_trunc_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_zero_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_zero_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__foreach_zero_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__native_batch_norm_legit_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__segment_reduce_offsets_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__segment_reduce_offsets_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__unsafe_masked_index_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__unsafe_masked_index_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__unsafe_masked_index_put_accumulate_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__unsafe_masked_index_put_accumulate_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace__upsample_bilinear2d_aa_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_abs_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_abs_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_acos_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_acos_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_acos_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_acos_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_acos_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_acosh_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_add_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_add_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_add_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_add_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_add_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_add_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_addcdiv_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_addcdiv_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_addcmul_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_addcmul_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_addcmul_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_addcmul_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_addmm_decomposed_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_addmv_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_addmv_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_addr_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_alias_copy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_alias_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_alias_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides___rdiv___cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides___rxor___cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides__chunk_cat_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides__foreach_add_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides__foreach_erfc_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides__foreach_log10_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides__foreach_max_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides__foreach_mul_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides__foreach_norm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides__foreach_reciprocal_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides__foreach_sigmoid_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides__foreach_sin_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides__foreach_tanh_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides__unsafe_masked_index_put_accumulate_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_acosh_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_aminmax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_argmax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_argsort_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_as_strided_partial_views_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_as_strided_scatter_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_atleast_3d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_cfloat_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_chalf_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_cholesky_inverse_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_clamp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_corrcoef_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_cos_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_cosh_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_count_nonzero_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_cov_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_cumulative_trapezoid_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_deg2rad_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_diagonal_scatter_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_diff_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_div_floor_rounding_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_dot_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_dsplit_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_einsum_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_empty_like_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_empty_strided_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_fft_fft2_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_fft_fft_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_fft_ihfft2_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_fft_ihfftn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_fft_irfft2_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_fft_irfftn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_fft_rfft2_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_fill_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_flatten_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_float_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_gcd_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_gt_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_half_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_hash_tensor_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_isclose_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_isreal_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_ldexp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_linalg_cholesky_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_linalg_det_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_linalg_eigvals_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_linalg_lu_factor_ex_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_linalg_svdvals_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_linalg_vector_norm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_log_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_log_normal_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_logit_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_logspace_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_lu_solve_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_masked_logaddexp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_masked_logsumexp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_masked_select_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_matmul_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_msort_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_mv_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_mvlgamma_mvlgamma_p_1_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_native_layer_norm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_neg_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_avg_pool1d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_avg_pool2d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_binary_cross_entropy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_conv2d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_ctc_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_feature_alpha_dropout_with_train_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_fractional_max_pool2d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_gelu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_hardsigmoid_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_interpolate_linear_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_interpolate_nearest-exact_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_local_response_norm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_max_unpool3d_grad_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_multi_margin_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_one_hot_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_pad_reflect_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_pad_replicate_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_pdist_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_pixel_unshuffle_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_relu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_rrelu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_nn_functional_smooth_l1_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_norm_inf_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_ormqr_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_permute_copy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_pinverse_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_polygamma_polygamma_n_0_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_polygamma_polygamma_n_3_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_real_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_resize__cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_resolve_conj_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_resolve_neg_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_round_decimals_0_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_select_scatter_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_sgn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_signal_windows_gaussian_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_signal_windows_general_cosine_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_signal_windows_hamming_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_signal_windows_kaiser_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_sin_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_sparse_mm_reduce_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_special_entr_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_special_i1e_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_special_laguerre_polynomial_l_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_special_ndtri_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_special_polygamma_special_polygamma_n_0_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_special_xlog1py_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_split_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_split_with_sizes_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_square_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_sub_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_sum_to_size_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_trace_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_transpose_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_trapz_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_triu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_unsqueeze_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_vdot_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_all_strides_view_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_amax_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_amin_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_angle_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_angle_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_angle_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_any_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_any_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_arange_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_argmax_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_argmax_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_argmax_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_argmin_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_argmin_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_argsort_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_argsort_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_argwhere_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_argwhere_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_argwhere_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_as_strided_copy_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_as_strided_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_as_strided_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_as_strided_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_as_strided_partial_views_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_as_strided_partial_views_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_asinh_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_asinh_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_atan2_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_atan2_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_atanh_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_atanh_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_atanh_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_atleast_1d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_atleast_1d_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_atleast_1d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_atleast_1d_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_atleast_2d_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_atleast_3d_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_atleast_3d_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_baddbmm_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_baddbmm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bernoulli_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bernoulli_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bfloat16_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bfloat16_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bfloat16_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bincount_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bincount_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bincount_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bitwise_and_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bitwise_left_shift_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bitwise_not_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bitwise_not_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bitwise_not_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bitwise_xor_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_block_diag_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_block_diag_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_block_diag_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_block_diag_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bmm_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bmm_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bool_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bool_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bool_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_broadcast_tensors_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_broadcast_tensors_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_broadcast_to_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_broadcast_to_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_bucketize_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_byte_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_byte_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cartesian_prod_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cat_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cdist_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cdouble_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cdouble_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cdouble_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cdouble_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cdouble_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cdouble_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_ceil_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_ceil_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_ceil_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_ceil_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cfloat_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_chalf_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_chalf_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cholesky_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cholesky_inverse_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cholesky_solve_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_chunk_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_chunk_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_clamp_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_clamp_max_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_clamp_max_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_clamp_max_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_clamp_min_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_clamp_min_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_clamp_min_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_clamp_min_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_clamp_min_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_clone_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_column_stack_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_column_stack_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_combinations_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_combinations_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_combinations_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_conj_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_conj_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_conj_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_conj_physical_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_conj_physical_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_conj_physical_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_constant_pad_nd_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_contiguous_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_contiguous_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_contiguous_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_contiguous_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_copysign_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_copysign_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_corrcoef_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_corrcoef_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_corrcoef_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cos_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cosh_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cosh_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_count_nonzero_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_count_nonzero_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cov_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cov_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cross_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cross_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cross_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cummax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cummax_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cummax_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cummin_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cummin_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cummin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cumprod_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cumsum_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cumsum_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_cumsum_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_deg2rad_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_diag_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_diag_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_diag_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_diag_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_diag_embed_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_diag_embed_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_diagflat_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_diagonal_copy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_diagonal_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_diagonal_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_diagonal_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_diagonal_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_diagonal_scatter_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_diagonal_scatter_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_diff_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_diff_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_digamma_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_dist_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_div_floor_rounding_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_div_floor_rounding_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_div_floor_rounding_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_div_no_rounding_mode_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_div_no_rounding_mode_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_div_no_rounding_mode_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_div_no_rounding_mode_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_div_trunc_rounding_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_div_trunc_rounding_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_dot_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_double_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_double_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_double_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_double_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_dsplit_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_dsplit_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_dsplit_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_dsplit_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_dsplit_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_dstack_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_dstack_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_einsum_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_einsum_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_empty_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_empty_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_empty_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_empty_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_empty_like_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_empty_like_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_empty_like_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_empty_like_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_empty_like_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_empty_permuted_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_empty_permuted_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_empty_permuted_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_empty_permuted_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_empty_strided_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_eq_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_eq_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_eq_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_eq_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_equal_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_equal_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_erf_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_erf_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_erfc_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_erfc_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_erfc_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_erfinv_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_erfinv_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_erfinv_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_exp2_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_exp2_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_exp_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_exp_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_expand_as_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_expand_as_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_expand_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_expand_copy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_expand_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_expand_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_expand_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_eye_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_eye_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_fft2_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_fft_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_fft_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_fftn_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_fftn_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_fftn_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_fftn_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_fftshift_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_fftshift_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_hfft2_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_hfft2_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_hfft_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_hfftn_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_hfftn_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_ifft2_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_ifft_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_ifft_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_ifft_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_ifftn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_ifftn_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_ifftn_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_ifftshift_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_ifftshift_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_ifftshift_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_ihfft2_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_ihfft2_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_ihfft2_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_ihfft_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_ihfft_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_ihfftn_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_ihfftn_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_irfft2_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_irfft2_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_irfft_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_irfftn_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_rfft2_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_rfft2_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_rfft_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_rfft_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_rfftn_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fft_rfftn_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fill_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fill_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fill_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fill_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_flatten_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_flatten_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_flip_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_flip_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_flip_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_flip_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_flip_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_flip_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fliplr_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fliplr_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_flipud_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_flipud_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_flipud_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_float_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_float_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_float_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_float_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_float_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_float_power_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_float_power_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_floor_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_floor_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_floor_divide_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fmax_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fmax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fmin_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fmod_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fmod_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_fmod_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_frac_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_full_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_full_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_full_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_full_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_full_like_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_full_like_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_full_like_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_full_like_cuda_uint16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_gather_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_gather_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_ge_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_ge_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_geometric_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_geometric_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_geometric_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_gradient_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_gradient_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_grid_sampler_3d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_half_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_half_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_hash_tensor_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_hash_tensor_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_hash_tensor_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_heaviside_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_heaviside_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_heaviside_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_histc_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_histc_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_histc_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_histc_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_hsplit_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_hstack_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_hstack_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_hstack_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_hypot_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_i0_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_i0_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_imag_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_copy_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_copy_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_copy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_fill_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_fill_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_fill_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_put_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_put_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_put_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_put_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_reduce_amax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_reduce_amax_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_reduce_amin_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_reduce_mean_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_reduce_mean_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_reduce_prod_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_reduce_prod_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_index_select_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_inner_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_inner_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_inner_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_int_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_int_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_int_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_isclose_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_isfinite_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_isfinite_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_isin_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_isin_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_isinf_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_isinf_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_isnan_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_isnan_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_isneginf_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_isneginf_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_isposinf_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_isposinf_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_isreal_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_isreal_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_isreal_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_isreal_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_isreal_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_item_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_item_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_item_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_item_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_jiterator_2inputs_2outputs_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_jiterator_2inputs_2outputs_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_jiterator_2inputs_2outputs_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_jiterator_2inputs_2outputs_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_jiterator_2inputs_2outputs_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_jiterator_2inputs_2outputs_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_jiterator_4inputs_with_extra_args_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_jiterator_4inputs_with_extra_args_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_jiterator_4inputs_with_extra_args_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_jiterator_4inputs_with_extra_args_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_jiterator_4inputs_with_extra_args_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_jiterator_4inputs_with_extra_args_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_jiterator_binary_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_jiterator_binary_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_jiterator_binary_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_jiterator_binary_return_by_ref_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_jiterator_unary_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_jiterator_unary_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_jiterator_unary_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_kron_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_kron_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_kron_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_kthvalue_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_kthvalue_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_kthvalue_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_kthvalue_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_lcm_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_ldexp_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_le_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_le_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_le_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_le_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_lgamma_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_cholesky_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_cond_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_diagonal_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_diagonal_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_diagonal_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_diagonal_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_inv_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_inv_ex_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_ldl_factor_ex_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_ldl_solve_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_lstsq_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_lstsq_grad_oriented_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_lstsq_grad_oriented_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_lu_solve_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_matrix_norm_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_matrix_norm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_matrix_rank_hermitian_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_matrix_rank_hermitian_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_multi_dot_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_multi_dot_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_multi_dot_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_multi_dot_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_norm_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_norm_subgradients_at_zero_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_pinv_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_pinv_singular_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_pinv_singular_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_slogdet_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_slogdet_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_solve_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_solve_ex_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_solve_ex_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_solve_triangular_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_tensorinv_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_tensorsolve_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_tensorsolve_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_vander_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_vecdot_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_vector_norm_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linalg_vector_norm_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linspace_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linspace_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linspace_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linspace_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_linspace_tensor_overload_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_log1p_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_log1p_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_log2_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_log_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_log_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_log_softmax_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_log_softmax_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_log_softmax_with_dtype_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_log_softmax_with_dtype_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_log_softmax_with_dtype_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logaddexp2_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logaddexp_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logcumsumexp_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logical_and_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logical_and_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logical_and_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logical_not_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logical_not_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logical_not_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logical_not_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logical_or_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logical_or_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logical_xor_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logical_xor_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logspace_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logspace_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logspace_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logspace_tensor_overload_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logspace_tensor_overload_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logspace_tensor_overload_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logsumexp_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_logsumexp_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_lt_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_lt_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_lu_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_lu_unpack_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mH_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mH_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mH_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mT_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mT_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_amax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_amax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_amax_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_amax_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_argmax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_argmin_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_argmin_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_argmin_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_cumprod_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_cumprod_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_cumprod_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_cumprod_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_cumsum_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_cumsum_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_cumsum_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_cumsum_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_fill_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_logaddexp_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_logsumexp_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_logsumexp_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_logsumexp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_logsumexp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_logsumexp_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_mean_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_mean_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_norm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_prod_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_prod_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_prod_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_scatter_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_select_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_select_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_select_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_select_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_softmax_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_softmax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_std_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_std_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_std_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_sum_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_sum_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_sum_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_sum_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_var_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_masked_var_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_matmul_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_matmul_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_matmul_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_matrix_exp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_matrix_exp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_max_reduction_no_dim_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_max_reduction_with_dim_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_max_reduction_with_dim_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_max_reduction_with_dim_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_max_reduction_with_dim_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_max_reduction_with_dim_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_maximum_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_maximum_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_maximum_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mean_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_median_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_median_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_median_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_meshgrid_list_of_tensors_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_meshgrid_variadic_tensors_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_min_binary_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_min_binary_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_min_reduction_with_dim_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_min_reduction_with_dim_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_min_reduction_with_dim_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_minimum_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_minimum_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_minimum_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_minimum_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mode_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_msort_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mul_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mul_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mul_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_multinomial_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mv_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mvlgamma_mvlgamma_p_1_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mvlgamma_mvlgamma_p_1_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mvlgamma_mvlgamma_p_1_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mvlgamma_mvlgamma_p_3_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mvlgamma_mvlgamma_p_3_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mvlgamma_mvlgamma_p_3_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mvlgamma_mvlgamma_p_5_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_mvlgamma_mvlgamma_p_5_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nanmean_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nanmean_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nanmedian_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nanmedian_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nanmedian_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nansum_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nansum_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nansum_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_narrow_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_narrow_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_native_batch_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_native_batch_norm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_ne_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_ne_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_neg_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_neg_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_new_empty_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_new_empty_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_new_empty_strided_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_new_empty_strided_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_new_full_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_new_full_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_new_full_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_new_full_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_new_full_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_new_ones_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_new_ones_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_new_ones_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_new_ones_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_new_zeros_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_adaptive_avg_pool3d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_adaptive_max_pool1d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_adaptive_max_pool3d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_alpha_dropout_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_avg_pool1d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_avg_pool2d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_avg_pool2d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_avg_pool3d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_avg_pool3d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_batch_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_batch_norm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_batch_norm_without_cudnn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_batch_norm_without_cudnn_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_binary_cross_entropy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_binary_cross_entropy_with_logits_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_binary_cross_entropy_with_logits_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_channel_shuffle_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_channel_shuffle_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_conv1d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_conv2d_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_conv2d_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_conv_transpose1d_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_conv_transpose2d_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_conv_transpose3d_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_conv_transpose3d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_cosine_embedding_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_cosine_embedding_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_cross_entropy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_dropout2d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_dropout_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_feature_alpha_dropout_with_train_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_feature_alpha_dropout_with_train_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_feature_alpha_dropout_without_train_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_feature_alpha_dropout_without_train_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_fractional_max_pool2d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_fractional_max_pool2d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_fractional_max_pool3d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_gaussian_nll_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_gelu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_hardsigmoid_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_hardswish_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_hardswish_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_hardtanh_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_huber_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_interpolate_area_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_interpolate_bilinear_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_interpolate_linear_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_interpolate_nearest-exact_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_l1_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_l1_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_layer_norm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_leaky_relu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_max_pool1d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_max_pool1d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_max_pool2d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_max_pool3d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_max_unpool1d_grad_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_max_unpool2d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_max_unpool2d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_max_unpool3d_grad_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_max_unpool3d_grad_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_mish_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_mse_loss_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_multi_head_attention_forward_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_multi_head_attention_forward_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_multi_margin_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_normalize_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pad_circular_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pad_circular_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pad_constant_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pad_constant_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pad_constant_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pad_replicate_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pad_replicate_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pad_replicate_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pad_replicate_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pad_replicate_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pad_replicate_negative_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pad_replicate_negative_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pairwise_distance_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pairwise_distance_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pairwise_distance_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pdist_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pixel_shuffle_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pixel_shuffle_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pixel_shuffle_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pixel_unshuffle_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pixel_unshuffle_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pixel_unshuffle_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_pixel_unshuffle_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_relu_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_rms_norm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_rrelu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_smooth_l1_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_soft_margin_loss_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_soft_margin_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_soft_margin_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_softmin_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_softmin_with_dtype_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_softmin_with_dtype_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_softplus_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_softshrink_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_tanhshrink_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_tanhshrink_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_threshold_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_threshold_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_triplet_margin_loss_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_triplet_margin_with_distance_loss_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_triplet_margin_with_distance_loss_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_triplet_margin_with_distance_loss_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_triplet_margin_with_distance_loss_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_triplet_margin_with_distance_loss_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_unfold_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_unfold_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_unfold_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nn_functional_upsample_bilinear_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nonzero_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nonzero_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nonzero_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_nonzero_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_norm_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_norm_fro_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_norm_fro_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_norm_fro_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_norm_inf_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_norm_inf_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_norm_nuc_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_normal_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_normal_in_place_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_normal_number_mean_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_ones_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_ones_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_ones_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_ones_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_ones_like_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_ones_like_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_ormqr_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_outer_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_outer_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_pca_lowrank_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_pca_lowrank_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_permute_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_permute_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_permute_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_pinverse_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_pinverse_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_pinverse_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_polar_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_polygamma_polygamma_n_0_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_polygamma_polygamma_n_0_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_polygamma_polygamma_n_1_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_polygamma_polygamma_n_3_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_polygamma_polygamma_n_3_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_polygamma_polygamma_n_3_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_polygamma_polygamma_n_4_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_positive_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_positive_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_pow_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_pow_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_prod_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_prod_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_prod_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_put_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_put_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_qr_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_rad2deg_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_rad2deg_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_rad2deg_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_rad2deg_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_rand_like_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_rand_like_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_randint_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_randint_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_randint_like_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_randint_like_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_randint_like_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_randn_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_randn_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_randn_like_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_ravel_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_ravel_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_ravel_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_real_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_real_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_remainder_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_renorm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_repeat_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_repeat_interleave_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_repeat_interleave_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_repeat_interleave_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_repeat_interleave_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_reshape_as_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_reshape_as_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_reshape_as_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_reshape_as_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_reshape_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_reshape_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_reshape_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_reshape_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_reshape_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_resize__cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_resize__cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_resize_as__cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_resize_as__cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_resolve_conj_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_resolve_conj_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_resolve_conj_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_roll_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_roll_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_roll_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_roll_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_rot90_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_rot90_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_rot90_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_rot90_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_round_decimals_3_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_round_decimals_neg_3_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_rsqrt_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_rsqrt_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_rsub_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_rsub_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_scalar_tensor_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_scalar_tensor_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_scatter_add_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_scatter_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_scatter_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_scatter_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_scatter_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_scatter_reduce_amax_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_scatter_reduce_mean_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_scatter_reduce_mean_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_scatter_reduce_prod_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_scatter_reduce_prod_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_scatter_reduce_sum_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_scatter_reduce_sum_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_searchsorted_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_select_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_select_scatter_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_select_scatter_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_select_scatter_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sgn_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sgn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sgn_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sgn_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sigmoid_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sigmoid_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sign_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sign_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sign_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_signal_windows_bartlett_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_signal_windows_blackman_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_signal_windows_general_cosine_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_signal_windows_general_cosine_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_signal_windows_kaiser_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_signbit_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_signbit_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sin_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sin_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sin_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sinc_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sinc_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sinc_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sinh_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_slice_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_slice_scatter_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_softmax_with_dtype_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_softmax_with_dtype_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_softmax_with_dtype_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sort_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sparse_mm_reduce_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sparse_sampled_addmm_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_airy_ai_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_airy_ai_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_bessel_j0_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_bessel_j0_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_bessel_j1_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_bessel_j1_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_bessel_y0_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_bessel_y0_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_bessel_y0_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_bessel_y1_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_bessel_y1_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_chebyshev_polynomial_t_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_chebyshev_polynomial_v_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_chebyshev_polynomial_v_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_chebyshev_polynomial_v_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_entr_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_erfcx_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_erfcx_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_hermite_polynomial_h_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_hermite_polynomial_he_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_hermite_polynomial_he_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_hermite_polynomial_he_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_i0e_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_i0e_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_i0e_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_i0e_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_i1_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_i1_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_i1e_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_i1e_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_i1e_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_laguerre_polynomial_l_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_laguerre_polynomial_l_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_legendre_polynomial_p_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_log_ndtr_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_log_ndtr_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_modified_bessel_i0_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_modified_bessel_i1_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_modified_bessel_i1_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_modified_bessel_k0_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_modified_bessel_k0_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_modified_bessel_k1_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_modified_bessel_k1_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_ndtr_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_ndtr_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_ndtri_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_ndtri_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_polygamma_special_polygamma_n_0_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_polygamma_special_polygamma_n_0_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_scaled_modified_bessel_k0_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_scaled_modified_bessel_k1_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_scaled_modified_bessel_k1_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_shifted_chebyshev_polynomial_t_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_shifted_chebyshev_polynomial_u_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_shifted_chebyshev_polynomial_v_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_shifted_chebyshev_polynomial_v_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_shifted_chebyshev_polynomial_w_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_xlog1py_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_xlog1py_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_xlog1py_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_xlog1py_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_special_zeta_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_split_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_split_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_split_list_args_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_split_with_sizes_copy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_split_with_sizes_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_split_with_sizes_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_split_with_sizes_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sqrt_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sqrt_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_square_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_square_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_square_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_squeeze_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_squeeze_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_squeeze_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_squeeze_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_squeeze_multiple_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_squeeze_multiple_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_squeeze_multiple_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_squeeze_multiple_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_squeeze_multiple_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_stack_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_stack_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_stack_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_std_mean_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_std_mean_unbiased_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_std_unbiased_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_std_unbiased_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_std_unbiased_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_stft_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_stft_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sub_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sub_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sum_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sum_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sum_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sum_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sum_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sum_to_size_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sum_to_size_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sum_to_size_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_sum_to_size_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_svd_lowrank_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_t_copy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_t_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_t_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_t_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_take_along_dim_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_take_along_dim_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_take_along_dim_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_tan_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_tan_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_tanh_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_tanh_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_tanh_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_tensor_split_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_tensor_split_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_tensor_split_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_tensordot_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_tile_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_tile_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_tile_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_tile_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_to_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_to_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_topk_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_torch__scaled_mm_v2_cuda_float8_e4m3fn, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_torch_ops_aten__safe_softmax_default_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_torch_ops_aten__safe_softmax_default_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_trace_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_trace_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_transpose_copy_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_transpose_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_transpose_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_transpose_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_trapz_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_trapz_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_triangular_solve_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_tril_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_tril_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_tril_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_triu_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_triu_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_triu_indices_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_true_divide_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_trunc_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_trunc_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unbind_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unbind_copy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unbind_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unbind_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unbind_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unflatten_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unflatten_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unfold_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_uniform_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unique_consecutive_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unique_consecutive_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unique_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unravel_index_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unsafe_chunk_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unsafe_split_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unsafe_split_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unsafe_split_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unsafe_split_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unsqueeze_copy_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unsqueeze_copy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unsqueeze_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unsqueeze_copy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unsqueeze_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_unsqueeze_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_var_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_var_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_var_mean_unbiased_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_var_mean_unbiased_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_var_unbiased_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_vdot_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_view_as_complex_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_view_as_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_view_as_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_view_as_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_view_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_view_cuda_int16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_view_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_vsplit_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_vsplit_cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_vsplit_cuda_int32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_vstack_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_vstack_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_vstack_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_where_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_where_cuda_float16, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_where_cuda_int64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_where_cuda_int8, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_xlogy_cuda_bool, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_xlogy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_zero__cuda_float32, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_zero__cuda_float64, test/test_meta.py::TestMetaCUDA::test_dispatch_symbolic_meta_outplace_zero__cuda_int16, test/test_meta.py::TestMetaCUDA::test_embedding_bag_byte_prepack_cuda, test/test_meta.py::TestMetaCUDA::test_embedding_bag_dense_backward_mode_1_cuda, test/test_meta.py::TestMetaCUDA::test_embedding_bag_dense_backward_mode_2_cuda, test/test_meta.py::TestMetaCUDA::test_group_norm_backward_output_mask2_cuda, test/test_meta.py::TestMetaCUDA::test_group_norm_backward_output_mask3_cuda, test/test_meta.py::TestMetaCUDA::test_meta_inplace_H_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_H_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_T_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_T_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_T_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_T_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace___getitem___cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace___getitem___cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace___getitem___cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace___getitem___cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace___getitem___cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace___getitem___cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace___radd___cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace___radd___cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace___radd___cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace___rdiv___cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace___rdiv___cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace___rdiv___cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace___rdiv___cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace___rdiv___cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace___rmod___cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace___rmul___cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace___rmul___cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace___rmul___cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace___ror___cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace___rpow___cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace___rpow___cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace___rpow___cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace___rsub___cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace___rxor___cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__batch_norm_with_update_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__chunk_cat_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace__chunk_cat_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_abs_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_acos_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_acos_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_acos_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_add_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_add_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_add_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_add_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_addcdiv_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_addcdiv_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_addcdiv_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_addcdiv_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_addcmul_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_addcmul_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_addcmul_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_addcmul_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_addcmul_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_addcmul_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_asin_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_atan_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_atan_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_atan_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_ceil_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_clamp_max_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_clamp_max_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_clamp_max_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_clamp_min_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_clamp_min_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_copy_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_cos_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_cos_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_cos_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_cosh_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_cosh_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_div_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_div_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_div_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_div_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_erf_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_erf_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_erfc_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_erfc_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_erfc_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_erfc_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_erfc_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_exp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_exp_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_expm1_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_expm1_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_expm1_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_floor_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_floor_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_frac_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_frac_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_frac_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_lerp_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_lerp_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_lerp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_lgamma_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_lgamma_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_log10_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_log10_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_log1p_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_log1p_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_log1p_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_log2_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_log2_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_log_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_log_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_max_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_maximum_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_maximum_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_minimum_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_mul_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_mul_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_mul_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_mul_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_neg_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_neg_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_neg_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_neg_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_norm_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_pow_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_pow_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_reciprocal_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_reciprocal_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_reciprocal_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_round_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_round_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_round_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_round_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_rsqrt_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_sigmoid_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_sigmoid_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_sign_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_sign_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_sign_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_sign_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_sin_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_sin_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_sin_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_sinh_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_sinh_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_sinh_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_sqrt_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_sqrt_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_sqrt_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_sub_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_sub_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_tan_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_tan_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_tanh_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_trunc_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_trunc_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_trunc_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_trunc_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_zero_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_zero_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_zero_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_zero_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_zero_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_zero_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__foreach_zero_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace__segment_reduce_offsets_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace__softmax_backward_data_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace__unsafe_masked_index_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace__upsample_bilinear2d_aa_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_abs_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_abs_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_add_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_add_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_addbmm_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_addcdiv_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_addcdiv_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_addcmul_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_addmm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_addmm_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_addmm_decomposed_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_addmv_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_addmv_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_addr_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_addr_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_alias_copy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_alias_copy_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_alias_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_alias_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_alias_copy_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_all_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_allclose_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_amax_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_amax_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_amin_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_amin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_aminmax_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_aminmax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_angle_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_angle_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_any_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_any_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_arange_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_arange_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_arange_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_arange_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_argmax_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_argmin_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_argsort_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_argsort_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_argwhere_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_argwhere_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_as_strided_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_as_strided_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_as_strided_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_as_strided_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_as_strided_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_as_strided_partial_views_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_as_strided_partial_views_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_as_strided_partial_views_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_as_strided_scatter_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_as_strided_scatter_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_asin_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_asin_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_asinh_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_asinh_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_asinh_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_asinh_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_asinh_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_atan2_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_atan2_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_atan_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_atan_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_atan_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_atan_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_atanh_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_atleast_1d_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_atleast_1d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_atleast_1d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_atleast_1d_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_atleast_2d_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_atleast_3d_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_atleast_3d_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_atleast_3d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_atleast_3d_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_atleast_3d_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_bernoulli_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_bfloat16_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_bfloat16_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_bitwise_and_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_bitwise_and_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_bitwise_and_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_bitwise_or_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_bitwise_or_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_bitwise_right_shift_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_bitwise_right_shift_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_bitwise_xor_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_block_diag_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_block_diag_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_block_diag_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_block_diag_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_block_diag_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_bmm_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_bool_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_bool_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_broadcast_shapes_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_broadcast_to_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_bucketize_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_bucketize_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_bucketize_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cartesian_prod_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cartesian_prod_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cartesian_prod_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cartesian_prod_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cartesian_prod_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cat_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cat_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cat_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cat_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cauchy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cdist_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cdouble_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cdouble_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cdouble_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_ceil_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_ceil_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cfloat_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cfloat_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cfloat_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_chalf_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_chalf_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_chalf_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_char_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_char_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_char_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cholesky_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cholesky_inverse_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cholesky_solve_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_chunk_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_chunk_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_chunk_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_chunk_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_clamp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_clamp_max_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_clamp_max_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_clamp_max_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_clamp_max_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_clamp_max_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_clamp_min_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_clone_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_clone_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_clone_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_combinations_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_complex_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_conj_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_conj_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_conj_physical_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_conj_physical_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_constant_pad_nd_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_constant_pad_nd_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_contiguous_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_contiguous_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_contiguous_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_contiguous_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_contiguous_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_contiguous_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_copysign_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_copysign_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_corrcoef_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cos_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cos_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cos_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cosh_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cosh_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cosh_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_count_nonzero_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_count_nonzero_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_count_nonzero_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_count_nonzero_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cov_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cross_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cross_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cross_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cross_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cummax_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cummin_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cumprod_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cumprod_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cumprod_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_cumulative_trapezoid_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_deg2rad_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_deg2rad_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_deg2rad_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_deg2rad_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_deg2rad_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_diag_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_diag_embed_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_diag_embed_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_diag_embed_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_diagonal_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_diagonal_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_diagonal_copy_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_diagonal_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_diagonal_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_diagonal_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_diagonal_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_diagonal_scatter_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_diagonal_scatter_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_diagonal_scatter_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_diff_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_diff_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_diff_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_digamma_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_div_floor_rounding_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_div_floor_rounding_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_div_floor_rounding_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_div_no_rounding_mode_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_div_no_rounding_mode_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_div_no_rounding_mode_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_div_no_rounding_mode_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_div_trunc_rounding_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_dot_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_double_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_double_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_double_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_dsplit_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_dsplit_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_dsplit_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_dsplit_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_dsplit_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_dstack_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_dstack_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_dstack_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_einsum_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_einsum_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_empty_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_empty_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_empty_like_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_empty_like_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_empty_like_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_empty_permuted_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_empty_permuted_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_empty_permuted_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_empty_strided_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_empty_strided_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_empty_strided_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_empty_strided_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_empty_strided_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_empty_strided_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_empty_strided_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_equal_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_equal_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_erf_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_erf_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_erf_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_erf_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_erfinv_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_exp2_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_exp2_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_exp2_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_exp_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_exp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_expand_as_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_expand_copy_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_expand_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_expand_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_expand_copy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_expm1_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_expm1_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_exponential_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_eye_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_eye_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_eye_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_eye_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_fft2_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_fft2_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_fft_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_fftn_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_fftshift_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_hfft2_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_hfft2_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_hfft_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_hfft_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_hfft_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_hfftn_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_ifft_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_ifftn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_ifftn_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_ifftshift_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_ifftshift_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_ihfft_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_ihfft_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_ihfft_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_ihfft_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_ihfftn_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_ihfftn_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_irfft2_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_irfft_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_irfft_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_irfftn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_irfftn_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_rfft2_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_rfft2_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_rfft2_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_rfft2_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fft_rfft_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fill_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fill_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_flatten_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_flatten_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_flatten_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_flip_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fliplr_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fliplr_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_flipud_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_flipud_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_flipud_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_flipud_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_flipud_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_float_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_float_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_float_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_float_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_float_power_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_floor_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_floor_divide_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_floor_divide_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fmin_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fmin_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fmod_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fmod_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_fmod_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_frac_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_frexp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_full_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_full_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_full_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_full_like_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_full_like_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_gather_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_gather_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_gather_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_ge_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_geqrf_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_gradient_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_gradient_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_gradient_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_grid_sampler_3d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_gt_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_gt_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_half_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_half_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_half_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_half_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_histc_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_hsplit_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_hsplit_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_hsplit_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_hstack_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_hstack_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_hstack_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_hypot_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_i0_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_i0_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_i0_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_i0_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_imag_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_add_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_add_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_add_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_add_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_add_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_copy_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_copy_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_fill_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_fill_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_fill_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_put_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_put_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_put_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_reduce_amax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_reduce_amax_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_reduce_amin_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_reduce_amin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_reduce_mean_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_reduce_mean_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_reduce_prod_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_reduce_prod_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_reduce_prod_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_select_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_select_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_index_select_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_inner_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_inner_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_int_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_isclose_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_isfinite_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_isfinite_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_isin_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_isin_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_isinf_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_isinf_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_isnan_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_isnan_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_isnan_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_isnan_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_isposinf_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_isreal_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_isreal_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_isreal_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_item_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_item_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_jiterator_4inputs_with_extra_args_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_jiterator_4inputs_with_extra_args_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_jiterator_binary_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_jiterator_binary_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_jiterator_binary_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_jiterator_binary_return_by_ref_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_jiterator_binary_return_by_ref_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_jiterator_binary_return_by_ref_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_jiterator_unary_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_jiterator_unary_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_jiterator_unary_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_kron_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_kthvalue_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_lcm_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_le_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_le_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_lerp_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_lerp_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_lerp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_lgamma_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_lgamma_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_lgamma_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_cholesky_ex_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_cholesky_ex_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_cond_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_cond_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_cross_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_cross_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_cross_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_det_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_diagonal_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_eigh_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_eigvals_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_inv_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_inv_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_inv_ex_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_ldl_factor_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_ldl_solve_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_lstsq_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_lstsq_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_lstsq_grad_oriented_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_lu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_lu_factor_ex_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_lu_factor_ex_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_matrix_rank_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_multi_dot_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_multi_dot_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_norm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_pinv_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_pinv_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_pinv_hermitian_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_pinv_singular_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_qr_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_qr_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_slogdet_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_solve_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_solve_ex_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_solve_triangular_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_svdvals_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_tensorinv_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_vecdot_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_vecdot_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_vecdot_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linalg_vecdot_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linspace_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linspace_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linspace_tensor_overload_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linspace_tensor_overload_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_linspace_tensor_overload_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_log10_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_log1p_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_log1p_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_log2_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_log2_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_log2_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_log2_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_log_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_log_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_log_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_log_normal_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_log_softmax_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_log_softmax_with_dtype_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_log_softmax_with_dtype_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_logaddexp2_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_logaddexp_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_logaddexp_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_logcumsumexp_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_logdet_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_logical_and_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_logical_not_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_logical_not_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_logical_not_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_logical_or_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_logical_xor_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_logical_xor_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_logical_xor_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_logit_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_logit_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_logspace_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_logspace_tensor_overload_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_long_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_long_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_lt_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_lt_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_lt_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_lt_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_lt_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_lt_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_lu_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_lu_solve_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_lu_unpack_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_mH_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_mH_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_mT_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_mT_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_mT_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_amax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_amin_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_amin_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_amin_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_argmax_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_argmax_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_argmin_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_cumsum_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_cumsum_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_cumsum_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_fill_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_fill_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_fill_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_logsumexp_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_logsumexp_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_logsumexp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_logsumexp_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_mean_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_median_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_normalize_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_prod_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_prod_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_scatter_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_select_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_select_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_std_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_sum_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_sum_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_sum_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_sum_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_var_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_masked_var_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_matmul_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_matrix_exp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_max_binary_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_max_binary_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_max_binary_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_max_binary_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_max_reduction_no_dim_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_max_reduction_with_dim_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_mean_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_median_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_meshgrid_list_of_tensors_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_meshgrid_list_of_tensors_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_meshgrid_list_of_tensors_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_meshgrid_list_of_tensors_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_meshgrid_variadic_tensors_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_meshgrid_variadic_tensors_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_min_binary_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_min_reduction_no_dim_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_min_reduction_no_dim_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_min_reduction_with_dim_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_minimum_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_minimum_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_mode_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_movedim_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_mul_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_mul_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_mul_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_mul_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_multinomial_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_mv_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_mvlgamma_mvlgamma_p_1_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_mvlgamma_mvlgamma_p_1_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nan_to_num_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nan_to_num_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nanmedian_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nanmedian_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nansum_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nansum_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nansum_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_narrow_copy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_narrow_copy_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_narrow_copy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_narrow_copy_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_narrow_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_narrow_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_native_dropout_backward_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_ne_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_ne_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_ne_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_neg_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_neg_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_neg_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_empty_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_empty_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_empty_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_empty_strided_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_empty_strided_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_empty_strided_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_empty_strided_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_full_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_full_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_full_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_full_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_ones_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_ones_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_ones_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_ones_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_ones_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_zeros_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_zeros_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_zeros_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_new_zeros_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nextafter_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_adaptive_avg_pool1d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_adaptive_max_pool2d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_adaptive_max_pool2d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_alpha_dropout_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_alpha_dropout_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_avg_pool2d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_batch_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_batch_norm_without_cudnn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_conv1d_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_conv3d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_conv3d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_conv_transpose1d_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_conv_transpose3d_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_conv_transpose3d_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_conv_transpose3d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_conv_transpose3d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_cosine_embedding_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_cosine_embedding_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_cosine_embedding_loss_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_cosine_embedding_loss_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_cosine_similarity_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_cross_entropy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_cross_entropy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_dropout2d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_dropout2d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_embedding_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_feature_alpha_dropout_without_train_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_feature_alpha_dropout_without_train_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_glu_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_grid_sample_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_group_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_hardshrink_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_hardswish_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_hardswish_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_hardswish_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_hardtanh_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_hardtanh_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_hardtanh_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_hardtanh_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_hinge_embedding_loss_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_hinge_embedding_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_instance_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_instance_norm_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_interpolate_area_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_interpolate_bilinear_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_interpolate_bilinear_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_interpolate_nearest-exact_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_interpolate_nearest-exact_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_interpolate_trilinear_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_kl_div_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_kl_div_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_l1_loss_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_leaky_relu_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_linear_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_linear_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_linear_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_linear_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_margin_ranking_loss_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_max_unpool1d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_max_unpool1d_grad_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_max_unpool1d_grad_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_max_unpool2d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_max_unpool2d_grad_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_max_unpool3d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_max_unpool3d_grad_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_max_unpool3d_grad_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_multi_head_attention_forward_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_multi_head_attention_forward_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_multi_margin_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_multilabel_margin_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_multilabel_margin_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_nll_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_normalize_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pad_circular_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pad_circular_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pad_circular_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pad_constant_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pad_constant_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pad_reflect_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pad_reflect_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pad_replicate_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pad_replicate_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pad_replicate_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pad_replicate_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pad_replicate_negative_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pad_replicate_negative_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pad_replicate_negative_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pad_replicate_negative_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pairwise_distance_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pairwise_distance_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pairwise_distance_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pairwise_distance_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pdist_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pixel_shuffle_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pixel_shuffle_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pixel_unshuffle_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pixel_unshuffle_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pixel_unshuffle_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pixel_unshuffle_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_pixel_unshuffle_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_poisson_nll_loss_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_poisson_nll_loss_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_poisson_nll_loss_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_prelu_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_prelu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_prelu_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_relu6_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_relu_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_relu_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_rms_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_rms_norm_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_scaled_dot_product_attention_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_scaled_dot_product_attention_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_smooth_l1_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_softmin_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_softmin_with_dtype_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_softmin_with_dtype_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_softmin_with_dtype_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_softmin_with_dtype_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_softshrink_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_softshrink_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_softsign_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_softsign_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_tanhshrink_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_tanhshrink_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_threshold_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_threshold_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_triplet_margin_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_triplet_margin_loss_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_triplet_margin_with_distance_loss_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_unfold_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_upsample_nearest_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_upsample_nearest_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nn_functional_upsample_nearest_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nonzero_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nonzero_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nonzero_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nonzero_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nonzero_static_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nonzero_static_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nonzero_static_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_nonzero_static_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_norm_fro_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_norm_fro_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_norm_inf_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_norm_nuc_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_norm_nuc_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_normal_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_normal_in_place_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_normal_number_mean_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_normal_number_mean_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_ones_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_ones_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_ones_like_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_ormqr_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_ormqr_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_ormqr_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_pca_lowrank_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_permute_copy_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_permute_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_permute_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_permute_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_polygamma_polygamma_n_0_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_polygamma_polygamma_n_0_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_polygamma_polygamma_n_0_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_polygamma_polygamma_n_0_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_polygamma_polygamma_n_1_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_polygamma_polygamma_n_1_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_polygamma_polygamma_n_2_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_polygamma_polygamma_n_3_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_polygamma_polygamma_n_3_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_polygamma_polygamma_n_3_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_polygamma_polygamma_n_3_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_positive_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_pow_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_pow_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_prod_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_prod_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_prod_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_put_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_put_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_qr_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_rand_like_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_randint_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_randint_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_randint_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_randint_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_randint_like_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_randint_like_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_randn_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_randn_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_randn_like_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_randn_like_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_randn_like_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_ravel_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_ravel_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_ravel_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_real_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_real_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_real_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_real_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_real_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_reciprocal_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_reciprocal_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_remainder_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_repeat_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_repeat_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_repeat_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_repeat_interleave_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_repeat_interleave_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_repeat_interleave_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_repeat_interleave_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_reshape_as_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_reshape_as_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_reshape_as_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_reshape_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_reshape_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_resize__cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_resize__cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_resize_as__cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_resolve_conj_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_resolve_conj_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_resolve_conj_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_resolve_conj_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_resolve_conj_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_resolve_neg_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_roll_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_roll_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_roll_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_roll_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_rot90_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_rot90_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_round_decimals_3_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_round_decimals_neg_3_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_rsqrt_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_rsqrt_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_rsqrt_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_rsub_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_rsub_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_rsub_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_rsub_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_rsub_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_rsub_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_scalar_tensor_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_scatter_add_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_scatter_add_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_scatter_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_scatter_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_scatter_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_scatter_reduce_amax_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_scatter_reduce_amax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_scatter_reduce_amin_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_scatter_reduce_amin_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_scatter_reduce_amin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_scatter_reduce_mean_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_scatter_reduce_prod_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_scatter_reduce_prod_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_scatter_reduce_prod_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_scatter_reduce_prod_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_scatter_reduce_sum_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_scatter_reduce_sum_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_scatter_reduce_sum_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_select_scatter_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_select_scatter_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_select_scatter_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_select_scatter_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sgn_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sgn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sgn_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_short_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_short_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sigmoid_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sigmoid_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sign_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sign_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sign_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_signal_windows_bartlett_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_signal_windows_blackman_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_signal_windows_general_cosine_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_signal_windows_hamming_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_signal_windows_kaiser_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_signbit_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_signbit_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_signbit_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_signbit_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sin_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sin_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sin_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sinc_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sinc_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sinh_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sinh_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sinh_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_slice_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_slice_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_slice_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_slice_scatter_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_slice_scatter_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_softmax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_softmax_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_softmax_with_dtype_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sort_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sparse_mm_reduce_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_airy_ai_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_airy_ai_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_bessel_j1_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_bessel_y0_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_bessel_y0_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_bessel_y0_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_bessel_y1_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_bessel_y1_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_bessel_y1_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_bessel_y1_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_bessel_y1_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_chebyshev_polynomial_u_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_chebyshev_polynomial_u_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_chebyshev_polynomial_v_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_chebyshev_polynomial_v_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_chebyshev_polynomial_v_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_chebyshev_polynomial_w_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_chebyshev_polynomial_w_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_hermite_polynomial_h_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_hermite_polynomial_h_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_i0e_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_i0e_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_i1_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_i1_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_i1e_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_laguerre_polynomial_l_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_laguerre_polynomial_l_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_legendre_polynomial_p_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_legendre_polynomial_p_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_log_ndtr_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_modified_bessel_i0_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_modified_bessel_i1_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_modified_bessel_i1_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_modified_bessel_k0_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_modified_bessel_k0_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_modified_bessel_k1_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_modified_bessel_k1_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_modified_bessel_k1_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_modified_bessel_k1_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_modified_bessel_k1_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_ndtr_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_ndtri_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_ndtri_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_polygamma_special_polygamma_n_0_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_scaled_modified_bessel_k0_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_scaled_modified_bessel_k0_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_scaled_modified_bessel_k0_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_scaled_modified_bessel_k1_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_scaled_modified_bessel_k1_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_scaled_modified_bessel_k1_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_shifted_chebyshev_polynomial_t_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_shifted_chebyshev_polynomial_u_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_shifted_chebyshev_polynomial_v_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_shifted_chebyshev_polynomial_w_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_shifted_chebyshev_polynomial_w_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_spherical_bessel_j0_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_spherical_bessel_j0_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_spherical_bessel_j0_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_xlog1py_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_special_xlog1py_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_split_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_split_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_split_list_args_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_split_list_args_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_split_with_sizes_copy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_split_with_sizes_copy_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_split_with_sizes_copy_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_split_with_sizes_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_split_with_sizes_copy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sqrt_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_square_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_square_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_square_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_squeeze_copy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_squeeze_copy_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_squeeze_copy_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_squeeze_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_squeeze_copy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_squeeze_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_squeeze_copy_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_squeeze_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_squeeze_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_squeeze_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_squeeze_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_squeeze_multiple_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_squeeze_multiple_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_stack_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_stack_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_stack_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_std_mean_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_std_mean_unbiased_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_std_unbiased_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_std_unbiased_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_std_unbiased_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sub_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sum_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sum_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sum_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sum_to_size_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sum_to_size_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sum_to_size_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_sum_to_size_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_svd_lowrank_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_svd_lowrank_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_t_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_t_copy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_t_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_t_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_t_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_take_along_dim_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_take_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_tan_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_tanh_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_tanh_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_tanh_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_tensor_split_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_tensordot_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_tile_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_tile_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_tile_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_to_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_to_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_to_sparse_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_topk_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_torch__scaled_mm_v2_cuda_float8_e4m3fn, test/test_meta.py::TestMetaCUDA::test_meta_inplace_torch_ops_aten__safe_softmax_default_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_trace_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_trace_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_transpose_copy_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_transpose_copy_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_transpose_copy_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_transpose_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_transpose_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_transpose_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_transpose_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_trapezoid_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_trapezoid_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_trapezoid_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_trapz_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_trapz_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_tril_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_tril_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_tril_indices_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_triu_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_true_divide_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_true_divide_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_true_divide_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_true_divide_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_true_divide_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_trunc_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_trunc_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_trunc_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_unbind_copy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_unbind_copy_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_unflatten_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_unflatten_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_unflatten_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_unflatten_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_unfold_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_uniform_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_uniform_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_unique_consecutive_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_unique_cuda_uint32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_unravel_index_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_unsafe_chunk_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_unsafe_chunk_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_unsqueeze_copy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_unsqueeze_copy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_unsqueeze_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_unsqueeze_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_unsqueeze_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_var_mean_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_var_mean_unbiased_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_var_unbiased_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_vdot_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_view_as_complex_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_view_as_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_view_as_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_view_as_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_view_as_real_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_view_copy_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_view_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_view_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_view_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_view_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_view_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_view_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_inplace_vsplit_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_vsplit_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_vsplit_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_vstack_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_vstack_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_vstack_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_where_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_where_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_where_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_xlogy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_xlogy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_zero__cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_zeros_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_zeros_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_inplace_zeros_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_zeros_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_zeros_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_inplace_zeros_like_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_inplace_zeros_like_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_inplace_zeros_like_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_inplace_zeros_like_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_H_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_T_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_T_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_T_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_T_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace___getitem___cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace___getitem___cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace___getitem___cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace___getitem___cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace___radd___cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace___radd___cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace___radd___cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace___rand___cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace___rdiv___cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace___rmatmul___cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace___rmatmul___cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace___rmod___cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace___rmod___cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace___rmod___cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace___rmul___cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace___rmul___cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace___rpow___cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace___rpow___cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace___rpow___cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace___rsub___cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace___rsub___cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace___rxor___cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace___rxor___cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace__batch_norm_with_update_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__batch_norm_with_update_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace__chunk_cat_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace__chunk_cat_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace__chunk_cat_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace__chunk_cat_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_abs_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_abs_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_acos_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_acos_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_add_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_add_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_addcdiv_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_addcdiv_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_addcmul_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_addcmul_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_asin_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_asin_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_asin_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_asin_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_atan_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_atan_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_ceil_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_ceil_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_ceil_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_clamp_max_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_clamp_max_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_clamp_max_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_clamp_min_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_cos_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_cos_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_cos_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_cos_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_cosh_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_cosh_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_cosh_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_div_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_div_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_div_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_erf_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_erf_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_erf_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_erfc_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_erfc_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_erfc_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_erfc_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_exp_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_exp_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_exp_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_expm1_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_expm1_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_floor_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_floor_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_floor_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_floor_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_frac_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_lerp_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_lerp_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_lgamma_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_lgamma_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_lgamma_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_lgamma_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_lgamma_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_log10_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_log10_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_log10_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_log10_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_log10_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_log1p_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_log1p_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_log1p_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_log2_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_log2_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_log_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_log_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_log_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_log_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_log_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_max_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_maximum_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_minimum_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_mul_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_mul_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_neg_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_neg_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_neg_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_neg_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_norm_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_pow_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_pow_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_pow_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_reciprocal_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_reciprocal_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_reciprocal_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_reciprocal_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_round_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_sigmoid_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_sigmoid_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_sign_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_sign_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_sin_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_sin_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_sin_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_sinh_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_sinh_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_sinh_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_sqrt_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_sub_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_tan_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_tan_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_tanh_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_tanh_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_tanh_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_trunc_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_trunc_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_trunc_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_zero_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__foreach_zero_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace__native_batch_norm_legit_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__native_batch_norm_legit_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace__segment_reduce_lengths_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace__segment_reduce_lengths_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__unsafe_masked_index_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace__unsafe_masked_index_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace__unsafe_masked_index_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace__unsafe_masked_index_put_accumulate_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_abs_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_acos_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_acosh_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_acosh_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_add_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_add_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_add_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_add_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_addbmm_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_addbmm_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_addbmm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_addcdiv_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_addcmul_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_addcmul_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_addcmul_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_addcmul_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_addmm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_addmm_decomposed_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_addmm_decomposed_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_addmm_decomposed_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_addmm_decomposed_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_addr_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_addr_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_alias_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_alias_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_all_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_all_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_all_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_allclose_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_allclose_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_amax_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_amax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_amax_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_amax_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_amin_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_aminmax_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_angle_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_angle_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_any_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_arange_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_argmax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_argmax_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_argsort_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_argwhere_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_as_strided_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_as_strided_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_as_strided_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_as_strided_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_as_strided_partial_views_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_as_strided_scatter_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_as_strided_scatter_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_asin_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_asin_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_asin_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_asin_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_asinh_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_atan2_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_atan2_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_atan_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_atan_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_atan_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_atanh_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_atanh_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_atleast_1d_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_atleast_2d_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_atleast_2d_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_atleast_3d_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_atleast_3d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_baddbmm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_baddbmm_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_baddbmm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_bernoulli_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_bernoulli_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_bernoulli_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_bfloat16_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_bfloat16_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_bfloat16_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_bfloat16_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_bfloat16_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_bitwise_not_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_bitwise_right_shift_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_bitwise_right_shift_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_block_diag_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_block_diag_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_block_diag_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_bool_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_bool_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_bool_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_broadcast_tensors_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_broadcast_tensors_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_broadcast_tensors_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_broadcast_tensors_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_broadcast_tensors_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_broadcast_to_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_broadcast_to_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_broadcast_to_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_broadcast_to_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_bucketize_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_bucketize_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_bucketize_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_byte_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_byte_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_byte_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cartesian_prod_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cartesian_prod_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cat_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cat_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cat_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cat_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cat_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cat_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cauchy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cdouble_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cdouble_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cdouble_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cdouble_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ceil_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ceil_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cfloat_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cfloat_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cfloat_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_chalf_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_chalf_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_char_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cholesky_inverse_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cholesky_solve_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cholesky_solve_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cholesky_solve_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_chunk_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_chunk_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_chunk_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_chunk_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_clamp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_clamp_max_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_clamp_min_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_clamp_min_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_clamp_min_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_clamp_min_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_clamp_min_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_clone_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_clone_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_clone_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_column_stack_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_column_stack_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_combinations_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_combinations_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_combinations_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_conj_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_conj_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_conj_physical_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_conj_physical_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_conj_physical_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_constant_pad_nd_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_constant_pad_nd_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_contiguous_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_contiguous_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_contiguous_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_contiguous_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_contiguous_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_copysign_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_corrcoef_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_corrcoef_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cosh_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_count_nonzero_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_count_nonzero_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cov_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cov_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cross_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cross_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cummax_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cummax_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cummin_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cummin_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cummin_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cumsum_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cumulative_trapezoid_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_cumulative_trapezoid_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_deg2rad_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_deg2rad_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_diag_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_diag_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_diag_embed_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_diag_embed_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_diag_embed_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_diagflat_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_diagflat_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_diagflat_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_diagonal_copy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_diagonal_copy_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_diagonal_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_diagonal_scatter_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_diagonal_scatter_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_diff_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_diff_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_diff_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_digamma_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_digamma_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_digamma_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_dist_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_dist_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_dist_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_div_no_rounding_mode_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_div_no_rounding_mode_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_div_no_rounding_mode_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_div_trunc_rounding_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_div_trunc_rounding_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_dot_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_dot_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_double_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_double_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_double_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_double_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_dsplit_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_dsplit_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_dstack_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_dstack_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_dstack_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_dstack_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_einsum_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_einsum_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_einsum_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_empty_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_empty_like_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_empty_like_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_empty_permuted_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_empty_permuted_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_empty_permuted_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_empty_strided_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_eq_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_eq_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_equal_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_erf_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_erf_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_erfc_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_erfinv_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_erfinv_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_exp2_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_exp2_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_exp_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_exp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_exp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_expand_as_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_expand_as_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_expand_as_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_expand_copy_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_expand_copy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_expand_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_expand_copy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_expm1_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_expm1_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_exponential_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_eye_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_eye_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_fft2_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_fft2_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_fft_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_fft_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_fftn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_fftn_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_fftshift_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_hfft2_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_hfft_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_hfftn_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_hfftn_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_ifft2_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_ifft2_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_ifft2_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_ifft_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_ifftn_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_ifftn_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_ifftn_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_ifftshift_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_ifftshift_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_ifftshift_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_ihfft2_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_ihfft2_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_ihfft2_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_ihfft2_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_ihfft2_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_ihfft2_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_ihfft_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_ihfftn_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_ihfftn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_irfft2_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_irfft_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_irfft_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_irfft_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_irfft_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_irfft_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_irfft_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_irfftn_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_irfftn_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_irfftn_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_rfft_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fft_rfftn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fill_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fill_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fill_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_flatten_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_flatten_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_flip_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_flip_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fliplr_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fliplr_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_flipud_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_flipud_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_flipud_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_flipud_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_float_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_float_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_float_power_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_floor_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_floor_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_floor_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_floor_divide_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_floor_divide_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fmax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fmax_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fmax_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fmin_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fmin_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_fmod_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_frexp_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_frexp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_full_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_full_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_full_like_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_full_like_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_full_like_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_gather_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_gather_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_gcd_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ge_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ge_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ge_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ge_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_geometric_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_geometric_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_gradient_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_grid_sampler_3d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_grid_sampler_3d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_grid_sampler_3d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_gt_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_half_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_half_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_hash_tensor_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_hash_tensor_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_heaviside_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_heaviside_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_histc_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_histc_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_histc_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_hsplit_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_hstack_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_hstack_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_hstack_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_hstack_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_i0_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_i0_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_i0_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_add_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_add_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_copy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_copy_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_copy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_copy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_fill_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_fill_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_fill_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_put_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_put_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_put_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_reduce_amax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_reduce_amax_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_reduce_amin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_reduce_mean_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_reduce_mean_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_reduce_mean_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_reduce_prod_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_select_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_index_select_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_inner_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_int_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_int_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isclose_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isclose_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isfinite_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isfinite_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isin_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isinf_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isnan_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isnan_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isneginf_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isposinf_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isposinf_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isposinf_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isposinf_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isposinf_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isreal_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isreal_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isreal_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isreal_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isreal_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_isreal_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_item_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_item_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_item_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_jiterator_2inputs_2outputs_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_jiterator_4inputs_with_extra_args_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_jiterator_4inputs_with_extra_args_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_jiterator_4inputs_with_extra_args_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_jiterator_4inputs_with_extra_args_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_jiterator_binary_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_jiterator_binary_return_by_ref_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_jiterator_binary_return_by_ref_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_jiterator_unary_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_jiterator_unary_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_jiterator_unary_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_kron_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_kron_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_kron_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_kron_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_kthvalue_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_kthvalue_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_lcm_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_lcm_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_lcm_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ldexp_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ldexp_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ldexp_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_le_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_le_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_lerp_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_cross_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_cross_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_diagonal_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_diagonal_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_diagonal_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_diagonal_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_ldl_factor_ex_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_ldl_factor_ex_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_ldl_solve_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_lstsq_grad_oriented_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_lstsq_grad_oriented_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_lu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_lu_factor_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_lu_factor_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_lu_solve_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_matrix_norm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_matrix_rank_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_matrix_rank_hermitian_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_multi_dot_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_multi_dot_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_norm_subgradients_at_zero_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_norm_subgradients_at_zero_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_pinv_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_solve_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_solve_ex_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_solve_triangular_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_solve_triangular_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_solve_triangular_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_svd_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_svdvals_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_svdvals_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_tensorinv_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_linalg_tensorsolve_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_log10_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_log1p_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_log1p_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_log1p_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_log2_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_log_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_log_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_log_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_log_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_log_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_log_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_log_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_log_softmax_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_log_softmax_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_log_softmax_with_dtype_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_log_softmax_with_dtype_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_log_softmax_with_dtype_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_logical_and_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_logical_and_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_logical_and_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_logical_and_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_logical_not_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_logical_or_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_logical_or_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_logical_or_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_logit_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_logit_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_logit_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_logspace_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_logspace_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_logspace_tensor_overload_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_logspace_tensor_overload_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_logsumexp_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_logsumexp_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_logsumexp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_long_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_long_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_lt_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_lu_unpack_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_mH_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_mT_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_amax_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_amin_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_amin_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_amin_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_argmax_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_argmax_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_cumprod_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_cumprod_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_cumprod_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_cumprod_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_cumsum_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_cumsum_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_fill_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_logaddexp_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_logsumexp_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_logsumexp_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_mean_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_median_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_median_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_median_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_normalize_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_normalize_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_normalize_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_prod_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_prod_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_prod_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_scatter_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_scatter_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_scatter_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_scatter_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_scatter_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_select_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_select_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_select_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_softmax_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_softmax_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_softmax_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_softmin_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_softmin_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_std_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_std_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_std_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_std_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_sum_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_sum_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_var_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_masked_var_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_matmul_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_matmul_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_matmul_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_matrix_exp_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_matrix_exp_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_matrix_exp_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_max_binary_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_max_pool2d_with_indices_backward_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_max_reduction_no_dim_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_max_reduction_no_dim_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_max_reduction_with_dim_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_max_reduction_with_dim_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_maximum_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_maximum_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_maximum_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_maximum_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_mean_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_mean_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_meshgrid_list_of_tensors_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_meshgrid_variadic_tensors_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_meshgrid_variadic_tensors_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_meshgrid_variadic_tensors_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_meshgrid_variadic_tensors_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_min_binary_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_min_reduction_no_dim_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_min_reduction_no_dim_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_min_reduction_with_dim_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_min_reduction_with_dim_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_min_reduction_with_dim_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_mm_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_mm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_mm_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_movedim_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_mul_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_mv_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_mvlgamma_mvlgamma_p_1_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_mvlgamma_mvlgamma_p_3_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_mvlgamma_mvlgamma_p_5_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_mvlgamma_mvlgamma_p_5_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nan_to_num_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nanmean_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nanmean_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nanmedian_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nansum_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nansum_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_narrow_copy_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_narrow_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_narrow_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_narrow_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_narrow_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_narrow_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_native_dropout_backward_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ne_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ne_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ne_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ne_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_neg_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_neg_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_neg_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_neg_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_new_empty_strided_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_new_full_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_new_ones_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_new_ones_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_new_ones_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_new_ones_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_new_ones_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_new_zeros_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_new_zeros_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_new_zeros_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nextafter_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nextafter_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nextafter_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_adaptive_avg_pool1d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_adaptive_avg_pool2d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_adaptive_max_pool2d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_adaptive_max_pool2d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_alpha_dropout_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_avg_pool1d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_avg_pool2d_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_avg_pool3d_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_avg_pool3d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_batch_norm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_batch_norm_without_cudnn_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_binary_cross_entropy_with_logits_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_channel_shuffle_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_channel_shuffle_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_channel_shuffle_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_conv1d_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_conv2d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_conv2d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_conv3d_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_conv_transpose1d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_conv_transpose3d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_cosine_embedding_loss_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_cosine_embedding_loss_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_cross_entropy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_dropout3d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_dropout_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_elu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_elu_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_embedding_bag_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_embedding_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_feature_alpha_dropout_without_train_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_feature_alpha_dropout_without_train_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_feature_alpha_dropout_without_train_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_feature_alpha_dropout_without_train_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_grid_sample_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_hardshrink_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_hardsigmoid_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_hardswish_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_instance_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_interpolate_bilinear_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_interpolate_linear_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_interpolate_linear_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_interpolate_nearest-exact_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_interpolate_nearest_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_interpolate_nearest_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_l1_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_layer_norm_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_linear_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_margin_ranking_loss_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_margin_ranking_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_max_pool1d_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_max_unpool3d_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_mse_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_multi_head_attention_forward_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_multi_head_attention_forward_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_multilabel_soft_margin_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_nll_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_nll_loss_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_nll_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_normalize_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_pad_circular_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_pad_circular_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_pad_circular_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_pad_constant_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_pad_constant_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_pad_reflect_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_pad_reflect_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_pad_replicate_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_pad_replicate_negative_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_pad_replicate_negative_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_pixel_shuffle_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_pixel_shuffle_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_pixel_unshuffle_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_poisson_nll_loss_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_poisson_nll_loss_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_prelu_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_relu_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_relu_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_rms_norm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_selu_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_silu_complex_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_silu_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_silu_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_silu_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_smooth_l1_loss_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_soft_margin_loss_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_softmin_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_softmin_with_dtype_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_softmin_with_dtype_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_softplus_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_softplus_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_softshrink_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_softsign_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_softsign_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_softsign_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_tanhshrink_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_threshold_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_triplet_margin_loss_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_triplet_margin_with_distance_loss_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_triplet_margin_with_distance_loss_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_unfold_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_upsample_bilinear_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nn_functional_upsample_nearest_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nonzero_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nonzero_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nonzero_static_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nonzero_static_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nonzero_static_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_nonzero_static_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_norm_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_norm_fro_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_norm_inf_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_norm_inf_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_norm_inf_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_norm_inf_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_normal_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_normal_in_place_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_normal_in_place_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_normal_in_place_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_normal_number_mean_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ones_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ones_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ones_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ones_like_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ones_like_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ones_like_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ones_like_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_ones_like_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_outer_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_outer_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_permute_copy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_permute_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_permute_copy_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_permute_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_permute_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_permute_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_permute_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_permute_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_polygamma_polygamma_n_0_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_polygamma_polygamma_n_0_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_polygamma_polygamma_n_0_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_polygamma_polygamma_n_1_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_polygamma_polygamma_n_1_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_polygamma_polygamma_n_2_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_polygamma_polygamma_n_4_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_polygamma_polygamma_n_4_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_polygamma_polygamma_n_4_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_positive_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_pow_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_prod_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_prod_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_prod_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_prod_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_put_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_put_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_rad2deg_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_rand_like_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_randint_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_randint_like_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_randn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_randn_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_randn_like_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_randn_like_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_real_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_real_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_reciprocal_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_reciprocal_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_remainder_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_remainder_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_renorm_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_renorm_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_repeat_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_repeat_interleave_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_repeat_interleave_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_repeat_interleave_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_repeat_interleave_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_reshape_as_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_reshape_as_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_reshape_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_reshape_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_resize__cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_resize_as__cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_resize_as__cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_resolve_conj_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_resolve_neg_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_resolve_neg_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_resolve_neg_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_roll_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_rot90_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_rot90_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_rot90_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_round_decimals_0_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_round_decimals_3_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_round_decimals_3_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_round_decimals_neg_3_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_rsqrt_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_rsqrt_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_rsqrt_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_rsqrt_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_rsub_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_rsub_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scalar_tensor_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scalar_tensor_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scalar_tensor_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_add_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_add_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_add_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_reduce_amax_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_reduce_amax_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_reduce_amin_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_reduce_mean_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_reduce_mean_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_reduce_prod_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_reduce_prod_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_reduce_prod_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_reduce_prod_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_reduce_sum_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_reduce_sum_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_scatter_reduce_sum_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_searchsorted_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_searchsorted_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_searchsorted_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_searchsorted_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_select_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_select_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_select_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_select_scatter_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_select_scatter_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_select_scatter_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sgn_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sgn_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sgn_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sgn_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sgn_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_short_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sigmoid_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sigmoid_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sign_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sign_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_signal_windows_blackman_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_signal_windows_cosine_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_signal_windows_gaussian_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_signal_windows_general_cosine_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_signal_windows_kaiser_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_signbit_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_signbit_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sin_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sin_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sin_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sinc_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sinh_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_slice_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_slice_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_slice_scatter_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_slice_scatter_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_softmax_with_dtype_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sparse_sampled_addmm_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_airy_ai_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_airy_ai_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_bessel_j0_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_bessel_j0_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_bessel_j1_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_bessel_j1_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_bessel_j1_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_bessel_y0_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_bessel_y1_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_bessel_y1_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_bessel_y1_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_chebyshev_polynomial_t_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_chebyshev_polynomial_t_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_chebyshev_polynomial_u_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_chebyshev_polynomial_u_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_chebyshev_polynomial_u_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_chebyshev_polynomial_v_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_chebyshev_polynomial_w_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_chebyshev_polynomial_w_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_entr_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_entr_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_erfcx_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_erfcx_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_hermite_polynomial_h_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_hermite_polynomial_h_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_hermite_polynomial_he_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_i0e_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_i1_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_i1_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_i1e_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_i1e_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_i1e_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_i1e_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_laguerre_polynomial_l_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_legendre_polynomial_p_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_legendre_polynomial_p_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_log_ndtr_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_log_ndtr_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_log_ndtr_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_log_ndtr_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_modified_bessel_i0_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_modified_bessel_i1_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_modified_bessel_i1_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_modified_bessel_k0_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_modified_bessel_k1_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_modified_bessel_k1_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_ndtr_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_ndtr_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_ndtr_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_ndtr_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_ndtr_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_ndtri_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_polygamma_special_polygamma_n_0_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_scaled_modified_bessel_k0_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_scaled_modified_bessel_k0_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_scaled_modified_bessel_k1_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_shifted_chebyshev_polynomial_t_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_shifted_chebyshev_polynomial_u_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_shifted_chebyshev_polynomial_w_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_spherical_bessel_j0_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_spherical_bessel_j0_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_zeta_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_zeta_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_zeta_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_special_zeta_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_split_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_split_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_split_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_split_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_split_list_args_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_split_list_args_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_split_with_sizes_copy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_split_with_sizes_copy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_split_with_sizes_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_split_with_sizes_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_split_with_sizes_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_split_with_sizes_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sqrt_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sqrt_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_square_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_square_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_squeeze_copy_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_squeeze_copy_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_squeeze_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_squeeze_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_squeeze_multiple_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_squeeze_multiple_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_squeeze_multiple_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_squeeze_multiple_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_squeeze_multiple_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_stack_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_std_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_std_mean_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_std_mean_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_std_mean_unbiased_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_std_mean_unbiased_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_std_mean_unbiased_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_std_mean_unbiased_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_stft_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sub_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sub_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sum_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sum_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sum_to_size_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_sum_to_size_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_svd_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_svd_lowrank_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_svd_lowrank_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_t_copy_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_t_copy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_t_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_t_copy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_t_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_take_along_dim_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_take_along_dim_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_take_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_take_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_take_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_take_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_tan_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_tan_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_tanh_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_tanh_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_tanh_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_tensordot_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_tile_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_to_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_to_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_to_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_to_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_to_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_to_sparse_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_to_sparse_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_to_sparse_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_to_sparse_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_topk_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_topk_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_topk_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_torch_ops_aten__safe_softmax_default_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_trace_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_trace_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_transpose_copy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_transpose_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_transpose_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_trapezoid_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_trapezoid_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_trapezoid_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_trapezoid_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_trapz_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_trapz_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_trapz_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_triangular_solve_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_tril_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_tril_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_tril_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_triu_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_triu_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_triu_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_true_divide_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_true_divide_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_true_divide_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_true_divide_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_true_divide_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_unflatten_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_unflatten_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_unfold_copy_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_unfold_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_uniform_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_uniform_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_unique_consecutive_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_unique_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_unique_cuda_uint32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_unsafe_chunk_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_unsafe_chunk_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_unsafe_chunk_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_unsafe_chunk_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_unsafe_split_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_unsafe_split_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_unsqueeze_copy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_unsqueeze_copy_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_unsqueeze_copy_cuda_uint8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_unsqueeze_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_var_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_var_mean_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_var_mean_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_var_mean_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_vdot_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_vdot_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_vdot_cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_view_as_complex_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_view_as_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_view_as_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_view_as_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_view_as_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_view_as_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_view_copy_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_view_copy_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_view_copy_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_view_copy_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_view_cuda_bfloat16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_view_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_view_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_view_cuda_int8, test/test_meta.py::TestMetaCUDA::test_meta_outplace_vstack_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_vstack_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_vstack_cuda_complex32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_vstack_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_where_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_where_cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_where_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_where_cuda_int32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_where_cuda_int64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_xlogy_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_xlogy_cuda_float32, test/test_meta.py::TestMetaCUDA::test_meta_outplace_zero__cuda_complex128, test/test_meta.py::TestMetaCUDA::test_meta_outplace_zero__cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_zero__cuda_float16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_zeros_cuda_complex64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_zeros_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_zeros_like_cuda_bool, test/test_meta.py::TestMetaCUDA::test_meta_outplace_zeros_like_cuda_float64, test/test_meta.py::TestMetaCUDA::test_meta_outplace_zeros_like_cuda_int16, test/test_meta.py::TestMetaCUDA::test_meta_outplace_zeros_like_cuda_int8, test/test_meta.py::TestMetaCUDA::test_mixed_dtype_for_native_layer_norm_backward_float16_bias_dtype2_cuda, test/test_meta.py::TestMetaCUDA::test_stride_for_index_Tensor_cuda 2025-12-04T13:47:41.0790422Z 2025-12-04T13:47:41.0790741Z Finished test_meta 1/5 ... [2025-12-04 13:47:40.421018][16444.111404908], took 24.62min 2025-12-04T13:47:41.0791804Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_meta/test_meta-029b8ed96c60bfb7.xml 2025-12-04T13:47:42.0473597Z Uploading artifacts took 1.31 seconds 2025-12-04T13:47:42.0479206Z Running test_ops_jit 1/2 ... [2025-12-04 13:47:42.047687][16445.738078537] 2025-12-04T13:47:42.0479982Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T13:47:42.0485524Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ops_jit.py', '--shard-id=1', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 13:47:42.048206] 2025-12-04T13:56:09.9600439Z 2025-12-04T13:56:09.9603200Z test_ops_jit 1/2 was successful, full logs can be found in artifacts with path test/test-reports/test_ops_jit_1.2_893406403e6f8fa5_.log 2025-12-04T13:56:09.9836657Z Running 546 items in this shard: test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_abs_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_acosh_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_asin_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_atan2_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_cat_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_clamp_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_digamma_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_div_no_rounding_mode_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_div_trunc_rounding_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_erfinv_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_i0_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_igamma_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_le_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_lgamma_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_linalg_householder_product_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_linalg_matrix_power_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_log1p_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_log_softmax_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_matmul_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_min_binary_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_movedim_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_mul_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_mvlgamma_mvlgamma_p_1_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_mvlgamma_mvlgamma_p_5_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_ne_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_nn_functional_conv1d_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_nn_functional_conv_transpose1d_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_nn_functional_conv_transpose2d_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_nn_functional_conv_transpose3d_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_nn_functional_group_norm_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_nn_functional_layer_norm_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_nn_functional_rms_norm_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_round_decimals_3_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_sigmoid_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_sinc_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_softmax_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_softmax_with_dtype_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_sub_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_tanh_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_vstack_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_jit_alias_remapping_xlogy_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_H_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit___getitem___cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit___rdiv___cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit___rmatmul___cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit___rmatmul___cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit___rpow___cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit___rpow___cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit__chunk_cat_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit__unsafe_masked_index_put_accumulate_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_acosh_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_acosh_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_add_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_addbmm_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_addcdiv_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_addcdiv_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_addcmul_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_addmm_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_addmm_decomposed_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_addmv_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_amax_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_aminmax_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_angle_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_argmin_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_argsort_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_argwhere_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_argwhere_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_as_strided_copy_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_asin_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_asin_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_asinh_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_atan2_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_atan_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_atleast_2d_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_atleast_3d_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_atleast_3d_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_baddbmm_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_baddbmm_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_bernoulli_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_block_diag_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_bool_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_broadcast_shapes_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_broadcast_tensors_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_broadcast_to_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_bucketize_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_byte_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cartesian_prod_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cat_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cat_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cauchy_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cdouble_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cfloat_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cfloat_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_char_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cholesky_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cholesky_inverse_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cholesky_solve_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_clamp_max_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_clone_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_clone_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_column_stack_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_combinations_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_conj_physical_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_constant_pad_nd_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_constant_pad_nd_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_contiguous_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_copysign_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_corrcoef_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_corrcoef_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cos_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cos_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cosh_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cosh_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_count_nonzero_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_count_nonzero_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cov_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cov_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cross_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cross_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cummax_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cumprod_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cumsum_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_cumulative_trapezoid_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_diag_embed_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_diag_embed_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_diagflat_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_diagonal_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_digamma_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_div_no_rounding_mode_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_div_no_rounding_mode_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_dot_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_dot_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_double_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_dsplit_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_dsplit_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_dstack_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_empty_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_empty_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_empty_like_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_empty_permuted_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_empty_strided_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_empty_strided_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_eq_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_eq_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_equal_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_erf_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_exp_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_expand_as_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_expand_copy_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_expand_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_expm1_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_exponential_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_eye_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fft_fft2_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fft_fftn_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fft_fftshift_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fft_fftshift_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fft_hfft2_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fft_hfft_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fft_hfft_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fft_hfftn_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fft_hfftn_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fft_ifft2_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fft_ifft2_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fft_ifft_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fft_ifftshift_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fft_ifftshift_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fft_ihfftn_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fft_irfft2_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fft_irfft_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fft_irfftn_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fft_rfft2_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_flatten_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_flip_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fliplr_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_flipud_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_float_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_float_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_float_power_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fmin_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_fmod_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_full_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_gradient_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_gradient_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_grid_sampler_2d_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_gt_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_half_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_histc_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_hsplit_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_hsplit_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_i0_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_igammac_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_imag_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_index_fill_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_index_fill_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_index_reduce_mean_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_inner_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_int_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_int_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_isclose_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_isfinite_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_isinf_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_isnan_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_isnan_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_jiterator_2inputs_2outputs_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_jiterator_binary_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_jiterator_binary_return_by_ref_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_jiterator_binary_return_by_ref_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_jiterator_unary_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_kthvalue_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_ldexp_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_lerp_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_lgamma_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_cholesky_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_cond_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_det_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_det_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_eig_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_eigh_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_eigh_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_eigvals_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_householder_product_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_inv_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_inv_ex_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_inv_ex_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_ldl_factor_ex_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_ldl_factor_ex_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_ldl_solve_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_lstsq_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_lu_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_lu_factor_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_lu_factor_ex_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_matrix_power_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_matrix_rank_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_matrix_rank_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_matrix_rank_hermitian_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_norm_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_norm_subgradients_at_zero_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_pinv_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_pinv_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_pinv_hermitian_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_qr_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_slogdet_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_slogdet_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_solve_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_solve_ex_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_svd_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_svdvals_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_tensorsolve_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_tensorsolve_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_vecdot_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_vector_norm_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_linalg_vector_norm_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_log10_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_log10_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_log1p_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_log2_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_log2_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_log_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_log_normal_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_log_softmax_with_dtype_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_log_softmax_with_dtype_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_logaddexp_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_logcumsumexp_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_logdet_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_logical_and_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_logical_not_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_logical_or_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_logspace_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_logspace_tensor_overload_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_logsumexp_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_lt_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_lu_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_lu_unpack_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_masked_amin_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_masked_argmax_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_masked_cumprod_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_masked_cumprod_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_masked_logsumexp_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_masked_median_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_masked_norm_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_masked_normalize_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_masked_normalize_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_masked_scatter_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_masked_scatter_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_masked_sum_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_masked_var_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_max_pool2d_with_indices_backward_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_max_reduction_no_dim_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_max_reduction_with_dim_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_mean_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_mean_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_meshgrid_list_of_tensors_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_meshgrid_variadic_tensors_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_min_binary_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_minimum_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_mm_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_movedim_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_movedim_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_msort_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_mul_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_mul_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_mv_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_mv_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_mvlgamma_mvlgamma_p_1_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_mvlgamma_mvlgamma_p_5_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nan_to_num_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nanmean_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nanquantile_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nansum_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_narrow_copy_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_narrow_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_narrow_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_native_layer_norm_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_ne_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_ne_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_neg_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_neg_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_new_empty_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_new_empty_strided_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_new_empty_strided_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_new_full_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_new_full_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_adaptive_avg_pool1d_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_adaptive_avg_pool3d_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_alpha_dropout_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_avg_pool1d_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_avg_pool3d_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_channel_shuffle_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_conv2d_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_conv2d_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_conv3d_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_conv3d_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_conv_transpose2d_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_conv_transpose3d_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_cosine_similarity_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_cross_entropy_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_dropout_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_elu_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_embedding_bag_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_gaussian_nll_loss_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_group_norm_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_hardsigmoid_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_hardswish_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_hardtanh_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_huber_loss_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_instance_norm_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_interpolate_bilinear_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_l1_loss_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_l1_loss_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_linear_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_local_response_norm_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_max_pool3d_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_max_unpool1d_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_max_unpool2d_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_max_unpool3d_grad_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_normalize_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_normalize_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_pad_circular_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_pad_constant_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_pad_reflect_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_pad_reflect_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_pad_replicate_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_pad_replicate_negative_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_pad_replicate_negative_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_pdist_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_pixel_shuffle_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_pixel_shuffle_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_pixel_unshuffle_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_poisson_nll_loss_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_prelu_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_rrelu_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_scaled_dot_product_attention_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_selu_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_silu_complex_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_smooth_l1_loss_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_softmin_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_softmin_with_dtype_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_softmin_with_dtype_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_softplus_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_softshrink_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_softsign_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_tanhshrink_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_threshold_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_triplet_margin_loss_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_triplet_margin_loss_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_triplet_margin_with_distance_loss_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_triplet_margin_with_distance_loss_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_unfold_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nn_functional_upsample_nearest_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_nonzero_static_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_norm_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_norm_fro_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_norm_inf_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_norm_inf_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_normal_in_place_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_normal_number_mean_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_ones_like_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_ormqr_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_permute_copy_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_permute_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_pinverse_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_polar_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_polygamma_polygamma_n_0_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_polygamma_polygamma_n_1_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_polygamma_polygamma_n_2_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_polygamma_polygamma_n_3_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_pow_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_prod_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_put_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_put_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_qr_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_rad2deg_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_rand_like_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_rand_like_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_randint_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_randn_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_randn_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_randn_like_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_randn_like_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_real_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_reciprocal_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_reshape_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_resize__cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_resize__cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_resize_as__cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_resolve_conj_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_roll_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_round_decimals_0_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_round_decimals_3_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_rsqrt_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_rsub_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_rsub_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_scalar_tensor_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_scatter_add_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_scatter_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_scatter_reduce_prod_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_scatter_reduce_sum_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_searchsorted_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_select_scatter_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_short_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_sigmoid_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_sign_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_signal_windows_bartlett_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_signal_windows_blackman_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_signal_windows_cosine_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_signal_windows_gaussian_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_signal_windows_hann_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_signal_windows_kaiser_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_signal_windows_nuttall_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_signbit_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_sinc_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_sinh_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_slice_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_softmax_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_softmax_with_dtype_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_sort_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_sparse_mm_reduce_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_special_airy_ai_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_special_bessel_j0_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_special_bessel_y0_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_special_chebyshev_polynomial_u_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_special_entr_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_special_erfcx_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_special_hermite_polynomial_h_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_special_i0e_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_special_i1e_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_special_legendre_polynomial_p_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_special_modified_bessel_i1_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_special_modified_bessel_k0_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_special_modified_bessel_k1_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_special_ndtr_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_special_scaled_modified_bessel_k1_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_special_shifted_chebyshev_polynomial_t_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_special_xlog1py_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_split_with_sizes_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_sqrt_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_square_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_squeeze_copy_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_squeeze_copy_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_squeeze_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_squeeze_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_squeeze_multiple_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_squeeze_multiple_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_std_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_std_mean_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_std_mean_unbiased_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_std_mean_unbiased_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_stft_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_stft_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_sum_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_sum_to_size_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_svd_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_t_copy_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_t_copy_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_t_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_take_along_dim_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_take_along_dim_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_tan_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_tan_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_tensor_split_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_tensordot_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_tensordot_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_tile_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_tile_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_to_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_to_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_to_sparse_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_to_sparse_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_transpose_copy_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_transpose_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_trapezoid_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_trapz_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_triangular_solve_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_tril_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_triu_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_true_divide_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_unbind_copy_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_unflatten_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_unfold_copy_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_unfold_copy_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_unfold_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_unfold_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_unsafe_split_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_unsafe_split_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_unsqueeze_copy_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_unsqueeze_copy_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_var_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_var_mean_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_var_mean_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_var_mean_unbiased_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_var_mean_unbiased_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_var_unbiased_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_vdot_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_vdot_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_vsplit_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_vstack_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_vstack_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_where_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_xlogy_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_zero__cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_zero__cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_zeros_cuda_float32, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_zeros_like_cuda_complex64, test/test_ops_jit.py::TestJitCUDA::test_variant_consistency_jit_zeros_like_cuda_float32 2025-12-04T13:56:10.0067797Z 2025-12-04T13:56:10.0068094Z Finished test_ops_jit 1/2 ... [2025-12-04 13:56:09.960494][16953.650885312], took 8.47min 2025-12-04T13:56:10.0069176Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops_jit/test_ops_jit-9f733effbdfc0369.xml 2025-12-04T13:56:10.0888030Z Running test_nestedtensor 2/4 ... [2025-12-04 13:56:10.088543][16953.778936285] 2025-12-04T13:56:10.0898814Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T13:56:10.0900055Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_nestedtensor.py', '--shard-id=2', '--num-shards=4', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 13:56:10.088936] 2025-12-04T14:06:02.2007902Z 2025-12-04T14:06:02.2008932Z test_nestedtensor 2/4 was successful, full logs can be found in artifacts with path test/test-reports/test_nestedtensor_2.4_5194ad5857dee29b_.log 2025-12-04T14:06:02.2250635Z Running 420 items in this shard: test/test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_4_max_seq_len_3_vocab_size_10, test/test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_4_max_seq_len_3_vocab_size_20, test/test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_4_max_seq_len_5_vocab_size_10, test/test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_4_max_seq_len_5_vocab_size_20, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_batch_size_4_max_seq_len_3_vocab_size_10, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_float_batch_size_4_max_seq_len_3_vocab_size_10, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_float_batch_size_4_max_seq_len_3_vocab_size_20, test/test_nestedtensor.py::TestNestedTensor::test_dim, test/test_nestedtensor.py::TestNestedTensor::test_fill_, test/test_nestedtensor.py::TestNestedTensor::test_like_functions_ones_like, test/test_nestedtensor.py::TestNestedTensor::test_nested_tensor_matching_dim, test/test_nestedtensor.py::TestNestedTensor::test_nested_view_from_buffer_overflow_errors, test/test_nestedtensor.py::TestNestedTensor::test_repr_string, test/test_nestedtensor.py::TestNestedTensor::test_to, test/test_nestedtensor.py::TestNestedTensor::test_to_padded_tensor_on_empty_tensor, test/test_nestedtensor.py::TestNestedTensor::test_unbind_0, test/test_nestedtensor.py::TestNestedTensor::test_unbind_1, test/test_nestedtensor.py::TestNestedTensor::test_zero_, test/test_nestedtensor.py::TestNestedInt::test_with_factor, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_bmm_cpu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_bmm_cuda_cuda_bfloat16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_bmm_cuda_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_bmm_noncontiguous_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_detach_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_detach_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_embedding_strided_cuda, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_empty_like_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_jagged_amax_dtypes_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_jagged_amax_dtypes_cuda_int16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_jagged_amin_dtypes_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_jagged_amin_dtypes_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_jagged_amin_dtypes_cuda_int8, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_jagged_argmax_dtypes_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_jagged_argmax_dtypes_cuda_int32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_jagged_argmax_dtypes_cuda_int64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_jagged_argmax_dtypes_cuda_uint8, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_jagged_max_dtypes_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_jagged_max_dtypes_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_jagged_max_dtypes_cuda_int64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_jagged_max_dtypes_cuda_uint8, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_jagged_min_dtypes_cuda_bfloat16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_jagged_min_dtypes_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_jagged_min_dtypes_cuda_int16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_linear_noncontiguous_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_linear_noncontiguous_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_matmul_noncontiguous_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_matmul_nt_with_broadcasted_t_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_narrow_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_add_in_place_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_indexing_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_indexing_noncontiguous_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_indexing_noncontiguous_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_mul_in_place_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_mul_in_place_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_split_with_sizes_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_sub_transpose_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_sum_dim_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_reshape_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_serialization_requires_grad_False_weights_only_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_serialization_requires_grad_True_weights_only_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_serialization_requires_grad_True_weights_only_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_to_padded_tensor_dim3_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_to_padded_tensor_noncontiguous_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_to_padded_tensor_output_size_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_to_padded_tensor_simple_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_to_padded_tensor_simple_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_transpose_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_transpose_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_unary_funcs_gelu__cuda, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_unary_funcs_isinf_cuda, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_unary_funcs_isneginf_cuda, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_unary_funcs_silu__cuda, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_unary_funcs_silu_cuda, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_unary_funcs_sin_cuda, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_view_inference_mode_interaction_cuda_float64, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_abs_backward_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_layer_norm_backward_size_1024_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_layer_norm_backward_size_256_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_layer_norm_backward_size_32_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_layer_norm_backward_size_513_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_nested_tensor_from_list_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_nested_tensor_from_padded_fused_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_nested_tensor_linear_plus_transpose_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_relu_backward_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_values_grad_with_broadcast_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_jagged_requires_grad_False_contiguous_False_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_jagged_requires_grad_True_contiguous_False_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_jagged_requires_grad_True_contiguous_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_strided_requires_grad_False_contiguous_False_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_strided_requires_grad_False_contiguous_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_strided_requires_grad_False_contiguous_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_strided_requires_grad_True_contiguous_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_strided_requires_grad_True_contiguous_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_strided_requires_grad_True_contiguous_True_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_1_layout_jagged_requires_grad_False_contiguous_True_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_1_layout_jagged_requires_grad_True_contiguous_True_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_1_layout_jagged_requires_grad_True_contiguous_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_1_layout_jagged_requires_grad_True_contiguous_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_1_layout_strided_requires_grad_False_contiguous_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_1_layout_strided_requires_grad_False_contiguous_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_1_layout_strided_requires_grad_True_contiguous_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_2_layout_jagged_requires_grad_False_contiguous_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_2_layout_jagged_requires_grad_False_contiguous_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_2_layout_jagged_requires_grad_True_contiguous_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_2_layout_jagged_requires_grad_True_contiguous_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_2_layout_jagged_requires_grad_True_contiguous_True_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_2_layout_strided_requires_grad_False_contiguous_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_2_layout_strided_requires_grad_False_contiguous_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_2_layout_strided_requires_grad_True_contiguous_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_jagged_requires_grad_False_contiguous_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_jagged_requires_grad_False_contiguous_True_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_jagged_requires_grad_False_contiguous_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_jagged_requires_grad_False_contiguous_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_jagged_requires_grad_True_contiguous_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_strided_requires_grad_False_contiguous_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_strided_requires_grad_True_contiguous_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_strided_requires_grad_True_contiguous_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_strided_requires_grad_True_contiguous_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_4_layout_jagged_requires_grad_False_contiguous_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_4_layout_strided_requires_grad_False_contiguous_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_binary_pointwise_transposed_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_broadcast_shapes_on_in_graph_constructed_njt_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_compile_preserves_metadata_cache_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_compile_with_dynamic_max_seq_len_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_compile_with_dynamic_min_seq_len_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_composite_op_with_custom_mode_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_device_dtype_transfer_updates_offsets_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_dummy_mha_with_nt_use_legacy_api_True_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_layout_construction_as_nested_tensor_components_require_grad_False_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_layout_construction_as_nested_tensor_components_require_grad_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_layout_construction_as_nested_tensor_components_require_grad_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_layout_construction_nested_tensor_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_layout_construction_nested_tensor_requires_grad_True_components_require_grad_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_layout_construction_nested_tensor_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_layout_construction_nested_tensor_requires_grad_True_components_require_grad_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_layout_construction_with_pinned_memory_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_op_different_output_shape_dim_mean_keepdim_True_requires_grad_True_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_op_different_output_shape_dim_sum_keepdim_False_requires_grad_True_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_op_different_output_shape_dim_sum_keepdim_False_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_op_different_output_shape_dim_sum_keepdim_True_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_op_different_output_shape_dim_sum_keepdim_True_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_padded_dense_conversion_kernels_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_padded_dense_conversion_kernels_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_view_from_values_offsets_requires_grad_True_values_is_view_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_layer_norm_operate_on_batch_dim_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_layer_norm_with_lengths_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_layout_under_torch_dispatch_mode_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_like_value_rand_like_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_linear_backward_memory_usage_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_nested_tensor_from_jagged_fx_trace_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_nested_tensor_from_jagged_pass_min_max_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_njt_cat_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_noncontiguous_to_noncontig_transposed_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_noncontiguous_to_noncontig_transposed_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_batch_only_different_output_shape_mean_keepdim_False_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_batch_only_different_output_shape_mean_keepdim_True_requires_grad_True_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_1_different_output_shape_mean_keepdim_True_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_1_different_output_shape_mean_keepdim_True_requires_grad_True_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_1_different_output_shape_mean_keepdim_True_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_1_different_output_shape_sum_keepdim_False_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_greater_than_1_different_output_shape_mean_transpose_offset_1_keepdim_False_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_greater_than_1_different_output_shape_mean_transpose_offset_1_keepdim_True_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_greater_than_1_different_output_shape_sum_transpose_offset_1_keepdim_True_requires_grad_True_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_greater_than_1_different_output_shape_sum_transpose_offset_2_keepdim_False_requires_grad_True_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_greater_than_1_different_output_shape_sum_transpose_offset_2_keepdim_False_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_transpose_non_ragged_dim_different_output_shape_mean_keepdim_True_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_transpose_non_ragged_dim_different_output_shape_sum_keepdim_False_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_transpose_non_ragged_dim_different_output_shape_sum_keepdim_True_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_with_lengths_different_output_shape_mean_keepdim_True_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_pin_memory_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sdpa_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sdpa_flop_counter_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sdpa_with_constant_sequence_length_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sdpa_with_packed_in_proj_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_serialization_noncontig_with_holes_weights_only_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_dim_reduce_ragged_idx_1_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_dim_reduce_ragged_idx_1_requires_grad_True_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_dim_reduce_ragged_idx_1_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_dim_reduce_ragged_idx_greater_than_1_same_output_shape_transpose_offset_2_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_dim_requires_grad_True_components_require_grad_True_log_softmax_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_dim_transpose_non_ragged_dim_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_dim_transpose_non_ragged_dim_requires_grad_True_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_dim_with_lengths_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_dim_with_lengths_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_reduce_batch_dim_requires_grad_False_components_require_grad_False_log_softmax_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_reduce_batch_dim_requires_grad_False_components_require_grad_False_softmax_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_reduce_batch_dim_requires_grad_True_components_require_grad_True_log_softmax_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_split_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sum_dim_reduce_batch_and_non_batch_keepdim_True_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sum_dim_reduce_ragged_and_non_batch_keepdim_False_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sum_dim_reduce_ragged_and_non_batch_keepdim_False_requires_grad_True_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sum_dim_reduce_ragged_and_non_batch_keepdim_True_requires_grad_True_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sum_dim_reduce_ragged_and_non_batch_keepdim_True_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_tensor_attributes_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_threshold_backward_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_copy_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_compile_nt_dim_2_requires_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_compile_nt_dim_2_requires_grad_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_compile_nt_dim_3_requires_grad_True_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_compile_nt_dim_4_requires_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_nt_dim_2_requires_grad_False_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_nt_dim_2_requires_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_nt_dim_2_requires_grad_True_cuda_bool, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_nt_dim_3_requires_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_nt_dim_4_requires_grad_False_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_nt_dim_4_requires_grad_True_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_nt_dim_4_requires_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_unary_pointwise_transposed_inputs_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_unbind_lengths_ragged_idx_2_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_unbind_lengths_ragged_idx_3_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_unbind_lengths_ragged_idx_equals_2_bad_dim_cuda, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_atanh_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_bfloat16_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_ceil_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_clamp_max_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_copysign_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_cos_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_div_floor_rounding_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_erf_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_float_power_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_fmod_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_hypot_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_lgamma_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_linalg_vector_norm_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_log1p_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_masked_amin_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_masked_select_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_maximum_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_mul_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_nan_to_num_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_nanmean_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_nansum_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_neg_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_nn_functional_celu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_nn_functional_hardsigmoid_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_nn_functional_prelu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_nn_functional_rrelu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_nn_functional_softshrink_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_polar_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_round_decimals_0_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_round_decimals_3_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_rsqrt_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_select_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_sinc_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_sinh_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_special_i0e_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_special_ndtr_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_true_divide_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_trunc_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_unsqueeze_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_var_unbiased_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_xlogy_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward___rmod___cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_atan2_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_clamp_max_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_conj_physical_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_deg2rad_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_div_floor_rounding_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_double_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_exp_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_fmin_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_log10_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_log1p_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_log_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_logaddexp_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_masked_logsumexp_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_masked_mean_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_masked_norm_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_max_binary_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_nan_to_num_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_nanmean_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_narrow_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_neg_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_nn_functional_celu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_nn_functional_hardsigmoid_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_nn_functional_softshrink_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_polygamma_polygamma_n_4_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_pow_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_reciprocal_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_remainder_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_round_decimals_3_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_sigmoid_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_special_i1e_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_split_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_squeeze_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_std_unbiased_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_trunc_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward___rmod___cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward___rmul___cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward___rpow___cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_acos_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_all_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_amin_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_angle_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_any_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_atan2_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_atanh_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_bfloat16_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_bmm_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_bool_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_cdouble_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_ceil_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_cfloat_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_chalf_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_chunk_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_div_no_rounding_mode_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_exp2_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_exp_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_float_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_frexp_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_hash_tensor_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_heaviside_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_hypot_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_i0_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_igamma_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_int_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_isclose_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_log_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_logit_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_long_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_lt_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_max_binary_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_min_reduction_with_dim_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_minimum_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_mvlgamma_mvlgamma_p_3_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_mvlgamma_mvlgamma_p_5_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_nan_to_num_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_nanmean_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_neg_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_nn_functional_elu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_nn_functional_embedding_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_nn_functional_prelu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_nn_functional_selu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_nn_functional_softplus_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_nn_functional_softsign_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_nn_functional_threshold_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_polar_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_polygamma_polygamma_n_2_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_positive_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_real_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_remainder_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_round_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_select_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_sign_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_sinc_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_sinh_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_bessel_y0_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_erfcx_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_i1_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_i1e_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_laguerre_polynomial_l_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_ndtr_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_scaled_modified_bessel_k0_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_shifted_chebyshev_polynomial_u_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_spherical_bessel_j0_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_split_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_sqrt_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_squeeze_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_sub_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_sum_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_to_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_var_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward___radd___cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward___rdiv___cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_asin_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_atanh_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_byte_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_cfloat_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_char_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_cosh_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_deg2rad_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_eq_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_erf_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_erfc_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_fmax_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_fmin_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_frexp_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_gt_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_half_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_heaviside_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_int_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_isfinite_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_lgamma_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_log1p_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_logaddexp_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_logical_or_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_masked_logsumexp_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_masked_mean_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_nanmean_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_narrow_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_ne_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_nn_functional_hardshrink_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_nn_functional_rms_norm_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_nn_functional_softshrink_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_nn_functional_threshold_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_polar_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_polygamma_polygamma_n_0_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_positive_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_round_decimals_3_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_round_decimals_neg_3_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_sinc_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_airy_ai_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_bessel_y0_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_chebyshev_polynomial_t_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_chebyshev_polynomial_u_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_chebyshev_polynomial_w_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_hermite_polynomial_h_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_hermite_polynomial_he_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_modified_bessel_k0_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_ndtr_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_ndtri_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_scaled_modified_bessel_k1_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_shifted_chebyshev_polynomial_t_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_xlog1py_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_zeta_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_split_with_sizes_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_sum_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_tanh_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_trunc_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_nested_tensor_non_contiguous_mutation_cuda 2025-12-04T14:06:02.2486810Z 2025-12-04T14:06:02.2487164Z Finished test_nestedtensor 2/4 ... [2025-12-04 14:06:02.201409][17545.891798418], took 9.87min 2025-12-04T14:06:02.2488599Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_nestedtensor/test_nestedtensor-8e0ff602fd37df16.xml 2025-12-04T14:06:02.3329205Z Running test_public_bindings 1/1 ... [2025-12-04 14:06:02.332252][17546.022641744] 2025-12-04T14:06:02.3329766Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T14:06:02.3330969Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_public_bindings.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 14:06:02.332737] 2025-12-04T14:06:20.3761317Z 2025-12-04T14:06:20.3762333Z test_public_bindings 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_public_bindings_1.1_c7580692556b690c_.log 2025-12-04T14:06:20.3764673Z Running 4 items in this shard: test/test_public_bindings.py::TestPublicBindings::test_correct_module_names, test/test_public_bindings.py::TestPublicBindings::test_modules_can_be_imported, test/test_public_bindings.py::TestPublicBindings::test_no_new_bindings, test/test_public_bindings.py::TestPublicBindings::test_no_new_reexport_callables 2025-12-04T14:06:20.3766376Z 2025-12-04T14:06:20.3766702Z Finished test_public_bindings 1/1 ... [2025-12-04 14:06:20.375911][17564.066304771], took 0.30min 2025-12-04T14:06:20.4150327Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_public_bindings/test_public_bindings-6fed1053bfc25f05.xml 2025-12-04T14:06:20.4890064Z Running test_ops 1/11 ... [2025-12-04 14:06:20.488698][17564.179089303] 2025-12-04T14:06:20.4890596Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T14:06:20.4893602Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ops.py', '--shard-id=1', '--num-shards=11', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 14:06:20.489130] 2025-12-04T14:28:08.5655657Z 2025-12-04T14:28:08.5658690Z test_ops 1/11 was successful, full logs can be found in artifacts with path test/test-reports/test_ops_1.11_b5934d9d9b184ff9_.log 2025-12-04T14:28:08.7065120Z Running 3172 items in this shard: test/test_ops.py::TestCommonCUDA::test_compare_cpu___rxor___cuda_int64, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs__conversions_bfloat16_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs__conversions_cfloat_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs__conversions_long_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs__conversions_short_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_as_strided_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_atleast_2d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_dstack_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_expand_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_flipud_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_hsplit_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_logspace_tensor_overload_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_narrow_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_new_empty_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_nn_functional_glu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_nn_functional_softmin_with_dtype_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_renorm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_stft_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_unfold_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__segment_reduce_lengths_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_addcdiv_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_cauchy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_column_stack_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_corrcoef_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_gather_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_inner_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_linalg_cholesky_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_linalg_cholesky_ex_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_linalg_inv_ex_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_linalg_lu_factor_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_linalg_vector_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_mT_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_masked_median_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_matmul_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_new_ones_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_adaptive_max_pool2d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_cosine_similarity_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_fractional_max_pool2d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_glu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_hardtanh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_interpolate_bilinear_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_max_pool1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_max_pool2d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_triplet_margin_with_distance_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_norm_inf_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_ormqr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_pca_lowrank_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_resize_as__cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_scatter_reduce_amax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_special_shifted_chebyshev_polynomial_u_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_special_xlog1py_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_trapezoid_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_unsafe_split_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_var_mean_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_view_as_cuda_float32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_as_strided_scatter_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_asinh_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_atan_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_atleast_3d_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_bfloat16_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_cdouble_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_diag_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_fft_ifftshift_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_new_full_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_new_ones_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_roll_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_sigmoid_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_squeeze_copy_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_unsqueeze_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_dtypes__native_batch_norm_legit_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs__conversions_int_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_as_strided_scatter_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_asin_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_bitwise_not_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_cat_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_chunk_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_clamp_min_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_constant_pad_nd_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_cosh_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_diag_embed_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_dsplit_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_dstack_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_fft_fft2_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_fft_ifftn_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_fill_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_flipud_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_isinf_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_linalg_matrix_norm_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_logical_not_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_lt_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_mean_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_meshgrid_list_of_tensors_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_new_empty_strided_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_nn_functional_gelu_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_nn_functional_nll_loss_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_nn_functional_threshold_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_positive_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_renorm_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_roll_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_select_scatter_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_special_logit_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_special_multigammaln_mvlgamma_p_5_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_special_xlog1py_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_special_zeta_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_std_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_std_mean_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_tensor_split_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_to_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_transpose_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_unflatten_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_unsqueeze_copy_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_view_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_argsort_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_atleast_3d_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_bmm_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_conj_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_cos_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_deg2rad_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_digamma_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_div_no_rounding_mode_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_einsum_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_equal_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_fft_fftshift_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_float_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_isin_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_istft_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_linalg_cholesky_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_linalg_det_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_linalg_eig_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_masked_log_softmax_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_masked_logsumexp_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_masked_sum_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_maximum_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_min_reduction_with_dim_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nanquantile_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_narrow_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_new_empty_strided_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_conv_transpose3d_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_cosine_embedding_loss_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_dropout3d_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_hardshrink_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_interpolate_nearest-exact_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_multi_head_attention_forward_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_multilabel_soft_margin_loss_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_upsample_bilinear_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nonzero_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_ones_like_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_qr_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_randn_like_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_reshape_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_scatter_add_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_sign_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_signal_windows_hamming_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_signal_windows_hann_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_signal_windows_kaiser_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_softmax_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_special_chebyshev_polynomial_w_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_special_shifted_chebyshev_polynomial_t_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_special_shifted_chebyshev_polynomial_v_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_squeeze_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_std_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_std_mean_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_torch__scaled_mm_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_torch_ops_aten__safe_softmax_default_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_tril_indices_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_triu_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_unsafe_split_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_unsqueeze_copy_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_view_as_complex_cuda, test/test_ops.py::TestCommonCUDA::test_errors___radd___cuda, test/test_ops.py::TestCommonCUDA::test_errors___rpow___cuda, test/test_ops.py::TestCommonCUDA::test_errors_logspace_tensor_overload_cuda, test/test_ops.py::TestCommonCUDA::test_errors_max_binary_cuda, test/test_ops.py::TestCommonCUDA::test_errors_mean_cuda, test/test_ops.py::TestCommonCUDA::test_errors_narrow_cuda, test/test_ops.py::TestCommonCUDA::test_errors_native_layer_norm_cuda, test/test_ops.py::TestCommonCUDA::test_errors_nn_functional_conv1d_cuda, test/test_ops.py::TestCommonCUDA::test_errors_nn_functional_huber_loss_cuda, test/test_ops.py::TestCommonCUDA::test_errors_nn_functional_max_pool3d_cuda, test/test_ops.py::TestCommonCUDA::test_errors_nn_functional_prelu_cuda, test/test_ops.py::TestCommonCUDA::test_errors_normal_in_place_cuda, test/test_ops.py::TestCommonCUDA::test_errors_polar_cuda, test/test_ops.py::TestCommonCUDA::test_errors_remainder_cuda, test/test_ops.py::TestCommonCUDA::test_errors_rot90_cuda, test/test_ops.py::TestCommonCUDA::test_errors_sparse_mul_layout0_cuda, test/test_ops.py::TestCommonCUDA::test_errors_t_copy_cuda, test/test_ops.py::TestCommonCUDA::test_errors_view_copy_cuda, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_expm1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_fft_fft2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_fft_ifft2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_frexp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_index_add_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_kron_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_linalg_cholesky_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_linalg_eigh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_linalg_norm_subgradients_at_zero_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_linalg_tensorsolve_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_logical_not_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_lt_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_lu_solve_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_nn_functional_softplus_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_round_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_scatter_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_searchsorted_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_sinc_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_softmax_with_dtype_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_special_airy_ai_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_special_hermite_polynomial_he_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_unfold_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_vdot_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_zeros_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices___rpow___cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices__unsafe_masked_index_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_abs_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_abs_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_add_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_argmax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_as_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_asinh_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_atanh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_bincount_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_broadcast_tensors_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_cdouble_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_cfloat_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_chunk_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_combinations_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_copysign_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_cov_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_cumprod_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_cumsum_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_diag_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_diag_embed_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_diff_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_dist_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_einsum_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_empty_strided_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_eq_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_expand_as_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_expm1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_fft_ifftshift_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_fft_rfft2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_fft_rfft_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_fmax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_full_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_grid_sampler_2d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_i0_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_index_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_isfinite_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_isin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_isinf_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_isreal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_lcm_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_linalg_matrix_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_linalg_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_log1p_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_log_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_log_softmax_with_dtype_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_logical_and_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_logit_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_logspace_tensor_overload_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_logsumexp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_masked_amax_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_masked_logaddexp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_masked_scatter_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_maximum_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_meshgrid_list_of_tensors_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_minimum_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_mvlgamma_mvlgamma_p_3_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_new_zeros_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_adaptive_avg_pool3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_binary_cross_entropy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_group_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_interpolate_bicubic_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_logsigmoid_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_multi_head_attention_forward_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_multi_margin_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_nll_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_pairwise_distance_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_poisson_nll_loss_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_soft_margin_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_softmin_with_dtype_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_norm_nuc_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_polygamma_polygamma_n_1_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_polygamma_polygamma_n_2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_polygamma_polygamma_n_4_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_positive_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_put_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_rad2deg_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_real_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_reciprocal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_scalar_tensor_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_scatter_reduce_amin_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_signal_windows_bartlett_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_sinc_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_entr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_hermite_polynomial_h_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_modified_bessel_k1_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_ndtr_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_polygamma_special_polygamma_n_0_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_xlog1py_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_split_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_split_with_sizes_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_square_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_squeeze_multiple_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_std_mean_unbiased_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_std_unbiased_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_svd_lowrank_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_to_sparse_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_triu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_view_as_complex_cuda_float32, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_H_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values___getitem___cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values___ror___cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values__chunk_cat_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_add_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_amax_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_conj_physical_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_dsplit_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_fft_fftn_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_fft_ihfft_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_fft_irfft_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_flip_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_fmin_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_full_like_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_gather_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_ge_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_hash_tensor_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_isnan_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_log1p_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_log2_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_logical_not_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_max_reduction_no_dim_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_maximum_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_nn_functional_channel_shuffle_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_nn_functional_pad_constant_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_nn_functional_unfold_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_put_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_rad2deg_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_sinc_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_sort_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_special_chebyshev_polynomial_u_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_special_chebyshev_polynomial_v_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_special_erfcx_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_special_spherical_bessel_j0_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_squeeze_copy_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_sum_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_t_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_unbind_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_vsplit_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_zero__cuda_bool, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples___rand___cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples___rxor___cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_addbmm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_addcdiv_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_addcdiv_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_addcmul_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_addr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_amin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_arange_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_argmin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_argsort_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_as_strided_scatter_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_asinh_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_atan_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_bincount_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_broadcast_shapes_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_bucketize_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_byte_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_byte_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_char_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_clamp_min_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_clone_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_combinations_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_conj_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_cummin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_cumprod_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_deg2rad_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_diag_embed_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_diagonal_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_diagonal_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_div_trunc_rounding_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_double_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_expand_as_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_expand_as_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_expand_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_expm1_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_fft_fft_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_fft_hfft_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_fft_ifftshift_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_fft_rfft_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_fill_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_flipud_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_float_power_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_fmax_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_frac_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_full_like_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_geqrf_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_histc_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_hstack_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_index_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_index_fill_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_index_reduce_mean_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_index_select_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_inner_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_isneginf_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_isposinf_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_jiterator_2inputs_2outputs_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_jiterator_binary_return_by_ref_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_cholesky_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_eigh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_householder_product_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_ldl_factor_ex_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_lstsq_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_matrix_rank_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_qr_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_slogdet_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linspace_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_log_softmax_with_dtype_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_logcumsumexp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_long_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_mT_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_masked_fill_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_masked_log_softmax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_masked_scatter_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_masked_sum_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_max_pool2d_with_indices_backward_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_mean_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_meshgrid_list_of_tensors_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_meshgrid_list_of_tensors_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_meshgrid_variadic_tensors_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_min_binary_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_minimum_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_mm_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nan_to_num_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_narrow_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_narrow_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_native_dropout_backward_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_new_ones_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_batch_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_channel_shuffle_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_conv_transpose1d_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_feature_alpha_dropout_with_train_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_gelu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_glu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_interpolate_area_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_interpolate_bicubic_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_max_pool1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_max_pool3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_nll_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_one_hot_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_pad_constant_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_pad_reflect_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_pad_replicate_negative_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_softmin_with_dtype_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nonzero_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nonzero_static_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_permute_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_put_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_put_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_randint_like_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_renorm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_repeat_interleave_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_reshape_as_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_resolve_conj_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_resolve_neg_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_roll_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_rot90_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_rsub_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_scatter_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_searchsorted_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_select_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_short_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_signal_windows_nuttall_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_slice_scatter_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_special_chebyshev_polynomial_w_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_special_log_ndtr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_special_log_ndtr_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_special_modified_bessel_k0_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_special_modified_bessel_k1_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_special_ndtr_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_special_scaled_modified_bessel_k0_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_special_shifted_chebyshev_polynomial_u_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_special_spherical_bessel_j0_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_square_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_square_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_squeeze_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_squeeze_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_std_mean_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_take_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_tanh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_to_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_topk_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_topk_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_transpose_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_trapezoid_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_tril_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_tril_indices_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_trunc_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_unfold_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_unsafe_split_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_vdot_cuda_float32, test/test_ops.py::TestCommonCUDA::test_numpy_ref_broadcast_tensors_cuda_int64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_linalg_tensorsolve_cuda_float64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_nn_functional_l1_loss_cuda_float64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_ravel_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_numpy_ref_searchsorted_cuda_int64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_signal_windows_hann_cuda_float64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_squeeze_copy_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_numpy_ref_where_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_numpy_ref_where_cuda_int64, test/test_ops.py::TestCommonCUDA::test_out_T_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out___rmod___cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs__conversions_half_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_acos_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_alias_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_arange_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_broadcast_shapes_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_clone_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_cumprod_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_cumsum_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_fft_fft_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_fft_ifft2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_fliplr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_frexp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_geometric_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_i0_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_index_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_lerp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_linalg_cross_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_mul_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_nn_functional_alpha_dropout_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_nn_functional_celu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_nn_functional_mse_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_nn_functional_pdist_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_nn_functional_softmin_with_dtype_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_nn_functional_softshrink_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_special_bessel_j0_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_special_erfcx_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_special_log_ndtr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_special_logit_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_sqrt_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_triu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_var_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_view_as_complex_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_view_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_where_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__unsafe_masked_index_put_accumulate_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_addbmm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_aminmax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_any_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_atan2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_bfloat16_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_cov_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_deg2rad_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_diagflat_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_diff_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_einsum_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_empty_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_eye_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_fft_ifftn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_fft_ihfft_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_geometric_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_index_add_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_linalg_inv_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_linalg_vector_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_log1p_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_log_normal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_logical_xor_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_logspace_tensor_overload_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_logsumexp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_masked_median_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_masked_scatter_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_median_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_meshgrid_list_of_tensors_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_movedim_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_native_batch_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_native_dropout_backward_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_neg_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_adaptive_max_pool3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_batch_norm_without_cudnn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_conv_transpose3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_cosine_similarity_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_embedding_bag_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_hardsigmoid_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_interpolate_area_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_max_pool1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_multilabel_margin_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_normalize_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_pad_replicate_negative_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_relu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_softmin_with_dtype_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_upsample_nearest_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_ones_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_ormqr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_polygamma_polygamma_n_2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_polygamma_polygamma_n_3_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_positive_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_addmm_decomposed_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_bmm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_cholesky_inverse_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_cholesky_solve_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_cholesky_solve_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_cosh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_div_floor_rounding_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_expand_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_fft_ifft_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_index_reduce_amax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_ldexp_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_lgamma_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_cond_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_householder_product_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_lu_factor_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_multi_dot_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_solve_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_solve_triangular_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linspace_tensor_overload_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_log_softmax_with_dtype_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_logspace_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_mean_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_mm_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_mul_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_nn_functional_avg_pool3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_permute_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_polygamma_polygamma_n_3_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_round_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_scatter_add_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_sigmoid_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_sparse_sampled_addmm_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_sqrt_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_squeeze_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_stack_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_t_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_take_along_dim_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_triu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_vdot_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_view_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_vstack_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_scatter_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_scatter_reduce_prod_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_signal_windows_general_hamming_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_sin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_special_hermite_polynomial_h_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_special_modified_bessel_k1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_special_ndtri_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_special_scaled_modified_bessel_k1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_std_mean_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_torch__scaled_mm_v2_cuda_float8_e4m3fn, test/test_ops.py::TestCommonCUDA::test_out_trapezoid_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_triu_indices_cuda_int64, test/test_ops.py::TestCommonCUDA::test_out_unbind_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_warning___rmod___cuda, test/test_ops.py::TestCommonCUDA::test_out_warning___rpow___cuda, test/test_ops.py::TestCommonCUDA::test_out_warning___rsub___cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs__conversions_chalf_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs__conversions_long_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_amin_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_conj_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_constant_pad_nd_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_equal_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_erfinv_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_exp_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_expand_as_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_fft_fftn_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_fft_ihfft2_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_fft_ihfft_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_frexp_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_geometric_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_igammac_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_isinf_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_isreal_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_item_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_linalg_vecdot_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_logical_or_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_lt_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_masked_fill_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_nn_functional_hinge_embedding_loss_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_nn_functional_softmax_with_dtype_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_normal_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_positive_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_real_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_roll_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_rsqrt_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_sin_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_softmax_with_dtype_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_special_i0e_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_special_spherical_bessel_j0_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_special_zeta_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_triu_indices_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_unfold_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_addmm_decomposed_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_allclose_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_as_strided_scatter_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_atan_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_bfloat16_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_clamp_min_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_conj_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_constant_pad_nd_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_cross_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_cumulative_trapezoid_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_dot_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_erfc_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_fft_fft2_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_fft_rfftn_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_floor_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_gather_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_ge_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_grid_sampler_3d_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_histogramdd_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_hsplit_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_isin_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_ldexp_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_linalg_eig_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_linalg_eigvals_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_linalg_inv_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_linalg_ldl_solve_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_linalg_norm_subgradients_at_zero_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_linalg_tensorsolve_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_linspace_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_linspace_tensor_overload_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_log1p_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_masked_amin_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_masked_std_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_max_reduction_no_dim_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_mvlgamma_mvlgamma_p_5_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_adaptive_avg_pool2d_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_batch_norm_without_cudnn_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_binary_cross_entropy_with_logits_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_conv3d_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_conv_transpose2d_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_max_unpool3d_grad_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_mse_loss_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_multi_head_attention_forward_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_multilabel_soft_margin_loss_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_soft_margin_loss_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_softsign_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_threshold_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_upsample_bilinear_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_norm_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_polygamma_polygamma_n_2_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_roll_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_round_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_rsub_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_scatter_add_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_scatter_reduce_prod_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_signal_windows_exponential_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_sin_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_special_bessel_y0_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_special_hermite_polynomial_he_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_special_laguerre_polynomial_l_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_special_shifted_chebyshev_polynomial_t_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_special_shifted_chebyshev_polynomial_w_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_special_zeta_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_squeeze_multiple_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_tan_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_tensor_split_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_unsqueeze_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_var_mean_unbiased_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_xlogy_cuda, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_asin_cuda_bool, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_asin_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_atan2_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_atan_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_copysign_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_deg2rad_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_digamma_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_erfinv_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_exp_cuda_int64, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_exp_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_expm1_cuda_int32, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_expm1_cuda_int8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_log_cuda_int32, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_polygamma_polygamma_n_0_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_polygamma_polygamma_n_1_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_polygamma_polygamma_n_2_cuda_bool, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_rsqrt_cuda_bool, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_sinc_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_sinh_cuda_int64, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_chebyshev_polynomial_v_cuda_int64, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_hermite_polynomial_h_cuda_int8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_hermite_polynomial_he_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_laguerre_polynomial_l_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_laguerre_polynomial_l_cuda_int32, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_shifted_chebyshev_polynomial_u_cuda_int32, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_shifted_chebyshev_polynomial_v_cuda_int8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_shifted_chebyshev_polynomial_w_cuda_int32, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_zeta_cuda_int64, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_tan_cuda_bool, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_tan_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_xlogy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_T_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_bfloat16_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_bool_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_bool_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_bool_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_bool_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_cdouble_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_cfloat_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_cfloat_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_char_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_double_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_double_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_double_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_float_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_float_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_int_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_long_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_short_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_short_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_short_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_abs_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_abs_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_acos_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_acosh_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_acosh_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_acosh_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_acosh_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_add_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_addcdiv_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_addcmul_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_addr_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_alias_copy_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_alias_copy_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_allclose_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_amax_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_amax_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_amin_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_amin_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_any_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_any_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_arange_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_as_strided_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_as_strided_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_as_strided_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_as_strided_partial_views_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atan_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atan_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atanh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_bitwise_or_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_bitwise_or_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_broadcast_tensors_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_bucketize_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_cauchy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_clamp_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_clamp_max_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_clamp_min_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_clamp_min_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_clamp_min_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_column_stack_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_conj_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_conj_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_conj_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_conj_physical_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_conj_physical_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_copysign_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_cos_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_cosh_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_cosh_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_count_nonzero_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_cumprod_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_cumsum_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_cumsum_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_diag_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_diag_embed_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_diagonal_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_diagonal_scatter_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_div_floor_rounding_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_div_floor_rounding_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_div_no_rounding_mode_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_div_trunc_rounding_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_dot_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_dstack_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_empty_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_empty_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_empty_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_empty_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_empty_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_empty_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_empty_like_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_empty_like_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_empty_strided_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_empty_strided_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_empty_strided_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_equal_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_erfc_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_erfc_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_exp2_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_expand_as_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_expand_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_expand_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_expm1_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_expm1_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_eye_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_fft2_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_fft2_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_fft2_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_fft_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_fftn_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_fftn_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_fftshift_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_hfft2_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_hfft2_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_hfft2_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_hfft_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_hfftn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_ifft2_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_ifft_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_ifftn_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_ifftshift_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_ifftshift_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_ifftshift_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_irfft_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_irfft_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_rfft2_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_rfft2_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_rfftn_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fill_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fliplr_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_floor_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_floor_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_floor_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_floor_divide_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_floor_divide_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fmax_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fmod_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_gcd_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_ge_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_geometric_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_gt_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_heaviside_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_hsplit_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_hsplit_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_hsplit_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_hstack_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_i0_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_i0_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_imag_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_index_add_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_index_add_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_index_fill_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_index_fill_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_index_select_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_index_select_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isclose_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isnan_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isnan_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isneginf_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isposinf_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_item_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_cross_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_cross_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_cross_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_diagonal_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_matrix_norm_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_norm_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_svd_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_vector_norm_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_vector_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linspace_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linspace_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linspace_tensor_overload_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linspace_tensor_overload_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_log_softmax_with_dtype_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logaddexp2_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logical_and_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logical_and_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logical_and_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logical_or_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logical_or_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logical_xor_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logical_xor_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logspace_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logspace_tensor_overload_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logspace_tensor_overload_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_maximum_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_meshgrid_list_of_tensors_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_meshgrid_list_of_tensors_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_meshgrid_list_of_tensors_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_meshgrid_variadic_tensors_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_meshgrid_variadic_tensors_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_minimum_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_narrow_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_narrow_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_native_layer_norm_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_native_layer_norm_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_ne_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_new_empty_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_new_empty_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_new_full_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_new_ones_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_new_ones_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nextafter_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_channel_shuffle_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_channel_shuffle_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_gelu_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_hardtanh_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_hinge_embedding_loss_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_log_softmax_with_dtype_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_pairwise_distance_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_pixel_shuffle_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_pixel_shuffle_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_poisson_nll_loss_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_relu_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_softmax_with_dtype_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_softmax_with_dtype_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_softmax_with_dtype_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_softmin_with_dtype_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_softplus_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_softshrink_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_threshold_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_threshold_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_norm_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_normal_number_mean_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_ones_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_permute_copy_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_permute_copy_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_positive_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_prod_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_rad2deg_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_rad2deg_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_randn_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_randn_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_real_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_real_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_reciprocal_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_remainder_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_remainder_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_repeat_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_repeat_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_repeat_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_reshape_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_rot90_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_round_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_round_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_rsqrt_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_rsqrt_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_select_scatter_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sgn_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sigmoid_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sigmoid_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sin_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sin_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sinc_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sinc_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sinh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sinh_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_softmax_with_dtype_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_bessel_j0_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_bessel_j0_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_bessel_j1_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_entr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_i0e_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_i1_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_i1_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_i1_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_i1e_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_log_softmax_with_dtype_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_logit_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_multigammaln_mvlgamma_p_3_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_multigammaln_mvlgamma_p_5_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_ndtr_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_ndtr_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_ndtri_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_spherical_bessel_j0_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_spherical_bessel_j0_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_spherical_bessel_j0_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_split_with_sizes_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sqrt_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sqrt_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_square_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_squeeze_multiple_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_stack_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_std_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_std_mean_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_std_mean_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_stft_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sub_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sum_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sum_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_t_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_t_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_t_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_t_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_take_along_dim_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_tanh_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_tensor_split_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_to_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_to_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_transpose_copy_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_transpose_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_tril_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_tril_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_tril_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_tril_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_true_divide_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_unbind_copy_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_unbind_copy_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_unbind_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_unbind_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_unflatten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_unfold_copy_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_unsqueeze_copy_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_unsqueeze_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_view_as_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_view_copy_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_view_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_view_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_vsplit_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_where_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_where_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_xlogy_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_xlogy_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_zeros_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs__conversions_polar_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_exponential_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_hstack_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_lcm_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_linalg_diagonal_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_linspace_tensor_overload_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_logspace_tensor_overload_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_narrow_copy_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_nn_functional_softshrink_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_pow_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_remainder_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_renorm_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_T_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_T_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_bfloat16_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_bfloat16_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_cdouble_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_chalf_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_char_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_char_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_char_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_double_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_double_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_double_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_half_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_int_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_int_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_long_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_short_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_acos_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_acosh_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_add_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_add_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_alias_copy_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_all_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_amin_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_any_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_arange_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_as_strided_copy_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_as_strided_copy_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_as_strided_partial_views_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_as_strided_scatter_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_asin_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atan2_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atan2_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atanh_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atanh_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atleast_1d_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atleast_1d_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atleast_2d_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atleast_3d_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atleast_3d_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_bitwise_left_shift_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_bitwise_right_shift_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_block_diag_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_block_diag_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_block_diag_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_broadcast_tensors_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_broadcast_to_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cat_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cauchy_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_ceil_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_clamp_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_clamp_min_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_clone_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_column_stack_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_conj_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_conj_physical_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_contiguous_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_copysign_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_copysign_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cos_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cos_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cosh_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_count_nonzero_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_count_nonzero_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cumprod_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cumprod_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cumsum_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cumsum_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_deg2rad_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_diagonal_copy_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_diagonal_copy_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_diagonal_copy_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_diagonal_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_diagonal_scatter_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_div_floor_rounding_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_div_trunc_rounding_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_div_trunc_rounding_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_dot_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_dsplit_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_dstack_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_empty_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_empty_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_empty_like_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_empty_strided_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_eq_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_erf_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_erf_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_erfc_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_erfinv_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_erfinv_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_exp2_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_exp_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_exp_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_expand_as_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_expand_copy_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_expand_copy_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_eye_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_fft2_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_fftshift_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_fftshift_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_hfft2_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_hfft2_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_hfft2_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_hfft2_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_hfftn_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_hfftn_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ifft2_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ifft_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ifft_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ifftn_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ifftn_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ifftshift_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ifftshift_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ihfft2_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ihfft_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_irfft_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_irfft_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_irfftn_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_irfftn_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_rfftn_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fill_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fill_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fill_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_flatten_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_flatten_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_flipud_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_flipud_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_floor_divide_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_ge_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_geometric_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_geometric_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_gt_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_gt_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_gt_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_hstack_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_i0_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_igamma_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_index_add_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_index_add_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_index_copy_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_index_copy_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_index_fill_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_index_select_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_index_select_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_isclose_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_isclose_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_isfinite_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_isinf_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_isreal_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_isreal_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_item_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_item_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_item_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_le_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_le_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_lgamma_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_lgamma_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linalg_cross_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linalg_diagonal_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linalg_diagonal_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linalg_matrix_norm_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linalg_svdvals_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log10_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log10_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log1p_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log_softmax_with_dtype_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_and_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_and_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_not_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_or_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_or_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_or_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_or_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_xor_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_xor_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_xor_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logspace_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logspace_tensor_overload_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_lt_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_lt_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_masked_fill_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_masked_fill_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_maximum_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_movedim_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_mul_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nan_to_num_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_narrow_copy_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_narrow_copy_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_narrow_copy_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_narrow_copy_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_native_layer_norm_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_neg_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_new_empty_strided_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_new_full_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_new_full_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_new_ones_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_new_zeros_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_new_zeros_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_alpha_dropout_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_channel_shuffle_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_dropout_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_group_norm_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_hardshrink_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_hinge_embedding_loss_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_log_softmax_with_dtype_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_log_softmax_with_dtype_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_margin_ranking_loss_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_mish_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_pairwise_distance_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_pixel_shuffle_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_pixel_shuffle_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_pixel_unshuffle_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_pixel_unshuffle_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_poisson_nll_loss_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_relu_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_softmax_with_dtype_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_softmax_with_dtype_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_softmin_with_dtype_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_threshold_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_triplet_margin_loss_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_normal__in_place_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_normal_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_normal_number_mean_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_permute_copy_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_permute_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_permute_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_permute_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_prod_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_prod_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_rad2deg_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_ravel_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_real_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_reciprocal_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_reciprocal_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_renorm_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_renorm_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_repeat_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_repeat_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_reshape_as_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_reshape_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_roll_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_rsqrt_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_rsqrt_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_rsqrt_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_rsub_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sign_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sign_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_signbit_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sinc_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sinh_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sinh_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sinh_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_softmax_with_dtype_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_bessel_j0_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_bessel_j1_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_bessel_j1_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_entr_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_i0e_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_i1_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_i1e_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_log_ndtr_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_log_softmax_with_dtype_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_log_softmax_with_dtype_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_log_softmax_with_dtype_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_logit_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_logit_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_logit_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_multigammaln_mvlgamma_p_3_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_multigammaln_mvlgamma_p_5_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_multigammaln_mvlgamma_p_5_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_ndtr_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_ndtri_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_softmax_with_dtype_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_spherical_bessel_j0_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_spherical_bessel_j0_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_xlog1py_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sqrt_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sqrt_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_squeeze_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_squeeze_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_squeeze_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_squeeze_multiple_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_squeeze_multiple_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_stack_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_std_mean_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sum_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_t_copy_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_t_copy_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_t_copy_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_t_copy_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_t_copy_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_tanh_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_tensor_split_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_to_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_to_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_trace_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_trace_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_trace_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_transpose_copy_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_transpose_copy_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_transpose_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_tril_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_triu_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_true_divide_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_trunc_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_trunc_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_trunc_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unbind_copy_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unfold_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unfold_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unsqueeze_copy_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unsqueeze_copy_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_var_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_var_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_var_mean_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_vdot_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_view_as_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_view_as_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_view_copy_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_view_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_vsplit_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_vstack_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_where_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_where_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_where_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_zeros_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_bfloat16_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_bool_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_byte_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_cdouble_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_cdouble_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_chalf_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_char_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_float_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_float_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_half_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_long_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_long_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_short_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_abs_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_acos_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_acosh_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_add_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_addcmul_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_addcmul_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_addr_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_addr_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_addr_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_addr_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_alias_copy_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_allclose_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_amin_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_amin_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_any_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_arange_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_as_strided_copy_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_as_strided_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_as_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_as_strided_partial_views_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_as_strided_scatter_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_as_strided_scatter_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_asin_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_asin_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_asinh_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_atan_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_atanh_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_atanh_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_atleast_1d_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_atleast_1d_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_atleast_1d_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_atleast_2d_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_atleast_2d_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_bitwise_left_shift_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_bitwise_not_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_bitwise_not_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_bitwise_not_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_bitwise_xor_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_bitwise_xor_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_broadcast_tensors_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_broadcast_tensors_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_broadcast_tensors_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_broadcast_to_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_bucketize_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_bucketize_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cat_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cauchy_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_ceil_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_clamp_max_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_clamp_max_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_clamp_min_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_column_stack_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_column_stack_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_column_stack_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_conj_physical_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_conj_physical_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_conj_physical_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_constant_pad_nd_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_constant_pad_nd_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_contiguous_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_contiguous_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cos_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cosh_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cosh_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cosh_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_count_nonzero_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cumprod_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cumsum_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_diag_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_diag_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_diagonal_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_diagonal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_div_floor_rounding_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_div_no_rounding_mode_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_div_no_rounding_mode_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_dot_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_dot_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_dsplit_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_dsplit_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_dsplit_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_dstack_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_empty_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_empty_like_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_empty_like_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_eq_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_eq_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_eq_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_eq_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_erf_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_erfc_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_erfc_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_erfinv_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_exp2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_expand_as_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_expand_as_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_expand_copy_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_expand_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_expm1_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_eye_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_eye_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_fft2_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_fft_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_fft_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_hfft_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_hfft_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ifft2_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ifft_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ifftn_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ifftn_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ifftshift_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ihfft2_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ihfft_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ihfftn_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_irfft2_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_irfft_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_irfft_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_irfft_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_irfftn_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_irfftn_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_rfft_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fill_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_flatten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_flip_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_flip_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fliplr_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fliplr_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_flipud_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_float_power_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_float_power_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_floor_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fmax_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fmod_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_frexp_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_ge_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_ge_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_gt_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_gt_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_hsplit_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_hsplit_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_hsplit_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_hstack_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_hstack_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_hstack_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_index_copy_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_isfinite_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_isinf_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_isnan_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_isneginf_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_isneginf_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_isreal_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_item_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_item_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_item_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_item_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_le_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_le_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_lgamma_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_lgamma_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linalg_cross_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linalg_diagonal_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linalg_matrix_norm_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linalg_norm_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linalg_svd_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linalg_vector_norm_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log10_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log2_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log_softmax_with_dtype_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log_softmax_with_dtype_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log_softmax_with_dtype_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logaddexp2_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logical_and_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logical_not_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logical_not_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logical_not_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logical_or_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logical_or_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logical_or_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logical_xor_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logspace_tensor_overload_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logspace_tensor_overload_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_masked_fill_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_masked_fill_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_mean_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_meshgrid_variadic_tensors_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_minimum_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_movedim_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_mul_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_mul_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nan_to_num_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_narrow_copy_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_narrow_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_native_layer_norm_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_ne_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_new_empty_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_new_empty_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_new_ones_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nextafter_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_alpha_dropout_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_celu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_channel_shuffle_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_channel_shuffle_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_elu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_hardshrink_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_hardtanh_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_leaky_relu_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_log_softmax_with_dtype_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_log_softmax_with_dtype_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_mse_loss_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_mse_loss_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_pixel_shuffle_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_pixel_shuffle_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_pixel_unshuffle_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_poisson_nll_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_poisson_nll_loss_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_relu6_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_softmax_with_dtype_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_softmax_with_dtype_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_threshold_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_threshold_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_ones_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_ones_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_permute_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_permute_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_permute_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_pow_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_prod_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_rad2deg_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_rad2deg_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_randn_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_ravel_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_real_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_remainder_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_renorm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_repeat_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_repeat_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_reshape_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_reshape_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_roll_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_roll_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_rot90_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_round_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_rsqrt_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sigmoid_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_signbit_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sin_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sin_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sinh_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sinh_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_bessel_j0_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_bessel_j0_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_i1_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_i1_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_i1e_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_log_softmax_with_dtype_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_multigammaln_mvlgamma_p_1_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_multigammaln_mvlgamma_p_3_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_multigammaln_mvlgamma_p_5_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_multigammaln_mvlgamma_p_5_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_ndtri_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_softmax_with_dtype_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_spherical_bessel_j0_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_zeta_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_split_with_sizes_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sqrt_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sqrt_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_squeeze_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_squeeze_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_squeeze_multiple_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_squeeze_multiple_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_std_mean_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_std_mean_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sub_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sum_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sum_to_size_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_t_copy_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_t_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_t_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_t_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_take_along_dim_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_take_along_dim_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_take_along_dim_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tan_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tan_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tanh_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tensor_split_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_to_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_transpose_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_transpose_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_transpose_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tril_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tril_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tril_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tril_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tril_indices_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_triu_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_triu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_true_divide_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_true_divide_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_true_divide_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unbind_copy_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unbind_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unbind_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unflatten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unfold_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unsqueeze_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unsqueeze_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unsqueeze_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unsqueeze_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_var_mean_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_view_as_complex_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_view_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_view_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_vsplit_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_vsplit_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_vsplit_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_vstack_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_where_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_where_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_where_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_where_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_xlogy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_zeros_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_zeros_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_zeros_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_zeros_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_T_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_T_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_bfloat16_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_byte_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_byte_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_byte_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_cdouble_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_cdouble_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_cdouble_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_cdouble_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_cfloat_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_chalf_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_double_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_double_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_float_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_half_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_half_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_half_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_long_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_long_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_short_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_short_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_acos_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_acos_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_acosh_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_acosh_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_addcdiv_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_alias_copy_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_all_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_all_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_any_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_any_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_copy_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_copy_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_scatter_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_scatter_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_asin_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_asin_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_asin_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atan_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atanh_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atanh_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atleast_1d_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atleast_1d_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atleast_2d_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atleast_2d_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atleast_2d_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atleast_2d_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_bitwise_or_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_broadcast_to_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_broadcast_to_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cat_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_ceil_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_ceil_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_chunk_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_clamp_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_clamp_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_clamp_max_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_clamp_min_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_column_stack_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_conj_physical_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_constant_pad_nd_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_contiguous_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_contiguous_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_contiguous_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cos_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cos_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_count_nonzero_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cumsum_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_diag_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_diag_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_diagonal_copy_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_diagonal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_diagonal_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_diagonal_scatter_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_digamma_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_div_floor_rounding_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_div_trunc_rounding_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_empty_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_empty_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_empty_like_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_empty_like_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_empty_like_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_eq_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_eq_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_eq_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_equal_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_equal_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_erf_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_erfc_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_erfc_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_exp2_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_expand_as_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_expand_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_expand_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_expand_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_expm1_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_exponential_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_eye_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_eye_cuda_float8_e5m2, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_eye_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_fft2_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_fft2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_fftn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_fftn_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_fftshift_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_hfft2_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_hfft2_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_hfftn_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_ifft_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_ifft_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_ifft_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_ifftshift_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_ifftshift_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_irfft_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_irfftn_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_irfftn_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_rfftn_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fill_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_flip_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_flip_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fliplr_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fliplr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fliplr_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_float_power_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_floor_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_floor_divide_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fmax_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fmax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fmod_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_frexp_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_frexp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_ge_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_geometric_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_i0_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_i0_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_igamma_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_index_add_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_index_copy_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isclose_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isfinite_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isnan_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isnan_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isneginf_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isposinf_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isreal_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isreal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_item_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_item_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_le_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_le_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_lerp_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_linalg_diagonal_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_linalg_svd_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_linalg_vecdot_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_linspace_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_linspace_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log10_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log10_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log10_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log1p_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log2_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log2_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log_softmax_with_dtype_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log_softmax_with_dtype_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logaddexp_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logical_and_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logical_xor_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logspace_tensor_overload_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logspace_tensor_overload_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logsumexp_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logsumexp_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_lt_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_masked_fill_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_maximum_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_mean_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_meshgrid_list_of_tensors_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_minimum_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_mul_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_narrow_copy_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_narrow_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_narrow_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_narrow_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_neg_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_neg_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_empty_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_empty_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_full_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_zeros_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_zeros_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nextafter_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_alpha_dropout_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_gelu_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_glu_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_hardtanh_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_hinge_embedding_loss_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_l1_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_log_softmax_with_dtype_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_log_softmax_with_dtype_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_margin_ranking_loss_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_mish_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_mish_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_mse_loss_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_pairwise_distance_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_pdist_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_pixel_shuffle_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_pixel_shuffle_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_pixel_unshuffle_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_poisson_nll_loss_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_poisson_nll_loss_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_relu6_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_relu_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_relu_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_softmax_with_dtype_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_softmax_with_dtype_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_tanhshrink_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_threshold_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_threshold_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_triplet_margin_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_triplet_margin_loss_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_ones_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_ones_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_permute_copy_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_permute_copy_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_permute_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_positive_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_pow_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_prod_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_randn_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_randn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_real_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_reciprocal_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_repeat_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_repeat_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_reshape_as_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_reshape_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_reshape_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_roll_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_rot90_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_rsqrt_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_rsqrt_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_rsqrt_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_select_scatter_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sgn_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sgn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sgn_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sigmoid_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_signbit_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_signbit_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sinh_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_softmax_with_dtype_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_bessel_j0_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_entr_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_i0e_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_i0e_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_i0e_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_i1_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_log_ndtr_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_log_softmax_with_dtype_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_multigammaln_mvlgamma_p_5_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_ndtr_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_ndtri_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_ndtri_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_softmax_with_dtype_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_xlog1py_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_xlog1py_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_xlog1py_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_split_with_sizes_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sqrt_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_square_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_squeeze_multiple_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_stack_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_stack_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_stack_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_stft_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sum_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sum_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sum_to_size_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sum_to_size_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_take_along_dim_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_tan_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_tanh_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_tensor_split_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_tensor_split_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_to_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_trace_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_trace_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_trace_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_transpose_copy_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_triu_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_triu_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_true_divide_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_true_divide_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_true_divide_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_trunc_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_trunc_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_trunc_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unbind_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unbind_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unbind_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unbind_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unbind_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unbind_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unbind_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unfold_copy_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unsqueeze_copy_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unsqueeze_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unsqueeze_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unsqueeze_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_var_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_var_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_var_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_var_mean_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_view_as_complex_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_view_as_complex_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_view_as_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_view_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_view_copy_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_view_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_view_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_vstack_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_where_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_xlogy_cuda_float64, test/test_ops.py::TestCommonCUDA::test_reduction_ops_reduce_max_reduction_no_dim_cuda, test/test_ops.py::TestCommonCUDA::test_reduction_ops_reduce_nansum_cuda, test/test_ops.py::TestCommonCUDA::test_reduction_ops_reduce_var_cuda, test/test_ops.py::TestCommonCUDA::test_reduction_ops_reduce_var_unbiased_cuda, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager___rdiv___cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager___rmul___cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager__segment_reduce_lengths_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager__softmax_backward_data_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_addmm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_addr_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_all_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_atan2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_broadcast_to_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_cat_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_cfloat_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_eq_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_expand_as_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_expm1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_fft_ifft2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_fft_ihfft2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_fft_rfftn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_floor_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_gt_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_hstack_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_index_put_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_int_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_jiterator_4inputs_with_extra_args_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_jiterator_binary_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_cross_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_inv_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_lstsq_grad_oriented_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_lu_factor_ex_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_matrix_rank_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_matrix_rank_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_pinv_hermitian_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_svd_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_vander_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_vecdot_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linspace_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_log1p_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_logical_and_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_logspace_tensor_overload_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_logspace_tensor_overload_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_matmul_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_mode_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_mvlgamma_mvlgamma_p_5_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nansum_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_narrow_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_neg_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_new_empty_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_alpha_dropout_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_conv_transpose3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_feature_alpha_dropout_without_train_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_hardswish_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_instance_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_interpolate_bicubic_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_interpolate_bilinear_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_interpolate_nearest-exact_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_l1_loss_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_linear_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_max_pool1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_pad_reflect_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_pad_replicate_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_softmin_with_dtype_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_softsign_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nonzero_static_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_norm_inf_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_normal_in_place_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_positive_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_prod_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_put_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_reshape_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_round_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_round_decimals_neg_3_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_scatter_reduce_mean_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_sgn_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_signal_windows_kaiser_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_special_hermite_polynomial_h_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_special_polygamma_special_polygamma_n_0_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_split_list_args_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_square_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_stack_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_svd_lowrank_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_var_unbiased_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_vdot_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_vsplit_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_where_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_acos_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_alias_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_broadcast_tensors_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_cdist_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_cfloat_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_conj_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_cos_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_cumulative_trapezoid_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_div_no_rounding_mode_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_erfinv_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_fft_irfft_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_fft_rfftn_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_flatten_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_float_power_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_grid_sampler_2d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_index_put_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_inner_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_linalg_lu_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_linalg_lu_factor_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_linalg_pinv_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_linalg_vander_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_linalg_vector_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_logaddexp2_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_logsumexp_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_masked_amin_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_masked_normalize_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_masked_sum_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_masked_var_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_adaptive_max_pool2d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_avg_pool1d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_channel_shuffle_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_conv3d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_cross_entropy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_fractional_max_pool3d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_gaussian_nll_loss_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_hardswish_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_logsigmoid_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_max_pool1d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_soft_margin_loss_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_pca_lowrank_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_pow_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_renorm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_round_decimals_neg_3_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_rsqrt_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_sinh_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_special_xlog1py_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_split_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_split_list_args_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_std_mean_unbiased_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_trapezoid_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_unfold_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_unfold_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_unsqueeze_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_var_mean_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_var_mean_unbiased_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input___rmod___cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_addmv_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_allclose_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_as_strided_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_as_strided_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_asin_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_broadcast_to_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_cat_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_ceil_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_dstack_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_erf_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_expm1_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_exponential_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_geqrf_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_igamma_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_index_fill_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_index_reduce_amin_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_isclose_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_isneginf_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_jiterator_binary_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_ldexp_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_linalg_ldl_solve_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_linalg_multi_dot_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_linalg_solve_triangular_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_linspace_tensor_overload_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_log_normal_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_log_softmax_with_dtype_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_logical_and_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_logit_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_long_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_masked_fill_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_masked_select_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_masked_softmax_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_meshgrid_list_of_tensors_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_movedim_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_narrow_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_narrow_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_native_dropout_backward_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_native_layer_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_ne_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nextafter_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_celu_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_conv3d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_hardswish_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_hardtanh_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_interpolate_linear_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_max_unpool1d_grad_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_max_unpool2d_grad_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_pairwise_distance_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_pixel_unshuffle_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_scaled_dot_product_attention_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_polygamma_polygamma_n_4_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_qr_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_randn_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_real_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_reshape_as_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_reshape_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_resize_as__cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_scatter_reduce_mean_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_scatter_reduce_sum_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_searchsorted_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_signal_windows_exponential_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_signal_windows_general_hamming_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_slice_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_sort_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_special_bessel_j1_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_special_chebyshev_polynomial_t_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_special_log_ndtr_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_special_modified_bessel_k0_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_special_scaled_modified_bessel_k0_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_squeeze_multiple_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_trace_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_triangular_solve_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_unfold_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_unique_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_unsqueeze_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_zeros_like_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad___rdiv___cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_addcmul_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_any_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_asinh_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_atleast_1d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_atleast_3d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_bmm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_broadcast_to_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_cat_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_cdouble_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_constant_pad_nd_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_corrcoef_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_diag_embed_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_div_floor_rounding_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_fft_fftn_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_fft_fftshift_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_fft_ifftshift_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_ge_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_index_reduce_amin_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_index_reduce_prod_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_index_select_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_isneginf_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_jiterator_binary_return_by_ref_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_ldexp_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_linalg_eigvals_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_linalg_norm_subgradients_at_zero_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_linalg_pinv_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_linalg_pinv_singular_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_lt_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_lu_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_masked_median_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_masked_normalize_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_max_reduction_with_dim_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_min_reduction_with_dim_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_new_full_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_adaptive_max_pool1d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_batch_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_dropout3d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_embedding_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_feature_alpha_dropout_with_train_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_interpolate_nearest-exact_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_smooth_l1_loss_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_norm_inf_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_normal_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_scalar_tensor_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_scatter_reduce_sum_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_sign_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_signal_windows_exponential_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_special_airy_ai_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_special_bessel_y0_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_special_entr_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_special_log_ndtr_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_special_modified_bessel_i0_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_special_shifted_chebyshev_polynomial_w_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_split_list_args_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_split_with_sizes_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_squeeze_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_std_mean_unbiased_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_stft_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_to_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_unsafe_split_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_var_unbiased_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_view_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_zero__cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_zeros_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_zeros_like_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator___rsub___cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_acos_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_acosh_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_addcmul_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_asin_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_atanh_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_clamp_min_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_combinations_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_digamma_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_expand_as_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_fft_fft2_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_fft_ifftn_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_fft_ifftshift_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_fft_rfft_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_flatten_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_flipud_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_gather_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_geqrf_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_grid_sampler_2d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_index_reduce_amax_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_index_select_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_int_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_isinf_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_isneginf_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_jiterator_binary_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_linalg_det_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_linalg_lu_factor_ex_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_linalg_matrix_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_linalg_pinv_singular_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_log2_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_logsumexp_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_lt_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_masked_cumprod_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_masked_fill_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_masked_log_softmax_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_masked_median_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_max_binary_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_min_binary_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_mm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nanmean_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_avg_pool3d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_huber_loss_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_instance_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_interpolate_linear_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_leaky_relu_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_local_response_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_max_unpool1d_grad_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_mish_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_poisson_nll_loss_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_upsample_bilinear_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_norm_inf_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_normal_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_normal_number_mean_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_ones_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_outer_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_polygamma_polygamma_n_0_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_qr_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_repeat_interleave_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_resolve_conj_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_scatter_reduce_mean_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_signal_windows_exponential_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_sinc_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_special_hermite_polynomial_h_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_special_polygamma_special_polygamma_n_0_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_sum_to_size_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_trunc_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_unique_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_unsafe_chunk_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_view_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_view_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay___rpow___cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay__softmax_backward_data_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_acosh_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_arange_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_as_strided_partial_views_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_cauchy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_chunk_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_cumprod_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_diag_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_diagonal_scatter_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_digamma_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_dist_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_double_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_empty_like_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_empty_strided_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_erfinv_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_eye_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_fft_hfft2_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_fft_ihfft2_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_fft_rfft_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_frac_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_frexp_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_hsplit_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_index_reduce_amax_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_isreal_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_jiterator_2inputs_2outputs_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_linalg_cholesky_ex_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_linalg_cross_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_linalg_matrix_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_linalg_vander_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_masked_fill_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_masked_mean_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_neg_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_new_full_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nextafter_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_adaptive_max_pool1d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_avg_pool1d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_cosine_embedding_loss_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_grid_sample_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_max_pool1d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_mish_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_nll_loss_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_pdist_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_softplus_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_ones_like_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_put_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_real_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_repeat_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_reshape_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_resize__cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_rsqrt_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_sigmoid_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_signal_windows_gaussian_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_signbit_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_special_chebyshev_polynomial_u_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_special_chebyshev_polynomial_v_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_special_erfcx_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_special_i1e_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_special_modified_bessel_k1_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_special_shifted_chebyshev_polynomial_w_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_std_mean_unbiased_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_to_sparse_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_trapz_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_true_divide_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_view_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_zeros_cuda_float32, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_diagonal_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_expand_copy_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_fft_fft_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_fft_fftshift_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_fill_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_float_power_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_hstack_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_isinf_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_logical_xor_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_logspace_tensor_overload_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_narrow_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_new_empty_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_nn_functional_log_softmax_with_dtype_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_pow_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_prod_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_randn_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_rot90_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_sigmoid_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_squeeze_copy_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_triu_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_vsplit_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_zeros_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_add_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_addcdiv_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_addr_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_as_strided_copy_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_cat_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_combinations_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_conj_physical_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_cross_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_cumulative_trapezoid_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_div_no_rounding_mode_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_double_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_equal_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_expand_copy_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_fft_fftshift_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_fft_ifftshift_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_geqrf_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_index_select_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_item_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_linalg_cholesky_ex_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_linalg_pinv_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_linalg_solve_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_linalg_tensorinv_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_logaddexp_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_matmul_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_mv_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_nansum_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_narrow_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_ne_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_nn_functional_conv_transpose3d_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_nn_functional_rms_norm_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_nn_functional_softsign_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_nn_functional_tanhshrink_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_pca_lowrank_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_ravel_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_repeat_interleave_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_rsub_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_softmax_with_dtype_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_triangular_solve_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_unbind_copy_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_unsafe_split_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_vstack_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view___rmul___cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view___rsub___cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs__conversions_cfloat_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs__conversions_double_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_as_strided_partial_views_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_as_strided_scatter_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_asin_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_atanh_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_column_stack_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_contiguous_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_hstack_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_isfinite_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_meshgrid_list_of_tensors_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_softmax_with_dtype_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_sub_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_unsqueeze_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_vdot_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_atan_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_constant_pad_nd_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_diag_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_einsum_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_flip_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_gather_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_int_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_item_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_jiterator_4inputs_with_extra_args_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_jiterator_binary_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_kron_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_linalg_cholesky_ex_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_logical_and_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_mean_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_narrow_copy_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_new_empty_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_new_ones_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_nn_functional_channel_shuffle_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_ones_like_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_sigmoid_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_split_list_args_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_squeeze_copy_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_stack_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_std_mean_unbiased_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_std_unbiased_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_t_copy_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_take_along_dim_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_triangular_solve_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_triu_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_view_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_vsplit_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_zeros_like_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_view___rmod___cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_T_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs__conversions_short_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_amin_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_bucketize_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_cat_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_clone_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_diagonal_copy_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_diagonal_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_div_floor_rounding_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_dstack_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_expand_as_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_fft_ifft2_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_fft_rfftn_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_flatten_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_index_copy_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_lerp_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_log10_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_logaddexp_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_minimum_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_mul_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_narrow_copy_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_native_layer_norm_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_new_full_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_new_ones_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_nn_functional_hardshrink_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_nn_functional_pdist_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_nn_functional_pixel_shuffle_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_nn_functional_softplus_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_ravel_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_sigmoid_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_special_entr_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_special_i1e_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_special_multigammaln_mvlgamma_p_1_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_std_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_unfold_copy_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__unsafe_masked_index_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_abs_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_add_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_addmm_decomposed_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_angle_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_as_strided_copy_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_asinh_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_cummax_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_cummin_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_fft_hfftn_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_fft_ifft2_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_fft_ihfft_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_flatten_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_frac_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_heaviside_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_linalg_ldl_factor_ex_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_linalg_ldl_solve_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_linalg_matrix_power_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_linalg_vander_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_logdet_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_lu_unpack_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_masked_cumprod_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_masked_cumsum_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_masked_logaddexp_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_masked_softmin_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_max_reduction_with_dim_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_median_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_mode_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_narrow_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_new_zeros_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nextafter_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_bilinear_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_celu_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_conv2d_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_conv_transpose2d_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_elu_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_group_norm_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_hardtanh_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_max_unpool2d_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_max_unpool3d_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_normalize_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_pad_replicate_negative_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_pairwise_distance_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_relu6_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_rms_norm_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_silu_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_softmin_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_softmin_with_dtype_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_softplus_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_softshrink_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_triplet_margin_loss_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_pca_lowrank_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_permute_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_polar_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_randint_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_reciprocal_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_round_decimals_neg_3_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_sgn_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_signal_windows_general_cosine_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_sort_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_special_chebyshev_polynomial_w_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_special_modified_bessel_k0_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_special_xlog1py_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_split_with_sizes_copy_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_std_mean_unbiased_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_std_unbiased_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_svd_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_torch_ops_aten__safe_softmax_default_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_unbind_copy_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_unfold_copy_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_zeros_like_cuda_float64, test/test_ops.py::TestFakeTensorCUDA::test_fake___rand___cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_fake___rxor___cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_fake__batch_norm_with_update_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake__chunk_cat_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_abs_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_H_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_addcdiv_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_amin_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_argsort_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_as_strided_partial_views_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_atanh_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_bitwise_xor_cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_block_diag_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_combinations_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_cosh_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_cross_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_cumprod_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_diag_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_diagflat_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_diagonal_scatter_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_fft_fft_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_fft_ihfft2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_half_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_isfinite_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_lcm_cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_linalg_eigvalsh_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_linalg_slogdet_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_linalg_vander_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_logaddexp2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_logical_not_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_lt_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_lu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_masked_cumprod_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_masked_mean_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_masked_softmax_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_masked_std_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_min_binary_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nanquantile_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_new_ones_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_channel_shuffle_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_conv3d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_dropout2d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_group_norm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_hardshrink_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_interpolate_trilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_max_pool3d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_one_hot_cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_scaled_dot_product_attention_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_softplus_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nonzero_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_normal_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_ones_like_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_polar_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_prod_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_put_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_remainder_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_rsqrt_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_rsub_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_signal_windows_blackman_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_special_modified_bessel_k1_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_special_scaled_modified_bessel_k0_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_special_shifted_chebyshev_polynomial_u_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_split_with_sizes_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_sum_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_transpose_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_unique_consecutive_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_unsafe_split_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_vstack_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_conj_physical_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_constant_pad_nd_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_count_nonzero_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_cov_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_addbmm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_as_strided_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_atleast_1d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_constant_pad_nd_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_copysign_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_digamma_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_double_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_einsum_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_fft_rfftn_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_fliplr_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_float_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_gradient_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_grid_sampler_3d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_half_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_index_reduce_amin_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_linalg_matrix_power_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_linalg_pinv_singular_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_lu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_lu_solve_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_masked_cumsum_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_masked_logaddexp_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_masked_scatter_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nan_to_num_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_neg_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_dropout_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_gelu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_glu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_max_pool1d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_pad_replicate_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_polygamma_polygamma_n_3_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_polygamma_polygamma_n_4_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_positive_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_slice_scatter_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_split_with_sizes_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_std_mean_unbiased_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_svd_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp___rmul___cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp___rsub___cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_abs_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_as_strided_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_atanh_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_cfloat_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_chalf_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_deg2rad_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_diagonal_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_fft_ifftshift_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_fft_ihfft2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_fft_rfftn_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_floor_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_gradient_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_half_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_index_put_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_index_select_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_linalg_cholesky_ex_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_linalg_eigh_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_linalg_pinv_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_linalg_pinv_singular_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_linalg_slogdet_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_linalg_svd_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_logaddexp_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_masked_logsumexp_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_masked_softmax_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_conv_transpose3d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_cosine_embedding_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_dropout2d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_embedding_bag_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_gaussian_nll_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_huber_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_multi_head_attention_forward_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_multi_margin_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_pixel_shuffle_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_poisson_nll_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_relu6_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_tanhshrink_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_threshold_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_norm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_norm_inf_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_normal_number_mean_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_polar_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_polygamma_polygamma_n_3_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_positive_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_repeat_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_select_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_sgn_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_sparse_sampled_addmm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_special_i1_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_t_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_tile_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_tril_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_view_as_complex_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_div_trunc_rounding_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_dot_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_erf_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_expand_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_expm1_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_fft_fft_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_fft_ifftshift_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_fft_irfft_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_igamma_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_jiterator_2inputs_2outputs_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_jiterator_binary_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_jiterator_binary_return_by_ref_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_kron_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_lgamma_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_linalg_cholesky_ex_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_linalg_lu_factor_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_linalg_vander_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_linalg_vecdot_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_log10_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_log_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_lu_solve_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_masked_prod_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_masked_select_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_max_pool2d_with_indices_backward_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_min_binary_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_msort_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_multinomial_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_elu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_hardshrink_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_interpolate_linear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_layer_norm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_multi_margin_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_rrelu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_selu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_silu_complex_cuda_complex64, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_threshold_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_pca_lowrank_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_polygamma_polygamma_n_0_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_polygamma_polygamma_n_2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_prod_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_rand_like_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_randn_like_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_reciprocal_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_repeat_interleave_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_resolve_neg_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_scatter_reduce_amax_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_select_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_special_airy_ai_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_special_entr_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_special_hermite_polynomial_he_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_sqrt_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_sub_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_tan_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_tanh_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_torch__scaled_mm_cuda_float8_e4m3fn, test/test_ops.py::TestFakeTensorCUDA::test_fake_trace_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_transpose_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_unbind_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_unsafe_chunk_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops__segment_reduce_lengths_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_addmm_decomposed_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_amin_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_bool_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_char_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_constant_pad_nd_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_div_trunc_rounding_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_fft_ihfft2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_fft_irfft2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_fft_irfftn_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_geometric_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_grid_sampler_3d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_isneginf_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_le_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_linalg_lu_factor_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_linalg_pinv_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_linalg_slogdet_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_linalg_svd_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_linalg_tensorinv_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_linalg_vector_norm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_log2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_log_softmax_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_logcumsumexp_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_long_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_lu_solve_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_masked_prod_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_masked_select_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_max_reduction_no_dim_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_maximum_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_min_binary_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_min_reduction_no_dim_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_min_reduction_with_dim_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_new_ones_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_adaptive_avg_pool1d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_alpha_dropout_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_glu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_interpolate_trilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_margin_ranking_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_max_unpool3d_grad_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_nll_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_pad_constant_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_pdist_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_pixel_shuffle_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_softsign_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nonzero_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_polygamma_polygamma_n_2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_remainder_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_repeat_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_signal_windows_gaussian_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_special_bessel_j1_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_special_bessel_y0_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_special_chebyshev_polynomial_t_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_squeeze_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_sum_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_take_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_torch__scaled_mm_v2_cuda_float8_e4m3fn, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_triangular_solve_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_triu_indices_cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_unique_consecutive_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_arange_cuda_bfloat16, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_linspace_cuda_float64, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_linspace_cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_linspace_tensor_overload_cuda_float64, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_linspace_tensor_overload_cuda_int8, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_logspace_cuda_bfloat16, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_logspace_cuda_int32, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_logspace_tensor_overload_cuda_complex128, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_logspace_tensor_overload_cuda_complex64, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_logspace_tensor_overload_cuda_float16, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_logspace_tensor_overload_cuda_float64, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_ones_cuda_bfloat16, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_zeros_cuda_bfloat16, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_zeros_cuda_int16, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_zeros_cuda_int32, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_arange_cuda_int8, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_linspace_cuda_complex128, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_linspace_tensor_overload_cuda_bfloat16, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_linspace_tensor_overload_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_linspace_tensor_overload_cuda_int32, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_linspace_tensor_overload_cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_linspace_tensor_overload_cuda_int8, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_logspace_tensor_overload_cuda_bfloat16, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_logspace_tensor_overload_cuda_float16, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_zeros_cuda_bfloat16, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_zeros_cuda_complex128, test/test_ops.py::TestTagsCUDA::test_tags__refs__conversions_bfloat16_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_alias_copy_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_amin_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_as_strided_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_as_strided_scatter_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_atan2_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_ceil_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_clamp_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_clamp_max_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_count_nonzero_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_cumsum_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_digamma_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_div_trunc_rounding_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_fft_hfft2_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_geometric_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_gt_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_linalg_matrix_norm_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_narrow_copy_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_nn_functional_dropout_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_nn_functional_huber_loss_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_nn_functional_nll_loss_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_nn_functional_smooth_l1_loss_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_nn_functional_softmax_with_dtype_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_normal_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_special_ndtr_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_special_spherical_bessel_j0_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_squeeze_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_unsqueeze_copy_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_var_mean_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_view_as_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_zeros_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_add_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_addmm_decomposed_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_amax_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_amin_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_any_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_argmin_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_ceil_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_corrcoef_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_fft_ihfft2_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_fft_ihfft_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_fft_ihfftn_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_float_power_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_fmax_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_full_like_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_grid_sampler_3d_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_histc_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_index_reduce_prod_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_inner_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_int_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_isin_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_isreal_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_jiterator_4inputs_with_extra_args_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_linalg_eigvals_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_linalg_lstsq_grad_oriented_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_linalg_qr_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_linalg_vector_norm_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_logcumsumexp_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_lu_unpack_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_masked_amax_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_min_reduction_no_dim_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_msort_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_mvlgamma_mvlgamma_p_5_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_new_zeros_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_alpha_dropout_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_conv_transpose1d_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_conv_transpose3d_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_instance_norm_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_interpolate_area_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_interpolate_nearest_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_max_unpool2d_grad_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_mish_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_multi_head_attention_forward_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_poisson_nll_loss_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_positive_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_rad2deg_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_reciprocal_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_resize_as__cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_scatter_reduce_prod_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_signal_windows_cosine_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_signal_windows_gaussian_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_signal_windows_hann_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_signal_windows_kaiser_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_signal_windows_nuttall_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_special_chebyshev_polynomial_v_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_special_hermite_polynomial_he_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_special_i1e_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_special_shifted_chebyshev_polynomial_u_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_std_mean_unbiased_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_svd_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_torch__scaled_mm_cuda_float8_e4m3fn, test/test_ops.py::TestTagsCUDA::test_tags_transpose_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_trunc_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_unfold_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_view_copy_cuda_float32 2025-12-04T14:28:08.8344720Z 2025-12-04T14:28:08.8345035Z Finished test_ops 1/11 ... [2025-12-04 14:28:08.569779][18872.260168424], took 21.80min 2025-12-04T14:28:08.8346083Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops/test_ops-b6bd03fdf7fbcee0.xml 2025-12-04T14:28:09.9934485Z Uploading artifacts took 1.23 seconds 2025-12-04T14:28:09.9940294Z Running test_ops 6/11 ... [2025-12-04 14:28:09.993776][18873.684166212] 2025-12-04T14:28:09.9941041Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T14:28:09.9946827Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ops.py', '--shard-id=6', '--num-shards=11', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 14:28:09.994347] 2025-12-04T14:47:55.0161545Z 2025-12-04T14:47:55.0162462Z test_ops 6/11 was successful, full logs can be found in artifacts with path test/test-reports/test_ops_6.11_ab7dea6c6996f1c3_.log 2025-12-04T14:47:55.1440285Z Running 3066 items in this shard: test/test_ops.py::TestCommonCUDA::test_compare_cpu___rsub___cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_arange_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_as_strided_partial_views_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_bitwise_right_shift_cuda_int64, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_div_no_rounding_mode_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_exponential_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_fmin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_linalg_matrix_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_meshgrid_list_of_tensors_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_new_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_nn_functional_pixel_shuffle_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_nn_functional_relu6_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_reshape_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_special_log_softmax_with_dtype_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_tril_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_unfold_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_atleast_3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_bool_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_chalf_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_cholesky_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_contiguous_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_diag_embed_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_dstack_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_empty_permuted_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_expand_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_hsplit_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_ldexp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_linalg_det_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_linalg_eigh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_linalg_qr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_linalg_solve_ex_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_masked_cumsum_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_masked_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_meshgrid_list_of_tensors_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_min_reduction_no_dim_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_new_empty_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_adaptive_avg_pool1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_adaptive_avg_pool2d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_channel_shuffle_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_embedding_bag_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_grid_sample_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_hinge_embedding_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_huber_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_multi_margin_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_pad_replicate_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_scaled_dot_product_attention_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_normal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_polar_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_reshape_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_scatter_reduce_prod_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_short_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_special_laguerre_polynomial_l_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_sum_to_size_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_take_along_dim_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_torch_ops_aten__safe_softmax_default_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_unflatten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_clone_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_dsplit_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_fft_fftn_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_fft_irfftn_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_float_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_mul_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_permute_copy_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_randn_like_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_tanh_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_transpose_copy_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_dtypes___rpow___cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs__conversions_chalf_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs__conversions_double_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_acos_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_add_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_as_strided_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_atanh_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_ceil_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_div_floor_rounding_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_exp_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_fft_hfft_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_fft_ifft_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_fft_rfft_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_fmin_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_lcm_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_linalg_diagonal_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_linspace_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_logspace_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_maximum_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_nan_to_num_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_nn_functional_channel_shuffle_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_nn_functional_layer_norm_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_nn_functional_mish_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_special_entr_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_stack_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_triu_indices_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_unsqueeze_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_view_as_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__softmax_backward_data_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__upsample_bilinear2d_aa_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_addbmm_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_all_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_as_strided_copy_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_atan2_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_bernoulli_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_broadcast_tensors_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_clamp_min_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_clone_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_contiguous_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_diag_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_diagonal_scatter_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_double_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_eye_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_fft_fftn_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_floor_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_hsplit_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_item_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_jiterator_2inputs_2outputs_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_jiterator_binary_return_by_ref_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_linalg_cholesky_ex_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_linalg_lu_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_linalg_lu_factor_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_linalg_multi_dot_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_linalg_qr_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_linspace_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_log_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_log_softmax_with_dtype_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_logaddexp_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_logical_or_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_masked_logaddexp_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_masked_select_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_max_reduction_no_dim_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_multinomial_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_mvlgamma_mvlgamma_p_1_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_native_batch_norm_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_cross_entropy_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_huber_loss_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_interpolate_linear_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_interpolate_nearest_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_kl_div_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_max_unpool1d_grad_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_pad_constant_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_pixel_shuffle_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_poisson_nll_loss_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_relu6_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_smooth_l1_loss_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_pca_lowrank_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_resize__cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_round_decimals_0_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_rsqrt_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_rsub_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_searchsorted_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_slice_scatter_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_sort_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_special_erfcx_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_special_log_ndtr_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_special_polygamma_special_polygamma_n_0_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_special_xlog1py_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_std_mean_unbiased_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_sum_to_size_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_var_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_view_copy_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_zeros_like_cuda, test/test_ops.py::TestCommonCUDA::test_errors___ror___cuda, test/test_ops.py::TestCommonCUDA::test_errors__chunk_cat_cuda, test/test_ops.py::TestCommonCUDA::test_errors_arange_cuda, test/test_ops.py::TestCommonCUDA::test_errors_dsplit_cuda, test/test_ops.py::TestCommonCUDA::test_errors_empty_permuted_cuda, test/test_ops.py::TestCommonCUDA::test_errors_fft_rfft2_cuda, test/test_ops.py::TestCommonCUDA::test_errors_floor_divide_cuda, test/test_ops.py::TestCommonCUDA::test_errors_isclose_cuda, test/test_ops.py::TestCommonCUDA::test_errors_linalg_diagonal_cuda, test/test_ops.py::TestCommonCUDA::test_errors_logical_and_cuda, test/test_ops.py::TestCommonCUDA::test_errors_logspace_cuda, test/test_ops.py::TestCommonCUDA::test_errors_masked_select_cuda, test/test_ops.py::TestCommonCUDA::test_errors_movedim_cuda, test/test_ops.py::TestCommonCUDA::test_errors_nn_functional_adaptive_max_pool2d_cuda, test/test_ops.py::TestCommonCUDA::test_errors_nn_functional_hardtanh_cuda, test/test_ops.py::TestCommonCUDA::test_errors_nn_functional_softshrink_cuda, test/test_ops.py::TestCommonCUDA::test_errors_nn_functional_triplet_margin_with_distance_loss_cuda, test/test_ops.py::TestCommonCUDA::test_errors_rsub_cuda, test/test_ops.py::TestCommonCUDA::test_errors_signal_windows_general_cosine_cuda, test/test_ops.py::TestCommonCUDA::test_errors_take_cuda, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch__native_batch_norm_legit_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_add_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_addmv_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_bernoulli_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_cat_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_complex_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_eye_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_fft_fft_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_fft_hfft2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_fft_hfft_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_fft_rfft2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_frac_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_lerp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_linalg_cross_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_linalg_eigvals_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_linalg_lstsq_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_linalg_matrix_rank_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_linalg_solve_triangular_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_log2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_lu_unpack_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_min_reduction_with_dim_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_nn_functional_gelu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_polygamma_polygamma_n_4_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_rad2deg_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_round_decimals_neg_3_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_signbit_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_slice_scatter_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_sort_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_special_i1e_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_special_polygamma_special_polygamma_n_0_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_special_shifted_chebyshev_polynomial_v_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_stack_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_triangular_solve_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_unsqueeze_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_var_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_vstack_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_xlogy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices___rand___cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices___rmod___cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices___rmod___cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices___rxor___cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices__batch_norm_with_update_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_acos_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_addcmul_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_alias_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_all_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_angle_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_argsort_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_as_strided_partial_views_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_bfloat16_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_broadcast_to_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_cauchy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_ceil_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_cholesky_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_clamp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_cummax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_diagonal_scatter_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_div_no_rounding_mode_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_double_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_empty_like_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_empty_permuted_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_equal_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_erf_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_fft_fft_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_fft_rfft2_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_fill_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_flatten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_fliplr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_float_power_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_floor_divide_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_fmin_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_gather_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_gcd_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_gradient_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_histc_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_igamma_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_index_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_index_fill_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_index_reduce_amax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_index_reduce_amin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_index_reduce_mean_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_isclose_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_isreal_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_jiterator_binary_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_jiterator_unary_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_linalg_cond_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_linalg_eigvals_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_linalg_pinv_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_log1p_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_log_normal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_logaddexp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_masked_cumsum_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_masked_cumsum_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_masked_prod_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_mode_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nan_to_num_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_batch_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_dropout_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_fractional_max_pool3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_gelu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_l1_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_margin_ranking_loss_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_smooth_l1_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_softplus_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_triplet_margin_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nonzero_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_ones_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_ormqr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_polygamma_polygamma_n_3_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_polygamma_polygamma_n_4_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_prod_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_randint_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_randint_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_resolve_neg_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_round_decimals_3_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_scatter_reduce_sum_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_searchsorted_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_signal_windows_general_hamming_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_signbit_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_bessel_j1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_chebyshev_polynomial_t_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_i1e_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_laguerre_polynomial_l_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_legendre_polynomial_p_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_scaled_modified_bessel_k1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_shifted_chebyshev_polynomial_u_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_spherical_bessel_j0_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_stack_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_tan_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_tensor_split_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_torch_ops_aten__safe_softmax_default_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_transpose_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_true_divide_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_unfold_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_unique_consecutive_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_unsafe_chunk_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_var_mean_unbiased_cuda_float32, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_asinh_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_atan2_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_atleast_2d_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_bitwise_or_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_constant_pad_nd_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_cos_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_fft_ihfft2_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_fliplr_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_flipud_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_hsplit_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_index_copy_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_jiterator_binary_return_by_ref_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_logical_xor_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_mH_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_mode_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_movedim_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_new_empty_strided_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_new_ones_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_polygamma_polygamma_n_0_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_polygamma_polygamma_n_2_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_polygamma_polygamma_n_4_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_ravel_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_sin_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_special_entr_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_special_laguerre_polynomial_l_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_special_log_ndtr_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_special_scaled_modified_bessel_k0_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_special_shifted_chebyshev_polynomial_v_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_special_xlog1py_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_square_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_take_along_dim_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_tile_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_where_cuda_bool, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_H_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples___getitem___cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples___rdiv___cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples___ror___cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples___rsub___cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples__chunk_cat_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples__native_batch_norm_legit_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_acos_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_alias_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_any_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_any_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_arange_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_argwhere_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_baddbmm_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_broadcast_tensors_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_cat_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_chalf_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_clamp_max_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_conj_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_conj_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_cumulative_trapezoid_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_diagonal_scatter_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_div_floor_rounding_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_dot_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_dot_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_dsplit_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_empty_permuted_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_equal_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_erfinv_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_exp2_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_expm1_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_eye_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_fft_fft2_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_fft_fftn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_fft_hfft2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_fft_ifftshift_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_fft_ihfft2_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_fft_irfft_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_float_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_float_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_full_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_full_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_gcd_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_geometric_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_gradient_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_grid_sampler_3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_igamma_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_igammac_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_index_add_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_index_fill_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_isnan_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_jiterator_4inputs_with_extra_args_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_eigvalsh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_lstsq_grad_oriented_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_solve_ex_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_tensorsolve_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_log10_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_log_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_log_normal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_logdet_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_logical_not_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_logspace_tensor_overload_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_long_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_mH_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_masked_logsumexp_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_masked_select_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_matmul_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_matmul_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_matrix_exp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_meshgrid_list_of_tensors_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_min_reduction_no_dim_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_min_reduction_no_dim_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_movedim_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_multinomial_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nansum_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_new_empty_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_new_full_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_conv_transpose1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_conv_transpose2d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_fractional_max_pool2d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_huber_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_max_unpool1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_max_unpool3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_pixel_shuffle_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_pixel_unshuffle_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_poisson_nll_loss_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_tanhshrink_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nonzero_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_norm_fro_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_norm_inf_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_norm_nuc_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_ormqr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_permute_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_polar_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_prod_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_prod_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_randint_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_reciprocal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_reshape_as_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_resolve_conj_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_roll_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_scatter_reduce_prod_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_scatter_reduce_sum_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_sgn_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_sgn_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_sinc_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_slice_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_slice_scatter_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_softmax_with_dtype_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_special_hermite_polynomial_he_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_special_spherical_bessel_j0_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_squeeze_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_squeeze_multiple_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_std_unbiased_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_stft_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_sub_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_sub_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_sum_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_sum_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_t_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_tan_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_tensor_split_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_tile_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_tile_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_trace_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_triu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_true_divide_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_true_divide_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_unbind_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_unbind_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_unflatten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_unfold_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_uniform_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_unique_consecutive_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_unique_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_unsafe_chunk_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_unsafe_split_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_var_mean_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_var_unbiased_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_view_as_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_where_cuda_int64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_argwhere_cuda_int64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_clone_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_numpy_ref_diagflat_cuda_int64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_jiterator_2inputs_2outputs_cuda_float64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_linalg_cross_cuda_int64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_nn_functional_conv_transpose2d_cuda_float64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_permute_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_numpy_ref_ravel_cuda_int64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_repeat_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_numpy_ref_repeat_cuda_float64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_tensor_split_cuda_float64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_tile_cuda_int64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_transpose_cuda_int64, test/test_ops.py::TestCommonCUDA::test_out___rpow___cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_acosh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_asinh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_atleast_3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_bitwise_and_cuda_int64, test/test_ops.py::TestCommonCUDA::test_out__refs_bitwise_right_shift_cuda_int64, test/test_ops.py::TestCommonCUDA::test_out__refs_cos_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_deg2rad_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_expand_as_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_expand_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_exponential_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_fft_hfft_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_fft_ihfft2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_fft_rfft2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_flip_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_igammac_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_isclose_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_log1p_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_logical_xor_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_masked_fill_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_nn_functional_glu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_nn_functional_hinge_embedding_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_nn_functional_huber_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_nn_functional_mish_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_nn_functional_tanhshrink_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_normal__in_place_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_ones_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_permute_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_rsqrt_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_rsub_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_sin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_special_i0e_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_special_multigammaln_mvlgamma_p_3_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_abs_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_argsort_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_as_strided_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_chalf_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_chunk_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_conj_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_cummax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_diagonal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_dist_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_fft_rfftn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_floor_divide_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_igammac_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_index_reduce_amax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_index_reduce_amin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_int_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_lcm_cuda_int64, test/test_ops.py::TestCommonCUDA::test_out_linalg_lstsq_grad_oriented_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_linalg_vander_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_logical_or_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_masked_cumprod_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_masked_softmax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_masked_softmin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_matmul_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_alpha_dropout_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_avg_pool3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_celu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_conv3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_gelu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_glu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_hinge_embedding_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_max_pool3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_selu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_randn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_repeat_interleave_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_asin_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_atan_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_cholesky_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_cholesky_inverse_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_cos_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_cos_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_cosh_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_cumsum_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_div_trunc_rounding_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_dot_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_exp2_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_expm1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_inner_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_kthvalue_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_inv_ex_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_lu_factor_ex_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_matrix_power_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_qr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_solve_ex_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_svdvals_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_tensorsolve_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_vecdot_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_log_softmax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_lu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_maximum_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_msort_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_nn_functional_avg_pool2d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_norm_fro_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_rsqrt_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_sinc_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_sinc_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_sinh_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_special_entr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_special_erfcx_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_squeeze_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_svd_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_take_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_triangular_solve_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_tril_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_true_divide_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_unbind_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_unfold_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_var_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_resolve_conj_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_round_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_scatter_reduce_amax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_short_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_signal_windows_cosine_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_signal_windows_hann_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_sinh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_slice_scatter_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_softmax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_special_chebyshev_polynomial_v_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_special_i1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_special_log_ndtr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_special_spherical_bessel_j0_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_sqrt_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_to_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_unsqueeze_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_var_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_view_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_warning_T_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_T_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs__conversions_byte_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_addr_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_allclose_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_asin_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_asinh_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_bitwise_and_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_bitwise_or_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_cat_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_clamp_min_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_count_nonzero_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_div_no_rounding_mode_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_fft_fft2_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_fft_hfft2_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_fft_ifft2_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_fft_ihfftn_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_fmod_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_hstack_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_isclose_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_lerp_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_linalg_cross_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_log2_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_logical_not_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_new_ones_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_nn_functional_pairwise_distance_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_nn_functional_prelu_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_nn_functional_relu6_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_renorm_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_round_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_special_i1_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_stack_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_stft_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_t_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_tanh_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_tril_indices_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_true_divide_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_unsqueeze_copy_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_var_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_where_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_abs_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_add_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_all_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_argmax_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_as_strided_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_as_strided_partial_views_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_asin_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_bernoulli_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_bitwise_or_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_cos_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_diagflat_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_diff_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_digamma_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_div_no_rounding_mode_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_empty_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_empty_like_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_erfinv_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_fmax_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_fmin_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_geqrf_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_gt_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_linalg_cross_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_linalg_eigvalsh_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_linalg_matrix_rank_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_mH_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_mT_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_masked_logsumexp_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_masked_mean_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_max_pool2d_with_indices_backward_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_mean_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_minimum_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_mm_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_msort_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_mvlgamma_mvlgamma_p_1_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_new_empty_strided_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_new_full_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_new_zeros_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_adaptive_avg_pool1d_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_adaptive_avg_pool3d_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_adaptive_max_pool3d_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_avg_pool1d_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_cosine_embedding_loss_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_cross_entropy_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_ctc_loss_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_instance_norm_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_interpolate_linear_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_margin_ranking_loss_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_pixel_unshuffle_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_rrelu_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_silu_complex_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_triplet_margin_loss_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_triplet_margin_with_distance_loss_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_polygamma_polygamma_n_0_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_positive_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_pow_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_rand_like_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_resize_as__cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_resolve_neg_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_scatter_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_signal_windows_general_cosine_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_softmax_with_dtype_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_special_airy_ai_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_special_bessel_j1_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_special_entr_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_special_hermite_polynomial_h_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_special_i1e_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_special_modified_bessel_k1_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_special_shifted_chebyshev_polynomial_u_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_split_with_sizes_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_unflatten_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_var_unbiased_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_zero__cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_zeros_cuda, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float___rdiv___cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_asinh_cuda_int32, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_atan2_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_copysign_cuda_int64, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_div_no_rounding_mode_cuda_int8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_div_no_rounding_mode_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_erf_cuda_int8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_erfinv_cuda_bool, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_float_power_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_i0_cuda_bool, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_lgamma_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_log10_cuda_int32, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_log1p_cuda_int8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_masked_var_cuda_int32, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_mvlgamma_mvlgamma_p_1_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_mvlgamma_mvlgamma_p_3_cuda_int32, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_polygamma_polygamma_n_2_cuda_int32, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_polygamma_polygamma_n_4_cuda_int64, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_rad2deg_cuda_bool, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_sigmoid_cuda_bool, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_sigmoid_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_sigmoid_cuda_int64, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_sinh_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_chebyshev_polynomial_u_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_chebyshev_polynomial_v_cuda_int8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_hermite_polynomial_h_cuda_bool, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_hermite_polynomial_h_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_laguerre_polynomial_l_cuda_int64, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_shifted_chebyshev_polynomial_t_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_shifted_chebyshev_polynomial_v_cuda_int32, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_xlog1py_cuda_bool, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_xlog1py_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_zeta_cuda_int32, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_tan_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_true_divide_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_xlogy_cuda_int8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_xlogy_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_T_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_T_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_T_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_cfloat_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_chalf_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_char_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_double_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_float_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_float_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_half_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_half_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_int_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_long_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_acos_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_add_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_add_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_addcmul_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_addr_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_addr_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_alias_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_amin_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_amin_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_arange_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_arange_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_as_strided_copy_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_asin_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_asinh_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atan2_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atan_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atan_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atanh_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atleast_1d_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atleast_1d_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atleast_1d_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atleast_3d_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_bitwise_right_shift_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_block_diag_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_broadcast_tensors_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_broadcast_to_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_cat_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_ceil_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_ceil_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_ceil_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_clamp_max_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_clamp_min_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_column_stack_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_conj_physical_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_conj_physical_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_conj_physical_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_contiguous_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_copysign_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_cos_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_cos_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_cosh_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_cosh_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_cumprod_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_cumsum_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_deg2rad_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_deg2rad_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_diag_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_diagonal_copy_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_diagonal_copy_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_diagonal_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_diagonal_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_diagonal_scatter_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_div_floor_rounding_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_div_floor_rounding_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_div_no_rounding_mode_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_dsplit_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_empty_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_empty_like_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_empty_strided_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_equal_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_erf_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_exp2_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_exp2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_exp2_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_expand_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_expand_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_exponential_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_eye_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_fft2_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_fft2_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_fft2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_fft_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_fft_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_hfft_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_hfft_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_hfftn_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_ifft_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_ifftn_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_ifftn_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_ifftshift_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_irfft2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_irfft2_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_irfft2_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_irfft_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_irfftn_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_irfftn_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_irfftn_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_rfft2_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_rfft2_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fill_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fill_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_flatten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_flip_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_flip_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fliplr_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_flipud_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_float_power_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_float_power_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_floor_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_floor_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fmax_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fmod_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fmod_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_frexp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_frexp_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_ge_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_ge_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_ge_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_geometric_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_geometric_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_geometric_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_gt_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_heaviside_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_heaviside_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_hsplit_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_hstack_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_i0_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_i0_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_index_copy_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_index_copy_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_index_copy_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_index_select_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isclose_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isinf_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isnan_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isneginf_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isneginf_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isneginf_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isreal_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isreal_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_item_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_item_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_le_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_le_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_le_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_lerp_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_lgamma_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_lgamma_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_cross_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_diagonal_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_diagonal_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_matrix_norm_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_svd_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_vecdot_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linspace_tensor_overload_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_log10_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_log10_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_log10_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_log10_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_log1p_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_log1p_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_log2_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_log2_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_log_softmax_with_dtype_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logical_and_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logical_and_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logical_not_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logical_not_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logsumexp_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_lt_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_masked_fill_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_maximum_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_maximum_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_meshgrid_list_of_tensors_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_meshgrid_list_of_tensors_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_meshgrid_variadic_tensors_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_movedim_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_mul_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_mul_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_mul_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_mul_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_narrow_copy_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_narrow_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_narrow_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_narrow_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_neg_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_neg_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_new_empty_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_new_empty_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_new_empty_strided_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_new_ones_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_new_ones_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_new_zeros_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nextafter_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_celu_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_channel_shuffle_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_dropout_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_gelu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_hardshrink_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_huber_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_huber_loss_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_l1_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_log_softmax_with_dtype_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_pairwise_distance_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_pixel_unshuffle_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_pixel_unshuffle_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_prelu_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_relu6_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_relu6_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_smooth_l1_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_softmin_with_dtype_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_softplus_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_tanhshrink_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_normal__in_place_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_normal__in_place_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_normal_number_mean_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_permute_copy_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_permute_copy_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_permute_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_positive_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_positive_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_pow_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_pow_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_prod_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_prod_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_ravel_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_ravel_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_ravel_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_real_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_remainder_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_remainder_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_repeat_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_reshape_as_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_reshape_as_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_reshape_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_roll_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_roll_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_roll_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_rot90_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_select_scatter_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sgn_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sgn_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sigmoid_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_signbit_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_signbit_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sin_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sin_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_erfcx_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_i0e_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_i0e_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_i1e_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_i1e_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_log_ndtr_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_log_ndtr_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_log_softmax_with_dtype_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_log_softmax_with_dtype_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_log_softmax_with_dtype_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_log_softmax_with_dtype_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_log_softmax_with_dtype_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_logit_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_multigammaln_mvlgamma_p_1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_multigammaln_mvlgamma_p_1_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_multigammaln_mvlgamma_p_1_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_ndtr_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_softmax_with_dtype_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_spherical_bessel_j0_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_spherical_bessel_j0_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_spherical_bessel_j0_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_xlog1py_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sqrt_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_square_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_square_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_squeeze_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_squeeze_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_squeeze_multiple_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_squeeze_multiple_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_stack_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_stack_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_std_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_std_mean_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sum_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_t_copy_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_take_along_dim_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_tan_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_tan_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_tanh_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_tanh_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_tanh_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_tensor_split_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_tensor_split_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_to_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_to_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_to_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_trace_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_trace_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_transpose_copy_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_transpose_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_triu_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_triu_indices_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_trunc_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_unbind_copy_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_unflatten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_unflatten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_unfold_copy_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_var_mean_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_vdot_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_view_as_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_view_as_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_view_as_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_view_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_view_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_vstack_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_vstack_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_where_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_zeros_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_diag_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_diag_embed_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_div_floor_rounding_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_heaviside_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_maximum_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_native_layer_norm_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_neg_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_vsplit_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_xlogy_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_bool_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_bool_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_byte_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_cdouble_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_cfloat_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_cfloat_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_cfloat_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_cfloat_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_chalf_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_chalf_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_float_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_half_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_half_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_acos_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_acosh_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_acosh_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_add_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_add_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_addcmul_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_addr_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_alias_copy_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_alias_copy_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_alias_copy_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_all_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_all_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_allclose_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_amax_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_amin_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_amin_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_as_strided_copy_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_as_strided_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_as_strided_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_as_strided_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_as_strided_partial_views_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_asin_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_asinh_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atan_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atan_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atleast_1d_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atleast_1d_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atleast_3d_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_bitwise_left_shift_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_bitwise_not_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_bitwise_or_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_block_diag_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_block_diag_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_bucketize_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cat_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cat_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_chunk_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_clamp_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_column_stack_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_column_stack_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_column_stack_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_conj_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_conj_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_conj_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_conj_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_conj_physical_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_contiguous_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_contiguous_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_copysign_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cos_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cos_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cos_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cumsum_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cumsum_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_deg2rad_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_deg2rad_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_diag_embed_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_diagonal_copy_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_diagonal_copy_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_diagonal_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_diagonal_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_digamma_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_div_no_rounding_mode_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_div_no_rounding_mode_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_div_trunc_rounding_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_div_trunc_rounding_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_dsplit_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_dsplit_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_dstack_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_empty_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_empty_like_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_empty_like_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_equal_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_erf_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_erfc_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_erfc_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_exp_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_exp_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_exp_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_expand_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_expand_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_expm1_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_expm1_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_expm1_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_exponential_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_eye_executor_aten_cuda_float8_e4m3fn, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_eye_executor_aten_cuda_float8_e4m3fnuz, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_fft2_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_fftshift_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_hfft2_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_hfft_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_hfftn_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_hfftn_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ifft2_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ifft2_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ifft_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ifftshift_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ifftshift_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ihfftn_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_irfft2_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_irfft_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_irfft_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_irfft_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_rfft2_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fill_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fill_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_flatten_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_flatten_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_flipud_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_flipud_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_flipud_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_float_power_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_float_power_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_floor_divide_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fmax_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_frac_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_gcd_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_geometric_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_gt_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_gt_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_heaviside_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_hsplit_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_hsplit_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_hsplit_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_hsplit_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_imag_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_index_add_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_index_copy_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_index_copy_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_index_copy_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_index_fill_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_index_fill_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_index_select_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_index_select_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_index_select_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_isclose_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_isclose_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_isclose_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_isfinite_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_isfinite_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_isinf_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_isneginf_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_le_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_lgamma_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_lgamma_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_lgamma_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linalg_cross_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linalg_diagonal_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linalg_norm_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linalg_vector_norm_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linspace_tensor_overload_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linspace_tensor_overload_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linspace_tensor_overload_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linspace_tensor_overload_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linspace_tensor_overload_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log10_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log1p_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log2_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log2_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log_softmax_with_dtype_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log_softmax_with_dtype_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_not_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_or_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_or_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logsumexp_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_lt_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_masked_fill_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_meshgrid_list_of_tensors_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_meshgrid_variadic_tensors_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_minimum_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_movedim_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_mul_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_mul_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nan_to_num_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_narrow_copy_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_narrow_copy_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_narrow_copy_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_ne_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_neg_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_neg_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_new_empty_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_new_empty_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_new_full_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_new_full_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_new_zeros_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_new_zeros_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_new_zeros_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_alpha_dropout_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_channel_shuffle_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_elu_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_elu_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_hardtanh_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_huber_loss_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_log_softmax_with_dtype_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_margin_ranking_loss_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_nll_loss_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_pairwise_distance_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_pairwise_distance_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_pixel_unshuffle_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_relu_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_softmax_with_dtype_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_softmin_with_dtype_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_tanhshrink_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_threshold_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_triplet_margin_loss_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_normal__in_place_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_permute_copy_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_rad2deg_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_randn_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_ravel_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_ravel_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_ravel_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_ravel_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_real_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_real_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_reciprocal_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_remainder_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_remainder_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_remainder_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_remainder_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_renorm_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_reshape_as_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_reshape_as_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_reshape_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_reshape_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_round_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_round_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_rsqrt_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_rsqrt_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_rsub_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_select_scatter_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_select_scatter_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sgn_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sgn_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sigmoid_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sign_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sign_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_signbit_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sin_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sin_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sinc_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_softmax_with_dtype_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_bessel_j0_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_entr_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_erfcx_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_i0e_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_i1_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_log_ndtr_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_log_softmax_with_dtype_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_logit_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_multigammaln_mvlgamma_p_1_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_multigammaln_mvlgamma_p_5_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_softmax_with_dtype_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_spherical_bessel_j0_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_xlog1py_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_zeta_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sqrt_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sqrt_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_square_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_squeeze_copy_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_squeeze_multiple_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_stack_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_std_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_std_mean_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sub_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sub_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_t_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_t_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_t_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_t_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_tan_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_tan_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_tensor_split_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_transpose_copy_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_transpose_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_tril_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_tril_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unbind_copy_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unbind_copy_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unbind_copy_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unbind_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unflatten_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unflatten_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unfold_copy_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unfold_copy_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unfold_copy_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unfold_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unfold_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unsqueeze_copy_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unsqueeze_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_var_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_var_mean_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_view_as_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_view_copy_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_view_copy_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_view_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_vsplit_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_where_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_T_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_T_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_T_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_T_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_T_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_bfloat16_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_bool_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_bool_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_cdouble_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_char_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_double_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_double_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_float_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_float_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_half_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_int_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_int_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_long_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_abs_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_acos_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_acosh_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_add_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_alias_copy_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_alias_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_all_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_amax_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_any_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_any_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_arange_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_as_strided_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_as_strided_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_as_strided_partial_views_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_as_strided_scatter_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_as_strided_scatter_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_asin_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_asin_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_asinh_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_asinh_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_atan_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_atleast_2d_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_atleast_2d_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_atleast_2d_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_bitwise_or_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_bitwise_xor_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_bitwise_xor_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_bitwise_xor_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_block_diag_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_block_diag_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_broadcast_to_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_broadcast_to_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cauchy_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_ceil_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_chunk_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_chunk_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_clamp_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_clamp_max_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_clamp_max_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_clamp_min_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_clone_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_conj_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_conj_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_conj_physical_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_copysign_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_copysign_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_copysign_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cos_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cos_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cosh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cosh_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_deg2rad_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_diag_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_diag_embed_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_diag_embed_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_diagonal_copy_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_diagonal_copy_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_diagonal_copy_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_diagonal_scatter_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_diagonal_scatter_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_div_trunc_rounding_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_dot_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_dot_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_dstack_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_empty_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_empty_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_empty_like_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_empty_like_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_empty_like_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_eq_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_equal_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_equal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_equal_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_erf_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_erfc_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_exp2_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_exp2_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_exp2_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_exp2_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_exp_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_expand_copy_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_expand_copy_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_expm1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_eye_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_fft2_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_fft_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_hfft2_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_hfft2_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_hfft_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_hfft_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_hfftn_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_hfftn_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ifft2_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ifft_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ihfft_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ihfft_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_irfft2_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_irfft_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_irfftn_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_irfftn_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_rfft_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_rfftn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_rfftn_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_rfftn_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fill_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_flatten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_flip_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_flip_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_float_power_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_floor_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_floor_divide_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fmax_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fmin_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fmin_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fmod_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_ge_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_geometric_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_heaviside_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_hstack_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_hstack_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_hstack_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_i0_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_igammac_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_index_add_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_index_add_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_index_add_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_index_add_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_index_fill_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_index_fill_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_isclose_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_isfinite_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_isinf_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_isreal_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_isreal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_isreal_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_item_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_item_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_lcm_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_lerp_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_lgamma_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linalg_cross_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linalg_cross_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linalg_diagonal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linalg_matrix_norm_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linalg_norm_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linalg_vecdot_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linalg_vecdot_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linspace_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log10_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log10_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log10_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log1p_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logical_and_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logical_and_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logical_not_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logical_xor_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logspace_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logsumexp_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_masked_fill_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_maximum_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_maximum_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_meshgrid_list_of_tensors_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_meshgrid_list_of_tensors_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_meshgrid_list_of_tensors_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_meshgrid_variadic_tensors_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_meshgrid_variadic_tensors_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_minimum_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_movedim_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_mul_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nan_to_num_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nan_to_num_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_native_layer_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_neg_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_new_empty_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_new_full_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_new_ones_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_new_ones_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_new_ones_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_channel_shuffle_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_channel_shuffle_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_dropout_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_gelu_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_glu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_hardtanh_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_layer_norm_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_log_softmax_with_dtype_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_log_softmax_with_dtype_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_pairwise_distance_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_pixel_shuffle_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_pixel_shuffle_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_pixel_shuffle_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_pixel_unshuffle_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_poisson_nll_loss_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_prelu_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_relu6_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_softmax_with_dtype_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_tanhshrink_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_tanhshrink_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_threshold_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_triplet_margin_loss_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_normal__in_place_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_normal__in_place_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_normal_number_mean_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_normal_number_mean_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_permute_copy_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_permute_copy_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_permute_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_positive_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_pow_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_pow_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_pow_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_prod_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_prod_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_rad2deg_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_ravel_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_ravel_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_real_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_reciprocal_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_reciprocal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_renorm_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_repeat_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_repeat_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_reshape_as_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_reshape_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_reshape_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_rot90_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_rot90_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_round_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_round_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_rsqrt_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_rsqrt_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_rsqrt_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sigmoid_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sign_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sin_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sin_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sinc_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_entr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_entr_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_erfcx_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_i0e_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_i1_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_i1e_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_log_ndtr_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_log_softmax_with_dtype_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_logit_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_logit_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_multigammaln_mvlgamma_p_5_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_multigammaln_mvlgamma_p_5_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_ndtr_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_ndtr_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_ndtr_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_ndtr_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_xlog1py_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_xlog1py_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_zeta_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_split_with_sizes_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sqrt_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_square_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_squeeze_copy_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_squeeze_copy_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_squeeze_multiple_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_stack_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_std_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sum_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sum_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sum_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_t_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tan_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tan_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tan_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tanh_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tensor_split_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tensor_split_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tensor_split_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_trace_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_transpose_copy_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_transpose_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_transpose_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tril_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_triu_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_triu_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_triu_indices_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_true_divide_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_true_divide_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_trunc_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_trunc_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unflatten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unflatten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unflatten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unfold_copy_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unsqueeze_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_vdot_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_view_as_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_view_as_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_view_copy_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_view_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_view_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_view_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_vsplit_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_vsplit_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_vstack_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_where_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_where_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_xlogy_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_zeros_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_T_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_bfloat16_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_bool_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_cfloat_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_double_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_float_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_float_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_short_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_short_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_abs_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_abs_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_acos_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_acos_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_add_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_addcmul_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_addcmul_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_addcmul_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_addr_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_any_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_any_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_arange_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_copy_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_partial_views_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_scatter_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_scatter_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_scatter_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_asinh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_asinh_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atan2_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atan2_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atan_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atanh_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atleast_2d_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atleast_2d_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atleast_3d_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_bitwise_left_shift_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_bitwise_right_shift_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_bitwise_xor_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_block_diag_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_block_diag_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_broadcast_tensors_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_broadcast_to_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_broadcast_to_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_broadcast_to_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cat_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cat_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_ceil_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_chunk_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_clamp_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_clamp_max_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_clamp_min_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_clone_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_clone_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_column_stack_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_conj_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_conj_physical_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_conj_physical_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_conj_physical_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_conj_physical_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_copysign_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cos_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cosh_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cosh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cosh_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_count_nonzero_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_count_nonzero_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cumsum_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cumsum_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cumsum_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_deg2rad_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_diag_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_diag_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_diagonal_copy_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_diagonal_copy_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_diagonal_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_diagonal_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_diagonal_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_diagonal_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_div_no_rounding_mode_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_div_no_rounding_mode_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_div_no_rounding_mode_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_dstack_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_empty_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_empty_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_empty_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_empty_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_empty_like_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_empty_strided_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_eq_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_eq_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_eq_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_eq_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_equal_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_erf_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_erfinv_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_erfinv_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_exp_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_expand_as_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_expand_copy_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_expand_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_expand_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_expm1_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_eye_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_fft2_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_fft_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_fftn_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_fftn_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_fftshift_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_ifft_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_ifftshift_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_ihfft2_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_ihfft_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_ihfft_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_irfft2_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_irfft_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_flip_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_flipud_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_flipud_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_float_power_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_floor_divide_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fmax_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fmin_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_frac_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_ge_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_ge_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_geometric_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_gt_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_gt_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_hstack_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_i0_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_i0_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_i0_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_index_add_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_index_fill_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_index_fill_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_index_select_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isclose_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isfinite_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isfinite_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isfinite_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isneginf_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isposinf_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_istft_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_lcm_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_lcm_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_le_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_lgamma_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_linalg_cross_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_linalg_cross_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_linalg_cross_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_linalg_diagonal_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_linalg_svd_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_linalg_svdvals_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_linalg_vecdot_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_linspace_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_linspace_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_linspace_tensor_overload_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log10_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log1p_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log2_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log2_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log2_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log_normal_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log_softmax_with_dtype_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logical_and_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logical_or_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logical_xor_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logical_xor_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logspace_tensor_overload_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logspace_tensor_overload_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_lt_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_lt_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_masked_fill_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_masked_fill_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_masked_fill_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_maximum_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_maximum_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_maximum_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_mean_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_meshgrid_list_of_tensors_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_meshgrid_list_of_tensors_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_movedim_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_movedim_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_movedim_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_movedim_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_narrow_copy_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_narrow_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_native_layer_norm_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_native_layer_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_neg_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_empty_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_empty_strided_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_full_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_full_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_full_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_ones_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_ones_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_ones_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_zeros_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_zeros_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nextafter_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_channel_shuffle_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_channel_shuffle_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_channel_shuffle_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_group_norm_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_group_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_hardtanh_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_hardtanh_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_huber_loss_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_l1_loss_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_log_softmax_with_dtype_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_log_softmax_with_dtype_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_log_softmax_with_dtype_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_log_softmax_with_dtype_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_pairwise_distance_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_pairwise_distance_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_pdist_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_pixel_shuffle_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_pixel_shuffle_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_pixel_unshuffle_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_prelu_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_softmax_with_dtype_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_softmax_with_dtype_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_softmin_with_dtype_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_softmin_with_dtype_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_softmin_with_dtype_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_softshrink_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_softshrink_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_tanhshrink_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_threshold_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_threshold_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_threshold_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_threshold_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_triplet_margin_loss_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_triplet_margin_loss_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_norm_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_normal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_normal_number_mean_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_permute_copy_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_permute_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_permute_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_positive_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_positive_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_prod_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_prod_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_randn_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_ravel_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_real_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_renorm_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_renorm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_repeat_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_reshape_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_reshape_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_roll_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_rot90_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_rsqrt_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sgn_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sigmoid_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sin_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sin_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sinc_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sinh_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_bessel_j0_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_bessel_j0_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_entr_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_erfcx_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_i1_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_i1e_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_i1e_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_log_ndtr_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_log_ndtr_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_log_ndtr_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_logit_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_multigammaln_mvlgamma_p_1_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_multigammaln_mvlgamma_p_3_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_ndtri_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_xlog1py_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_zeta_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_split_with_sizes_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_split_with_sizes_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_split_with_sizes_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_split_with_sizes_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_split_with_sizes_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_split_with_sizes_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_split_with_sizes_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sqrt_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sqrt_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_square_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_square_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_squeeze_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_squeeze_multiple_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_squeeze_multiple_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_stack_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sub_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sum_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sum_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sum_to_size_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sum_to_size_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sum_to_size_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_t_copy_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_t_copy_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_take_along_dim_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_tan_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_tanh_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_tanh_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_trace_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_trace_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_trace_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_transpose_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_tril_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_triu_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_true_divide_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_trunc_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unbind_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unbind_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unfold_copy_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unfold_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unfold_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unfold_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unsqueeze_copy_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_vdot_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_view_as_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_view_as_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_view_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_view_copy_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_view_copy_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_view_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_vsplit_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_vsplit_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_where_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_reduction_ops_reduce_any_cuda, test/test_ops.py::TestCommonCUDA::test_reduction_ops_reduce_min_reduction_with_dim_cuda, test/test_ops.py::TestCommonCUDA::test_reduction_ops_reduce_prod_cuda, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_T_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager___getitem___cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager___rpow___cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager__unsafe_masked_index_put_accumulate_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_amax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_as_strided_partial_views_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_as_strided_scatter_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_asin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_atleast_1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_atleast_3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_bool_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_bucketize_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_cfloat_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_cholesky_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_cholesky_inverse_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_column_stack_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_copysign_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_cross_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_empty_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_exp2_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_expand_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_expand_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_fft_fft2_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_fft_fft_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_fft_fftn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_flatten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_flipud_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_half_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_i0_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_isposinf_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_pinv_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_pinv_singular_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_qr_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_slogdet_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_svdvals_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linspace_tensor_overload_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_log_softmax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_logdet_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_masked_amin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_masked_cumprod_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_masked_prod_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_masked_select_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_new_full_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_avg_pool2d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_binary_cross_entropy_with_logits_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_dropout3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_embedding_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_group_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_interpolate_linear_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_kl_div_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_margin_ranking_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_max_unpool2d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_pad_constant_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_pad_replicate_negative_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_pairwise_distance_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_pairwise_distance_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_prelu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_rrelu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_triplet_margin_loss_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_triplet_margin_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nonzero_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_norm_nuc_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_ones_like_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_pca_lowrank_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_randn_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_randn_like_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_ravel_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_remainder_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_repeat_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_resolve_neg_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_select_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_select_scatter_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_signal_windows_bartlett_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_slice_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_special_airy_ai_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_special_bessel_j1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_special_chebyshev_polynomial_t_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_special_i1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_special_log_ndtr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_special_modified_bessel_k1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_special_ndtr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_special_spherical_bessel_j0_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_stft_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_sum_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_t_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_t_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_t_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_to_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_torch_ops_aten__efficient_attention_forward_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_tril_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_unfold_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_unsqueeze_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_unsqueeze_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_view_as_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_view_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_vstack_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_addmv_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_amax_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_atleast_1d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_cartesian_prod_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_chalf_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_clamp_max_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_complex_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_cummax_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_diag_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_diagonal_scatter_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_fft_rfft_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_fmax_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_grid_sampler_3d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_index_add_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_kron_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_ldexp_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_lerp_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_linalg_diagonal_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_linalg_eig_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_linalg_lu_solve_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_log_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_masked_log_softmax_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_meshgrid_list_of_tensors_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_mode_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nanquantile_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_cosine_embedding_loss_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_ctc_loss_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_hardshrink_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_hardtanh_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_hinge_embedding_loss_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_interpolate_area_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_interpolate_bicubic_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_max_unpool3d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_mish_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_pad_circular_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_pad_reflect_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_pad_replicate_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_prelu_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_selu_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_polar_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_polygamma_polygamma_n_2_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_qr_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_quantile_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_repeat_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_scatter_reduce_sum_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_sin_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_squeeze_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_svd_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_svd_lowrank_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_triangular_solve_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_tril_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_vdot_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input___rpow___cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_arange_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_argmin_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_atanh_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_atleast_2d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_bernoulli_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_broadcast_shapes_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_broadcast_tensors_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_cartesian_prod_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_cholesky_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_cholesky_solve_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_clamp_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_complex_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_cov_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_cummax_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_diagonal_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_diagonal_scatter_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_diff_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_einsum_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_empty_like_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_fft_rfft_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_float_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_int_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_isinf_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_jiterator_4inputs_with_extra_args_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_linalg_diagonal_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_linalg_eig_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_linalg_inv_ex_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_linalg_pinv_singular_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_masked_logsumexp_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_masked_softmin_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_max_reduction_with_dim_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_maximum_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_avg_pool1d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_binary_cross_entropy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_channel_shuffle_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_ctc_loss_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_embedding_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_huber_loss_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_interpolate_area_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_interpolate_bicubic_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_interpolate_bilinear_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_max_unpool2d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_soft_margin_loss_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_norm_fro_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_normal_number_mean_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_polygamma_polygamma_n_1_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_rand_like_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_resolve_conj_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_sin_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_special_i0e_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_split_list_args_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_svd_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_tensor_split_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_to_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_transpose_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_unsafe_split_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_H_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_T_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad__batch_norm_with_update_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad__native_batch_norm_legit_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad__unsafe_masked_index_put_accumulate_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad__upsample_bilinear2d_aa_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_addmm_decomposed_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_addmv_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_all_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_asin_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_bernoulli_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_deg2rad_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_equal_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_erfinv_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_fft_fft_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_floor_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_frac_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_gt_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_igamma_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_linalg_eig_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_linalg_inv_ex_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_linalg_lstsq_grad_oriented_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_linalg_lu_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_log1p_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_log_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_log_normal_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_logspace_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_masked_cumprod_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_masked_cumsum_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_masked_select_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_msort_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_new_zeros_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_adaptive_avg_pool3d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_adaptive_max_pool3d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_elu_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_embedding_bag_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_fractional_max_pool2d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_fractional_max_pool3d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_instance_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_margin_ranking_loss_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_max_unpool1d_grad_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_pad_reflect_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_pad_replicate_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_prelu_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_scaled_dot_product_attention_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_normal_number_mean_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_outer_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_reshape_as_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_reshape_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_resolve_neg_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_short_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_signal_windows_blackman_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_signal_windows_general_hamming_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_special_erfcx_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_special_modified_bessel_k1_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_special_ndtr_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_special_shifted_chebyshev_polynomial_v_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_special_spherical_bessel_j0_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_split_with_sizes_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_squeeze_multiple_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_tensor_split_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_to_sparse_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_trunc_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_unsqueeze_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_view_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator__unsafe_masked_index_put_accumulate_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_argmin_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_atleast_3d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_baddbmm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_cholesky_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_cumulative_trapezoid_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_dstack_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_einsum_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_fft_hfft_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_fft_ihfft2_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_i0_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_jiterator_4inputs_with_extra_args_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_linalg_eigh_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_linalg_eigvalsh_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_linalg_ldl_factor_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_linalg_matrix_rank_hermitian_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_log1p_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_masked_logsumexp_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_masked_std_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_masked_sum_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_max_reduction_no_dim_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nan_to_num_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nextafter_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_adaptive_avg_pool2d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_avg_pool1d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_conv2d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_conv3d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_embedding_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_l1_loss_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_logsigmoid_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_max_pool3d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_pad_reflect_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_scaled_dot_product_attention_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_tanhshrink_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nonzero_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_polygamma_polygamma_n_2_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_ravel_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_sort_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_squeeze_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_svd_lowrank_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_t_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_torch_ops_aten__safe_softmax_default_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_true_divide_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_view_as_complex_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_vsplit_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_addbmm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_addmm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_asin_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_atan2_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_atanh_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_char_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_erf_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_exp_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_full_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_full_like_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_gradient_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_gt_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_isposinf_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_linalg_eigh_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_linalg_matrix_power_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_linalg_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_log10_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_logspace_tensor_overload_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_logsumexp_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_lu_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_max_pool2d_with_indices_backward_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_max_reduction_with_dim_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nanmean_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_adaptive_max_pool3d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_channel_shuffle_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_conv_transpose1d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_dropout2d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_fractional_max_pool3d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_hardsigmoid_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_interpolate_bicubic_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_max_unpool3d_grad_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_multi_margin_loss_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_rms_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_normal_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_permute_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_polygamma_polygamma_n_1_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_prod_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_randint_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_reshape_as_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_scatter_reduce_sum_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_special_legendre_polynomial_p_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_special_modified_bessel_i0_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_special_scaled_modified_bessel_k1_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_special_shifted_chebyshev_polynomial_t_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_transpose_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_transpose_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_triangular_solve_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_view_as_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_view_copy_cuda_float32, test/test_ops.py::TestMathBitsCUDA::test_conj_view___rdiv___cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs__conversions_short_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_add_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_as_strided_scatter_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_atleast_2d_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_conj_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_constant_pad_nd_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_fft_fftn_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_fft_ifft_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_fft_irfft_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_flip_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_linalg_matrix_norm_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_linalg_vector_norm_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_logical_not_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_ones_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_reciprocal_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_tensor_split_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_trace_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_addcmul_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_as_strided_partial_views_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_asin_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_atan_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_cfloat_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_contiguous_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_cosh_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_diag_embed_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_diagonal_scatter_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_expand_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_fft_hfft2_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_fft_ifftn_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_float_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_gradient_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_index_add_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_jiterator_binary_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_ldexp_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_linalg_det_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_linalg_eigh_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_linalg_householder_product_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_linalg_inv_ex_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_linalg_lu_factor_ex_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_linalg_matrix_rank_hermitian_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_linalg_norm_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_linalg_pinv_singular_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_linalg_solve_ex_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_linalg_svdvals_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_log1p_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_nn_functional_conv2d_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_nn_functional_conv3d_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_nn_functional_feature_alpha_dropout_without_train_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_reshape_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_sin_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_split_list_args_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_split_with_sizes_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_sqrt_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_std_mean_unbiased_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_svd_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_transpose_copy_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_tril_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_triu_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_vdot_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_view_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs__conversions_bool_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_addr_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_as_strided_copy_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_atleast_3d_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_diag_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_dot_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_fft_ifft_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_linalg_cross_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_linalg_matrix_norm_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_neg_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_nn_functional_channel_shuffle_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_nn_functional_softmax_with_dtype_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_squeeze_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_t_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_transpose_copy_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_where_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__unsafe_masked_index_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_bool_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_cholesky_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_cholesky_inverse_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_chunk_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_cov_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_diagflat_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_diff_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_div_no_rounding_mode_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_fft_hfft_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_fft_ifft_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_fft_irfft2_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_fill_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_isclose_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_linalg_det_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_linalg_lu_factor_ex_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_linalg_pinv_hermitian_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_linalg_pinv_singular_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_linalg_vecdot_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_log_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_logical_not_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_logical_or_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_logspace_tensor_overload_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_mH_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_masked_normalize_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_masked_std_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_neg_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_nn_functional_rms_norm_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_nonzero_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_norm_fro_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_normal_in_place_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_ormqr_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_outer_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_put_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_rsub_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_scatter_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_sinh_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_split_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_sqrt_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_sum_to_size_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_take_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_trapezoid_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_unsafe_split_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_view_as_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_view_copy_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_zero__cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_addr_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_allclose_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_atan_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_atleast_2d_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_clamp_min_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_conj_physical_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_dot_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_eye_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_fft_fft2_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_fft_fftshift_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_fft_ifftn_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_fft_ifftshift_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_fft_ihfft_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_fft_irfftn_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_geometric_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_isinf_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_logspace_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_movedim_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_nan_to_num_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_nn_functional_huber_loss_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_nn_functional_poisson_nll_loss_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_nn_functional_relu_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_normal__in_place_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_permute_copy_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_randn_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_real_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_remainder_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_roll_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_special_logit_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_special_ndtr_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_special_zeta_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_squeeze_multiple_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_take_along_dim_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_tan_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_zeros_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_bernoulli_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_broadcast_to_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_cdist_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_cholesky_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_cholesky_inverse_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_contiguous_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_corrcoef_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_count_nonzero_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_cumsum_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_digamma_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_dist_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_dot_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_exponential_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_fft_irfft2_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_float_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_geometric_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_histc_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_index_fill_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_isreal_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_kron_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_linalg_lstsq_grad_oriented_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_linalg_lu_factor_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_linalg_lu_factor_ex_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_lu_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_masked_amax_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_masked_argmax_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_masked_select_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_meshgrid_variadic_tensors_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_minimum_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nan_to_num_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_avg_pool3d_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_ctc_loss_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_huber_loss_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_interpolate_area_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_interpolate_linear_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_interpolate_trilinear_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_local_response_norm_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_max_unpool1d_grad_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_max_unpool2d_grad_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_mse_loss_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_multilabel_margin_loss_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_pad_reflect_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_soft_margin_loss_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_ones_like_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_polygamma_polygamma_n_1_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_randint_like_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_randn_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_randn_like_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_remainder_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_rot90_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_round_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_round_decimals_3_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_select_scatter_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_signal_windows_nuttall_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_sinh_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_special_hermite_polynomial_h_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_special_laguerre_polynomial_l_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_special_shifted_chebyshev_polynomial_w_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_squeeze_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_t_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_transpose_copy_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_trapezoid_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_tril_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_unsqueeze_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_var_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_vsplit_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_where_cuda_float64, test/test_ops.py::TestFakeTensorCUDA::test_fake___rmod___cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake___rsub___cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_addcmul_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_arange_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_asin_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_atanh_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast___rand___cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast___rmod___cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast__batch_norm_with_update_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast__native_batch_norm_legit_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast__unsafe_masked_index_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast__unsafe_masked_index_put_accumulate_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_addbmm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_alias_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_byte_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_cat_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_ceil_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_count_nonzero_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_einsum_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_eq_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_erfc_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_expand_as_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_expand_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_fft_ihfftn_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_fft_rfft2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_full_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_grid_sampler_3d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_hsplit_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_isin_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_linalg_eig_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_linalg_lstsq_grad_oriented_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_linalg_lu_factor_ex_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_linalg_lu_solve_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_linalg_solve_ex_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_linalg_solve_triangular_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_linspace_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_masked_scatter_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_max_reduction_no_dim_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_min_reduction_with_dim_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_multinomial_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_mvlgamma_mvlgamma_p_1_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nan_to_num_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nansum_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_neg_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_new_empty_strided_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_adaptive_max_pool2d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_binary_cross_entropy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_fractional_max_pool2d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_fractional_max_pool3d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_gelu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_multilabel_margin_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_triplet_margin_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_norm_fro_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_pca_lowrank_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_pinverse_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_renorm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_signal_windows_general_cosine_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_sparse_sampled_addmm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_special_bessel_y0_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_std_mean_unbiased_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_std_unbiased_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_t_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_tensor_split_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_torch__scaled_mm_v2_cuda_float8_e4m3fn, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_torch_ops_aten__safe_softmax_default_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_triangular_solve_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_unsafe_chunk_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_var_mean_unbiased_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_baddbmm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_cat_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_cauchy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_cholesky_inverse_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_column_stack_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp___rmul___cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_addmm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_addr_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_amax_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_angle_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_asin_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_atleast_2d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_cdouble_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_corrcoef_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_expand_as_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_fft_fft2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_fft_ifftn_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_fft_ifftshift_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_fft_ihfft2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_fft_ihfftn_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_fft_irfft2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_gather_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_hstack_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_index_add_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_linalg_eig_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_linalg_solve_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_mH_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_masked_cumprod_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_masked_var_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_mv_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_mvlgamma_mvlgamma_p_1_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nanmean_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nanmedian_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_adaptive_avg_pool2d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_alpha_dropout_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_cross_entropy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_group_norm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_margin_ranking_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_multi_margin_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_rrelu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_polygamma_polygamma_n_2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_resolve_conj_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_resolve_neg_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_rot90_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_round_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_scatter_reduce_prod_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_slice_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_special_xlog1py_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_square_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_std_mean_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_torch_ops_aten__safe_softmax_default_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_var_mean_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_var_mean_unbiased_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp__segment_reduce_lengths_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp__unsafe_masked_index_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp__upsample_bilinear2d_aa_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_alias_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_cdouble_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_ceil_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_combinations_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_contiguous_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_cov_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_div_no_rounding_mode_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_expand_as_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_fft_ifft2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_float_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_gather_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_index_add_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_index_reduce_amin_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_linalg_lstsq_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_linalg_lu_factor_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_linalg_svdvals_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_linalg_tensorsolve_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_lu_solve_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_masked_log_softmax_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_masked_prod_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_max_pool2d_with_indices_backward_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_mode_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nan_to_num_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_adaptive_max_pool1d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_batch_norm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_batch_norm_without_cudnn_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_binary_cross_entropy_with_logits_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_local_response_norm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_logsigmoid_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_multilabel_margin_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_selu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_silu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_soft_margin_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_normal_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_polygamma_polygamma_n_4_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_rad2deg_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_real_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_rsub_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_sqrt_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_stack_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_to_sparse_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_trapezoid_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_trunc_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_view_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_view_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_where_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_cumulative_trapezoid_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_fft_fftn_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_fft_ifft2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_fft_ifft_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_fft_irfft2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_fmin_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_gradient_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_index_add_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_index_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_linalg_inv_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_linalg_ldl_factor_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_linalg_lu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_linalg_pinv_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_linalg_tensorinv_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_linspace_tensor_overload_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_logaddexp2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_logical_not_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_logsumexp_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_long_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_masked_fill_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_max_binary_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_native_layer_norm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_conv_transpose1d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_hardsigmoid_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_hardtanh_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_interpolate_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_local_response_norm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_mish_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_multilabel_margin_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_one_hot_cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_pixel_unshuffle_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_rms_norm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_softshrink_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_pinverse_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_polygamma_polygamma_n_3_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_rsqrt_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_scatter_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_signal_windows_gaussian_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_signal_windows_general_hamming_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_sinc_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_special_bessel_y1_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_special_erfcx_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_special_ndtr_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_special_shifted_chebyshev_polynomial_u_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_special_xlog1py_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_stack_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_take_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_to_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_torch_ops_aten__safe_softmax_default_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_unique_consecutive_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_var_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_addr_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_argmax_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_as_strided_partial_views_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_asinh_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_atleast_1d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_atleast_2d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_bitwise_xor_cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_broadcast_tensors_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_cholesky_inverse_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_clamp_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_complex_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_cos_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_cumprod_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_cumulative_trapezoid_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_exp_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_fft_hfft_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_fft_rfft_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_fill_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_gt_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_hsplit_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_isposinf_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_linalg_cross_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_linalg_eigvals_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_linalg_lu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_linalg_lu_solve_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_linalg_matrix_rank_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_linalg_multi_dot_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_log_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_logical_not_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_logit_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_logspace_tensor_overload_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_masked_amax_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_masked_fill_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_native_batch_norm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_adaptive_max_pool1d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_avg_pool3d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_binary_cross_entropy_with_logits_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_fractional_max_pool2d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_hinge_embedding_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_interpolate_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_max_unpool2d_grad_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_softshrink_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_tanhshrink_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_unfold_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_upsample_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nonzero_static_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_positive_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_round_decimals_neg_3_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_scatter_reduce_amin_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_signal_windows_cosine_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_signal_windows_nuttall_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_softmax_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_sparse_sampled_addmm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_special_scaled_modified_bessel_k1_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_special_shifted_chebyshev_polynomial_v_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_split_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_svd_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_take_along_dim_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_topk_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_unsqueeze_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_var_mean_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_view_as_complex_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_arange_cuda_int32, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_linspace_cuda_bfloat16, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_linspace_cuda_int8, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_logspace_cuda_complex64, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_logspace_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_logspace_tensor_overload_cuda_int32, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_ones_cuda_bool, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_zeros_cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_linspace_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_linspace_cuda_int16, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_linspace_tensor_overload_cuda_complex128, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_logspace_cuda_float64, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_logspace_cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_ones_cuda_complex32, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_zeros_cuda_int64, test/test_ops.py::TestTagsCUDA::test_tags__refs_acos_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_add_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_bitwise_and_cuda_int64, test/test_ops.py::TestTagsCUDA::test_tags__refs_conj_physical_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_cosh_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_diagonal_scatter_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_fft_ihfft_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_fft_rfftn_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_float_power_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_hstack_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_isneginf_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_istft_cuda_complex64, test/test_ops.py::TestTagsCUDA::test_tags__refs_log1p_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_logaddexp_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_maximum_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_minimum_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_narrow_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_nn_functional_log_softmax_with_dtype_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_nn_functional_pairwise_distance_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_permute_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_positive_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_prod_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_sign_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_sin_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_sinh_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_special_bessel_j1_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_squeeze_multiple_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_transpose_copy_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_var_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_view_as_complex_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_atanh_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_bfloat16_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_bmm_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_constant_pad_nd_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_count_nonzero_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_diagonal_copy_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_digamma_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_dstack_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_empty_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_floor_divide_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_gather_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_geometric_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_index_select_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_kthvalue_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_linalg_cross_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_linalg_lu_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_linalg_vander_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_linalg_vecdot_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_linspace_tensor_overload_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_log_softmax_with_dtype_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_logical_or_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_masked_log_softmax_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_masked_softmax_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_max_reduction_no_dim_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_meshgrid_list_of_tensors_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_adaptive_max_pool3d_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_avg_pool3d_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_feature_alpha_dropout_with_train_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_hardshrink_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_interpolate_linear_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_linear_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_max_pool2d_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_max_pool3d_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_max_unpool1d_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_max_unpool3d_grad_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_relu6_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_scaled_dot_product_attention_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_silu_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_softmin_with_dtype_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_softsign_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_permute_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_pow_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_quantile_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_roll_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_round_decimals_neg_3_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_sigmoid_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_signal_windows_bartlett_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_signal_windows_general_hamming_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_sparse_mm_reduce_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_special_chebyshev_polynomial_t_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_special_chebyshev_polynomial_u_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_special_entr_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_squeeze_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_squeeze_multiple_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_sub_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_torch_ops_aten__efficient_attention_forward_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_transpose_copy_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_view_as_cuda_float32, test/test_ops.py::TestForwardADWithScalarsCUDA::test_0d_tensor_with_python_scalar_mul_cuda_float32 2025-12-04T14:47:55.2690472Z 2025-12-04T14:47:55.2690787Z Finished test_ops 6/11 ... [2025-12-04 14:47:55.020544][20058.710932104], took 19.75min 2025-12-04T14:47:55.2691829Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops/test_ops-e04c41c7bbbdce7a.xml 2025-12-04T14:47:55.2692789Z Running test_ops 11/11 ... [2025-12-04 14:47:55.221852][20058.912242595] 2025-12-04T14:47:55.2693296Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T14:47:55.2694509Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ops.py', '--shard-id=11', '--num-shards=11', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 14:47:55.222297] 2025-12-04T15:08:17.3432843Z 2025-12-04T15:08:17.3435467Z test_ops 11/11 was successful, full logs can be found in artifacts with path test/test-reports/test_ops_11.11_c9baa1a26aa25252_.log 2025-12-04T15:08:17.4695824Z Running 3042 items in this shard: test/test_ops.py::TestCommonCUDA::test_compare_cpu___radd___cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu___rand___cuda_int64, test/test_ops.py::TestCommonCUDA::test_compare_cpu__batch_norm_with_update_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs__conversions_byte_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_as_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_bitwise_left_shift_cuda_int64, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_div_floor_rounding_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_fft_ifftshift_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_fmax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_lerp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_linalg_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_mul_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_new_zeros_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_nn_functional_alpha_dropout_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_nn_functional_channel_shuffle_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_nn_functional_poisson_nll_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_nn_functional_softshrink_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_view_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu__refs_zeros_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_addmm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_argsort_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_bitwise_right_shift_cuda_int64, test/test_ops.py::TestCommonCUDA::test_compare_cpu_block_diag_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_bucketize_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_byte_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_cdist_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_cross_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_igamma_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_index_fill_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_index_reduce_amin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_linalg_ldl_solve_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_linalg_lu_solve_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_linalg_matrix_power_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_logaddexp2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_masked_fill_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_masked_softmax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_masked_softmin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_matrix_exp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_new_zeros_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_cosine_embedding_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_fractional_max_pool3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_hardswish_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_max_unpool1d_grad_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_max_unpool3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_multilabel_soft_margin_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_soft_margin_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_upsample_bilinear_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_upsample_nearest_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nonzero_static_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_ones_like_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_pinverse_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_reshape_as_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_resize__cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_sort_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_special_hermite_polynomial_he_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_std_mean_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_svd_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_torch__scaled_mm_cuda_float8_e4m3fn, test/test_ops.py::TestCommonCUDA::test_compare_cpu_triu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_unique_consecutive_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_var_mean_unbiased_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_zero__cuda_float32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_chalf_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_fft_irfft_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_flatten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_movedim_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_new_empty_strided_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_ones_like_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_ravel_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_reshape_as_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_scalar_tensor_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_split_with_sizes_copy_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_sqrt_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_stack_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_transpose_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_unbind_copy_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_unflatten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_complex_half_reference_testing_vstack_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_dtypes___rmod___cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_amin_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_block_diag_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_broadcast_shapes_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_bucketize_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_clamp_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_column_stack_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_count_nonzero_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_cumprod_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_expand_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_expm1_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_fft_irfft2_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_fft_irfft_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_hsplit_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_i0_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_igammac_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_le_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_linalg_svdvals_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_neg_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_norm_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_pow_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_randn_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_sin_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_special_i1_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_sum_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_unfold_copy_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_var_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes__refs_vsplit_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_abs_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_acosh_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_add_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_any_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_argmin_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_bitwise_or_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_block_diag_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_cdouble_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_cholesky_inverse_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_combinations_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_complex_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_cumulative_trapezoid_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_erf_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_histogram_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_igamma_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_index_reduce_mean_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_isreal_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_kthvalue_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_linalg_matrix_norm_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_linalg_pinv_hermitian_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_log1p_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_lu_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_lu_solve_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_max_reduction_with_dim_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_mean_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_meshgrid_list_of_tensors_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_min_binary_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nanmedian_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_adaptive_avg_pool2d_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_avg_pool1d_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_avg_pool2d_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_binary_cross_entropy_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_dropout2d_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_embedding_bag_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_grid_sample_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_hardswish_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_interpolate_trilinear_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_logsigmoid_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_max_unpool3d_grad_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_multi_margin_loss_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_pixel_unshuffle_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_softmin_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_softsign_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_nn_functional_upsample_nearest_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_polygamma_polygamma_n_2_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_roll_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_round_decimals_neg_3_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_scalar_tensor_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_short_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_sigmoid_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_sinh_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_slice_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_special_modified_bessel_k0_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_special_ndtri_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_tan_cuda, test/test_ops.py::TestCommonCUDA::test_dtypes_view_as_real_cuda, test/test_ops.py::TestCommonCUDA::test_errors___rmul___cuda, test/test_ops.py::TestCommonCUDA::test_errors_copysign_cuda, test/test_ops.py::TestCommonCUDA::test_errors_div_floor_rounding_cuda, test/test_ops.py::TestCommonCUDA::test_errors_div_no_rounding_mode_cuda, test/test_ops.py::TestCommonCUDA::test_errors_eye_cuda, test/test_ops.py::TestCommonCUDA::test_errors_fft_fft2_cuda, test/test_ops.py::TestCommonCUDA::test_errors_fft_ihfft2_cuda, test/test_ops.py::TestCommonCUDA::test_errors_fft_ihfftn_cuda, test/test_ops.py::TestCommonCUDA::test_errors_fft_irfft_cuda, test/test_ops.py::TestCommonCUDA::test_errors_heaviside_cuda, test/test_ops.py::TestCommonCUDA::test_errors_igamma_cuda, test/test_ops.py::TestCommonCUDA::test_errors_logcumsumexp_cuda, test/test_ops.py::TestCommonCUDA::test_errors_logical_or_cuda, test/test_ops.py::TestCommonCUDA::test_errors_nn_functional_conv3d_cuda, test/test_ops.py::TestCommonCUDA::test_errors_nn_functional_hinge_embedding_loss_cuda, test/test_ops.py::TestCommonCUDA::test_errors_reshape_as_cuda, test/test_ops.py::TestCommonCUDA::test_errors_roll_cuda, test/test_ops.py::TestCommonCUDA::test_errors_signal_windows_gaussian_cuda, test/test_ops.py::TestCommonCUDA::test_errors_sparse_mul_layout1_cuda, test/test_ops.py::TestCommonCUDA::test_errors_sparse_mul_layout2_cuda, test/test_ops.py::TestCommonCUDA::test_errors_sparse_sum_layout2_cuda, test/test_ops.py::TestCommonCUDA::test_errors_sparse_zeros_like_layout1_cuda, test/test_ops.py::TestCommonCUDA::test_errors_special_chebyshev_polynomial_v_cuda, test/test_ops.py::TestCommonCUDA::test_errors_special_legendre_polynomial_p_cuda, test/test_ops.py::TestCommonCUDA::test_errors_special_shifted_chebyshev_polynomial_v_cuda, test/test_ops.py::TestCommonCUDA::test_errors_t_cuda, test/test_ops.py::TestCommonCUDA::test_errors_tril_cuda, test/test_ops.py::TestCommonCUDA::test_errors_unbind_copy_cuda, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_addbmm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_angle_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_arange_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_argmax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_asinh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_baddbmm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_fft_ihfft2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_fft_irfftn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_fft_rfftn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_fmod_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_heaviside_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_index_reduce_amin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_index_reduce_mean_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_index_reduce_prod_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_inner_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_linalg_det_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_linalg_svdvals_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_linalg_vecdot_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_log_softmax_with_dtype_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_lu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_minimum_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_narrow_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_nonzero_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_ones_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_quantile_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_sigmoid_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_special_bessel_j1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_special_i1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_special_laguerre_polynomial_l_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_special_modified_bessel_i0_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_t_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_topk_cuda_float32, test/test_ops.py::TestCommonCUDA::test_meta_consistency_out_dtype_mismatch_tril_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices___radd___cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices___rmul___cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_acosh_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_add_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_aminmax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_as_strided_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_asin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_atleast_2d_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_bitwise_left_shift_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_bitwise_right_shift_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_bitwise_xor_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_bool_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_cat_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_cfloat_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_chalf_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_complex_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_contiguous_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_cosh_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_cross_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_deg2rad_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_diagflat_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_digamma_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_empty_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_expand_as_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_expand_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_expand_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_fft_hfft2_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_fft_hfftn_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_fft_ifftn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_fft_irfft_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_fill_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_flatten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_flip_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_flipud_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_frac_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_index_select_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_kron_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_linalg_diagonal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_linalg_eig_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_linalg_eigh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_linalg_ldl_factor_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_linalg_ldl_factor_ex_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_linalg_lstsq_grad_oriented_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_linalg_lu_factor_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_linspace_tensor_overload_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_log2_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_logical_and_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_lt_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_mH_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_mT_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_masked_logsumexp_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_masked_var_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_maximum_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_min_reduction_no_dim_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_movedim_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_narrow_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_native_layer_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_batch_norm_without_cudnn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_conv2d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_cross_entropy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_feature_alpha_dropout_without_train_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_hardsigmoid_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_interpolate_linear_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_interpolate_nearest-exact_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_interpolate_trilinear_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_linear_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_local_response_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_one_hot_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_pad_replicate_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_pad_replicate_negative_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nn_functional_softshrink_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_nonzero_static_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_norm_inf_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_normal_number_mean_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_pca_lowrank_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_permute_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_polygamma_polygamma_n_0_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_ravel_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_roll_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_scatter_add_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_scatter_reduce_sum_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_sigmoid_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_signal_windows_cosine_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_signal_windows_hann_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_signbit_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_sin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_slice_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_softmax_with_dtype_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_airy_ai_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_chebyshev_polynomial_v_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_i0e_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_log_ndtr_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_modified_bessel_i0_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_special_shifted_chebyshev_polynomial_t_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_split_list_args_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_square_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_squeeze_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_squeeze_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_true_divide_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_trunc_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_trunc_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_unbind_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_unflatten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_unfold_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_vdot_cuda_float32, test/test_ops.py::TestCommonCUDA::test_multiple_devices_view_cuda_int64, test/test_ops.py::TestCommonCUDA::test_multiple_devices_vsplit_cuda_float32, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values__unsafe_masked_index_put_accumulate_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_as_strided_copy_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_bfloat16_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_cdouble_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_cfloat_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_chalf_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_cosh_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_cummax_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_double_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_dstack_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_empty_like_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_erfc_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_exp_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_fft_ifft2_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_fft_irfft2_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_fft_rfft_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_fft_rfftn_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_i0_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_isposinf_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_log_softmax_with_dtype_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_long_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_ones_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_real_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_repeat_interleave_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_rsqrt_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_scatter_add_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_signbit_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_special_i0e_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_special_modified_bessel_i0_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_special_polygamma_special_polygamma_n_0_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_to_sparse_cuda_bool, test/test_ops.py::TestCommonCUDA::test_non_standard_bool_values_view_copy_cuda_bool, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_H_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_T_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples___rsub___cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_acosh_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_allclose_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_amax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_aminmax_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_as_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_atan2_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_atleast_1d_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_bitwise_not_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_bmm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_cdist_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_cfloat_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_combinations_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_contiguous_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_corrcoef_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_cos_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_diag_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_diag_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_diagonal_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_einsum_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_empty_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_empty_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_expand_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_exponential_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_fft_hfft_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_fft_hfftn_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_fft_ifft2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_fft_ifft_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_fft_irfftn_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_float_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_floor_divide_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_gather_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_half_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_hash_tensor_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_heaviside_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_hsplit_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_hstack_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_hypot_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_isfinite_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_isnan_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_isposinf_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_item_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_jiterator_unary_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_lcm_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_ldexp_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_lerp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_det_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_eigvals_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_inv_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_lstsq_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_matrix_rank_hermitian_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_solve_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_vander_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_vecdot_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_linalg_vector_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_log2_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_logcumsumexp_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_logical_or_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_logical_xor_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_lt_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_mT_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_masked_cumsum_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_masked_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_masked_prod_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_masked_select_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_masked_softmax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_masked_var_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_max_binary_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_max_reduction_no_dim_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_min_binary_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_mv_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_mvlgamma_mvlgamma_p_5_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nanmean_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nanquantile_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_celu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_channel_shuffle_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_conv3d_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_conv_transpose2d_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_cross_entropy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_gaussian_nll_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_hardshrink_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_interpolate_nearest-exact_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_layer_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_max_pool2d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_max_unpool1d_grad_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_pad_circular_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_relu6_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_relu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_nn_functional_rms_norm_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_normal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_reshape_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_resize__cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_round_decimals_3_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_rsqrt_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_signal_windows_blackman_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_signal_windows_general_cosine_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_signal_windows_hann_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_signbit_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_slice_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_softmax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_special_modified_bessel_i1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_special_modified_bessel_k1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_split_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_split_with_sizes_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_stack_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_std_mean_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_sum_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_svd_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_take_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_tan_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_tensor_split_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_to_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_trapz_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_tril_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_unbind_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_unfold_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_unsafe_chunk_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_var_unbiased_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_vdot_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_view_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_view_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_vsplit_cuda_float32, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_vstack_cuda_int64, test/test_ops.py::TestCommonCUDA::test_noncontiguous_samples_xlogy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_addbmm_cuda_float64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_broadcast_to_cuda_float64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_cat_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_numpy_ref_diff_cuda_int64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_jiterator_4inputs_with_extra_args_cuda_float64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_linalg_tensorinv_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_numpy_ref_nn_functional_conv_transpose3d_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_numpy_ref_permute_cuda_int64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_repeat_cuda_int64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_signal_windows_kaiser_cuda_float64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_squeeze_copy_cuda_int64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_squeeze_multiple_cuda_int64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_tensor_split_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_numpy_ref_triu_indices_cuda_int64, test/test_ops.py::TestCommonCUDA::test_numpy_ref_unbind_cuda_int64, test/test_ops.py::TestCommonCUDA::test_out___rdiv___cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs__conversions_byte_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs__conversions_cdouble_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs__conversions_char_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs__conversions_polar_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_addr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_amax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_amin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_as_strided_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_atanh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_bitwise_not_cuda_int64, test/test_ops.py::TestCommonCUDA::test_out__refs_clamp_max_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_conj_physical_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_copysign_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_diag_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_div_trunc_rounding_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_erfc_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_erfinv_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_fft_fftn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_fft_ifftshift_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_fft_rfft_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_isfinite_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_isnan_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_linspace_tensor_overload_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_log_softmax_with_dtype_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_mean_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_native_layer_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_nn_functional_elu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_nn_functional_group_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_nn_functional_layer_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_nn_functional_triplet_margin_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_randn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_t_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_tanh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__refs_triu_indices_cuda_int64, test/test_ops.py::TestCommonCUDA::test_out__refs_xlogy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out__softmax_backward_data_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_acos_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_addmm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_addr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_asinh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_atanh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_bitwise_left_shift_cuda_int64, test/test_ops.py::TestCommonCUDA::test_out_bitwise_or_cuda_int64, test/test_ops.py::TestCommonCUDA::test_out_cfloat_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_contiguous_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_cos_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_cosh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_cummin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_fill_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_hstack_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_inner_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_isin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_linalg_lstsq_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_linalg_lu_factor_ex_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_linalg_matrix_power_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_linalg_multi_dot_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_linalg_solve_ex_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_long_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_mT_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_masked_std_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_meshgrid_variadic_tensors_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_new_empty_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nextafter_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_conv_transpose1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_dropout3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_fractional_max_pool3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_interpolate_bilinear_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_l1_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_leaky_relu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_nn_functional_soft_margin_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_norm_nuc_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_normal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_permute_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_polygamma_polygamma_n_0_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_real_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_repeat_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_addmm_decomposed_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_addr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_angle_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_cat_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_conj_physical_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_diag_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_diag_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_div_no_rounding_mode_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_fft_fftn_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_fmod_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_frac_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_hstack_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_index_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_index_reduce_mean_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_index_select_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_lerp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_det_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_eigvals_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_eigvalsh_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_eigvalsh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_householder_product_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_slogdet_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_slogdet_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_linalg_svdvals_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_logaddexp2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_logspace_tensor_overload_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_min_reduction_no_dim_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_mode_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_mv_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_round_decimals_neg_3_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_softmax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_special_log_ndtr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_std_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_requires_grad_error_true_divide_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_resize_as__cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_scatter_reduce_amin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_signal_windows_exponential_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_signal_windows_gaussian_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_sinc_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_sort_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_special_bessel_y1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_special_modified_bessel_i0_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_special_polygamma_special_polygamma_n_0_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_special_xlog1py_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_special_zeta_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_take_along_dim_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_take_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_tensordot_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_torch_ops_aten__flash_attention_forward_cuda_float16, test/test_ops.py::TestCommonCUDA::test_out_unfold_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_out_view_as_real_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_out_warning_H_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs__conversions_short_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_add_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_addcmul_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_bitwise_not_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_bitwise_right_shift_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_bucketize_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_chunk_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_clamp_max_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_dstack_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_erfc_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_fft_irfftn_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_linspace_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_log_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_neg_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_nn_functional_group_norm_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_nn_functional_pixel_unshuffle_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_ravel_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_repeat_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_square_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__refs_view_as_complex_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__unsafe_masked_index_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning__upsample_bilinear2d_aa_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_acos_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_argwhere_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_atan2_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_conj_physical_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_contiguous_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_diagonal_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_div_trunc_rounding_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_exp_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_fft_hfft2_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_fft_ifft2_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_full_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_heaviside_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_imag_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_index_reduce_amax_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_index_reduce_amin_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_int_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_isfinite_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_jiterator_4inputs_with_extra_args_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_jiterator_binary_return_by_ref_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_lcm_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_linalg_cholesky_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_linalg_householder_product_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_linalg_ldl_factor_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_linalg_pinv_singular_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_linalg_svdvals_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_linalg_vander_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_log10_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_log_normal_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_masked_var_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_conv_transpose3d_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_gaussian_nll_loss_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_glu_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_hardshrink_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_hardswish_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_hinge_embedding_loss_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_interpolate_area_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_interpolate_bicubic_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_one_hot_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_pdist_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_prelu_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_relu_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nn_functional_smooth_l1_loss_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_nonzero_static_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_normal_in_place_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_ones_like_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_polar_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_reshape_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_searchsorted_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_sgn_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_sign_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_signal_windows_bartlett_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_special_spherical_bessel_j0_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_split_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_split_list_args_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_t_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_var_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_var_mean_cuda, test/test_ops.py::TestCommonCUDA::test_out_warning_view_as_real_cuda, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_atan_cuda_int32, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_atanh_cuda_bool, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_copysign_cuda_bool, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_copysign_cuda_int32, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_cosh_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_deg2rad_cuda_int64, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_deg2rad_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_erfinv_cuda_int8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_exp2_cuda_int8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_exp_cuda_int32, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_expm1_cuda_bool, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_float_power_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_ldexp_cuda_bool, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_log_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_mvlgamma_mvlgamma_p_3_cuda_int64, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_mvlgamma_mvlgamma_p_5_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_polygamma_polygamma_n_1_cuda_int32, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_polygamma_polygamma_n_3_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_rad2deg_cuda_int32, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_reciprocal_cuda_int64, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_rsqrt_cuda_int64, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_rsqrt_cuda_int8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_sinh_cuda_bool, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_chebyshev_polynomial_t_cuda_int8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_chebyshev_polynomial_v_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_chebyshev_polynomial_w_cuda_int64, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_chebyshev_polynomial_w_cuda_int8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_hermite_polynomial_he_cuda_int64, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_shifted_chebyshev_polynomial_t_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_shifted_chebyshev_polynomial_u_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_shifted_chebyshev_polynomial_v_cuda_bool, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_xlog1py_cuda_int16, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_special_zeta_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_tan_cuda_int32, test/test_ops.py::TestCommonCUDA::test_promotes_int_to_float_tanh_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_T_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_bool_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_cdouble_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_chalf_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_chalf_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_chalf_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_double_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_float_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_float_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_half_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_int_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_int_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_int_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_long_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_long_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_polar_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs__conversions_short_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_abs_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_acos_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_acosh_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_acosh_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_add_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_add_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_alias_copy_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_all_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_amin_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_any_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_any_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_any_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_as_strided_copy_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_as_strided_partial_views_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_asin_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_asin_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_asin_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_asinh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atan2_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atan_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atan_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atanh_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atanh_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atleast_1d_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atleast_1d_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atleast_2d_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atleast_3d_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_atleast_3d_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_bitwise_and_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_bitwise_and_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_bitwise_or_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_bitwise_right_shift_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_bitwise_xor_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_broadcast_tensors_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_broadcast_tensors_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_broadcast_tensors_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_broadcast_tensors_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_broadcast_to_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_broadcast_to_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_bucketize_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_cat_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_cat_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_clamp_max_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_clamp_min_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_clamp_min_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_clone_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_column_stack_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_conj_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_conj_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_copysign_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_cos_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_cos_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_count_nonzero_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_cumsum_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_diag_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_diagonal_copy_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_diagonal_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_diagonal_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_div_floor_rounding_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_div_no_rounding_mode_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_div_no_rounding_mode_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_div_trunc_rounding_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_dot_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_dsplit_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_empty_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_empty_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_empty_like_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_empty_like_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_eq_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_eq_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_eq_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_erfc_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_erfc_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_exp2_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_exp_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_exp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_expand_copy_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_expm1_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_expm1_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_exponential_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_eye_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_eye_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_eye_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_fft_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_fft_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_fft_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_fftn_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_hfftn_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_hfftn_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_ifft2_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_ihfft2_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_ihfft_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_ihfft_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_ihfftn_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_irfft2_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_irfft_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_irfftn_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_rfft2_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_rfft_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fft_rfftn_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_flatten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_flatten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_flip_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_flip_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_flip_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fliplr_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_flipud_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_flipud_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_float_power_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_float_power_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_floor_divide_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_floor_divide_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fmax_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_fmod_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_frexp_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_gcd_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_geometric_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_gt_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_gt_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_gt_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_heaviside_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_heaviside_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_hsplit_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_hstack_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_hstack_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_hstack_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_hstack_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_index_fill_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_index_select_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isfinite_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isinf_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isinf_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isnan_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isneginf_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isposinf_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_isreal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_item_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_le_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_le_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_lerp_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_lgamma_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_cross_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_matrix_norm_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_norm_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_svdvals_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_vecdot_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linalg_vector_norm_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linspace_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_linspace_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_log10_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_log1p_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_log_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_log_softmax_with_dtype_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_log_softmax_with_dtype_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logaddexp_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logical_or_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logical_or_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logspace_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logspace_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logspace_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_logspace_tensor_overload_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_lt_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_lt_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_maximum_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_minimum_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_minimum_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_minimum_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_mul_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_mul_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nan_to_num_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_narrow_copy_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_narrow_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_narrow_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_ne_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_new_empty_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_new_empty_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_new_empty_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_new_empty_strided_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_new_empty_strided_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_new_full_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_alpha_dropout_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_celu_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_dropout_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_elu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_group_norm_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_hardshrink_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_hardtanh_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_l1_loss_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_l1_loss_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_leaky_relu_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_log_softmax_with_dtype_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_log_softmax_with_dtype_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_margin_ranking_loss_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_pairwise_distance_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_pairwise_distance_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_pairwise_distance_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_pairwise_distance_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_pixel_shuffle_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_pixel_unshuffle_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_pixel_unshuffle_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_pixel_unshuffle_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_pixel_unshuffle_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_poisson_nll_loss_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_smooth_l1_loss_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_softmax_with_dtype_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_softmin_with_dtype_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_softplus_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_tanhshrink_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_threshold_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_threshold_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_triplet_margin_loss_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_nn_functional_triplet_margin_loss_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_normal_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_ones_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_permute_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_pow_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_prod_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_prod_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_prod_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_rad2deg_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_ravel_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_reciprocal_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_reciprocal_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_reciprocal_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_reshape_as_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_reshape_as_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_rot90_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_rsqrt_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_rsub_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_select_scatter_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sgn_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_signbit_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sin_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sin_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sinc_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sinc_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_bessel_j1_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_entr_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_entr_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_i0e_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_i0e_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_i1_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_log_softmax_with_dtype_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_log_softmax_with_dtype_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_logit_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_multigammaln_mvlgamma_p_1_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_multigammaln_mvlgamma_p_5_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_ndtr_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_ndtri_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_ndtri_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_softmax_with_dtype_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_spherical_bessel_j0_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_xlog1py_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_xlog1py_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_special_zeta_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_split_with_sizes_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sqrt_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_square_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_square_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_squeeze_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_squeeze_multiple_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_stack_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_std_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sum_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sum_to_size_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_sum_to_size_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_t_copy_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_take_along_dim_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_take_along_dim_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_tan_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_tan_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_trace_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_tril_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_triu_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_true_divide_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_true_divide_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_unflatten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_unflatten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_unfold_copy_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_unfold_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_unfold_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_var_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_var_mean_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_var_mean_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_view_as_complex_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_view_as_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_view_copy_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_view_copy_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_view_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_vsplit_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_vstack_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_vstack_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_where_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_where_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref__refs_where_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_as_strided_scatter_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_clamp_max_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_eye_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_gcd_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_hsplit_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_hypot_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_log_normal_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_mul_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_nn_functional_group_norm_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_rot90_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_sum_to_size_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_errors__refs_view_copy_cuda, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_T_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_T_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_T_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_T_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_bfloat16_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_byte_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_cdouble_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_cfloat_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_chalf_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_double_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_float_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_half_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_half_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_half_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_long_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_long_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_long_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_short_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs__conversions_short_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_abs_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_abs_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_acos_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_acos_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_acosh_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_acosh_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_acosh_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_addcmul_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_addr_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_addr_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_alias_copy_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_all_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_all_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_allclose_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_allclose_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_amin_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_amin_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_any_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_arange_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_arange_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_as_strided_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_as_strided_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_as_strided_partial_views_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_as_strided_partial_views_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_as_strided_scatter_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_as_strided_scatter_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_asin_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_asin_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_asinh_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atanh_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atanh_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atleast_1d_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atleast_2d_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atleast_2d_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_atleast_3d_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_bitwise_left_shift_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_bitwise_not_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_bitwise_or_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_bitwise_right_shift_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_block_diag_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_broadcast_tensors_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_broadcast_to_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_bucketize_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cat_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_ceil_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_chunk_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_clamp_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_clamp_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_clamp_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_clamp_max_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_clamp_max_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_clamp_min_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_clone_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_clone_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_clone_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_clone_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_conj_physical_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_constant_pad_nd_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_constant_pad_nd_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_copysign_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cosh_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cosh_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cosh_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_cumprod_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_diag_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_diagonal_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_diagonal_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_diagonal_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_diagonal_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_diagonal_scatter_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_div_trunc_rounding_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_empty_like_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_empty_like_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_empty_strided_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_eq_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_equal_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_erf_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_erf_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_erfinv_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_exp2_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_expand_copy_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_expand_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_expand_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_expm1_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_expm1_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_eye_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_eye_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_eye_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_fft2_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_fft_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_fftn_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_fftshift_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_hfft2_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_hfft_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_hfft_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_hfftn_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ifftn_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ifftshift_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ihfft_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ihfftn_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_ihfftn_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_irfft2_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_irfft2_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_irfft_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_irfft_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_rfft2_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_rfft_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fft_rfftn_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fill_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_flip_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fliplr_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fliplr_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_flipud_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_float_power_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_floor_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_floor_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fmax_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fmod_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_fmod_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_ge_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_hsplit_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_hypot_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_igamma_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_index_fill_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_isfinite_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_isnan_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_isneginf_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_isreal_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_isreal_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_isreal_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_istft_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_item_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_item_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_item_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_lcm_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_lerp_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_lerp_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linalg_cross_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linalg_matrix_norm_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linalg_matrix_norm_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linalg_norm_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linalg_vecdot_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linspace_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_linspace_tensor_overload_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log10_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log1p_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log1p_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log2_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log2_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log2_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_log_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logaddexp_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_and_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_and_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_not_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_not_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_or_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_or_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_or_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logical_xor_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logspace_tensor_overload_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logspace_tensor_overload_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_logsumexp_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_lt_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_maximum_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_maximum_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_meshgrid_list_of_tensors_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_meshgrid_variadic_tensors_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_minimum_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_minimum_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_movedim_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_movedim_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nan_to_num_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_narrow_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_native_layer_norm_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_ne_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_ne_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_neg_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_neg_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_new_empty_strided_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_new_full_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_new_full_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_new_ones_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_new_zeros_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_channel_shuffle_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_hardtanh_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_l1_loss_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_log_softmax_with_dtype_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_log_softmax_with_dtype_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_log_softmax_with_dtype_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_margin_ranking_loss_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_margin_ranking_loss_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_pixel_shuffle_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_pixel_shuffle_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_poisson_nll_loss_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_relu6_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_relu6_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_selu_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_softmax_with_dtype_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_softmin_with_dtype_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_softmin_with_dtype_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_tanhshrink_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_triplet_margin_loss_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_triplet_margin_loss_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_triplet_margin_loss_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_nn_functional_triplet_margin_loss_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_normal_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_permute_copy_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_permute_copy_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_permute_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_positive_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_prod_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_prod_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_randn_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_renorm_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_repeat_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_repeat_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_repeat_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_repeat_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_reshape_as_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_roll_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_roll_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_rot90_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_rot90_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_rsqrt_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_rsqrt_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_rsqrt_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_rsub_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sgn_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sigmoid_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sign_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_signbit_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_signbit_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sin_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sinc_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sinh_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_softmax_with_dtype_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_bessel_j1_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_erfcx_executor_aten_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_i0e_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_i1_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_i1e_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_log_ndtr_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_log_softmax_with_dtype_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_multigammaln_mvlgamma_p_3_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_multigammaln_mvlgamma_p_3_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_softmax_with_dtype_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_softmax_with_dtype_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_spherical_bessel_j0_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_special_zeta_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sqrt_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_square_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_square_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_square_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_square_executor_aten_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_squeeze_copy_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_squeeze_copy_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_squeeze_copy_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_squeeze_multiple_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_squeeze_multiple_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_std_mean_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sub_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sum_to_size_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sum_to_size_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_sum_to_size_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_t_copy_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_tan_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_tan_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_tanh_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_tanh_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_tensor_split_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_tensor_split_executor_aten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_to_executor_aten_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_trace_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_trace_executor_aten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_transpose_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_tril_executor_aten_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_triu_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_triu_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_true_divide_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_true_divide_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_true_divide_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unbind_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unbind_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unbind_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_unsqueeze_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_var_mean_executor_aten_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_view_as_complex_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_view_as_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_view_copy_executor_aten_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_view_executor_aten_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_vsplit_executor_aten_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_vstack_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_where_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_xlogy_executor_aten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_xlogy_executor_aten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_executor__refs_xlogy_executor_aten_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_T_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_bfloat16_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_bfloat16_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_bfloat16_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_bool_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_cdouble_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_cfloat_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_chalf_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_chalf_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_char_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_double_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_float_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_half_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_half_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_int_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_int_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_long_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_long_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs__conversions_polar_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_abs_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_acosh_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_acosh_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_acosh_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_addr_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_addr_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_addr_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_alias_copy_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_all_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_all_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_all_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_allclose_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_amax_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_any_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_arange_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_arange_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_as_strided_copy_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_atan_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_atleast_1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_atleast_1d_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_atleast_3d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_atleast_3d_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_bitwise_left_shift_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_bitwise_left_shift_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_bitwise_not_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_bitwise_or_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_block_diag_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_broadcast_to_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_bucketize_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_bucketize_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cat_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_ceil_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_ceil_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_clamp_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_clone_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_constant_pad_nd_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_contiguous_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_copysign_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cos_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_count_nonzero_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_count_nonzero_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_count_nonzero_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cumsum_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cumsum_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cumsum_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cumsum_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_cumsum_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_deg2rad_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_deg2rad_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_diag_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_diag_embed_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_diag_embed_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_diagonal_copy_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_diagonal_copy_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_digamma_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_digamma_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_div_no_rounding_mode_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_div_no_rounding_mode_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_div_no_rounding_mode_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_div_trunc_rounding_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_div_trunc_rounding_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_dstack_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_empty_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_empty_strided_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_equal_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_equal_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_erf_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_erfinv_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_exp_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_exp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_expand_as_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_expand_copy_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_expand_copy_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_expand_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_expand_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_eye_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_fft2_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_fftn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_fftshift_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_hfft2_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_hfft_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_hfftn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ifft2_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ifftn_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ifftshift_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ifftshift_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ihfft2_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ihfft_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ihfft_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_ihfftn_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_irfft2_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_irfft2_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_irfft_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_irfftn_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_rfft_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_rfft_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_rfftn_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fft_rfftn_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fill_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fliplr_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_flipud_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_flipud_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_float_power_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_float_power_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_floor_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_floor_divide_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fmax_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_fmin_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_frac_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_gcd_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_gt_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_gt_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_heaviside_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_hsplit_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_hstack_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_i0_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_i0_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_igamma_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_imag_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_index_add_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_index_copy_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_index_select_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_index_select_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_isclose_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_isnan_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_lcm_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_le_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_lerp_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_lerp_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_lerp_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linalg_cross_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linalg_cross_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linalg_cross_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linalg_diagonal_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linspace_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linspace_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linspace_tensor_overload_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_linspace_tensor_overload_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log10_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log10_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log1p_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log1p_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log2_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log2_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log_normal_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_log_softmax_with_dtype_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logaddexp_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logical_not_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logical_not_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logical_not_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logical_not_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logical_xor_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logspace_tensor_overload_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logspace_tensor_overload_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logsumexp_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_logsumexp_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_masked_fill_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_masked_fill_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_masked_fill_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_maximum_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_maximum_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_maximum_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_mean_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_mean_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_meshgrid_list_of_tensors_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_minimum_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_movedim_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_mul_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_mul_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nan_to_num_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_native_layer_norm_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_ne_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_new_empty_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_new_empty_strided_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_new_empty_strided_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_new_full_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_new_full_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_new_full_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_new_ones_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_new_zeros_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nextafter_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_elu_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_gelu_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_l1_loss_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_layer_norm_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_log_softmax_with_dtype_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_margin_ranking_loss_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_margin_ranking_loss_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_margin_ranking_loss_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_pairwise_distance_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_pixel_shuffle_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_pixel_unshuffle_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_pixel_unshuffle_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_poisson_nll_loss_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_relu6_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_relu_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_smooth_l1_loss_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_tanhshrink_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_tanhshrink_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_threshold_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_threshold_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_threshold_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_triplet_margin_loss_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_triplet_margin_loss_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_nn_functional_triplet_margin_loss_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_norm_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_ones_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_ones_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_permute_copy_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_positive_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_prod_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_prod_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_rad2deg_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_rad2deg_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_randn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_real_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_reciprocal_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_reciprocal_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_remainder_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_remainder_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_renorm_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_repeat_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_repeat_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_reshape_as_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_reshape_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_rsqrt_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_rsub_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_select_scatter_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_select_scatter_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sgn_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sigmoid_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sigmoid_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sign_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_signbit_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sin_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sinc_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sinh_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sinh_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_softmax_with_dtype_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_softmax_with_dtype_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_softmax_with_dtype_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_bessel_j0_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_bessel_j1_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_entr_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_erfcx_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_i0e_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_i0e_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_i1_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_i1e_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_i1e_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_log_ndtr_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_log_softmax_with_dtype_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_log_softmax_with_dtype_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_logit_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_multigammaln_mvlgamma_p_3_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_multigammaln_mvlgamma_p_5_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_ndtri_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_special_softmax_with_dtype_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_split_with_sizes_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sqrt_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_square_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_square_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_squeeze_multiple_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_squeeze_multiple_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_stack_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_stack_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_stack_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_std_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_stft_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sub_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sub_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sub_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sub_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sub_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sum_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sum_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sum_to_size_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_sum_to_size_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_t_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_t_copy_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_take_along_dim_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tan_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tanh_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_to_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_to_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_trace_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_trace_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_transpose_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_transpose_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_tril_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_triu_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_triu_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_true_divide_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_true_divide_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unbind_copy_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unbind_copy_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unbind_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unflatten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unflatten_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unflatten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unfold_copy_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unfold_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unfold_copy_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unfold_copy_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unfold_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unfold_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_unsqueeze_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_var_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_var_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_view_as_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_view_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_vsplit_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_vstack_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_vstack_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_vstack_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_where_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_xlogy_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_xlogy_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_meta__refs_xlogy_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_T_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_T_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_bool_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_bool_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_byte_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_byte_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_cdouble_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_cdouble_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_cfloat_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_chalf_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_chalf_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_chalf_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_chalf_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_char_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_char_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_char_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_float_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_int_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_int_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_long_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_short_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs__conversions_short_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_abs_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_acos_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_acosh_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_addcdiv_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_addcdiv_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_addcdiv_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_addr_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_alias_copy_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_alias_copy_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_allclose_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_amin_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_amin_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_any_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_copy_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_partial_views_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_partial_views_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_scatter_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_scatter_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_as_strided_scatter_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_asin_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_asin_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_asinh_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_asinh_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atanh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atanh_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atleast_1d_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atleast_1d_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atleast_1d_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atleast_1d_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atleast_2d_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atleast_2d_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atleast_3d_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_atleast_3d_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_bitwise_left_shift_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_block_diag_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_broadcast_tensors_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_broadcast_tensors_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_broadcast_tensors_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_broadcast_to_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_broadcast_to_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_bucketize_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cat_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cat_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cat_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_chunk_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_chunk_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_clamp_max_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_clamp_min_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_column_stack_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_conj_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_conj_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_constant_pad_nd_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_copysign_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cos_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cosh_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_count_nonzero_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_count_nonzero_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cumprod_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cumprod_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_cumsum_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_deg2rad_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_diag_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_diag_embed_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_diag_embed_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_diag_embed_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_digamma_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_digamma_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_div_floor_rounding_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_div_trunc_rounding_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_div_trunc_rounding_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_dot_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_dsplit_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_dstack_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_dstack_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_empty_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_empty_like_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_empty_strided_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_eq_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_equal_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_erf_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_exp2_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_exp2_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_exp_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_exp_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_expand_as_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_expand_copy_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_expand_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_expand_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_expm1_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_fft2_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_fft2_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_fft2_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_fft_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_fft_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_fftn_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_fftshift_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_fftshift_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_hfft_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_hfft_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_ifft2_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_ifftn_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_ifftshift_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_ifftshift_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_ifftshift_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_ifftshift_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_ihfft2_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_ihfftn_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_ihfftn_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_irfft_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_irfftn_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_rfft2_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_rfft2_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_rfft2_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_rfft_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_rfft_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fft_rfftn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fill_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fill_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fill_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_flatten_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_flip_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fliplr_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fliplr_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_float_power_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_fmax_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_gcd_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_ge_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_gt_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_gt_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_hstack_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_hstack_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_index_copy_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_index_copy_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_index_copy_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_index_fill_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_index_select_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isclose_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isclose_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isclose_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isinf_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isneginf_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isneginf_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isneginf_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isposinf_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isposinf_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_isreal_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_item_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_lcm_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_lerp_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_lgamma_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_linalg_cross_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_linalg_vector_norm_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_linalg_vector_norm_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log1p_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log1p_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log2_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log2_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log2_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_log_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logaddexp_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logical_and_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logical_not_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logical_or_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logical_or_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logical_xor_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_logspace_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_masked_fill_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_maximum_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_mean_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_meshgrid_variadic_tensors_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_minimum_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_mul_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_mul_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_mul_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nan_to_num_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nan_to_num_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_narrow_copy_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_neg_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_empty_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_full_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_full_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_full_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_ones_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_ones_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_new_zeros_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_alpha_dropout_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_celu_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_channel_shuffle_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_gelu_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_hardtanh_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_layer_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_leaky_relu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_log_softmax_with_dtype_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_mse_loss_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_pairwise_distance_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_pixel_shuffle_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_pixel_unshuffle_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_smooth_l1_loss_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_softmax_with_dtype_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_softmin_with_dtype_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_nn_functional_softplus_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_permute_copy_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_permute_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_positive_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_prod_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_rad2deg_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_rad2deg_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_real_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_reciprocal_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_reciprocal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_reciprocal_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_repeat_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_repeat_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_reshape_as_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_reshape_as_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_reshape_as_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_reshape_as_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_roll_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_roll_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_round_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_rsqrt_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_rsub_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sgn_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sgn_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_signbit_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sin_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sin_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sinc_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sinh_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sinh_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_bessel_j1_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_entr_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_entr_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_erfcx_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_i0e_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_i0e_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_i1_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_log_softmax_with_dtype_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_log_softmax_with_dtype_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_multigammaln_mvlgamma_p_1_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_multigammaln_mvlgamma_p_1_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_multigammaln_mvlgamma_p_3_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_multigammaln_mvlgamma_p_3_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_ndtr_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_softmax_with_dtype_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_softmax_with_dtype_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_special_spherical_bessel_j0_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_square_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_squeeze_copy_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_squeeze_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_squeeze_multiple_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_squeeze_multiple_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_squeeze_multiple_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_squeeze_multiple_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_std_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sub_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sum_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_sum_to_size_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_t_copy_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_take_along_dim_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_tan_cuda_bool, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_tan_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_tan_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_tanh_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_tensor_split_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_tensor_split_cuda_uint8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_to_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_transpose_copy_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_transpose_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_tril_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_tril_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_triu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_triu_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_true_divide_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_trunc_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unbind_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unflatten_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unfold_copy_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unsqueeze_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_unsqueeze_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_var_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_var_mean_cuda_float16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_view_as_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_view_cuda_bfloat16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_view_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_view_cuda_int8, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_vsplit_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_vsplit_cuda_int32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_vstack_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_vstack_cuda_float64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_vstack_cuda_int64, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_where_cuda_complex128, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_where_cuda_complex32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_where_cuda_int16, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_xlogy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_python_ref_torch_fallback__refs_xlogy_cuda_int16, test/test_ops.py::TestCommonCUDA::test_reduction_ops_reduce_sum_cuda, test/test_ops.py::TestCommonCUDA::test_reduction_ops_reduce_var_mean_cuda, test/test_ops.py::TestCommonCUDA::test_reduction_ops_reduce_var_mean_unbiased_cuda, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager___rpow___cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager___rsub___cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager__unsafe_masked_index_put_accumulate_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_abs_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_acosh_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_allclose_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_arange_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_argsort_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_as_strided_copy_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_atan_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_atanh_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_baddbmm_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_baddbmm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_broadcast_tensors_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_chalf_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_conj_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_contiguous_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_cummax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_deg2rad_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_diagonal_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_double_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_expand_as_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_fft_hfftn_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_fft_ifftn_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_fft_rfft2_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_fliplr_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_flipud_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_fmod_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_gradient_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_hsplit_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_index_add_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_index_reduce_amax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_jiterator_2inputs_2outputs_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_le_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_lerp_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_lgamma_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_diagonal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_eig_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_eigvals_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_ldl_solve_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_slogdet_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_solve_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_linalg_tensorsolve_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_log10_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_log1p_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_log_softmax_with_dtype_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_logical_or_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_long_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_mH_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_masked_mean_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_max_pool2d_with_indices_backward_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_max_reduction_no_dim_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_meshgrid_variadic_tensors_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_min_reduction_with_dim_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_minimum_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_multinomial_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nanquantile_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_narrow_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_narrow_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_native_layer_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_ne_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_channel_shuffle_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_channel_shuffle_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_glu_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_hardsigmoid_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_hardtanh_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_interpolate_area_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_local_response_norm_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_normalize_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_pad_replicate_negative_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_poisson_nll_loss_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_softmin_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_softplus_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_nn_functional_upsample_nearest_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_normal_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_ones_like_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_permute_copy_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_permute_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_polygamma_polygamma_n_3_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_rad2deg_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_repeat_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_scatter_add_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_scatter_reduce_amax_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_signal_windows_general_cosine_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_signal_windows_general_hamming_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_signal_windows_hamming_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_special_chebyshev_polynomial_v_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_special_ndtri_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_std_mean_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_trunc_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_unbind_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_unflatten_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_unfold_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_uniform_cuda_complex64, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_vsplit_cuda_float32, test/test_ops.py::TestCommonCUDA::test_variant_consistency_eager_zero__cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward___radd___cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward__batch_norm_with_update_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward__unsafe_masked_index_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward__unsafe_masked_index_put_accumulate_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_addcmul_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_addmm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_block_diag_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_constant_pad_nd_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_contiguous_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_diagonal_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_diff_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_erf_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_erfc_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_fft_rfft2_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_float_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_floor_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_fmod_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_frac_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_half_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_hypot_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_kthvalue_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_linalg_eigvalsh_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_linalg_householder_product_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_linalg_pinv_hermitian_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_linalg_solve_ex_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_linalg_tensorsolve_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_linalg_vecdot_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_msort_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_mvlgamma_mvlgamma_p_3_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nansum_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_adaptive_avg_pool1d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_batch_norm_without_cudnn_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_elu_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_interpolate_linear_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_interpolate_nearest_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_local_response_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_margin_ranking_loss_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_max_unpool2d_grad_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_normalize_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_rrelu_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_nn_functional_silu_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_pinverse_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_polygamma_polygamma_n_0_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_put_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_reciprocal_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_sparse_mm_reduce_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_std_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_stft_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_sub_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_take_along_dim_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_tensor_split_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_triu_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_unsafe_split_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_backward_xlogy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_addcmul_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_addmm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_argmax_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_bfloat16_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_bool_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_bucketize_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_conj_physical_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_copysign_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_diag_embed_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_diagonal_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_div_floor_rounding_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_div_trunc_rounding_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_empty_permuted_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_expand_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_eye_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_fft_fftshift_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_fmod_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_full_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_index_reduce_mean_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_isposinf_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_linalg_lu_solve_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_linalg_matrix_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_linalg_matrix_power_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_logical_not_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_lu_solve_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_masked_amin_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_masked_prod_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_meshgrid_variadic_tensors_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nan_to_num_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_neg_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_adaptive_max_pool2d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_cosine_embedding_loss_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_fractional_max_pool3d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_hardshrink_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_kl_div_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_layer_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_logsigmoid_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_max_pool2d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_mish_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_multilabel_margin_loss_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_pad_replicate_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_pixel_shuffle_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_rms_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_tanhshrink_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_nn_functional_unfold_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_polygamma_polygamma_n_0_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_rad2deg_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_renorm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_scatter_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_scatter_reduce_amax_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_short_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_special_chebyshev_polynomial_w_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_split_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_split_with_sizes_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_squeeze_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_std_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_std_mean_unbiased_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_sum_to_size_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_unfold_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_cow_input_unsqueeze_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad___radd___cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad___rmul___cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad__segment_reduce_offsets_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad__unsafe_masked_index_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_acosh_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_addr_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_alias_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_argmax_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_as_strided_scatter_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_cfloat_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_clone_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_conj_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_contiguous_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_diag_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_digamma_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_dist_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_div_no_rounding_mode_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_fft_ifft_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_fft_irfftn_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_histc_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_index_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_index_reduce_mean_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_isfinite_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_isin_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_le_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_lerp_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_linalg_matrix_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_linalg_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_linalg_solve_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_logaddexp_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_lu_solve_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_mT_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_masked_argmin_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_masked_prod_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_meshgrid_list_of_tensors_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_meshgrid_variadic_tensors_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_min_binary_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nanmedian_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_adaptive_avg_pool2d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_avg_pool3d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_celu_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_dropout_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_hardswish_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_logsigmoid_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_max_pool3d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_max_unpool2d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_mish_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_nn_functional_softshrink_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_polygamma_polygamma_n_1_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_prod_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_real_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_scatter_reduce_amax_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_select_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_signal_windows_hann_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_signbit_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_sinc_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_slice_scatter_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_sqrt_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_unique_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_forward_ad_where_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_addcdiv_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_bernoulli_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_bmm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_clamp_max_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_erfc_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_erfinv_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_fft_ihfft_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_fft_irfft2_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_fft_rfftn_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_frac_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_frexp_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_gradient_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_index_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_isfinite_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_item_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_kron_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_linalg_cholesky_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_linalg_lu_solve_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_linalg_vecdot_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_lu_solve_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_masked_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_masked_normalize_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_masked_scatter_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_masked_softmin_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_matmul_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_mode_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_new_empty_strided_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_new_ones_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_adaptive_avg_pool1d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_dropout2d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_dropout_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_gelu_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_interpolate_area_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_max_pool1d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_pad_circular_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_nn_functional_upsample_nearest_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_renorm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_resize__cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_scatter_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_scatter_reduce_prod_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_signal_windows_nuttall_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_softmax_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_special_shifted_chebyshev_polynomial_v_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_special_xlog1py_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_square_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_t_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_tan_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_var_mean_unbiased_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_operator_zeros_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_T_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay___rsub___cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_abs_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_addcdiv_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_atleast_1d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_cat_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_cdouble_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_copysign_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_diag_embed_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_diagonal_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_dstack_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_fft_ihfftn_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_fmax_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_fmin_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_igamma_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_index_fill_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_index_reduce_amin_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_jiterator_4inputs_with_extra_args_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_lgamma_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_linalg_cholesky_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_linalg_ldl_factor_ex_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_linalg_pinv_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_linalg_slogdet_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_linalg_svdvals_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_log_normal_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_logaddexp_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_logspace_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_masked_select_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_masked_var_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_matrix_exp_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_multinomial_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nanquantile_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_adaptive_avg_pool1d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_celu_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_conv2d_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_glu_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_max_unpool2d_grad_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_pad_replicate_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_threshold_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_nn_functional_unfold_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_norm_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_ones_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_scatter_reduce_mean_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_select_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_select_scatter_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_signal_windows_bartlett_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_signal_windows_general_cosine_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_signal_windows_hamming_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_sort_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_special_bessel_j0_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_special_modified_bessel_i1_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_split_list_args_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_squeeze_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_t_copy_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_uniform_cuda_float32, test/test_ops.py::TestCompositeComplianceCUDA::test_view_replay_where_cuda_float32, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_T_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs__conversions_float_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs__conversions_int_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs__conversions_long_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_alias_copy_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_as_strided_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_asinh_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_cat_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_exp_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_fft_fft2_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_fft_irfftn_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_imag_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_index_select_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_isnan_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_linalg_diagonal_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_linalg_svdvals_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_linalg_vecdot_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_log_softmax_with_dtype_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_logaddexp_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_mean_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_meshgrid_list_of_tensors_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_meshgrid_variadic_tensors_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_mul_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_neg_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_permute_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_roll_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_sum_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_unsqueeze_copy_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view__refs_var_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_acos_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_cartesian_prod_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_char_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_cholesky_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_cholesky_inverse_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_diagonal_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_dstack_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_expand_as_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_fft_fft2_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_fft_irfft2_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_float_power_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_hsplit_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_imag_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_index_fill_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_index_put_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_istft_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_jiterator_4inputs_with_extra_args_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_linalg_eig_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_linalg_inv_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_linalg_lu_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_linalg_multi_dot_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_logical_xor_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_lu_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_neg_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_nn_functional_pad_replicate_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_nn_functional_pad_replicate_negative_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_nn_functional_pixel_unshuffle_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_nn_functional_softmin_with_dtype_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_norm_nuc_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_outer_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_positive_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_renorm_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_reshape_as_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_sgn_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_squeeze_multiple_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_std_unbiased_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_transpose_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_where_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_conj_view_zeros_cuda_complex64, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_add_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_conj_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_exp2_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_expand_as_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_fft_fft_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_fft_hfftn_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_fill_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_flatten_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_hsplit_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_index_select_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_logical_not_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_logical_xor_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_ne_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_new_full_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_nn_functional_l1_loss_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_t_copy_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_unbind_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_unfold_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_view_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view__refs_vstack_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_as_strided_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_cdouble_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_char_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_diagonal_copy_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_diagonal_scatter_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_dot_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_fft_ifft2_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_flatten_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_imag_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_index_put_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_inner_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_linalg_ldl_solve_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_linalg_multi_dot_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_linalg_slogdet_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_linalg_solve_triangular_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_linalg_vander_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_linspace_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_log2_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_logaddexp_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_logsumexp_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_masked_logsumexp_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_masked_prod_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_masked_var_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_meshgrid_list_of_tensors_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_new_empty_strided_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_nn_functional_conv3d_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_nn_functional_linear_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_nn_functional_pairwise_distance_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_conj_view_unsqueeze_copy_cuda_complex128, test/test_ops.py::TestMathBitsCUDA::test_neg_view___radd___cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs__conversions_double_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs__conversions_polar_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_addcmul_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_any_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_chunk_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_constant_pad_nd_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_deg2rad_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_erfc_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_expm1_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_fft_fft_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_fft_rfft2_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_flipud_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_frac_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_index_fill_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_isfinite_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_isneginf_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_linalg_svd_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_linspace_tensor_overload_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_nn_functional_pixel_unshuffle_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_nn_functional_prelu_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_nn_functional_triplet_margin_loss_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_pow_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_signbit_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_special_log_softmax_with_dtype_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_special_multigammaln_mvlgamma_p_5_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_squeeze_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_transpose_copy_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_unsqueeze_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_vdot_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__refs_where_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view__upsample_bilinear2d_aa_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_acos_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_amax_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_argmax_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_argmin_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_atleast_2d_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_baddbmm_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_ceil_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_chalf_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_clamp_min_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_clone_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_cos_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_empty_permuted_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_erf_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_fft_hfft2_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_fft_irfft_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_gather_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_ge_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_index_put_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_isposinf_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_ldexp_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_linalg_householder_product_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_linalg_ldl_factor_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_linalg_lu_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_linalg_lu_solve_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_linalg_multi_dot_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_linalg_pinv_hermitian_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_log10_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_masked_amin_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_masked_log_softmax_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_masked_normalize_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_mean_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_min_reduction_no_dim_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_min_reduction_with_dim_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_adaptive_avg_pool3d_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_avg_pool2d_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_dropout3d_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_feature_alpha_dropout_with_train_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_fractional_max_pool2d_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_gelu_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_max_pool3d_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_multi_head_attention_forward_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_pad_constant_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_pixel_shuffle_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_rrelu_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_selu_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_nn_functional_tanhshrink_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_permute_copy_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_scatter_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_signal_windows_blackman_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_special_chebyshev_polynomial_u_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_special_modified_bessel_i1_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_split_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_split_list_args_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_split_with_sizes_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_std_mean_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_sum_to_size_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_tan_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_uniform_cuda_float64, test/test_ops.py::TestMathBitsCUDA::test_neg_view_unique_cuda_float64, test/test_ops.py::TestFakeTensorCUDA::test_fake___getitem___cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake___rmatmul___cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake___rmul___cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake___rpow___cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake__unsafe_masked_index_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake__unsafe_masked_index_put_accumulate_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_addmv_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_any_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast___rmul___cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast___rxor___cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast__upsample_bilinear2d_aa_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_acosh_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_argmin_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_asin_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_bitwise_and_cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_bitwise_left_shift_cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_chalf_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_cumsum_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_dstack_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_empty_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_fft_ifft2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_fmax_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_frexp_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_full_like_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_gather_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_index_reduce_amin_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_index_select_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_isnan_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_item_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_ldexp_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_linalg_ldl_factor_ex_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_linalg_ldl_solve_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_linalg_lu_factor_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_linalg_matrix_power_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_linalg_vecdot_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_linalg_vector_norm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_log10_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_log2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_log_normal_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_logical_or_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_masked_amax_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_mode_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nanmedian_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_ctc_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_embedding_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_hardswish_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_interpolate_nearest-exact_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_smooth_l1_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_softmin_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_nn_functional_tanhshrink_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_permute_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_polygamma_polygamma_n_3_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_rad2deg_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_repeat_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_repeat_interleave_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_reshape_as_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_signal_windows_kaiser_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_slice_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_softmax_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_special_modified_bessel_i0_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_special_modified_bessel_k0_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_topk_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_tril_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_triu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_unbind_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_zero__cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_autocast_zeros_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_bfloat16_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_bitwise_left_shift_cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_fake_bmm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_bool_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_broadcast_tensors_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_cfloat_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_cholesky_solve_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp___rmatmul___cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_block_diag_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_bmm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_broadcast_tensors_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_cat_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_column_stack_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_conj_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_cosh_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_cummax_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_cummin_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_dstack_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_expand_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_fft_fft_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_fft_hfftn_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_fft_irfftn_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_floor_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_frac_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_hypot_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_i0_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_lerp_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_linalg_matrix_norm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_linalg_pinv_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_linalg_pinv_hermitian_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_linalg_svdvals_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_linalg_vecdot_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_log1p_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_log_softmax_with_dtype_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_masked_normalize_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_masked_std_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_maximum_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_mvlgamma_mvlgamma_p_3_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_avg_pool2d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_batch_norm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_cosine_embedding_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_hardswish_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_mish_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_relu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_triplet_margin_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_nn_functional_upsample_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_ormqr_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_permute_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_qr_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_quantile_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_real_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_remainder_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_round_decimals_3_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_rsub_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_scatter_add_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_softmax_with_dtype_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_std_unbiased_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_take_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_trace_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_true_divide_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_unfold_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_var_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_view_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_view_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_amp_where_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp__batch_norm_with_update_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp__unsafe_masked_index_put_accumulate_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_as_strided_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_atleast_1d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_cat_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_complex_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_conj_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_cross_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_cummax_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_cummin_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_cumulative_trapezoid_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_diagflat_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_div_trunc_rounding_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_dstack_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_exp_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_expm1_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_fft_fft_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_fft_ihfft_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_log_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_logdet_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_lu_unpack_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_masked_scatter_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_max_binary_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_min_reduction_with_dim_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_msort_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_mvlgamma_mvlgamma_p_3_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_mvlgamma_mvlgamma_p_5_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_glu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_interpolate_bicubic_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_interpolate_nearest_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_max_pool2d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_multilabel_soft_margin_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_pixel_unshuffle_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_rrelu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_softmin_with_dtype_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_softplus_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_triplet_margin_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_nn_functional_unfold_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_norm_nuc_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_permute_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_prod_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_round_decimals_0_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_round_decimals_neg_3_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_scatter_add_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_sigmoid_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_softmax_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_softmax_with_dtype_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_special_entr_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_special_ndtr_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_split_with_sizes_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_svd_lowrank_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_take_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_trapz_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_unsafe_chunk_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_view_as_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_crossref_backward_no_amp_xlogy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_diagonal_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_exp2_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_float_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_frexp_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_gcd_cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_fake_histc_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_inner_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_int_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_linalg_lu_factor_ex_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_linalg_matrix_norm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_linalg_matrix_rank_hermitian_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_log_softmax_with_dtype_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_logical_or_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_masked_softmin_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_max_reduction_no_dim_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_max_reduction_with_dim_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_native_batch_norm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_neg_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_avg_pool3d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_fractional_max_pool2d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_glu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_grid_sample_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_interpolate_bicubic_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_interpolate_nearest-exact_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_l1_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_prelu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_relu_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_triplet_margin_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_nn_functional_triplet_margin_with_distance_loss_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_qr_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_ravel_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_remainder_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_reshape_as_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_signbit_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_slice_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_softmax_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_special_bessel_j0_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_special_chebyshev_polynomial_u_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_special_log_ndtr_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_special_scaled_modified_bessel_k0_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_sum_to_size_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_t_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_tril_indices_cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_fake_unsafe_split_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_vdot_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_vstack_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_fake_zeros_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops___rmatmul___cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops___ror___cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_allclose_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_as_strided_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_as_strided_scatter_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_bitwise_right_shift_cuda_int64, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_broadcast_shapes_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_cdist_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_cfloat_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_cholesky_solve_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_chunk_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_column_stack_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_conj_physical_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_count_nonzero_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_expand_as_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_fft_rfftn_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_flip_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_hypot_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_index_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_jiterator_unary_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_linalg_eigh_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_linalg_norm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_linalg_norm_subgradients_at_zero_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_linalg_pinv_singular_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_log1p_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_masked_argmin_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_masked_cumsum_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_mean_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_mv_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nan_to_num_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nanquantile_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_native_layer_norm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_adaptive_avg_pool2d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_avg_pool2d_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_batch_norm_without_cudnn_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_bilinear_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_embedding_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_local_response_norm_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_pixel_unshuffle_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_nn_functional_threshold_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_ones_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_polygamma_polygamma_n_1_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_polygamma_polygamma_n_3_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_qr_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_ravel_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_resize_as__cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_resolve_conj_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_signal_windows_general_hamming_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_special_chebyshev_polynomial_w_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_special_i1e_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_special_modified_bessel_i1_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_special_shifted_chebyshev_polynomial_t_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_split_with_sizes_copy_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_split_with_sizes_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_trunc_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_unfold_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_pointwise_ops_uniform_cuda_float32, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_linspace_tensor_overload_cuda_complex128, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_linspace_tensor_overload_cuda_int32, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_logspace_cuda_complex128, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_logspace_tensor_overload_cuda_bfloat16, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_logspace_tensor_overload_cuda_int16, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_logspace_tensor_overload_cuda_uint8, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout__refs_ones_cuda_complex64, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_full_cuda_float16, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_linspace_tensor_overload_cuda_uint8, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_logspace_cuda_bfloat16, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_ones_cuda_int32, test/test_ops.py::TestFakeTensorCUDA::test_strided_layout_ones_cuda_int64, test/test_ops.py::TestTagsCUDA::test_tags___radd___cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs__conversions_cfloat_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs__conversions_float_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_addcdiv_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_amax_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_asin_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_bitwise_xor_cuda_int64, test/test_ops.py::TestTagsCUDA::test_tags__refs_clamp_min_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_diagonal_copy_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_dot_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_dsplit_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_empty_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_empty_like_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_exp2_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_expand_as_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_eye_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_fft_fft2_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_fft_ifft2_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_fft_irfft_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_le_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_meshgrid_variadic_tensors_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_nn_functional_celu_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_nn_functional_hardshrink_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_nn_functional_leaky_relu_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_nn_functional_softplus_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_normal__in_place_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_reshape_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_special_erfcx_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_tensor_split_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_tril_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_unbind_copy_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__refs_unsqueeze_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags__upsample_bilinear2d_aa_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_acos_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_alias_copy_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_all_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_allclose_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_arange_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_atleast_1d_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_bool_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_char_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_div_floor_rounding_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_exp_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_exponential_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_eye_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_fft_fft_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_fft_fftn_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_fft_ifft_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_fft_irfftn_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_fft_rfft_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_fft_rfftn_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_frac_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_ge_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_half_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_item_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_linalg_cholesky_ex_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_linalg_diagonal_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_linalg_inv_ex_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_logspace_tensor_overload_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_mH_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_mT_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_masked_cumprod_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_masked_normalize_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_masked_scatter_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_masked_var_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_multinomial_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nansum_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_native_layer_norm_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_dropout_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_max_unpool2d_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_multilabel_margin_loss_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_pad_replicate_negative_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_pixel_unshuffle_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_nn_functional_upsample_nearest_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_norm_inf_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_ormqr_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_polygamma_polygamma_n_2_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_polygamma_polygamma_n_3_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_ravel_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_round_decimals_0_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_rsqrt_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_scalar_tensor_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_scatter_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_sgn_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_signal_windows_hamming_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_special_erfcx_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_special_log_ndtr_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_special_modified_bessel_k1_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_special_shifted_chebyshev_polynomial_v_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_special_xlog1py_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_square_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_svd_lowrank_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_t_copy_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_tensordot_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_tile_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_to_sparse_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_topk_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_torch_ops_aten__safe_softmax_default_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_vstack_cuda_float32, test/test_ops.py::TestTagsCUDA::test_tags_where_cuda_float32 2025-12-04T15:08:17.5922567Z 2025-12-04T15:08:17.5922869Z Finished test_ops 11/11 ... [2025-12-04 15:08:17.347877][21281.038263802], took 20.37min 2025-12-04T15:08:17.5923914Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_ops/test_ops-6f4cd93b78224a58.xml 2025-12-04T15:08:19.0333242Z Uploading artifacts took 1.46 seconds 2025-12-04T15:08:19.0337267Z Running functorch/test_ops 3/7 ... [2025-12-04 15:08:19.033545][21282.723937901] 2025-12-04T15:08:19.0337827Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T15:08:19.0342141Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/test_ops.py', '--shard-id=3', '--num-shards=7', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 15:08:19.033975] 2025-12-04T15:17:18.5196477Z 2025-12-04T15:17:18.5197438Z functorch/test_ops 3/7 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_ops_3.7_580309c6de0de560_.log 2025-12-04T15:17:18.5864732Z Running 1415 items in this shard: test/functorch/test_ops.py::TestOperatorsCUDA::test_data_write_errors_under_transform_cuda, test/functorch/test_ops.py::TestOperatorsCUDA::test_extremal_numerics_layer_norm_cuda, test/functorch/test_ops.py::TestOperatorsCUDA::test_extremal_numerics_mse_loss_cuda, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_NumpyExpMarkDirtyAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_NumpyTakeAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_SelectGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_T_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_argmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_as_strided_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_as_strided_partial_views_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_atan_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_atanh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_atleast_1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_atleast_2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_bool_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_broadcast_to_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_cauchy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_cfloat_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_conj_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_constant_pad_nd_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_cosh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_cummin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_diagonal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_double_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_erf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_exp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_fft_rfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_flatten_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_float_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_fmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_geqrf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_hstack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_index_fill_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_inner_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_int_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_item_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_lgamma_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_linalg_lu_factor_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_linalg_matrix_rank_hermitian_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_linalg_multi_dot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_linalg_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_linalg_slogdet_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_linalg_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_linspace_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_log_softmax_with_dtype_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_logaddexp2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_logcumsumexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_logical_and_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_logit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_logspace_tensor_overload_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_masked_amin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_masked_median_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_median_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_meshgrid_list_of_tensors_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_minimum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_ne_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_neg_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_adaptive_avg_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_conv2d_stride_padding_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_conv2d_strided_padding_dilation_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_cosine_similarity_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_fractional_max_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_max_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_multi_head_attention_forward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_poisson_nll_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_norm_fro_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_ops_aten__new_zeros_with_same_feature_meta_functorchonly_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_ormqr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_outer_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_polygamma_polygamma_n_2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_positive_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_quantile_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_ravel_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_round_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_sigmoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_signal_windows_general_cosine_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_signbit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_sparse_mm_reduce_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_special_bessel_j0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_special_erfcx_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_special_modified_bessel_k0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_special_modified_bessel_k1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_special_ndtr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_special_ndtri_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_special_shifted_chebyshev_polynomial_u_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_special_shifted_chebyshev_polynomial_w_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_stack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_t_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_transpose_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_trapz_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_true_divide_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_unflatten_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_unsafe_chunk_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_var_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_vdot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_zeros_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_MulGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_SelectGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_T_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp__unsafe_masked_index_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_acosh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_addbmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_addmm_decomposed_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_atleast_3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_bernoulli_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_cdist_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_clamp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_complex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_conj_physical_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_cumprod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_diag_embed_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_erfinv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_fft_hfft2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_fft_hfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_gather_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_half_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_hash_tensor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_heaviside_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_index_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_inner_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_int_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_isreal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_cond_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_eigh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_ldl_factor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_matrix_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_matrix_rank_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_multi_dot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_qr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_svd_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_masked_cumsum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_masked_fill_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_masked_fill_functorch_Scalar_only_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_masked_log_softmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_masked_logsumexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_masked_std_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_meshgrid_variadic_tensors_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_multinomial_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_mvlgamma_mvlgamma_p_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_alpha_dropout_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_binary_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_conv1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_conv2d_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_embedding_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_gelu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_hinge_embedding_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_huber_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_instance_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_interpolate_bicubic_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_interpolate_bilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_linear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_max_unpool3d_grad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_multi_head_attention_forward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_multilabel_soft_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_nll_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_pad_circular_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_pad_replicate_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_relu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_scaled_dot_product_attention_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_silu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_softmin_with_dtype_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_softplus_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_softshrink_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_softsign_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_threshold_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_pinverse_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_quantile_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_rad2deg_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_randn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_repeat_interleave_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_reshape_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_resolve_neg_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_scatter_add_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_scatter_reduce_amax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_searchsorted_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_select_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_sgn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_sigmoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_signal_windows_general_cosine_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_signal_windows_hamming_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_signal_windows_nuttall_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_sinh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_sparse_mm_reduce_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_special_chebyshev_polynomial_v_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_special_laguerre_polynomial_l_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_special_legendre_polynomial_p_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_special_modified_bessel_k0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_special_modified_bessel_k1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_special_shifted_chebyshev_polynomial_v_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_split_with_sizes_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_t_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_tanh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_tile_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_torch_ops_aten__efficient_attention_forward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_tril_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_unbind_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_var_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_where_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_ForwardHasDefaultArgsAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_SelectAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_SelectGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp___rpow___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp__softmax_backward_data_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp__upsample_bilinear2d_aa_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_addbmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_amax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_arange_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_argmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_as_strided_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_asin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_asinh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_atleast_2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_atleast_3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_bool_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_cholesky_inverse_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_cholesky_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_clamp_max_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_cummax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_cummin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_diagonal_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_div_trunc_rounding_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_dstack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_erfinv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_expm1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_eye_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_fft_fftn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_fft_irfft2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_fft_irfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_fft_rfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_frac_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_hash_tensor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_isnan_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_kron_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_lgamma_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_linalg_cross_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_linalg_ldl_factor_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_linalg_lu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_linalg_lu_factor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_linalg_lu_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_linalg_svdvals_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_logsumexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_mH_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_masked_logsumexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_masked_var_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_matrix_exp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_max_pool2d_with_indices_backward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_meshgrid_list_of_tensors_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_mul_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_mv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nan_to_num_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nanquantile_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nansum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_new_empty_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_adaptive_avg_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_avg_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_channel_shuffle_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_conv2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_conv2d_stride_padding_no_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_conv3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_dropout2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_embedding_bag_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_hardsigmoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_max_pool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_max_unpool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_max_unpool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_multilabel_soft_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_softplus_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_upsample_bilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nonzero_static_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_norm_inf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_permute_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_polygamma_polygamma_n_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_put_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_randn_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_ravel_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_real_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_reshape_as_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_reshape_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_rsqrt_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_rsub_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_short_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_sigmoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_signal_windows_gaussian_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_signal_windows_general_hamming_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_signal_windows_nuttall_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_special_chebyshev_polynomial_w_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_special_i0e_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_special_ndtri_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_special_shifted_chebyshev_polynomial_u_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_sum_to_size_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_t_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_take_along_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_tanh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_to_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_to_sparse_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_transpose_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_transpose_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_tril_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_triu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_unbind_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_var_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_view_as_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_where_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_zeros_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjpvmap_ForwardHasDefaultArgsAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvmap_NumpyCubeNotComposableAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvmap_ScaleGradGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvmapvmap_NumpyMulAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvmapvmap_NumpySortAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_ordered_complex_raises_amax_cuda_complex128, test/functorch/test_ops.py::TestOperatorsCUDA::test_ordered_complex_raises_amin_cuda_complex64, test/functorch/test_ops.py::TestOperatorsCUDA::test_ordered_complex_raises_argmax_cuda_complex64, test/functorch/test_ops.py::TestOperatorsCUDA::test_ordered_complex_raises_ceil_cuda_complex128, test/functorch/test_ops.py::TestOperatorsCUDA::test_ordered_complex_raises_floor_cuda_complex32, test/functorch/test_ops.py::TestOperatorsCUDA::test_ordered_complex_raises_gt_cuda_complex64, test/functorch/test_ops.py::TestOperatorsCUDA::test_ordered_complex_raises_le_cuda_complex32, test/functorch/test_ops.py::TestOperatorsCUDA::test_ordered_complex_raises_maximum_cuda_complex128, test/functorch/test_ops.py::TestOperatorsCUDA::test_ordered_complex_raises_maximum_cuda_complex32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_T_grad_op_jvp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_diagonal_grad_op_jvp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_diagonal_grad_op_vjp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_list_return_hsplit_grad_op_jvp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_list_return_unbind_grad_op_vjp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_mT_grad_op_vjp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_reshape_grad_op_vjp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_squeeze_grad_op_jvp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_unfold_grad_op_jvp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_unsqueeze_grad_op_vjp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp___rdiv___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp___rsub___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp__batch_norm_with_update_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp__segment_reduce_offsets_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_abs_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_any_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_argmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_as_strided_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_bmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_contiguous_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_cos_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_cummin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_double_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_erf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_fft_fft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_fft_fftshift_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_flatten_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_float_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_float_power_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_gradient_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_hsplit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_index_reduce_amax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_isin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_isnan_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_le_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_lerp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_linalg_cholesky_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_linalg_lstsq_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_linalg_lu_factor_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_linalg_multi_dot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_linalg_tensorinv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_linalg_vecdot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_linalg_vector_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_linspace_tensor_overload_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_logaddexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_logical_xor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_masked_fill_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_max_binary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_max_pool2d_with_indices_backward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_max_reduction_no_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_meshgrid_list_of_tensors_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_min_reduction_no_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_mv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_native_batch_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_native_dropout_backward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_native_layer_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_conv2d_no_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_conv2d_stride_padding_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_conv2d_strided_padding_dilation_no_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_fractional_max_pool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_grid_sample_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_hinge_embedding_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_interpolate_area_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_l1_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_layer_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_leaky_relu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_logsigmoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_max_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_max_unpool2d_grad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_mse_loss_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_pad_replicate_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_prelu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_soft_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_normal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_ops_aten_index_put_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_polar_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_repeat_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_round_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_round_decimals_neg_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_scatter_reduce_sum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_signal_windows_cosine_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_slice_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_special_entr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_special_modified_bessel_i0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_special_ndtr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_special_shifted_chebyshev_polynomial_t_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_special_shifted_chebyshev_polynomial_w_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_squeeze_multiple_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_svd_lowrank_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_topk_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_transpose_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_unfold_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_unique_consecutive_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_unique_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_unsqueeze_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_view_as_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_where_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_MulGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_SelectAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_ZeroGradientsGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp__batch_norm_with_update_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp__softmax_backward_data_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp__unsafe_masked_index_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_addbmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_addmv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_all_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_allclose_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_amax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_amin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_argmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_argmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_as_strided_partial_views_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_atan2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_baddbmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_bernoulli_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_block_diag_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_bool_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_ceil_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_char_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_constant_pad_nd_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_contiguous_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_count_nonzero_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_cumprod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_cumsum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_div_no_rounding_mode_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_double_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_empty_permuted_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_eq_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_equal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_exp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_expand_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_flatten_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_fliplr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_float_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_frac_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_geometric_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_half_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_heaviside_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_index_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_index_fill_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_index_put_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_index_select_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_int_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_isclose_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_isfinite_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_kron_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_linalg_cholesky_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_linalg_lstsq_grad_oriented_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_linalg_matrix_rank_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_linalg_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_linalg_pinv_singular_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_linalg_vander_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_logical_and_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_logspace_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_long_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_mT_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_masked_fill_functorch_Scalar_only_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_max_reduction_with_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_maximum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_median_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_min_reduction_no_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_mvlgamma_mvlgamma_p_1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nanmean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_narrow_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_ne_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_avg_pool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_batch_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_binary_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_conv2d_stride_padding_no_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_cosine_similarity_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_ctc_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_fractional_max_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_layer_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_max_unpool1d_grad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_mish_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_multi_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_softmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_softmin_with_dtype_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_triplet_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_upsample_bilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_ones_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_outer_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_polygamma_polygamma_n_1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_polygamma_polygamma_n_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_polygamma_polygamma_n_4_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_resize__cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_rot90_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_round_decimals_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_rsub_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_scatter_reduce_amin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_select_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_signal_windows_cosine_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_signal_windows_kaiser_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_slice_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_special_bessel_j0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_special_bessel_j1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_special_bessel_y1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_special_chebyshev_polynomial_v_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_special_i0e_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_special_ndtri_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_split_with_sizes_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_squeeze_multiple_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_std_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_take_along_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_tan_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_trace_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_unsqueeze_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_var_mean_unbiased_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_vdot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_view_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjpvmap_NumpyMulAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjpvmap_ScaleGradGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjpvmap_SelectAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_NumpySortAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_ScaleGradGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap__chunk_cat_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap__native_batch_norm_legit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap__segment_reduce_lengths_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap__unsafe_masked_index_put_accumulate_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_alias_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_as_strided_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_asin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_atan_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_atleast_3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_broadcast_to_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_cat_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_cdist_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_column_stack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_contiguous_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_cos_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_cummax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_dist_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_double_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_dstack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_empty_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_empty_strided_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_expand_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_fft_hfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_fft_ifftn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_fft_ifftshift_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_fft_ihfft2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_fft_ihfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_flatten_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_frexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_hsplit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_int_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_jiterator_binary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_le_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_linalg_det_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_linalg_eigh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_log2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_log_softmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_log_softmax_with_dtype_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_logcumsumexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_logdet_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_logical_and_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_logical_xor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_long_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_lu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_masked_argmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_masked_normalize_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_masked_softmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_minimum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_mode_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_narrow_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_native_dropout_backward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_native_layer_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_neg_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_avg_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_batch_norm_without_cudnn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_binary_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_conv1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_conv2d_stride_depthwise_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_conv2d_stride_padding_no_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_fractional_max_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_gaussian_nll_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_grid_sample_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_group_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_hardshrink_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_hardtanh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_l1_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_margin_ranking_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_mish_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_pdist_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_selu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_ops_aten__new_zeros_with_same_feature_meta_functorchonly_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_polygamma_polygamma_n_1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_put_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_rand_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_ravel_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_real_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_reshape_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_resolve_conj_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_round_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_round_decimals_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_short_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_sign_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_signal_windows_exponential_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_special_bessel_j0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_special_chebyshev_polynomial_t_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_special_i1e_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_special_laguerre_polynomial_l_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_special_modified_bessel_i1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_special_ndtri_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_special_scaled_modified_bessel_k1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_special_shifted_chebyshev_polynomial_t_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_split_list_args_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_tensordot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_trace_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_triu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_uniform_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_unsqueeze_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_vsplit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmapvmap_NumpyMulAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmapvmap_ScaleGradGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmapvmap_SelectGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_CubeGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_NumpyMulAutogradFunction_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_SelectAutogradFunction_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_T_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad___getitem___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad___getitem___functorch_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad___rpow___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad__chunk_cat_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad__segment_reduce_lengths_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_acos_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_acosh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_addcmul_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_addmm_decomposed_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_addr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_all_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_amin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_aminmax_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_arange_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_as_strided_copy_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_asin_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_atan_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_atan_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_bernoulli_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_bfloat16_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_bool_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_cartesian_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_cat_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_ceil_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_char_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_clamp_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_clamp_max_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_column_stack_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_contiguous_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_cummax_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_cumsum_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_cumulative_trapezoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_diagonal_copy_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_dstack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_empty_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_empty_permuted_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_fft_fft2_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_fft_ihfft_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_fft_irfft_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_fft_rfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_flatten_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_float_power_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_floor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_fmod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_fmod_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_frexp_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_full_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_grid_sampler_2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_gt_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_hash_tensor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_heaviside_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_hstack_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_index_fill_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_index_fill_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_index_put_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_index_reduce_amin_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_isclose_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_isposinf_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_lgamma_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_cholesky_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_cond_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_cross_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_det_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_det_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_inv_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_ldl_factor_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_matrix_power_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_matrix_power_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_norm_subgradients_at_zero_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_slogdet_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_log_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_log_normal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_log_softmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_logdet_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_logdet_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_logical_or_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_logical_xor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_logit_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_logspace_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_logspace_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_logsumexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_lu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_mH_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_masked_argmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_masked_logsumexp_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_masked_median_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_masked_normalize_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_masked_softmax_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_max_reduction_no_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_maximum_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_minimum_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_movedim_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_mvlgamma_mvlgamma_p_3_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nanmedian_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_narrow_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_native_dropout_backward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_native_layer_norm_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_new_empty_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_new_empty_strided_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_adaptive_avg_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_adaptive_max_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_adaptive_max_pool1d_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_alpha_dropout_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_avg_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_avg_pool3d_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_channel_shuffle_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_conv2d_stride_groups_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_conv2d_stride_no_bias_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_conv2d_strided_padding_dilation_no_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_conv2d_strided_padding_dilation_with_bias_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_conv2d_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_cosine_similarity_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_dropout_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_embedding_bag_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_embedding_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_feature_alpha_dropout_without_train_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_fractional_max_pool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_gaussian_nll_loss_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_gelu_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_hardshrink_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_interpolate_area_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_interpolate_nearest-exact_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_linear_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_margin_ranking_loss_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_max_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_max_pool1d_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_max_pool2d_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_max_unpool2d_grad_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_max_unpool3d_grad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_mse_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_pad_replicate_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_prelu_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_relu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_rrelu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_soft_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_soft_margin_loss_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_softplus_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_unfold_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nonzero_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_ones_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_ones_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_ormqr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_polygamma_polygamma_n_1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_positive_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_put_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_rand_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_randint_like_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_randn_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_randn_like_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_renorm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_resolve_neg_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_roll_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_round_decimals_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_round_decimals_3_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_rsqrt_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_scatter_reduce_sum_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_select_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_select_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_sgn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_sgn_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_signal_windows_bartlett_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_signal_windows_blackman_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_signal_windows_nuttall_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_sin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_slice_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_softmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_softmax_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_softmax_with_dtype_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_bessel_j0_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_bessel_y0_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_bessel_y1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_chebyshev_polynomial_t_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_chebyshev_polynomial_v_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_entr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_i0e_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_modified_bessel_k1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_polygamma_special_polygamma_n_0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_polygamma_special_polygamma_n_0_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_xlog1py_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_xlog1py_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_squeeze_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_squeeze_multiple_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_std_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_std_mean_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_stft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_sum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_svd_lowrank_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_take_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_tan_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_tile_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_to_sparse_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_transpose_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_trapz_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_true_divide_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_unbind_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_unfold_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_unique_consecutive_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_unique_consecutive_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_unsafe_split_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_view_as_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_vsplit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_zero__cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_zeros_like_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_SortGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall__unsafe_masked_index_put_accumulate_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_abs_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_add_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_addmm_decomposed_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_addr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_argmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_atan_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_atanh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_baddbmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_bfloat16_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_broadcast_shapes_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_clamp_min_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_contiguous_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_copysign_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_cumsum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_diagonal_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_diagonal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_digamma_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_empty_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_empty_permuted_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_exp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_exponential_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_fft_hfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_fft_irfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_float_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_floor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_floor_divide_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_fmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_ge_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_gt_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_H_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_NumpyCubeNotComposableAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_ScaleGradGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_acos_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_alias_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_amax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_amin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_argwhere_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_as_strided_partial_views_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_as_strided_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_asin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_broadcast_tensors_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_byte_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_cat_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_cdist_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_conj_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_contiguous_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_diagonal_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_dsplit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_erfinv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_expand_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_eye_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_fft_ifftshift_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_fft_ihfftn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_floor_divide_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_geometric_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_grid_sampler_2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_index_put_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_inner_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_int_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_isinf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_isnan_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_isreal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_jiterator_unary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_linalg_cholesky_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_linalg_eigvals_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_linalg_eigvalsh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_linalg_inv_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_linalg_ldl_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_linalg_multi_dot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_log_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_mT_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_masked_log_softmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_masked_logaddexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_masked_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_masked_var_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_min_binary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_mvlgamma_mvlgamma_p_5_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nansum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_narrow_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_new_empty_strided_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_adaptive_avg_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_binary_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_celu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_conv2d_stride_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_conv_transpose1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_conv_transpose3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_ctc_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_dropout2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_embedding_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_embedding_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_feature_alpha_dropout_with_train_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_fractional_max_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_hardtanh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_interpolate_linear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_linear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_local_response_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_mish_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_multilabel_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_prelu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_relu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_softmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_threshold_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_upsample_nearest_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nonzero_static_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_normal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_qr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_signal_windows_cosine_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_special_bessel_j1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_special_chebyshev_polynomial_t_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_special_modified_bessel_i1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_special_ndtri_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_square_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_squeeze_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_triu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_unflatten_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_var_mean_unbiased_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_view_as_complex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_zeros_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_histc_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_int_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_isinf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_item_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_jiterator_binary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_jiterator_binary_return_by_ref_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_kthvalue_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_linalg_eigvalsh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_linalg_lu_factor_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_linalg_matrix_power_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_linalg_matrix_rank_hermitian_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_linalg_norm_subgradients_at_zero_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_linspace_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_log_softmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_logaddexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_logcumsumexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_logical_or_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_logit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_lu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_masked_cumsum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_masked_logsumexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_masked_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_masked_select_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_masked_var_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_min_reduction_no_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_minimum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_mul_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_multinomial_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_mvlgamma_mvlgamma_p_1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_native_dropout_backward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_neg_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_binary_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_cosine_similarity_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_fractional_max_pool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_fractional_max_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_interpolate_area_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_interpolate_bilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_interpolate_nearest-exact_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_l1_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_margin_ranking_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_max_pool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_max_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_max_unpool2d_grad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_max_unpool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_multilabel_soft_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_normalize_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_pad_replicate_negative_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_pixel_shuffle_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_upsample_nearest_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nonzero_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_norm_fro_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_normal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_ormqr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_permute_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_polygamma_polygamma_n_0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_polygamma_polygamma_n_2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_reshape_as_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_roll_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_rot90_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_scatter_reduce_amax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_sigmoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_signal_windows_bartlett_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_signal_windows_general_hamming_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_signal_windows_hamming_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_signbit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_sin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_sinh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_special_chebyshev_polynomial_u_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_special_i0e_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_special_legendre_polynomial_p_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_special_modified_bessel_k1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_special_ndtr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_special_scaled_modified_bessel_k1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_split_with_sizes_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_squeeze_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_squeeze_multiple_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_sum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_svd_lowrank_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_take_along_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_tan_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_to_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_ForwardHasDefaultArgsAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_NumpyMulAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_NumpySortAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_T_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp___rdiv___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp___rmod___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp___rpow___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp___rsub___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp__unsafe_masked_index_put_accumulate_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_allclose_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_arange_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_as_strided_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_atan2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_bfloat16_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_bucketize_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_cholesky_inverse_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_column_stack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_diag_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_einsum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_empty_strided_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_exp2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_expand_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_fft_irfftn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_hypot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_isinf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_isnan_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_isreal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_kthvalue_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_lerp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_linalg_det_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_linalg_eigh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_linalg_inv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_linalg_lu_factor_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_linalg_pinv_hermitian_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_linalg_vector_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_linspace_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_log2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_log_normal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_logdet_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_logical_xor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_logspace_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_lu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_masked_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_masked_var_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_mode_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_movedim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_mvlgamma_mvlgamma_p_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_native_dropout_backward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_native_layer_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nextafter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_adaptive_max_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_avg_pool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_binary_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_conv1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_conv2d_stride_padding_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_conv2d_strided_padding_dilation_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_cosine_embedding_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_feature_alpha_dropout_with_train_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_grid_sample_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_hinge_embedding_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_huber_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_interpolate_linear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_local_response_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_margin_ranking_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_max_unpool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_max_unpool3d_grad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_pad_replicate_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_pixel_unshuffle_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_smooth_l1_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_triplet_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nonzero_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_normal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_polygamma_polygamma_n_2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_qr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_repeat_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_repeat_interleave_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_sign_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_signal_windows_cosine_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_special_bessel_j0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_special_chebyshev_polynomial_v_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_special_entr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_special_erfcx_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_special_modified_bessel_i0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_special_ndtri_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_split_with_sizes_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_sqrt_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_std_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_trapz_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_unique_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_unsafe_split_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_unsqueeze_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_zero__cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvmap_NumpyMulAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvmap_NumpyTakeAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvmap_ScaleGradGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_ScaleGradGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp__unsafe_masked_index_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp__unsafe_masked_index_put_accumulate_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_addbmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_atleast_1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_bernoulli_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_bfloat16_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_chalf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_column_stack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_diag_embed_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_diagflat_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_digamma_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_div_floor_rounding_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_empty_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_erfc_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_fft_ifftn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_flatten_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_NumpyCubeAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_NumpyExpMarkDirtyAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_SelectGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule___rsub___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule__segment_reduce_lengths_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_abs_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_acos_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_addcdiv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_addr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_amin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_argwhere_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_as_strided_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_atanh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_bfloat16_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_bool_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_byte_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_clamp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_clamp_max_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_count_nonzero_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_deg2rad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_diagonal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_double_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_empty_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_equal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_expand_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_fft_ifft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_fft_irfft2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_fft_rfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_fliplr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_grid_sampler_2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_gt_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_hstack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_index_reduce_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_jiterator_2inputs_2outputs_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_lerp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_lgamma_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_linalg_inv_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_linalg_lu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_linalg_pinv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_linalg_slogdet_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_linalg_solve_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_log1p_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_log_normal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_logical_and_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_logical_not_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_mT_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_masked_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_masked_softmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_masked_var_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_max_binary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_max_reduction_with_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_mul_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_mv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_mvlgamma_mvlgamma_p_1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nan_to_num_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_new_empty_strided_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_adaptive_max_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_avg_pool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_conv2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_conv2d_stride_depthwise_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_conv_transpose2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_cosine_embedding_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_cosine_similarity_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_dropout2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_embedding_bag_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_fractional_max_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_hardshrink_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_hardswish_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_hardtanh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_interpolate_nearest_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_margin_ranking_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_max_unpool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_relu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_smooth_l1_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_triplet_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_unfold_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_upsample_bilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_norm_inf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_normal_in_place_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_ones_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_reshape_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_rot90_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_round_decimals_0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_scalar_tensor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_sgn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_signal_windows_gaussian_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_signal_windows_nuttall_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_special_i1e_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_special_modified_bessel_k1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_square_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_sum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_t_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_topk_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_transpose_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_vsplit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_histc_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_jiterator_unary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_lerp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_linalg_cross_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_linalg_ldl_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_linalg_matrix_rank_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_linalg_norm_subgradients_at_zero_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_linalg_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_linalg_tensorsolve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_linspace_tensor_overload_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_log_softmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_logit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_masked_argmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_masked_cumprod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_masked_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_masked_sum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_matmul_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_max_pool2d_with_indices_backward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_meshgrid_variadic_tensors_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nansum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_new_empty_strided_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_new_ones_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nextafter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_binary_cross_entropy_with_logits_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_ctc_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_dropout2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_group_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_hinge_embedding_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_huber_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_interpolate_bilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_max_unpool2d_grad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_multi_head_attention_forward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_pad_circular_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_pdist_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_pixel_unshuffle_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_triplet_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nonzero_static_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_norm_inf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_permute_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_put_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_rand_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_renorm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_round_decimals_neg_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_signal_windows_cosine_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_signal_windows_gaussian_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_signal_windows_general_cosine_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_signal_windows_kaiser_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_signal_windows_nuttall_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_sparse_mm_reduce_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_special_airy_ai_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_special_chebyshev_polynomial_t_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_special_chebyshev_polynomial_u_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_special_hermite_polynomial_he_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_special_laguerre_polynomial_l_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_special_polygamma_special_polygamma_n_0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_special_shifted_chebyshev_polynomial_t_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_special_shifted_chebyshev_polynomial_v_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_squeeze_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_tile_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_tril_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_uniform_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_unsqueeze_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_NumpyExpMarkDirtyAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_SelectGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_SortGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_ZeroGradientsGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp___getitem___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp___getitem___functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp___rdiv___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp___rmul___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_alias_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_amax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_aminmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_any_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_atleast_2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_bfloat16_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_block_diag_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_bucketize_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_cauchy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_char_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_clone_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_complex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_count_nonzero_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_cumsum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_cumulative_trapezoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_diagonal_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_dsplit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_eq_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_erf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_fft_irfftn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_fft_rfftn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_grid_sampler_2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_hypot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_igammac_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_index_put_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_isposinf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_le_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_linalg_eig_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_linalg_eigvalsh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_log_normal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_logical_not_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_logspace_tensor_overload_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_logsumexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_mH_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_masked_var_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_matrix_exp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_min_binary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_mul_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_narrow_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_new_empty_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_adaptive_max_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_adaptive_max_pool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_binary_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_conv2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_conv2d_stride_no_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_conv2d_stride_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_conv2d_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_cosine_embedding_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_embedding_bag_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_embedding_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_gelu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_hardtanh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_interpolate_trilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_layer_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_max_unpool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_multilabel_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_normalize_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_pad_reflect_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_pairwise_distance_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_prelu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_scaled_dot_product_attention_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_softmin_with_dtype_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_triplet_margin_with_distance_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_permute_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_randint_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_round_decimals_neg_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_sigmoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_signal_windows_general_hamming_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_sin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_sinh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_slice_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_special_bessel_j0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_special_hermite_polynomial_h_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_special_modified_bessel_k0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_split_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_stack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_svd_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_tanh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_transpose_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_trapezoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_tril_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvmap_SortGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvmap_ZeroGradientsGenVmapAutogradFunction_cuda_float32 2025-12-04T15:17:18.6515763Z 2025-12-04T15:17:18.6516134Z Finished functorch/test_ops 3/7 ... [2025-12-04 15:17:18.521673][21822.212063385], took 8.99min 2025-12-04T15:17:18.6517528Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_ops/functorch.test_ops-37d651ead30b362d.xml 2025-12-04T15:17:18.6823464Z Running functorch/test_vmap 1/1 ... [2025-12-04 15:17:18.682043][21822.37243318] 2025-12-04T15:17:18.6824015Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T15:17:18.6827199Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/test_vmap.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 15:17:18.682457] 2025-12-04T15:25:36.2368265Z 2025-12-04T15:25:36.2369380Z PRINTING LOG FILE of functorch/test_vmap 1/1 (test/test-reports/functorch.test_vmap_1.1_151d84486b0c8e0c_.log) 2025-12-04T15:25:36.2370711Z Test results will be stored in test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-2355f4f75c6e0706.xml 2025-12-04T15:25:36.2371575Z ============================= test session starts ============================== 2025-12-04T15:25:36.2372467Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T15:25:36.2373306Z cachedir: .pytest_cache 2025-12-04T15:25:36.2374094Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T15:25:36.2375159Z rootdir: /var/lib/jenkins/workspace 2025-12-04T15:25:36.2375711Z configfile: pytest.ini 2025-12-04T15:25:36.2376891Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T15:25:36.2378134Z collecting ... collected 2134 items 2025-12-04T15:25:36.2378648Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T15:25:36.3571487Z Running 2134 items in this shard: test/functorch/test_vmap.py::TestVmapAPI::test_accepts_nested_inputs, test/functorch/test_vmap.py::TestVmapAPI::test_backward_unsupported_interaction, test/functorch/test_vmap.py::TestVmapAPI::test_batch_rule_does_not_need_to_handle_no_batched_input, test/functorch/test_vmap.py::TestVmapAPI::test_batched_gradient_basic, test/functorch/test_vmap.py::TestVmapAPI::test_checkpoint, test/functorch/test_vmap.py::TestVmapAPI::test_constant_function, test/functorch/test_vmap.py::TestVmapAPI::test_data_attribute, test/functorch/test_vmap.py::TestVmapAPI::test_data_dependent_control_flow_throws, test/functorch/test_vmap.py::TestVmapAPI::test_decomposition_under_python_dispatcher, test/functorch/test_vmap.py::TestVmapAPI::test_different_map_dim_size_raises, test/functorch/test_vmap.py::TestVmapAPI::test_fallback_does_not_warn_by_default, test/functorch/test_vmap.py::TestVmapAPI::test_fallback_masked_fill, test/functorch/test_vmap.py::TestVmapAPI::test_fallback_multiple_returns, test/functorch/test_vmap.py::TestVmapAPI::test_fallback_warning, test/functorch/test_vmap.py::TestVmapAPI::test_fallback_warns_when_warnings_are_enabled, test/functorch/test_vmap.py::TestVmapAPI::test_fallback_with_undefined_grad, test/functorch/test_vmap.py::TestVmapAPI::test_fallback_zero_dim, test/functorch/test_vmap.py::TestVmapAPI::test_func_with_no_inputs, test/functorch/test_vmap.py::TestVmapAPI::test_func_with_no_tensors, test/functorch/test_vmap.py::TestVmapAPI::test_functools_partial, test/functorch/test_vmap.py::TestVmapAPI::test_grad_unsupported_interaction, test/functorch/test_vmap.py::TestVmapAPI::test_in_dim_not_in_tensor_err_msg, test/functorch/test_vmap.py::TestVmapAPI::test_in_dims_wrong_type_err_msg, test/functorch/test_vmap.py::TestVmapAPI::test_inplace_fallback_nary_different_levels, test/functorch/test_vmap.py::TestVmapAPI::test_inplace_fallback_nary_same_levels, test/functorch/test_vmap.py::TestVmapAPI::test_inplace_fallback_unary, test/functorch/test_vmap.py::TestVmapAPI::test_integer_in_dim_but_not_tensor_input_err_msg, test/functorch/test_vmap.py::TestVmapAPI::test_item_throws, test/functorch/test_vmap.py::TestVmapAPI::test_multiple_inputs, test/functorch/test_vmap.py::TestVmapAPI::test_multiple_out_dims, test/functorch/test_vmap.py::TestVmapAPI::test_multiple_outputs, test/functorch/test_vmap.py::TestVmapAPI::test_multiple_outputs2, test/functorch/test_vmap.py::TestVmapAPI::test_nested_negative_in_dims, test/functorch/test_vmap.py::TestVmapAPI::test_nested_non_default_in_dims, test/functorch/test_vmap.py::TestVmapAPI::test_nested_out_dims, test/functorch/test_vmap.py::TestVmapAPI::test_nested_with_diag_embed, test/functorch/test_vmap.py::TestVmapAPI::test_nested_with_different_map_dim, test/functorch/test_vmap.py::TestVmapAPI::test_nested_with_same_map_dim, test/functorch/test_vmap.py::TestVmapAPI::test_nn_module, test/functorch/test_vmap.py::TestVmapAPI::test_non_default_in_dims_out_dims, test/functorch/test_vmap.py::TestVmapAPI::test_non_tensor_output_raises, test/functorch/test_vmap.py::TestVmapAPI::test_non_zero_in_dims, test/functorch/test_vmap.py::TestVmapAPI::test_none_in_dims, test/functorch/test_vmap.py::TestVmapAPI::test_nonzero_out_dims, test/functorch/test_vmap.py::TestVmapAPI::test_noop_in_inner_vmap, test/functorch/test_vmap.py::TestVmapAPI::test_not_enough_in_dims_err_msg, test/functorch/test_vmap.py::TestVmapAPI::test_out_dim_out_of_bounds_err_msg, test/functorch/test_vmap.py::TestVmapAPI::test_out_dims_and_num_outputs_mismatch_err_msg, test/functorch/test_vmap.py::TestVmapAPI::test_out_dims_edge_case, test/functorch/test_vmap.py::TestVmapAPI::test_out_dims_must_be_int_or_collection_of_int_err_msg, test/functorch/test_vmap.py::TestVmapAPI::test_out_dims_none, test/functorch/test_vmap.py::TestVmapAPI::test_out_dims_none_tuple, test/functorch/test_vmap.py::TestVmapAPI::test_out_dims_normal_tensor, test/functorch/test_vmap.py::TestVmapAPI::test_pytree_odict_returns, test/functorch/test_vmap.py::TestVmapAPI::test_pytree_returns, test/functorch/test_vmap.py::TestVmapAPI::test_pytree_returns_broadcast_nested, test/functorch/test_vmap.py::TestVmapAPI::test_pytree_returns_broadcast_simple, test/functorch/test_vmap.py::TestVmapAPI::test_pytree_returns_outdims, test/functorch/test_vmap.py::TestVmapAPI::test_reshape_dim_into, test/functorch/test_vmap.py::TestVmapAPI::test_reshape_dim_outof, test/functorch/test_vmap.py::TestVmapAPI::test_restore_vmap_no_vmapped_inputs, test/functorch/test_vmap.py::TestVmapAPI::test_restore_vmap_pytree_input_output, test/functorch/test_vmap.py::TestVmapAPI::test_restore_vmap_unexpanded_outputs, test/functorch/test_vmap.py::TestVmapAPI::test_single_input, test/functorch/test_vmap.py::TestVmapAPI::test_unsupported_op_err_msg, test/functorch/test_vmap.py::TestVmapAPI::test_vmap_autocast_cpu, test/functorch/test_vmap.py::TestVmapAPI::test_vmap_autocast_cuda, test/functorch/test_vmap.py::TestVmapOperators::test_T_numpy, test/functorch/test_vmap.py::TestVmapOperators::test_adaptive_avg_pool2d, test/functorch/test_vmap.py::TestVmapOperators::test_argmax_dim, test/functorch/test_vmap.py::TestVmapOperators::test_arithmetic_add, test/functorch/test_vmap.py::TestVmapOperators::test_arithmetic_add_dunder, test/functorch/test_vmap.py::TestVmapOperators::test_arithmetic_div, test/functorch/test_vmap.py::TestVmapOperators::test_arithmetic_div_dunder, test/functorch/test_vmap.py::TestVmapOperators::test_arithmetic_mul, test/functorch/test_vmap.py::TestVmapOperators::test_arithmetic_mul_dunder, test/functorch/test_vmap.py::TestVmapOperators::test_arithmetic_pow, test/functorch/test_vmap.py::TestVmapOperators::test_arithmetic_pow_dunder, test/functorch/test_vmap.py::TestVmapOperators::test_arithmetic_sub, test/functorch/test_vmap.py::TestVmapOperators::test_arithmetic_sub_dunder, test/functorch/test_vmap.py::TestVmapOperators::test_as_strided, test/functorch/test_vmap.py::TestVmapOperators::test_bmm, test/functorch/test_vmap.py::TestVmapOperators::test_cat, test/functorch/test_vmap.py::TestVmapOperators::test_chunk, test/functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_0_out_dim_0_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_0_out_dim_0_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_0_out_dim_1_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_0_out_dim_1_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_0_out_dim_2_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_0_out_dim_2_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_1_out_dim_0_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_1_out_dim_0_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_1_out_dim_1_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_1_out_dim_1_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_1_out_dim_2_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_1_out_dim_2_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_2_out_dim_0_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_2_out_dim_0_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_2_out_dim_1_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_2_out_dim_1_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_2_out_dim_2_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_2_out_dim_2_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_clamp, test/functorch/test_vmap.py::TestVmapOperators::test_clamp_inplace_variant_clamp_max_, test/functorch/test_vmap.py::TestVmapOperators::test_clamp_inplace_variant_clamp_min_, test/functorch/test_vmap.py::TestVmapOperators::test_clamp_variant_clamp_max, test/functorch/test_vmap.py::TestVmapOperators::test_clamp_variant_clamp_min, test/functorch/test_vmap.py::TestVmapOperators::test_clone, test/functorch/test_vmap.py::TestVmapOperators::test_comparison_ops, test/functorch/test_vmap.py::TestVmapOperators::test_conj, test/functorch/test_vmap.py::TestVmapOperators::test_conj_bit, test/functorch/test_vmap.py::TestVmapOperators::test_contiguous, test/functorch/test_vmap.py::TestVmapOperators::test_conv2d, test/functorch/test_vmap.py::TestVmapOperators::test_copy_, test/functorch/test_vmap.py::TestVmapOperators::test_cross_batch_size_three, test/functorch/test_vmap.py::TestVmapOperators::test_diagonal, test/functorch/test_vmap.py::TestVmapOperators::test_dot, test/functorch/test_vmap.py::TestVmapOperators::test_expand_as, test/functorch/test_vmap.py::TestVmapOperators::test_fill_and_zero_inplace, test/functorch/test_vmap.py::TestVmapOperators::test_imag, test/functorch/test_vmap.py::TestVmapOperators::test_is_complex, test/functorch/test_vmap.py::TestVmapOperators::test_is_contiguous, test/functorch/test_vmap.py::TestVmapOperators::test_is_floating_point, test/functorch/test_vmap.py::TestVmapOperators::test_mean, test/functorch/test_vmap.py::TestVmapOperators::test_mean_dim, test/functorch/test_vmap.py::TestVmapOperators::test_mm, test/functorch/test_vmap.py::TestVmapOperators::test_mode_key, test/functorch/test_vmap.py::TestVmapOperators::test_movedim, test/functorch/test_vmap.py::TestVmapOperators::test_mv, test/functorch/test_vmap.py::TestVmapOperators::test_narrow, test/functorch/test_vmap.py::TestVmapOperators::test_new_empty, test/functorch/test_vmap.py::TestVmapOperators::test_new_empty_strided, test/functorch/test_vmap.py::TestVmapOperators::test_new_zeros, test/functorch/test_vmap.py::TestVmapOperators::test_nll_loss, test/functorch/test_vmap.py::TestVmapOperators::test_one_hot, test/functorch/test_vmap.py::TestVmapOperators::test_real, test/functorch/test_vmap.py::TestVmapOperators::test_repeat, test/functorch/test_vmap.py::TestVmapOperators::test_reshape, test/functorch/test_vmap.py::TestVmapOperators::test_reshape_as, test/functorch/test_vmap.py::TestVmapOperators::test_result_type, test/functorch/test_vmap.py::TestVmapOperators::test_roll_no_dims, test/functorch/test_vmap.py::TestVmapOperators::test_select, test/functorch/test_vmap.py::TestVmapOperators::test_silu_backward, test/functorch/test_vmap.py::TestVmapOperators::test_slice, test/functorch/test_vmap.py::TestVmapOperators::test_slogdet, test/functorch/test_vmap.py::TestVmapOperators::test_split, test/functorch/test_vmap.py::TestVmapOperators::test_squeeze, test/functorch/test_vmap.py::TestVmapOperators::test_stack, test/functorch/test_vmap.py::TestVmapOperators::test_stride, test/functorch/test_vmap.py::TestVmapOperators::test_sum, test/functorch/test_vmap.py::TestVmapOperators::test_sum_dim, test/functorch/test_vmap.py::TestVmapOperators::test_t, test/functorch/test_vmap.py::TestVmapOperators::test_tensor_split, test/functorch/test_vmap.py::TestVmapOperators::test_to, test/functorch/test_vmap.py::TestVmapOperators::test_trace, test/functorch/test_vmap.py::TestVmapOperators::test_transpose, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_abs, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_acos, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_asin, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_atan, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_ceil, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_cos, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_cosh, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_digamma, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_exp, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_expm1, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_floor, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_frac, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_lgamma, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_log, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_log10, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_log1p, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_log2, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_neg, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_reciprocal, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_relu, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_round, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_rsqrt, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_sigmoid, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_sign, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_sin, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_sinh, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_sqrt, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_tan, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_tanh, test/functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_trunc, test/functorch/test_vmap.py::TestVmapOperators::test_unbind, test/functorch/test_vmap.py::TestVmapOperators::test_unfold, test/functorch/test_vmap.py::TestVmapOperators::test_unsafe_view, test/functorch/test_vmap.py::TestVmapOperators::test_unsqueeze, test/functorch/test_vmap.py::TestVmapOperators::test_view, test/functorch/test_vmap.py::TestVmapOperators::test_view_as, test/functorch/test_vmap.py::TestVmapOperators::test_view_as_complex, test/functorch/test_vmap.py::TestVmapOperators::test_view_as_real, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_composition_in_dim_0_out_dim_0_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_composition_in_dim_0_out_dim_0_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_composition_in_dim_0_out_dim_1_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_composition_in_dim_0_out_dim_1_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_composition_in_dim_1_out_dim_0_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_composition_in_dim_1_out_dim_0_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_composition_in_dim_1_out_dim_1_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_composition_in_dim_1_out_dim_1_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_error_in_dim_0_out_dim_0_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_error_in_dim_0_out_dim_0_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_error_in_dim_0_out_dim_1_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_error_in_dim_0_out_dim_1_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_error_in_dim_1_out_dim_0_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_error_in_dim_1_out_dim_0_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_error_in_dim_1_out_dim_1_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_error_in_dim_1_out_dim_1_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_0_out_dim_0_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_0_out_dim_0_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_0_out_dim_1_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_0_out_dim_1_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_0_out_dim_2_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_0_out_dim_2_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_1_out_dim_0_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_1_out_dim_0_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_1_out_dim_1_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_1_out_dim_1_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_1_out_dim_2_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_1_out_dim_2_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_2_out_dim_0_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_2_out_dim_0_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_2_out_dim_1_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_2_out_dim_1_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_2_out_dim_2_randomness_error, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_2_out_dim_2_randomness_same, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_fallback_check, test/functorch/test_vmap.py::TestVmapOperators::test_vmap_fallback_check_ok, test/functorch/test_vmap.py::TestVmapOperators::test_weird_matmul_case, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_0d_tensor_index_put_inplace_False_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_0d_tensor_index_put_inplace_True_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_advanced_indexing_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_batch_norm_training_False_track_running_stats_False_affine_False_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_batch_norm_training_False_track_running_stats_False_affine_True_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_batch_norm_training_False_track_running_stats_True_affine_False_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_batch_norm_training_False_track_running_stats_True_affine_True_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_batch_norm_training_True_track_running_stats_False_affine_False_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_batch_norm_training_True_track_running_stats_False_affine_True_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_batch_norm_training_True_track_running_stats_True_affine_False_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_batch_norm_training_True_track_running_stats_True_affine_True_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_conv_double_backward_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_fill__Tensor_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_flatten_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_foo_like_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_group_norm_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_index_fill_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_index_put_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_inplace_on_view_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_isinf_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_isnan_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_linalg_eigh_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_linalg_svd_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_namedtuple_returns_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_nested_advanced_indexing_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_CubeGenVmapAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ForwardHasDefaultArgsAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_H_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_MulGenVmapAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyCatCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyCubeAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyCubeCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyCubeNotComposableAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyExpMarkDirtyAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyMulAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyMulCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyMulScalarCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyNMSCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyNonzeroCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpySortAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpySortCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpySplitCopyCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpySplitCopyWithIntCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyTakeAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyTakeCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyViewCopyCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ScaleGradGenVmapAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_SelectAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_SelectGenVmapAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_SortGenVmapAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_T_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ZeroGradientsGenVmapAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___getitem___cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___getitem___functorch_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___radd___cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___rand___cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___rdiv___cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___rmatmul___cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___rmod___cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___rmul___cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___ror___cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___rpow___cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___rsub___cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___rxor___cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule__batch_norm_with_update_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule__chunk_cat_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule__native_batch_norm_legit_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule__segment_reduce_lengths_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule__segment_reduce_offsets_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule__softmax_backward_data_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule__unsafe_masked_index_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule__unsafe_masked_index_put_accumulate_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule__upsample_bilinear2d_aa_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_abs_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_acos_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_acosh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_add_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_addbmm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_addcdiv_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_addcmul_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_addmm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_addmm_decomposed_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_addmv_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_addr_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_alias_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_all_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_allclose_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_amax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_amin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_aminmax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_angle_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_any_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_arange_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_argmax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_argmin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_argsort_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_argwhere_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_as_strided_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_as_strided_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_as_strided_partial_views_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_as_strided_scatter_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_asin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_asinh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_atan2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_atan_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_atanh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_atleast_1d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_atleast_2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_atleast_3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_baddbmm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bernoulli_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bfloat16_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bfloat16_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bincount_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bitwise_and_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bitwise_left_shift_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bitwise_not_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bitwise_or_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bitwise_right_shift_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bitwise_xor_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_block_diag_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bmm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bool_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bool_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_broadcast_shapes_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_broadcast_tensors_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_broadcast_to_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bucketize_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_byte_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_byte_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cartesian_prod_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cat_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cauchy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cdist_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cdouble_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ceil_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cfloat_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_chalf_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_char_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_char_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cholesky_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cholesky_inverse_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cholesky_solve_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_chunk_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_clamp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_clamp_max_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_clamp_min_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_clone_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_column_stack_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_combinations_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_complex_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_conj_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_conj_physical_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_constant_pad_nd_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_contiguous_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_copysign_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_corrcoef_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cos_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cosh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_count_nonzero_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cov_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cross_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cummax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cummin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cumprod_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cumsum_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cumulative_trapezoid_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_deg2rad_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_diag_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_diag_embed_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_diagflat_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_diagonal_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_diagonal_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_diagonal_scatter_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_diff_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_digamma_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_dist_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_div_floor_rounding_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_div_no_rounding_mode_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_div_trunc_rounding_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_dot_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_double_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_double_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_dsplit_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_dstack_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_einsum_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_empty_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_empty_like_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_empty_permuted_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_empty_strided_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_eq_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_equal_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_erf_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_erfc_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_erfinv_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_exp2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_exp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_expand_as_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_expand_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_expand_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_expm1_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_exponential_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_eye_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_fft2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_fft_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_fftn_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_fftshift_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_hfft2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_hfft_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_hfftn_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_ifft2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_ifft_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_ifftn_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_ifftshift_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_ihfft2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_ihfft_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_ihfftn_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_irfft2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_irfft_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_irfftn_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_rfft2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_rfft_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_rfftn_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fill_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_flatten_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_flip_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fliplr_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_flipud_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_float_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_float_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_float_power_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_floor_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_floor_divide_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fmax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fmin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fmod_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_frac_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_frexp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_full_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_full_like_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_gather_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_gcd_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ge_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_geometric_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_geqrf_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_gradient_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_grid_sampler_2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_grid_sampler_3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_gt_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_half_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_half_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_hash_tensor_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_heaviside_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_histc_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_hsplit_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_hstack_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_hypot_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_i0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_igamma_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_igammac_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_imag_cuda_complex64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_add_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_fill_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_put_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_put_functorch_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_reduce_amax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_reduce_amin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_reduce_mean_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_reduce_prod_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_select_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_inner_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_int_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_int_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_isclose_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_isfinite_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_isin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_isinf_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_isnan_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_isneginf_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_isposinf_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_isreal_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_istft_cuda_complex64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_item_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_jiterator_2inputs_2outputs_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_jiterator_4inputs_with_extra_args_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_jiterator_binary_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_jiterator_binary_return_by_ref_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_jiterator_unary_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_kron_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_kthvalue_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_lcm_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ldexp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_le_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_lerp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_lgamma_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_cholesky_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_cholesky_ex_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_cond_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_cross_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_det_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_diagonal_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_eig_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_eigh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_eigvals_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_eigvalsh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_householder_product_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_inv_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_inv_ex_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_ldl_factor_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_ldl_factor_ex_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_ldl_solve_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_lstsq_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_lstsq_grad_oriented_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_lu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_lu_factor_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_lu_factor_ex_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_lu_solve_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_matrix_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_matrix_power_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_matrix_rank_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_matrix_rank_hermitian_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_multi_dot_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_norm_subgradients_at_zero_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_pinv_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_pinv_hermitian_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_pinv_singular_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_qr_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_slogdet_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_solve_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_solve_ex_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_solve_triangular_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_svd_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_svdvals_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_tensorinv_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_tensorsolve_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_vander_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_vecdot_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_vector_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linspace_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linspace_tensor_overload_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_log10_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_log1p_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_log2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_log_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_log_normal_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_log_softmax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_log_softmax_with_dtype_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logaddexp2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logaddexp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logcumsumexp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logdet_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logical_and_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logical_not_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logical_or_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logical_xor_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logit_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logspace_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logspace_tensor_overload_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logsumexp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_long_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_long_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_lt_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_lu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_lu_solve_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_lu_unpack_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mH_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mT_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_amax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_amin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_argmax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_argmin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_cumprod_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_cumsum_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_fill_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_fill_functorch_Scalar_only_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_log_softmax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_logaddexp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_logsumexp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_mean_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_median_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_normalize_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_prod_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_scatter_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_select_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_softmax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_softmin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_std_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_sum_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_var_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_matmul_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_matrix_exp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_max_binary_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_max_pool2d_with_indices_backward_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_max_reduction_no_dim_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_max_reduction_with_dim_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_maximum_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mean_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_median_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_meshgrid_list_of_tensors_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_meshgrid_variadic_tensors_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_min_binary_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_min_reduction_no_dim_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_min_reduction_with_dim_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_minimum_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mode_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_movedim_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_msort_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mul_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_multinomial_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mv_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mvlgamma_mvlgamma_p_1_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mvlgamma_mvlgamma_p_3_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mvlgamma_mvlgamma_p_5_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nan_to_num_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nanmean_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nanmedian_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nanquantile_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nansum_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_narrow_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_narrow_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_native_batch_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_native_dropout_backward_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_native_layer_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ne_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_neg_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_new_empty_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_new_empty_strided_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_new_full_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_new_ones_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_new_zeros_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nextafter_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_adaptive_avg_pool1d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_adaptive_avg_pool2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_adaptive_avg_pool3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_adaptive_max_pool1d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_adaptive_max_pool2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_adaptive_max_pool3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_alpha_dropout_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_avg_pool1d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_avg_pool2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_avg_pool3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_batch_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_batch_norm_without_cudnn_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_bilinear_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_binary_cross_entropy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_binary_cross_entropy_with_logits_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_celu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_channel_shuffle_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv1d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_no_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_stride_depthwise_with_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_stride_groups_with_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_stride_no_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_stride_padding_no_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_stride_padding_with_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_stride_with_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_strided_padding_dilation_no_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_strided_padding_dilation_with_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_with_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv_transpose1d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv_transpose2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv_transpose3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_cosine_embedding_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_cosine_similarity_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_cross_entropy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_ctc_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_dropout2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_dropout3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_dropout_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_elu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_embedding_bag_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_embedding_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_embedding_functorch_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_feature_alpha_dropout_with_train_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_fractional_max_pool2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_fractional_max_pool3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_gaussian_nll_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_gelu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_glu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_grid_sample_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_group_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_hardshrink_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_hardsigmoid_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_hardswish_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_hardtanh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_hinge_embedding_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_huber_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_instance_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_interpolate_area_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_interpolate_bicubic_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_interpolate_bilinear_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_interpolate_linear_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_interpolate_nearest-exact_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_interpolate_nearest_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_interpolate_trilinear_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_kl_div_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_l1_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_layer_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_leaky_relu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_linear_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_local_response_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_logsigmoid_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_margin_ranking_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_max_pool1d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_max_pool2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_max_pool3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_max_unpool1d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_max_unpool1d_grad_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_max_unpool2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_max_unpool2d_grad_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_max_unpool3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_max_unpool3d_grad_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_mish_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_mse_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_mse_loss_functorch_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_multi_head_attention_forward_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_multi_margin_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_multilabel_margin_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_multilabel_soft_margin_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_nll_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_normalize_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_one_hot_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_pad_circular_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_pad_constant_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_pad_reflect_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_pad_replicate_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_pad_replicate_negative_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_pairwise_distance_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_pdist_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_pixel_shuffle_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_pixel_unshuffle_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_poisson_nll_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_prelu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_relu6_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_relu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_rms_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_rrelu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_scaled_dot_product_attention_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_selu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_silu_complex_cuda_complex64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_silu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_smooth_l1_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_soft_margin_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_softmin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_softmin_with_dtype_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_softplus_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_softshrink_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_softsign_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_tanhshrink_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_threshold_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_triplet_margin_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_triplet_margin_with_distance_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_unfold_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_upsample_bilinear_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_upsample_nearest_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nonzero_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nonzero_static_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_norm_fro_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_norm_inf_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_norm_nuc_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_normal_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_normal_in_place_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_normal_number_mean_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ones_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ones_like_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ops_aten__new_zeros_with_same_feature_meta_functorchonly_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ops_aten_index_put_functorch_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ormqr_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_outer_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_pca_lowrank_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_permute_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_permute_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_pinverse_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_polar_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_polygamma_polygamma_n_0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_polygamma_polygamma_n_1_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_polygamma_polygamma_n_2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_polygamma_polygamma_n_3_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_polygamma_polygamma_n_4_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_positive_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_pow_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_prod_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_put_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_qr_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_quantile_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_rad2deg_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_rand_like_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_randint_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_randint_like_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_randn_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_randn_like_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ravel_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_real_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_reciprocal_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_remainder_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_renorm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_repeat_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_repeat_interleave_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_reshape_as_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_reshape_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_resize__cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_resize_as__cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_resolve_conj_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_resolve_neg_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_roll_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_rot90_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_round_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_round_decimals_0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_round_decimals_3_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_round_decimals_neg_3_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_rsqrt_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_rsub_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_scalar_tensor_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_scatter_add_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_scatter_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_scatter_reduce_amax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_scatter_reduce_amin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_scatter_reduce_mean_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_scatter_reduce_prod_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_scatter_reduce_sum_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_searchsorted_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_select_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_select_scatter_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sgn_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_short_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_short_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sigmoid_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sign_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_bartlett_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_blackman_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_cosine_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_exponential_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_gaussian_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_general_cosine_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_general_hamming_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_hamming_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_hann_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_kaiser_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_nuttall_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signbit_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sinc_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sinh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_slice_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_slice_scatter_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_softmax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_softmax_with_dtype_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sort_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sparse_mm_reduce_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sparse_sampled_addmm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_airy_ai_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_bessel_j0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_bessel_j1_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_bessel_y0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_bessel_y1_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_chebyshev_polynomial_t_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_chebyshev_polynomial_u_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_chebyshev_polynomial_v_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_chebyshev_polynomial_w_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_entr_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_erfcx_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_hermite_polynomial_h_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_hermite_polynomial_he_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_i0e_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_i1_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_i1e_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_laguerre_polynomial_l_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_legendre_polynomial_p_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_log_ndtr_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_modified_bessel_i0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_modified_bessel_i1_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_modified_bessel_k0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_modified_bessel_k1_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_ndtr_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_ndtri_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_polygamma_special_polygamma_n_0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_scaled_modified_bessel_k0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_scaled_modified_bessel_k1_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_shifted_chebyshev_polynomial_t_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_shifted_chebyshev_polynomial_u_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_shifted_chebyshev_polynomial_v_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_shifted_chebyshev_polynomial_w_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_spherical_bessel_j0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_xlog1py_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_zeta_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_split_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_split_list_args_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_split_with_sizes_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_split_with_sizes_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sqrt_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_square_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_squeeze_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_squeeze_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_squeeze_multiple_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_stack_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_std_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_std_mean_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_std_mean_unbiased_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_std_unbiased_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_stft_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sub_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sum_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sum_to_size_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_svd_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_svd_lowrank_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_t_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_t_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_take_along_dim_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_take_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_tan_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_tanh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_tensor_split_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_tensordot_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_tile_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_to_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_to_sparse_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_topk_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_torch__scaled_mm_cuda_float8_e4m3fn, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_torch__scaled_mm_v2_cuda_float8_e4m3fn, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_torch_ops_aten__efficient_attention_forward_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_torch_ops_aten__flash_attention_forward_cuda_float16, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_torch_ops_aten__safe_softmax_default_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_trace_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_transpose_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_transpose_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_trapezoid_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_trapz_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_triangular_solve_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_tril_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_tril_indices_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_triu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_triu_indices_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_true_divide_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_trunc_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unbind_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unbind_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unflatten_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unfold_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unfold_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_uniform_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unique_consecutive_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unique_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unravel_index_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unsafe_chunk_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unsafe_split_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unsqueeze_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unsqueeze_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_var_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_var_mean_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_var_mean_unbiased_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_var_unbiased_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_vdot_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_view_as_complex_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_view_as_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_view_as_real_cuda_complex64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_view_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_view_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_vsplit_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_vstack_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_where_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_xlogy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_zero__cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_zeros_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_zeros_like_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_searchsorted_bucketize_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_slogdet_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_sum_scalar_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_torch_return_types_returns_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_escaped_error_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_CubeGenVmapAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ForwardHasDefaultArgsAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_H_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_MulGenVmapAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyCatCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyCubeAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyCubeCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyCubeNotComposableAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyExpMarkDirtyAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyMulAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyMulCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyMulScalarCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyNMSCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyNonzeroCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpySortAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpySortCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpySplitCopyCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpySplitCopyWithIntCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyTakeAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyTakeCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyViewCopyCustomOp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ScaleGradGenVmapAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_SelectAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_SelectGenVmapAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_SortGenVmapAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_T_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ZeroGradientsGenVmapAutogradFunction_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___getitem___cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___getitem___functorch_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___radd___cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___rand___cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___rdiv___cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___rmatmul___cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___rmod___cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___rmul___cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___ror___cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___rpow___cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___rsub___cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___rxor___cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive__batch_norm_with_update_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive__chunk_cat_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive__native_batch_norm_legit_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive__segment_reduce_lengths_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive__segment_reduce_offsets_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive__softmax_backward_data_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive__unsafe_masked_index_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive__unsafe_masked_index_put_accumulate_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive__upsample_bilinear2d_aa_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_abs_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_acos_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_acosh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_add_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_addbmm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_addcdiv_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_addcmul_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_addmm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_addmm_decomposed_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_addmv_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_addr_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_alias_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_all_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_allclose_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_amax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_amin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_aminmax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_angle_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_any_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_arange_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_argmax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_argmin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_argsort_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_argwhere_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_as_strided_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_as_strided_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_as_strided_partial_views_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_as_strided_scatter_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_asin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_asinh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_atan2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_atan_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_atanh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_atleast_1d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_atleast_2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_atleast_3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_baddbmm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bernoulli_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bfloat16_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bfloat16_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bincount_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bitwise_and_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bitwise_left_shift_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bitwise_not_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bitwise_or_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bitwise_right_shift_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bitwise_xor_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_block_diag_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bmm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bool_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bool_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_broadcast_shapes_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_broadcast_tensors_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_broadcast_to_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bucketize_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_byte_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_byte_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cartesian_prod_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cat_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cauchy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cdist_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cdouble_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ceil_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cfloat_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_chalf_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_char_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_char_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cholesky_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cholesky_inverse_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cholesky_solve_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_chunk_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_clamp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_clamp_max_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_clamp_min_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_clone_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_column_stack_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_combinations_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_complex_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_conj_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_conj_physical_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_constant_pad_nd_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_contiguous_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_copysign_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_corrcoef_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cos_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cosh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_count_nonzero_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cov_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cross_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cummax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cummin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cumprod_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cumsum_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cumulative_trapezoid_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_deg2rad_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_diag_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_diag_embed_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_diagflat_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_diagonal_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_diagonal_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_diagonal_scatter_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_diff_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_digamma_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_dist_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_div_floor_rounding_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_div_no_rounding_mode_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_div_trunc_rounding_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_dot_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_double_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_double_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_dsplit_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_dstack_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_einsum_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_empty_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_empty_like_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_empty_permuted_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_empty_strided_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_eq_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_equal_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_erf_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_erfc_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_erfinv_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_exp2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_exp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_expand_as_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_expand_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_expand_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_expm1_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_exponential_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_eye_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_fft2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_fft_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_fftn_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_fftshift_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_hfft2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_hfft_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_hfftn_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_ifft2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_ifft_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_ifftn_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_ifftshift_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_ihfft2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_ihfft_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_ihfftn_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_irfft2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_irfft_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_irfftn_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_rfft2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_rfft_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_rfftn_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fill_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_flatten_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_flip_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fliplr_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_flipud_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_float_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_float_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_float_power_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_floor_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_floor_divide_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fmax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fmin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fmod_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_frac_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_frexp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_full_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_full_like_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_gather_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_gcd_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ge_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_geometric_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_geqrf_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_gradient_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_grid_sampler_2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_grid_sampler_3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_gt_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_half_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_half_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_hash_tensor_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_heaviside_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_histc_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_hsplit_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_hstack_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_hypot_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_i0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_igamma_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_igammac_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_imag_cuda_complex64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_add_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_fill_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_put_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_put_functorch_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_reduce_amax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_reduce_amin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_reduce_mean_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_reduce_prod_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_select_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_inner_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_int_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_int_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_isclose_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_isfinite_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_isin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_isinf_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_isnan_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_isneginf_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_isposinf_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_isreal_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_istft_cuda_complex64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_item_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_jiterator_2inputs_2outputs_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_jiterator_4inputs_with_extra_args_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_jiterator_binary_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_jiterator_binary_return_by_ref_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_jiterator_unary_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_kron_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_kthvalue_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_lcm_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ldexp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_le_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_lerp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_lgamma_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_cholesky_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_cholesky_ex_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_cond_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_cross_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_det_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_diagonal_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_eig_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_eigh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_eigvals_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_eigvalsh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_householder_product_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_inv_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_inv_ex_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_ldl_factor_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_ldl_factor_ex_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_ldl_solve_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_lstsq_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_lstsq_grad_oriented_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_lu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_lu_factor_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_lu_factor_ex_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_lu_solve_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_matrix_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_matrix_power_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_matrix_rank_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_matrix_rank_hermitian_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_multi_dot_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_norm_subgradients_at_zero_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_pinv_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_pinv_hermitian_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_pinv_singular_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_qr_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_slogdet_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_solve_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_solve_ex_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_solve_triangular_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_svd_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_svdvals_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_tensorinv_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_tensorsolve_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_vander_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_vecdot_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_vector_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linspace_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linspace_tensor_overload_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_log10_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_log1p_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_log2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_log_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_log_normal_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_log_softmax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_log_softmax_with_dtype_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logaddexp2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logaddexp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logcumsumexp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logdet_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logical_and_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logical_not_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logical_or_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logical_xor_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logit_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logspace_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logspace_tensor_overload_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logsumexp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_long_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_long_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_lt_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_lu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_lu_solve_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_lu_unpack_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mH_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mT_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_amax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_amin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_argmax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_argmin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_cumprod_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_cumsum_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_fill_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_fill_functorch_Scalar_only_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_log_softmax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_logaddexp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_logsumexp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_mean_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_median_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_normalize_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_prod_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_scatter_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_select_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_softmax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_softmin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_std_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_sum_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_var_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_matmul_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_matrix_exp_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_max_binary_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_max_pool2d_with_indices_backward_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_max_reduction_no_dim_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_max_reduction_with_dim_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_maximum_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mean_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_median_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_meshgrid_list_of_tensors_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_meshgrid_variadic_tensors_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_min_binary_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_min_reduction_no_dim_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_min_reduction_with_dim_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_minimum_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mode_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_movedim_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_msort_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mul_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_multinomial_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mv_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mvlgamma_mvlgamma_p_1_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mvlgamma_mvlgamma_p_3_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mvlgamma_mvlgamma_p_5_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nan_to_num_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nanmean_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nanmedian_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nanquantile_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nansum_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_narrow_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_narrow_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_native_batch_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_native_dropout_backward_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_native_layer_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ne_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_neg_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_new_empty_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_new_empty_strided_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_new_full_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_new_ones_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_new_zeros_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nextafter_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_adaptive_avg_pool1d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_adaptive_avg_pool2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_adaptive_avg_pool3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_adaptive_max_pool1d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_adaptive_max_pool2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_adaptive_max_pool3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_alpha_dropout_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_avg_pool1d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_avg_pool2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_avg_pool3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_batch_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_batch_norm_without_cudnn_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_bilinear_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_binary_cross_entropy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_binary_cross_entropy_with_logits_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_celu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_channel_shuffle_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv1d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_no_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_stride_depthwise_with_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_stride_groups_with_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_stride_no_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_stride_padding_no_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_stride_padding_with_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_stride_with_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_strided_padding_dilation_no_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_strided_padding_dilation_with_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_with_bias_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv_transpose1d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv_transpose2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv_transpose3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_cosine_embedding_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_cosine_similarity_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_cross_entropy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_ctc_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_dropout2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_dropout3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_dropout_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_elu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_embedding_bag_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_embedding_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_embedding_functorch_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_feature_alpha_dropout_with_train_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_fractional_max_pool2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_fractional_max_pool3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_gaussian_nll_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_gelu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_glu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_grid_sample_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_group_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_hardshrink_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_hardsigmoid_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_hardswish_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_hardtanh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_hinge_embedding_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_huber_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_instance_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_interpolate_area_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_interpolate_bicubic_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_interpolate_bilinear_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_interpolate_linear_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_interpolate_nearest-exact_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_interpolate_nearest_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_interpolate_trilinear_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_kl_div_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_l1_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_layer_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_leaky_relu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_linear_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_local_response_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_logsigmoid_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_margin_ranking_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_max_pool1d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_max_pool2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_max_pool3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_max_unpool1d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_max_unpool1d_grad_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_max_unpool2d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_max_unpool2d_grad_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_max_unpool3d_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_max_unpool3d_grad_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_mish_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_mse_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_mse_loss_functorch_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_multi_head_attention_forward_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_multi_margin_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_multilabel_margin_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_multilabel_soft_margin_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_nll_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_normalize_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_one_hot_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_pad_circular_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_pad_constant_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_pad_reflect_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_pad_replicate_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_pad_replicate_negative_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_pairwise_distance_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_pdist_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_pixel_shuffle_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_pixel_unshuffle_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_poisson_nll_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_prelu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_relu6_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_relu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_rms_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_rrelu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_scaled_dot_product_attention_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_selu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_silu_complex_cuda_complex64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_silu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_smooth_l1_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_soft_margin_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_softmin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_softmin_with_dtype_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_softplus_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_softshrink_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_softsign_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_tanhshrink_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_threshold_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_triplet_margin_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_triplet_margin_with_distance_loss_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_unfold_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_upsample_bilinear_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_upsample_nearest_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nonzero_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nonzero_static_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_norm_fro_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_norm_inf_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_norm_nuc_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_normal_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_normal_in_place_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_normal_number_mean_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ones_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ones_like_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ops_aten__new_zeros_with_same_feature_meta_functorchonly_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ops_aten_index_put_functorch_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ormqr_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_outer_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_pca_lowrank_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_permute_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_permute_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_pinverse_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_polar_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_polygamma_polygamma_n_0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_polygamma_polygamma_n_1_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_polygamma_polygamma_n_2_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_polygamma_polygamma_n_3_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_polygamma_polygamma_n_4_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_positive_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_pow_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_prod_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_put_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_qr_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_quantile_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_rad2deg_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_rand_like_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_randint_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_randint_like_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_randn_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_randn_like_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ravel_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_real_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_reciprocal_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_remainder_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_renorm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_repeat_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_repeat_interleave_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_reshape_as_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_reshape_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_resize__cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_resize_as__cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_resolve_conj_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_resolve_neg_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_roll_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_rot90_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_round_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_round_decimals_0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_round_decimals_3_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_round_decimals_neg_3_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_rsqrt_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_rsub_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_scalar_tensor_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_scatter_add_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_scatter_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_scatter_reduce_amax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_scatter_reduce_amin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_scatter_reduce_mean_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_scatter_reduce_prod_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_scatter_reduce_sum_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_searchsorted_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_select_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_select_scatter_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sgn_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_short_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_short_functorch_no_channels_last_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sigmoid_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sign_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_bartlett_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_blackman_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_cosine_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_exponential_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_gaussian_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_general_cosine_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_general_hamming_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_hamming_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_hann_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_kaiser_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_nuttall_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signbit_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sin_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sinc_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sinh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_slice_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_slice_scatter_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_softmax_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_softmax_with_dtype_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sort_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sparse_mm_reduce_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sparse_sampled_addmm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_airy_ai_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_bessel_j0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_bessel_j1_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_bessel_y0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_bessel_y1_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_chebyshev_polynomial_t_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_chebyshev_polynomial_u_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_chebyshev_polynomial_v_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_chebyshev_polynomial_w_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_entr_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_erfcx_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_hermite_polynomial_h_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_hermite_polynomial_he_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_i0e_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_i1_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_i1e_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_laguerre_polynomial_l_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_legendre_polynomial_p_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_log_ndtr_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_modified_bessel_i0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_modified_bessel_i1_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_modified_bessel_k0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_modified_bessel_k1_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_ndtr_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_ndtri_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_polygamma_special_polygamma_n_0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_scaled_modified_bessel_k0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_scaled_modified_bessel_k1_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_shifted_chebyshev_polynomial_t_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_shifted_chebyshev_polynomial_u_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_shifted_chebyshev_polynomial_v_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_shifted_chebyshev_polynomial_w_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_spherical_bessel_j0_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_xlog1py_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_zeta_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_split_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_split_list_args_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_split_with_sizes_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_split_with_sizes_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sqrt_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_square_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_squeeze_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_squeeze_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_squeeze_multiple_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_stack_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_std_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_std_mean_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_std_mean_unbiased_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_std_unbiased_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_stft_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sub_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sum_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sum_to_size_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_svd_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_svd_lowrank_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_t_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_t_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_take_along_dim_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_take_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_tan_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_tanh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_tensor_split_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_tensordot_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_tile_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_to_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_to_sparse_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_topk_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_torch__scaled_mm_cuda_float8_e4m3fn, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_torch__scaled_mm_v2_cuda_float8_e4m3fn, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_torch_ops_aten__efficient_attention_forward_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_torch_ops_aten__flash_attention_forward_cuda_float16, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_torch_ops_aten__safe_softmax_default_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_trace_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_transpose_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_transpose_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_trapezoid_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_trapz_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_triangular_solve_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_tril_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_tril_indices_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_triu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_triu_indices_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_true_divide_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_trunc_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unbind_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unbind_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unflatten_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unfold_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unfold_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_uniform_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unique_consecutive_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unique_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unravel_index_cuda_int64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unsafe_chunk_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unsafe_split_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unsqueeze_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unsqueeze_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_var_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_var_mean_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_var_mean_unbiased_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_var_unbiased_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_vdot_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_view_as_complex_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_view_as_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_view_as_real_cuda_complex64, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_view_copy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_view_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_vsplit_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_vstack_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_where_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_xlogy_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_zero__cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_zeros_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_zeros_like_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_cholesky_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_cholesky_ex_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_cond_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_cross_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_det_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_diagonal_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_eig_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_eigh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_eigvals_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_eigvalsh_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_householder_product_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_inv_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_inv_ex_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_ldl_factor_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_ldl_factor_ex_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_ldl_solve_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_lstsq_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_lstsq_grad_oriented_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_lu_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_lu_factor_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_lu_factor_ex_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_lu_solve_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_matrix_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_matrix_power_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_matrix_rank_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_matrix_rank_hermitian_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_multi_dot_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_norm_subgradients_at_zero_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_pinv_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_pinv_hermitian_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_pinv_singular_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_qr_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_slogdet_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_solve_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_solve_ex_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_solve_triangular_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_svd_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_svdvals_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_tensorinv_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_tensorsolve_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_vander_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_vecdot_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_vector_norm_cuda_float32, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_multi_dot_failure_1D_input_cuda, test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_with_anomaly_detection_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_add_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_binary_cross_entropy_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_diagonal_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_div_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_expand_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_index_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_inplace_manyview_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_inplace_view_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_lgamma_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_log1p_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_log_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_log_softmax_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_logsumexp_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_max_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_median_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_min_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_mul_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_permute_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_randomness_backend0_randomness_different_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_randomness_backend0_randomness_error_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_randomness_backend0_randomness_same_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_randomness_backend1_randomness_different_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_randomness_backend1_randomness_error_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_randomness_backend1_randomness_same_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_reshape_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_sdpa_backend0_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_sdpa_backend1_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_select_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_sigmoid_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_slice_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_stack_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_sub_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_threshold_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_trace_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_unrelated_output_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_unrelated_output_multiple_grad_cuda, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_vmap_fallback_check, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_vmap_fallback_check_ok, test/functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_where_cuda, test/functorch/test_vmap.py::TestTransformFailureCUDA::test_fails_with_autograd_function_transform_grad_and_value_cuda, test/functorch/test_vmap.py::TestTransformFailureCUDA::test_fails_with_autograd_function_transform_grad_cuda, test/functorch/test_vmap.py::TestTransformFailureCUDA::test_fails_with_autograd_function_transform_jacfwd_cuda, test/functorch/test_vmap.py::TestTransformFailureCUDA::test_fails_with_autograd_function_transform_jacrev_cuda, test/functorch/test_vmap.py::TestTransformFailureCUDA::test_fails_with_autograd_function_transform_jvp_cuda, test/functorch/test_vmap.py::TestTransformFailureCUDA::test_fails_with_autograd_function_transform_vjp_cuda, test/functorch/test_vmap.py::TestTransformFailureCUDA::test_fails_with_autograd_function_transform_vmap_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_alpha_dropout_randomness_different_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_alpha_dropout_randomness_different_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_alpha_dropout_randomness_different_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_alpha_dropout_randomness_error_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_alpha_dropout_randomness_error_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_alpha_dropout_randomness_error_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_alpha_dropout_randomness_same_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_alpha_dropout_randomness_same_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_alpha_dropout_randomness_same_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_different_batched_input_first_batched_probability_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_different_batched_input_first_batched_probability_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_different_batched_input_first_batched_probability_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_different_batched_input_last_batched_probability_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_different_batched_input_last_batched_probability_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_different_batched_input_last_batched_probability_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_different_batched_input_none_batched_probability_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_different_batched_input_none_batched_probability_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_different_batched_input_none_batched_probability_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_error_batched_input_first_batched_probability_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_error_batched_input_first_batched_probability_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_error_batched_input_first_batched_probability_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_error_batched_input_last_batched_probability_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_error_batched_input_last_batched_probability_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_error_batched_input_last_batched_probability_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_error_batched_input_none_batched_probability_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_error_batched_input_none_batched_probability_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_error_batched_input_none_batched_probability_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_same_batched_input_first_batched_probability_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_same_batched_input_first_batched_probability_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_same_batched_input_first_batched_probability_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_same_batched_input_last_batched_probability_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_same_batched_input_last_batched_probability_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_same_batched_input_last_batched_probability_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_same_batched_input_none_batched_probability_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_same_batched_input_none_batched_probability_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_same_batched_input_none_batched_probability_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_different_batched_input_first_batched_probability_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_different_batched_input_first_batched_probability_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_different_batched_input_first_batched_probability_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_different_batched_input_last_batched_probability_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_different_batched_input_last_batched_probability_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_different_batched_input_last_batched_probability_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_different_batched_input_none_batched_probability_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_different_batched_input_none_batched_probability_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_different_batched_input_none_batched_probability_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_error_batched_input_first_batched_probability_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_error_batched_input_first_batched_probability_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_error_batched_input_first_batched_probability_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_error_batched_input_last_batched_probability_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_error_batched_input_last_batched_probability_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_error_batched_input_last_batched_probability_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_error_batched_input_none_batched_probability_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_error_batched_input_none_batched_probability_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_error_batched_input_none_batched_probability_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_same_batched_input_first_batched_probability_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_same_batched_input_first_batched_probability_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_same_batched_input_first_batched_probability_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_same_batched_input_last_batched_probability_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_same_batched_input_last_batched_probability_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_same_batched_input_last_batched_probability_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_same_batched_input_none_batched_probability_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_same_batched_input_none_batched_probability_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_same_batched_input_none_batched_probability_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_chunk_vmap_in_dim_0_out_dim_0_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_chunk_vmap_in_dim_0_out_dim_1_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_chunk_vmap_in_dim_0_out_dim_2_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_chunk_vmap_in_dim_1_out_dim_0_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_chunk_vmap_in_dim_1_out_dim_1_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_chunk_vmap_in_dim_1_out_dim_2_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_chunk_vmap_in_dim_2_out_dim_0_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_chunk_vmap_in_dim_2_out_dim_1_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_chunk_vmap_in_dim_2_out_dim_2_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_randomness_different_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_randomness_different_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_randomness_different_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_randomness_error_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_randomness_error_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_randomness_error_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_randomness_same_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_randomness_same_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_randomness_same_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_unbatched_randomness_different_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_unbatched_randomness_error_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_unbatched_randomness_same_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_factory_ops_randomness_different_use_generator_False_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_factory_ops_randomness_different_use_generator_True_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_factory_ops_randomness_error_use_generator_False_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_factory_ops_randomness_error_use_generator_True_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_factory_ops_randomness_same_use_generator_False_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_factory_ops_randomness_same_use_generator_True_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_alpha_dropout_randomness_different_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_alpha_dropout_randomness_different_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_alpha_dropout_randomness_different_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_alpha_dropout_randomness_error_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_alpha_dropout_randomness_error_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_alpha_dropout_randomness_error_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_alpha_dropout_randomness_same_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_alpha_dropout_randomness_same_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_alpha_dropout_randomness_same_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_different_batched_input_first_dim_2_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_different_batched_input_first_dim_3_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_different_batched_input_last_dim_2_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_different_batched_input_last_dim_3_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_different_batched_input_none_dim_2_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_different_batched_input_none_dim_3_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_error_batched_input_first_dim_2_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_error_batched_input_first_dim_3_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_error_batched_input_last_dim_2_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_error_batched_input_last_dim_3_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_error_batched_input_none_dim_2_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_error_batched_input_none_dim_3_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_same_batched_input_first_dim_2_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_same_batched_input_first_dim_3_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_same_batched_input_last_dim_2_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_same_batched_input_last_dim_3_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_same_batched_input_none_dim_2_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_same_batched_input_none_dim_3_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_jacfwd_with_random_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_like_functions_randomness_different_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_like_functions_randomness_different_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_like_functions_randomness_different_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_like_functions_randomness_error_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_like_functions_randomness_error_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_like_functions_randomness_error_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_like_functions_randomness_same_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_like_functions_randomness_same_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_like_functions_randomness_same_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_different_batched_call_False_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_different_batched_call_False_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_different_batched_call_False_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_different_batched_call_True_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_different_batched_call_True_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_different_batched_call_True_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_error_batched_call_False_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_error_batched_call_False_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_error_batched_call_False_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_error_batched_call_True_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_error_batched_call_True_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_error_batched_call_True_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_same_batched_call_False_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_same_batched_call_False_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_same_batched_call_False_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_same_batched_call_True_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_same_batched_call_True_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_same_batched_call_True_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_different_batched_call_False_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_different_batched_call_False_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_different_batched_call_False_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_different_batched_call_True_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_different_batched_call_True_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_different_batched_call_True_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_error_batched_call_False_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_error_batched_call_False_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_error_batched_call_False_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_error_batched_call_True_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_error_batched_call_True_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_error_batched_call_True_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_same_batched_call_False_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_same_batched_call_False_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_same_batched_call_False_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_same_batched_call_True_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_same_batched_call_True_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_same_batched_call_True_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_different_batched_input_first_batched_other_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_different_batched_input_first_batched_other_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_different_batched_input_first_batched_other_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_different_batched_input_last_batched_other_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_different_batched_input_last_batched_other_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_different_batched_input_last_batched_other_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_different_batched_input_none_batched_other_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_different_batched_input_none_batched_other_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_different_batched_input_none_batched_other_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_error_batched_input_first_batched_other_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_error_batched_input_first_batched_other_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_error_batched_input_first_batched_other_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_error_batched_input_last_batched_other_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_error_batched_input_last_batched_other_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_error_batched_input_last_batched_other_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_error_batched_input_none_batched_other_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_error_batched_input_none_batched_other_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_error_batched_input_none_batched_other_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_same_batched_input_first_batched_other_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_same_batched_input_first_batched_other_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_same_batched_input_first_batched_other_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_same_batched_input_last_batched_other_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_same_batched_input_last_batched_other_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_same_batched_input_last_batched_other_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_same_batched_input_none_batched_other_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_same_batched_input_none_batched_other_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_same_batched_input_none_batched_other_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_different_batched_input_first_batched_other_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_different_batched_input_first_batched_other_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_different_batched_input_first_batched_other_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_different_batched_input_last_batched_other_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_different_batched_input_last_batched_other_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_different_batched_input_last_batched_other_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_different_batched_input_none_batched_other_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_different_batched_input_none_batched_other_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_different_batched_input_none_batched_other_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_error_batched_input_first_batched_other_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_error_batched_input_first_batched_other_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_error_batched_input_first_batched_other_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_error_batched_input_last_batched_other_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_error_batched_input_last_batched_other_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_error_batched_input_last_batched_other_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_error_batched_input_none_batched_other_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_error_batched_input_none_batched_other_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_error_batched_input_none_batched_other_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_same_batched_input_first_batched_other_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_same_batched_input_first_batched_other_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_same_batched_input_first_batched_other_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_same_batched_input_last_batched_other_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_same_batched_input_last_batched_other_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_same_batched_input_last_batched_other_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_same_batched_input_none_batched_other_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_same_batched_input_none_batched_other_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_same_batched_input_none_batched_other_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_False_randomness_different_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_False_randomness_different_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_False_randomness_different_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_False_randomness_error_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_False_randomness_error_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_False_randomness_error_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_False_randomness_same_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_False_randomness_same_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_False_randomness_same_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_True_randomness_different_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_True_randomness_different_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_True_randomness_different_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_True_randomness_error_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_True_randomness_error_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_True_randomness_error_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_True_randomness_same_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_True_randomness_same_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_True_randomness_same_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_False_randomness_different_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_False_randomness_different_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_False_randomness_different_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_False_randomness_error_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_False_randomness_error_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_False_randomness_error_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_False_randomness_same_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_False_randomness_same_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_False_randomness_same_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_True_randomness_different_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_True_randomness_different_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_True_randomness_different_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_True_randomness_error_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_True_randomness_error_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_True_randomness_error_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_True_randomness_same_batched_input_first_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_True_randomness_same_batched_input_last_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_True_randomness_same_batched_input_none_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_randperm_randomness_different_use_generator_False_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_randperm_randomness_different_use_generator_True_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_randperm_randomness_error_use_generator_False_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_randperm_randomness_error_use_generator_True_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_randperm_randomness_same_use_generator_False_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_randperm_randomness_same_use_generator_True_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_unsupported_random_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_vmap_chunksize_in_dim_0_out_dim_0_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_vmap_chunksize_in_dim_0_out_dim_1_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_vmap_chunksize_in_dim_0_out_dim_2_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_vmap_chunksize_in_dim_1_out_dim_0_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_vmap_chunksize_in_dim_1_out_dim_1_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_vmap_chunksize_in_dim_1_out_dim_2_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_vmap_chunksize_in_dim_2_out_dim_0_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_vmap_chunksize_in_dim_2_out_dim_1_cuda, test/functorch/test_vmap.py::TestRandomnessCUDA::test_vmap_chunksize_in_dim_2_out_dim_2_cuda, test/functorch/test_vmap.py::TestVmapDeviceTypeCUDA::test__is_all_true_cuda, test/functorch/test_vmap.py::TestVmapDeviceTypeCUDA::test__is_any_true_cuda, test/functorch/test_vmap.py::TestVmapDeviceTypeCUDA::test_check_tensor_cuda, test/functorch/test_vmap.py::TestVmapDeviceTypeCUDA::test_vmap_fallback_check, test/functorch/test_vmap.py::TestVmapDeviceTypeCUDA::test_vmap_fallback_check_ok, test/functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_cat_batching_rule_cuda, test/functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_fallback_binary_cuda, test/functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_fallback_binary_nt_and_batched_dense_cuda, test/functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_fallback_binary_nt_and_unbatched_dense_cuda, test/functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_fallback_unary_cuda, test/functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_fallback_with_nt_and_batched_dense_with_nonzero_bdim_raises_cuda, test/functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_multilevel_vmap_raises_cuda, test/functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_nt_acts_as_dense_in_vmap_cuda, test/functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_nt_with_nonzero_in_dim_raises_cuda, test/functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_nt_with_nonzero_out_dim_raises_cuda, test/functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_shape_call_cuda, test/functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_vmap_fallback_check, test/functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_vmap_fallback_check_ok 2025-12-04T15:25:36.4734865Z 2025-12-04T15:25:36.4735256Z functorch/test_vmap.py::TestVmapAPI::test_accepts_nested_inputs PASSED [0.0067s] [ 0%] 2025-12-04T15:25:36.4736508Z functorch/test_vmap.py::TestVmapAPI::test_backward_unsupported_interaction SKIPPED [0.0013s] (error: element 0 of tensors does not require grad and does not have a grad_fn) [ 0%] 2025-12-04T15:25:36.4737960Z functorch/test_vmap.py::TestVmapAPI::test_batch_rule_does_not_need_to_handle_no_batched_input PASSED [0.0019s] [ 0%] 2025-12-04T15:25:36.4738906Z functorch/test_vmap.py::TestVmapAPI::test_batched_gradient_basic PASSED [0.0019s] [ 0%] 2025-12-04T15:25:36.4739693Z functorch/test_vmap.py::TestVmapAPI::test_checkpoint PASSED [0.0185s] [ 0%] 2025-12-04T15:25:36.4740441Z functorch/test_vmap.py::TestVmapAPI::test_constant_function PASSED [0.0014s] [ 0%] 2025-12-04T15:25:36.4741184Z functorch/test_vmap.py::TestVmapAPI::test_data_attribute PASSED [0.0018s] [ 0%] 2025-12-04T15:25:36.4742015Z functorch/test_vmap.py::TestVmapAPI::test_data_dependent_control_flow_throws PASSED [0.0014s] [ 0%] 2025-12-04T15:25:36.4743724Z functorch/test_vmap.py::TestVmapAPI::test_decomposition_under_python_dispatcher PASSED [0.0021s] [ 0%] 2025-12-04T15:25:36.4744662Z functorch/test_vmap.py::TestVmapAPI::test_different_map_dim_size_raises PASSED [0.0015s] [ 0%] 2025-12-04T15:25:36.4745559Z functorch/test_vmap.py::TestVmapAPI::test_fallback_does_not_warn_by_default PASSED [0.0016s] [ 0%] 2025-12-04T15:25:36.4746413Z functorch/test_vmap.py::TestVmapAPI::test_fallback_masked_fill SKIPPED [0.0002s] [ 0%] 2025-12-04T15:25:36.4747233Z functorch/test_vmap.py::TestVmapAPI::test_fallback_multiple_returns PASSED [0.0073s] [ 0%] 2025-12-04T15:25:36.4748179Z functorch/test_vmap.py::TestVmapAPI::test_fallback_warning PASSED [0.0735s] [ 0%] 2025-12-04T15:25:36.4749107Z functorch/test_vmap.py::TestVmapAPI::test_fallback_warns_when_warnings_are_enabled XFAIL [0.0279s] [ 0%] 2025-12-04T15:25:36.4750028Z functorch/test_vmap.py::TestVmapAPI::test_fallback_with_undefined_grad PASSED [0.0012s] [ 0%] 2025-12-04T15:25:36.4750835Z functorch/test_vmap.py::TestVmapAPI::test_fallback_zero_dim PASSED [0.0025s] [ 0%] 2025-12-04T15:25:36.4751590Z functorch/test_vmap.py::TestVmapAPI::test_func_with_no_inputs PASSED [0.0012s] [ 0%] 2025-12-04T15:25:36.4752477Z functorch/test_vmap.py::TestVmapAPI::test_func_with_no_tensors PASSED [0.0012s] [ 0%] 2025-12-04T15:25:36.4753254Z functorch/test_vmap.py::TestVmapAPI::test_functools_partial PASSED [0.0015s] [ 0%] 2025-12-04T15:25:36.4754065Z functorch/test_vmap.py::TestVmapAPI::test_grad_unsupported_interaction XFAIL [0.0020s] [ 0%] 2025-12-04T15:25:36.4754910Z functorch/test_vmap.py::TestVmapAPI::test_in_dim_not_in_tensor_err_msg PASSED [0.0019s] [ 1%] 2025-12-04T15:25:36.4755748Z functorch/test_vmap.py::TestVmapAPI::test_in_dims_wrong_type_err_msg PASSED [0.0015s] [ 1%] 2025-12-04T15:25:36.4756650Z functorch/test_vmap.py::TestVmapAPI::test_inplace_fallback_nary_different_levels XFAIL [0.0073s] [ 1%] 2025-12-04T15:25:36.4757599Z functorch/test_vmap.py::TestVmapAPI::test_inplace_fallback_nary_same_levels PASSED [0.0056s] [ 1%] 2025-12-04T15:25:36.4758448Z functorch/test_vmap.py::TestVmapAPI::test_inplace_fallback_unary PASSED [0.0080s] [ 1%] 2025-12-04T15:25:36.4759349Z functorch/test_vmap.py::TestVmapAPI::test_integer_in_dim_but_not_tensor_input_err_msg PASSED [0.0016s] [ 1%] 2025-12-04T15:25:36.4760203Z functorch/test_vmap.py::TestVmapAPI::test_item_throws PASSED [0.0014s] [ 1%] 2025-12-04T15:25:36.4760925Z functorch/test_vmap.py::TestVmapAPI::test_multiple_inputs PASSED [0.0014s] [ 1%] 2025-12-04T15:25:36.4761676Z functorch/test_vmap.py::TestVmapAPI::test_multiple_out_dims PASSED [0.0024s] [ 1%] 2025-12-04T15:25:36.4762436Z functorch/test_vmap.py::TestVmapAPI::test_multiple_outputs PASSED [0.0016s] [ 1%] 2025-12-04T15:25:36.4763200Z functorch/test_vmap.py::TestVmapAPI::test_multiple_outputs2 PASSED [0.0014s] [ 1%] 2025-12-04T15:25:36.4763983Z functorch/test_vmap.py::TestVmapAPI::test_nested_negative_in_dims PASSED [0.0015s] [ 1%] 2025-12-04T15:25:36.4764813Z functorch/test_vmap.py::TestVmapAPI::test_nested_non_default_in_dims PASSED [0.0017s] [ 1%] 2025-12-04T15:25:36.4765600Z functorch/test_vmap.py::TestVmapAPI::test_nested_out_dims PASSED [0.0035s] [ 1%] 2025-12-04T15:25:36.4766376Z functorch/test_vmap.py::TestVmapAPI::test_nested_with_diag_embed PASSED [0.0017s] [ 1%] 2025-12-04T15:25:36.4767206Z functorch/test_vmap.py::TestVmapAPI::test_nested_with_different_map_dim PASSED [0.0023s] [ 1%] 2025-12-04T15:25:36.4768050Z functorch/test_vmap.py::TestVmapAPI::test_nested_with_same_map_dim PASSED [0.0019s] [ 1%] 2025-12-04T15:25:36.4768817Z functorch/test_vmap.py::TestVmapAPI::test_nn_module PASSED [0.0018s] [ 1%] 2025-12-04T15:25:36.4769589Z functorch/test_vmap.py::TestVmapAPI::test_non_default_in_dims_out_dims PASSED [0.0031s] [ 1%] 2025-12-04T15:25:36.4770433Z functorch/test_vmap.py::TestVmapAPI::test_non_tensor_output_raises PASSED [0.0016s] [ 1%] 2025-12-04T15:25:36.4771214Z functorch/test_vmap.py::TestVmapAPI::test_non_zero_in_dims PASSED [0.0023s] [ 1%] 2025-12-04T15:25:36.4771944Z functorch/test_vmap.py::TestVmapAPI::test_none_in_dims PASSED [0.0019s] [ 2%] 2025-12-04T15:25:36.4772662Z functorch/test_vmap.py::TestVmapAPI::test_nonzero_out_dims PASSED [0.0051s] [ 2%] 2025-12-04T15:25:36.4773425Z functorch/test_vmap.py::TestVmapAPI::test_noop_in_inner_vmap PASSED [0.0015s] [ 2%] 2025-12-04T15:25:36.4774227Z functorch/test_vmap.py::TestVmapAPI::test_not_enough_in_dims_err_msg PASSED [0.0017s] [ 2%] 2025-12-04T15:25:36.4775070Z functorch/test_vmap.py::TestVmapAPI::test_out_dim_out_of_bounds_err_msg PASSED [0.0037s] [ 2%] 2025-12-04T15:25:36.4776016Z functorch/test_vmap.py::TestVmapAPI::test_out_dims_and_num_outputs_mismatch_err_msg PASSED [0.0019s] [ 2%] 2025-12-04T15:25:36.4776962Z functorch/test_vmap.py::TestVmapAPI::test_out_dims_edge_case PASSED [0.0015s] [ 2%] 2025-12-04T15:25:36.4777911Z functorch/test_vmap.py::TestVmapAPI::test_out_dims_must_be_int_or_collection_of_int_err_msg PASSED [0.0013s] [ 2%] 2025-12-04T15:25:36.4778777Z functorch/test_vmap.py::TestVmapAPI::test_out_dims_none PASSED [0.0017s] [ 2%] 2025-12-04T15:25:36.4779523Z functorch/test_vmap.py::TestVmapAPI::test_out_dims_none_tuple PASSED [0.0019s] [ 2%] 2025-12-04T15:25:36.4780386Z functorch/test_vmap.py::TestVmapAPI::test_out_dims_normal_tensor PASSED [0.0018s] [ 2%] 2025-12-04T15:25:36.4781182Z functorch/test_vmap.py::TestVmapAPI::test_pytree_odict_returns PASSED [0.0017s] [ 2%] 2025-12-04T15:25:36.4781930Z functorch/test_vmap.py::TestVmapAPI::test_pytree_returns PASSED [0.0022s] [ 2%] 2025-12-04T15:25:36.4782745Z functorch/test_vmap.py::TestVmapAPI::test_pytree_returns_broadcast_nested PASSED [0.0018s] [ 2%] 2025-12-04T15:25:36.4783645Z functorch/test_vmap.py::TestVmapAPI::test_pytree_returns_broadcast_simple PASSED [0.0018s] [ 2%] 2025-12-04T15:25:36.4784496Z functorch/test_vmap.py::TestVmapAPI::test_pytree_returns_outdims PASSED [0.0019s] [ 2%] 2025-12-04T15:25:36.4785263Z functorch/test_vmap.py::TestVmapAPI::test_reshape_dim_into PASSED [0.0023s] [ 2%] 2025-12-04T15:25:36.4786017Z functorch/test_vmap.py::TestVmapAPI::test_reshape_dim_outof PASSED [0.0020s] [ 2%] 2025-12-04T15:25:36.4786840Z functorch/test_vmap.py::TestVmapAPI::test_restore_vmap_no_vmapped_inputs PASSED [0.0016s] [ 2%] 2025-12-04T15:25:36.4787732Z functorch/test_vmap.py::TestVmapAPI::test_restore_vmap_pytree_input_output PASSED [0.0021s] [ 2%] 2025-12-04T15:25:36.4788640Z functorch/test_vmap.py::TestVmapAPI::test_restore_vmap_unexpanded_outputs PASSED [0.0016s] [ 2%] 2025-12-04T15:25:36.4789450Z functorch/test_vmap.py::TestVmapAPI::test_single_input PASSED [0.0014s] [ 2%] 2025-12-04T15:25:36.4790215Z functorch/test_vmap.py::TestVmapAPI::test_unsupported_op_err_msg PASSED [0.0019s] [ 3%] 2025-12-04T15:25:36.4791210Z functorch/test_vmap.py::TestVmapAPI::test_vmap_autocast_cpu SKIPPED [0.0002s] (Somehow, vmap and autocast do not work on CPU) [ 3%] 2025-12-04T15:25:36.4792211Z functorch/test_vmap.py::TestVmapAPI::test_vmap_autocast_cuda PASSED [0.9263s] [ 3%] 2025-12-04T15:25:36.4792966Z functorch/test_vmap.py::TestVmapOperators::test_T_numpy PASSED [0.0084s] [ 3%] 2025-12-04T15:25:36.4793764Z functorch/test_vmap.py::TestVmapOperators::test_adaptive_avg_pool2d PASSED [0.0039s] [ 3%] 2025-12-04T15:25:36.4794571Z functorch/test_vmap.py::TestVmapOperators::test_argmax_dim PASSED [0.0078s] [ 3%] 2025-12-04T15:25:36.4795353Z functorch/test_vmap.py::TestVmapOperators::test_arithmetic_add PASSED [0.0233s] [ 3%] 2025-12-04T15:25:36.4796351Z functorch/test_vmap.py::TestVmapOperators::test_arithmetic_add_dunder PASSED [0.0231s] [ 3%] 2025-12-04T15:25:36.4797183Z functorch/test_vmap.py::TestVmapOperators::test_arithmetic_div PASSED [0.0229s] [ 3%] 2025-12-04T15:25:36.4798019Z functorch/test_vmap.py::TestVmapOperators::test_arithmetic_div_dunder PASSED [0.0238s] [ 3%] 2025-12-04T15:25:36.4798859Z functorch/test_vmap.py::TestVmapOperators::test_arithmetic_mul PASSED [0.0227s] [ 3%] 2025-12-04T15:25:36.4799692Z functorch/test_vmap.py::TestVmapOperators::test_arithmetic_mul_dunder PASSED [0.0228s] [ 3%] 2025-12-04T15:25:36.4800517Z functorch/test_vmap.py::TestVmapOperators::test_arithmetic_pow PASSED [0.0248s] [ 3%] 2025-12-04T15:25:36.4801353Z functorch/test_vmap.py::TestVmapOperators::test_arithmetic_pow_dunder PASSED [0.0250s] [ 3%] 2025-12-04T15:25:36.4802193Z functorch/test_vmap.py::TestVmapOperators::test_arithmetic_sub PASSED [0.0225s] [ 3%] 2025-12-04T15:25:36.4803031Z functorch/test_vmap.py::TestVmapOperators::test_arithmetic_sub_dunder PASSED [0.0227s] [ 3%] 2025-12-04T15:25:36.4803836Z functorch/test_vmap.py::TestVmapOperators::test_as_strided PASSED [0.0402s] [ 3%] 2025-12-04T15:25:36.4804657Z functorch/test_vmap.py::TestVmapOperators::test_bmm PASSED [0.0104s] [ 3%] 2025-12-04T15:25:36.4805392Z functorch/test_vmap.py::TestVmapOperators::test_cat PASSED [0.0095s] [ 3%] 2025-12-04T15:25:36.4806159Z functorch/test_vmap.py::TestVmapOperators::test_chunk PASSED [0.0200s] [ 3%] 2025-12-04T15:25:36.4807065Z functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_0_out_dim_0_randomness_error PASSED [0.0043s] [ 3%] 2025-12-04T15:25:36.4808138Z functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_0_out_dim_0_randomness_same PASSED [0.0051s] [ 4%] 2025-12-04T15:25:36.4809328Z functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_0_out_dim_1_randomness_error PASSED [0.0045s] [ 4%] 2025-12-04T15:25:36.4810389Z functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_0_out_dim_1_randomness_same PASSED [0.0054s] [ 4%] 2025-12-04T15:25:36.4811467Z functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_0_out_dim_2_randomness_error PASSED [0.0044s] [ 4%] 2025-12-04T15:25:36.4812539Z functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_0_out_dim_2_randomness_same PASSED [0.0053s] [ 4%] 2025-12-04T15:25:36.4813609Z functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_1_out_dim_0_randomness_error PASSED [0.0047s] [ 4%] 2025-12-04T15:25:36.4814668Z functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_1_out_dim_0_randomness_same PASSED [0.0054s] [ 4%] 2025-12-04T15:25:36.4815732Z functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_1_out_dim_1_randomness_error PASSED [0.0045s] [ 4%] 2025-12-04T15:25:36.4816862Z functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_1_out_dim_1_randomness_same PASSED [0.0056s] [ 4%] 2025-12-04T15:25:36.4817944Z functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_1_out_dim_2_randomness_error PASSED [0.0045s] [ 4%] 2025-12-04T15:25:36.4819011Z functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_1_out_dim_2_randomness_same PASSED [0.0057s] [ 4%] 2025-12-04T15:25:36.4820089Z functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_2_out_dim_0_randomness_error PASSED [0.0048s] [ 4%] 2025-12-04T15:25:36.4821159Z functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_2_out_dim_0_randomness_same PASSED [0.0056s] [ 4%] 2025-12-04T15:25:36.4822231Z functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_2_out_dim_1_randomness_error PASSED [0.0050s] [ 4%] 2025-12-04T15:25:36.4823287Z functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_2_out_dim_1_randomness_same PASSED [0.0060s] [ 4%] 2025-12-04T15:25:36.4824361Z functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_2_out_dim_2_randomness_error PASSED [0.0043s] [ 4%] 2025-12-04T15:25:36.4825427Z functorch/test_vmap.py::TestVmapOperators::test_chunk_vmap_in_dim_2_out_dim_2_randomness_same PASSED [0.0059s] [ 4%] 2025-12-04T15:25:36.4826347Z functorch/test_vmap.py::TestVmapOperators::test_clamp PASSED [0.0295s] [ 4%] 2025-12-04T15:25:36.4827212Z functorch/test_vmap.py::TestVmapOperators::test_clamp_inplace_variant_clamp_max_ PASSED [0.0093s] [ 4%] 2025-12-04T15:25:36.4828185Z functorch/test_vmap.py::TestVmapOperators::test_clamp_inplace_variant_clamp_min_ PASSED [0.0094s] [ 4%] 2025-12-04T15:25:36.4829134Z functorch/test_vmap.py::TestVmapOperators::test_clamp_variant_clamp_max PASSED [0.0132s] [ 4%] 2025-12-04T15:25:36.4830039Z functorch/test_vmap.py::TestVmapOperators::test_clamp_variant_clamp_min PASSED [0.0131s] [ 5%] 2025-12-04T15:25:36.4830856Z functorch/test_vmap.py::TestVmapOperators::test_clone PASSED [0.0187s] [ 5%] 2025-12-04T15:25:36.4831614Z functorch/test_vmap.py::TestVmapOperators::test_comparison_ops PASSED [0.1022s] [ 5%] 2025-12-04T15:25:36.4832388Z functorch/test_vmap.py::TestVmapOperators::test_conj PASSED [0.0124s] [ 5%] 2025-12-04T15:25:36.4833126Z functorch/test_vmap.py::TestVmapOperators::test_conj_bit PASSED [0.0015s] [ 5%] 2025-12-04T15:25:36.4833935Z functorch/test_vmap.py::TestVmapOperators::test_contiguous PASSED [0.0072s] [ 5%] 2025-12-04T15:25:36.4834672Z functorch/test_vmap.py::TestVmapOperators::test_conv2d PASSED [0.2128s] [ 5%] 2025-12-04T15:25:36.4835407Z functorch/test_vmap.py::TestVmapOperators::test_copy_ PASSED [0.0023s] [ 5%] 2025-12-04T15:25:36.4836245Z functorch/test_vmap.py::TestVmapOperators::test_cross_batch_size_three PASSED [0.0029s] [ 5%] 2025-12-04T15:25:36.4837054Z functorch/test_vmap.py::TestVmapOperators::test_diagonal PASSED [0.0065s] [ 5%] 2025-12-04T15:25:36.4837794Z functorch/test_vmap.py::TestVmapOperators::test_dot PASSED [0.0099s] [ 5%] 2025-12-04T15:25:36.4838595Z functorch/test_vmap.py::TestVmapOperators::test_expand_as PASSED [0.0094s] [ 5%] 2025-12-04T15:25:36.4839419Z functorch/test_vmap.py::TestVmapOperators::test_fill_and_zero_inplace PASSED [0.0153s] [ 5%] 2025-12-04T15:25:36.4840215Z functorch/test_vmap.py::TestVmapOperators::test_imag PASSED [0.0147s] [ 5%] 2025-12-04T15:25:36.4840964Z functorch/test_vmap.py::TestVmapOperators::test_is_complex PASSED [0.0016s] [ 5%] 2025-12-04T15:25:36.4841745Z functorch/test_vmap.py::TestVmapOperators::test_is_contiguous PASSED [0.2584s] [ 5%] 2025-12-04T15:25:36.4842557Z functorch/test_vmap.py::TestVmapOperators::test_is_floating_point PASSED [0.0019s] [ 5%] 2025-12-04T15:25:36.4843345Z functorch/test_vmap.py::TestVmapOperators::test_mean PASSED [0.0079s] [ 5%] 2025-12-04T15:25:36.4844081Z functorch/test_vmap.py::TestVmapOperators::test_mean_dim PASSED [0.0099s] [ 5%] 2025-12-04T15:25:36.4844812Z functorch/test_vmap.py::TestVmapOperators::test_mm PASSED [0.0109s] [ 5%] 2025-12-04T15:25:36.4845535Z functorch/test_vmap.py::TestVmapOperators::test_mode_key PASSED [0.0032s] [ 5%] 2025-12-04T15:25:36.4846264Z functorch/test_vmap.py::TestVmapOperators::test_movedim PASSED [0.0122s] [ 5%] 2025-12-04T15:25:36.4846988Z functorch/test_vmap.py::TestVmapOperators::test_mv PASSED [0.0105s] [ 6%] 2025-12-04T15:25:36.4847710Z functorch/test_vmap.py::TestVmapOperators::test_narrow PASSED [0.0061s] [ 6%] 2025-12-04T15:25:36.4848443Z functorch/test_vmap.py::TestVmapOperators::test_new_empty PASSED [0.0019s] [ 6%] 2025-12-04T15:25:36.4849242Z functorch/test_vmap.py::TestVmapOperators::test_new_empty_strided PASSED [0.0134s] [ 6%] 2025-12-04T15:25:36.4850038Z functorch/test_vmap.py::TestVmapOperators::test_new_zeros PASSED [0.0029s] [ 6%] 2025-12-04T15:25:36.4850773Z functorch/test_vmap.py::TestVmapOperators::test_nll_loss PASSED [0.0065s] [ 6%] 2025-12-04T15:25:36.4851506Z functorch/test_vmap.py::TestVmapOperators::test_one_hot PASSED [0.0053s] [ 6%] 2025-12-04T15:25:36.4852232Z functorch/test_vmap.py::TestVmapOperators::test_real PASSED [0.0151s] [ 6%] 2025-12-04T15:25:36.4852964Z functorch/test_vmap.py::TestVmapOperators::test_repeat PASSED [0.0027s] [ 6%] 2025-12-04T15:25:36.4853685Z functorch/test_vmap.py::TestVmapOperators::test_reshape PASSED [0.0066s] [ 6%] 2025-12-04T15:25:36.4854434Z functorch/test_vmap.py::TestVmapOperators::test_reshape_as PASSED [0.0080s] [ 6%] 2025-12-04T15:25:36.4855216Z functorch/test_vmap.py::TestVmapOperators::test_result_type PASSED [0.0064s] [ 6%] 2025-12-04T15:25:36.4855987Z functorch/test_vmap.py::TestVmapOperators::test_roll_no_dims PASSED [0.0065s] [ 6%] 2025-12-04T15:25:36.4856743Z functorch/test_vmap.py::TestVmapOperators::test_select PASSED [0.0053s] [ 6%] 2025-12-04T15:25:36.4857580Z functorch/test_vmap.py::TestVmapOperators::test_silu_backward PASSED [0.0042s] [ 6%] 2025-12-04T15:25:36.4858339Z functorch/test_vmap.py::TestVmapOperators::test_slice PASSED [0.0055s] [ 6%] 2025-12-04T15:25:36.4859052Z functorch/test_vmap.py::TestVmapOperators::test_slogdet PASSED [0.0066s] [ 6%] 2025-12-04T15:25:36.4859786Z functorch/test_vmap.py::TestVmapOperators::test_split PASSED [0.0394s] [ 6%] 2025-12-04T15:25:36.4860519Z functorch/test_vmap.py::TestVmapOperators::test_squeeze PASSED [0.0230s] [ 6%] 2025-12-04T15:25:36.4861251Z functorch/test_vmap.py::TestVmapOperators::test_stack PASSED [0.0067s] [ 6%] 2025-12-04T15:25:36.4862016Z functorch/test_vmap.py::TestVmapOperators::test_stride PASSED [0.0015s] [ 6%] 2025-12-04T15:25:36.4862751Z functorch/test_vmap.py::TestVmapOperators::test_sum PASSED [0.0072s] [ 7%] 2025-12-04T15:25:36.4863516Z functorch/test_vmap.py::TestVmapOperators::test_sum_dim PASSED [0.0086s] [ 7%] 2025-12-04T15:25:36.4864230Z functorch/test_vmap.py::TestVmapOperators::test_t PASSED [0.0055s] [ 7%] 2025-12-04T15:25:36.4865270Z functorch/test_vmap.py::TestVmapOperators::test_tensor_split SKIPPED [0.0005s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 7%] 2025-12-04T15:25:36.4866305Z functorch/test_vmap.py::TestVmapOperators::test_to PASSED [0.0066s] [ 7%] 2025-12-04T15:25:36.4867094Z functorch/test_vmap.py::TestVmapOperators::test_trace PASSED [0.0056s] [ 7%] 2025-12-04T15:25:36.4867824Z functorch/test_vmap.py::TestVmapOperators::test_transpose PASSED [0.0071s] [ 7%] 2025-12-04T15:25:36.4868631Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_abs PASSED [0.0116s] [ 7%] 2025-12-04T15:25:36.4869498Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_acos PASSED [0.0130s] [ 7%] 2025-12-04T15:25:36.4870362Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_asin PASSED [0.0148s] [ 7%] 2025-12-04T15:25:36.4871217Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_atan PASSED [0.0137s] [ 7%] 2025-12-04T15:25:36.4872074Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_ceil PASSED [0.0117s] [ 7%] 2025-12-04T15:25:36.4872930Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_cos PASSED [0.0128s] [ 7%] 2025-12-04T15:25:36.4873780Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_cosh PASSED [0.0126s] [ 7%] 2025-12-04T15:25:36.4874662Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_digamma PASSED [0.0117s] [ 7%] 2025-12-04T15:25:36.4875536Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_exp PASSED [0.0126s] [ 7%] 2025-12-04T15:25:36.4876409Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_expm1 PASSED [0.0129s] [ 7%] 2025-12-04T15:25:36.4877268Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_floor PASSED [0.0116s] [ 7%] 2025-12-04T15:25:36.4878138Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_frac PASSED [0.0114s] [ 7%] 2025-12-04T15:25:36.4879017Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_lgamma PASSED [0.0142s] [ 7%] 2025-12-04T15:25:36.4879879Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_log PASSED [0.0127s] [ 7%] 2025-12-04T15:25:36.4880727Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_log10 PASSED [0.0140s] [ 8%] 2025-12-04T15:25:36.4881600Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_log1p PASSED [0.0131s] [ 8%] 2025-12-04T15:25:36.4882467Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_log2 PASSED [0.0141s] [ 8%] 2025-12-04T15:25:36.4883309Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_neg PASSED [0.0114s] [ 8%] 2025-12-04T15:25:36.4884203Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_reciprocal PASSED [0.0119s] [ 8%] 2025-12-04T15:25:36.4885100Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_relu PASSED [0.0115s] [ 8%] 2025-12-04T15:25:36.4885965Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_round PASSED [0.0116s] [ 8%] 2025-12-04T15:25:36.4886825Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_rsqrt PASSED [0.0116s] [ 8%] 2025-12-04T15:25:36.4887709Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_sigmoid PASSED [0.0119s] [ 8%] 2025-12-04T15:25:36.4888593Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_sign PASSED [0.0116s] [ 8%] 2025-12-04T15:25:36.4889454Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_sin PASSED [0.0128s] [ 8%] 2025-12-04T15:25:36.4890306Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_sinh PASSED [0.0127s] [ 8%] 2025-12-04T15:25:36.4891201Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_sqrt PASSED [0.0128s] [ 8%] 2025-12-04T15:25:36.4892070Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_tan PASSED [0.0139s] [ 8%] 2025-12-04T15:25:36.4892963Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_tanh PASSED [0.0119s] [ 8%] 2025-12-04T15:25:36.4893820Z functorch/test_vmap.py::TestVmapOperators::test_unary_pointwise_trunc PASSED [0.0121s] [ 8%] 2025-12-04T15:25:36.4894625Z functorch/test_vmap.py::TestVmapOperators::test_unbind PASSED [0.3325s] [ 8%] 2025-12-04T15:25:36.4895422Z functorch/test_vmap.py::TestVmapOperators::test_unfold PASSED [0.0055s] [ 8%] 2025-12-04T15:25:36.4896333Z functorch/test_vmap.py::TestVmapOperators::test_unsafe_view PASSED [0.0030s] [ 8%] 2025-12-04T15:25:36.4897152Z functorch/test_vmap.py::TestVmapOperators::test_unsqueeze PASSED [0.0114s] [ 8%] 2025-12-04T15:25:36.4898027Z functorch/test_vmap.py::TestVmapOperators::test_view PASSED [0.0063s] [ 8%] 2025-12-04T15:25:36.4898762Z functorch/test_vmap.py::TestVmapOperators::test_view_as PASSED [0.0076s] [ 8%] 2025-12-04T15:25:36.4899519Z functorch/test_vmap.py::TestVmapOperators::test_view_as_complex PASSED [0.0272s] [ 9%] 2025-12-04T15:25:36.4900327Z functorch/test_vmap.py::TestVmapOperators::test_view_as_real PASSED [0.0146s] [ 9%] 2025-12-04T15:25:36.4901358Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_composition_in_dim_0_out_dim_0_randomness_error PASSED [0.0437s] [ 9%] 2025-12-04T15:25:36.4902612Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_composition_in_dim_0_out_dim_0_randomness_same PASSED [0.0497s] [ 9%] 2025-12-04T15:25:36.4903853Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_composition_in_dim_0_out_dim_1_randomness_error PASSED [0.0482s] [ 9%] 2025-12-04T15:25:36.4905103Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_composition_in_dim_0_out_dim_1_randomness_same PASSED [0.0548s] [ 9%] 2025-12-04T15:25:36.4906363Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_composition_in_dim_1_out_dim_0_randomness_error PASSED [0.0582s] [ 9%] 2025-12-04T15:25:36.4907616Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_composition_in_dim_1_out_dim_0_randomness_same PASSED [0.0671s] [ 9%] 2025-12-04T15:25:36.4908858Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_composition_in_dim_1_out_dim_1_randomness_error PASSED [0.0597s] [ 9%] 2025-12-04T15:25:36.4910108Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_composition_in_dim_1_out_dim_1_randomness_same PASSED [0.0706s] [ 9%] 2025-12-04T15:25:36.4911334Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_error_in_dim_0_out_dim_0_randomness_error PASSED [0.0019s] [ 9%] 2025-12-04T15:25:36.4912527Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_error_in_dim_0_out_dim_0_randomness_same PASSED [0.0015s] [ 9%] 2025-12-04T15:25:36.4913693Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_error_in_dim_0_out_dim_1_randomness_error PASSED [0.0014s] [ 9%] 2025-12-04T15:25:36.4914879Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_error_in_dim_0_out_dim_1_randomness_same PASSED [0.0015s] [ 9%] 2025-12-04T15:25:36.4916066Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_error_in_dim_1_out_dim_0_randomness_error PASSED [0.0014s] [ 9%] 2025-12-04T15:25:36.4917248Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_error_in_dim_1_out_dim_0_randomness_same PASSED [0.0015s] [ 9%] 2025-12-04T15:25:36.4918413Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_error_in_dim_1_out_dim_1_randomness_error PASSED [0.0014s] [ 9%] 2025-12-04T15:25:36.4919598Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_error_in_dim_1_out_dim_1_randomness_same PASSED [0.0015s] [ 9%] 2025-12-04T15:25:36.4920753Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_0_out_dim_0_randomness_error PASSED [0.0311s] [ 9%] 2025-12-04T15:25:36.4921977Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_0_out_dim_0_randomness_same PASSED [0.0334s] [ 9%] 2025-12-04T15:25:36.4923135Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_0_out_dim_1_randomness_error PASSED [0.0328s] [ 9%] 2025-12-04T15:25:36.4924270Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_0_out_dim_1_randomness_same PASSED [0.0355s] [ 10%] 2025-12-04T15:25:36.4925390Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_0_out_dim_2_randomness_error PASSED [0.0329s] [ 10%] 2025-12-04T15:25:36.4926642Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_0_out_dim_2_randomness_same PASSED [0.0355s] [ 10%] 2025-12-04T15:25:36.4927752Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_1_out_dim_0_randomness_error PASSED [0.0348s] [ 10%] 2025-12-04T15:25:36.4928881Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_1_out_dim_0_randomness_same PASSED [0.0376s] [ 10%] 2025-12-04T15:25:36.4930008Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_1_out_dim_1_randomness_error PASSED [0.0340s] [ 10%] 2025-12-04T15:25:36.4931135Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_1_out_dim_1_randomness_same PASSED [0.0370s] [ 10%] 2025-12-04T15:25:36.4932238Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_1_out_dim_2_randomness_error PASSED [0.0352s] [ 10%] 2025-12-04T15:25:36.4933355Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_1_out_dim_2_randomness_same PASSED [0.0386s] [ 10%] 2025-12-04T15:25:36.4934479Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_2_out_dim_0_randomness_error PASSED [0.0349s] [ 10%] 2025-12-04T15:25:36.4935598Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_2_out_dim_0_randomness_same PASSED [0.0380s] [ 10%] 2025-12-04T15:25:36.4936712Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_2_out_dim_1_randomness_error PASSED [0.0361s] [ 10%] 2025-12-04T15:25:36.4937895Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_2_out_dim_1_randomness_same PASSED [0.0394s] [ 10%] 2025-12-04T15:25:36.4939022Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_2_out_dim_2_randomness_error PASSED [0.0346s] [ 10%] 2025-12-04T15:25:36.4940143Z functorch/test_vmap.py::TestVmapOperators::test_vmap_chunksize_in_dim_2_out_dim_2_randomness_same PASSED [0.0386s] [ 10%] 2025-12-04T15:25:36.4941119Z functorch/test_vmap.py::TestVmapOperators::test_vmap_fallback_check XFAIL [0.0017s] [ 10%] 2025-12-04T15:25:36.4941993Z functorch/test_vmap.py::TestVmapOperators::test_vmap_fallback_check_ok PASSED [0.0015s] [ 10%] 2025-12-04T15:25:36.4942852Z functorch/test_vmap.py::TestVmapOperators::test_weird_matmul_case PASSED [0.0016s] [ 10%] 2025-12-04T15:25:36.4943849Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_0d_tensor_index_put_inplace_False_cuda PASSED [0.2076s] [ 10%] 2025-12-04T15:25:36.4944997Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_0d_tensor_index_put_inplace_True_cuda PASSED [0.0028s] [ 10%] 2025-12-04T15:25:36.4946085Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_advanced_indexing_cuda PASSED [0.1423s] [ 10%] 2025-12-04T15:25:36.4947330Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_batch_norm_training_False_track_running_stats_False_affine_False_cuda PASSED [0.0019s] [ 11%] 2025-12-04T15:25:36.4948791Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_batch_norm_training_False_track_running_stats_False_affine_True_cuda PASSED [0.0019s] [ 11%] 2025-12-04T15:25:36.4950236Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_batch_norm_training_False_track_running_stats_True_affine_False_cuda PASSED [0.0059s] [ 11%] 2025-12-04T15:25:36.4951679Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_batch_norm_training_False_track_running_stats_True_affine_True_cuda PASSED [0.2253s] [ 11%] 2025-12-04T15:25:36.4953170Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_batch_norm_training_True_track_running_stats_False_affine_False_cuda PASSED [0.0047s] [ 11%] 2025-12-04T15:25:36.4954662Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_batch_norm_training_True_track_running_stats_False_affine_True_cuda PASSED [0.0057s] [ 11%] 2025-12-04T15:25:36.4956102Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_batch_norm_training_True_track_running_stats_True_affine_False_cuda PASSED [0.0060s] [ 11%] 2025-12-04T15:25:36.4957529Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_batch_norm_training_True_track_running_stats_True_affine_True_cuda PASSED [0.0064s] [ 11%] 2025-12-04T15:25:36.4958834Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_conv_double_backward_cuda PASSED [1.7052s] [ 11%] 2025-12-04T15:25:36.4959852Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_fill__Tensor_cuda PASSED [0.0072s] [ 11%] 2025-12-04T15:25:36.4960809Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_flatten_cuda PASSED [0.0023s] [ 11%] 2025-12-04T15:25:36.4961724Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_foo_like_cuda PASSED [0.0021s] [ 11%] 2025-12-04T15:25:36.4962671Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_group_norm_cuda PASSED [0.0058s] [ 11%] 2025-12-04T15:25:36.4963624Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_index_fill_cuda PASSED [0.0249s] [ 11%] 2025-12-04T15:25:36.4964557Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_index_put_cuda PASSED [0.0103s] [ 11%] 2025-12-04T15:25:36.4965528Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_inplace_on_view_cuda PASSED [0.0180s] [ 11%] 2025-12-04T15:25:36.4966483Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_isinf_cuda PASSED [0.0022s] [ 11%] 2025-12-04T15:25:36.4967378Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_isnan_cuda PASSED [0.0021s] [ 11%] 2025-12-04T15:25:36.4968304Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_linalg_eigh_cuda PASSED [0.2238s] [ 11%] 2025-12-04T15:25:36.4969258Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_linalg_svd_cuda PASSED [0.9677s] [ 11%] 2025-12-04T15:25:36.4970254Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_namedtuple_returns_cuda PASSED [0.0025s] [ 11%] 2025-12-04T15:25:36.4971311Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_nested_advanced_indexing_cuda PASSED [0.0029s] [ 11%] 2025-12-04T15:25:36.4972546Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_CubeGenVmapAutogradFunction_cuda_float32 PASSED [0.0152s] [ 12%] 2025-12-04T15:25:36.4974006Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ForwardHasDefaultArgsAutogradFunction_cuda_float32 PASSED [0.0112s] [ 12%] 2025-12-04T15:25:36.4975311Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_H_cuda_float32 PASSED [0.0120s] [ 12%] 2025-12-04T15:25:36.4976553Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_MulGenVmapAutogradFunction_cuda_float32 PASSED [0.0188s] [ 12%] 2025-12-04T15:25:36.4977924Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyCatCustomOp_cuda_float32 PASSED [0.2131s] [ 12%] 2025-12-04T15:25:36.4979241Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyCubeAutogradFunction_cuda_float32 PASSED [0.0149s] [ 12%] 2025-12-04T15:25:36.4980562Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyCubeCustomOp_cuda_float32 PASSED [0.0130s] [ 12%] 2025-12-04T15:25:36.4981970Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyCubeNotComposableAutogradFunction_cuda_float32 XFAIL [0.0084s] [ 12%] 2025-12-04T15:25:36.4983456Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyExpMarkDirtyAutogradFunction_cuda_float32 PASSED [0.2159s] [ 12%] 2025-12-04T15:25:36.4984921Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyMulAutogradFunction_cuda_float32 PASSED [0.0187s] [ 12%] 2025-12-04T15:25:36.4986222Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyMulCustomOp_cuda_float32 PASSED [0.0148s] [ 12%] 2025-12-04T15:25:36.4987550Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyMulScalarCustomOp_cuda_float32 PASSED [0.0108s] [ 12%] 2025-12-04T15:25:36.4988839Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyNMSCustomOp_cuda_float32 XFAIL [0.0125s] [ 12%] 2025-12-04T15:25:36.4990092Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyNonzeroCustomOp_cuda_float32 XFAIL [0.2065s] [ 12%] 2025-12-04T15:25:36.4991477Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpySortAutogradFunction_cuda_float32 PASSED [0.2096s] [ 12%] 2025-12-04T15:25:36.4992791Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpySortCustomOp_cuda_float32 PASSED [0.0142s] [ 12%] 2025-12-04T15:25:36.4994092Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpySplitCopyCustomOp_cuda_float32 PASSED [0.0157s] [ 12%] 2025-12-04T15:25:36.4995462Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpySplitCopyWithIntCustomOp_cuda_float32 PASSED [0.0165s] [ 12%] 2025-12-04T15:25:36.4997042Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyTakeAutogradFunction_cuda_float32 PASSED [0.0307s] [ 12%] 2025-12-04T15:25:36.4998356Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyTakeCustomOp_cuda_float32 PASSED [0.0263s] [ 12%] 2025-12-04T15:25:36.4999659Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_NumpyViewCopyCustomOp_cuda_float32 PASSED [0.2071s] [ 12%] 2025-12-04T15:25:36.5001046Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ScaleGradGenVmapAutogradFunction_cuda_float32 PASSED [0.0131s] [ 13%] 2025-12-04T15:25:36.5002439Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_SelectAutogradFunction_cuda_float32 PASSED [0.2017s] [ 13%] 2025-12-04T15:25:36.5003813Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_SelectGenVmapAutogradFunction_cuda_float32 PASSED [0.0132s] [ 13%] 2025-12-04T15:25:36.5005211Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_SortGenVmapAutogradFunction_cuda_float32 PASSED [0.0171s] [ 13%] 2025-12-04T15:25:36.5006438Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_T_cuda_float32 PASSED [0.0136s] [ 13%] 2025-12-04T15:25:36.5007745Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ZeroGradientsGenVmapAutogradFunction_cuda_float32 PASSED [0.0218s] [ 13%] 2025-12-04T15:25:36.5009084Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___getitem___cuda_float32 XFAIL [0.0693s] [ 13%] 2025-12-04T15:25:36.5010311Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___getitem___functorch_cuda_float32 PASSED [0.4611s] [ 13%] 2025-12-04T15:25:36.5011525Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___radd___cuda_float32 PASSED [0.6100s] [ 13%] 2025-12-04T15:25:36.5012650Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___rand___cuda_int64 PASSED [0.4500s] [ 13%] 2025-12-04T15:25:36.5013784Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___rdiv___cuda_float32 PASSED [0.6376s] [ 13%] 2025-12-04T15:25:36.5014965Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___rmatmul___cuda_float32 PASSED [0.1001s] [ 13%] 2025-12-04T15:25:36.5016138Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___rmod___cuda_float32 PASSED [0.6165s] [ 13%] 2025-12-04T15:25:36.5017344Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___rmul___cuda_float32 PASSED [0.6098s] [ 13%] 2025-12-04T15:25:36.5018555Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___ror___cuda_int64 PASSED [0.4517s] [ 13%] 2025-12-04T15:25:36.5019748Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___rpow___cuda_float32 PASSED [0.6193s] [ 13%] 2025-12-04T15:25:36.5020899Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___rsub___cuda_float32 XFAIL [0.2491s] [ 13%] 2025-12-04T15:25:36.5022026Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule___rxor___cuda_int64 PASSED [0.7056s] [ 13%] 2025-12-04T15:25:36.5023329Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule__batch_norm_with_update_cuda_float32 XFAIL [0.0088s] [ 13%] 2025-12-04T15:25:36.5024568Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule__chunk_cat_cuda_float32 PASSED [0.2417s] [ 13%] 2025-12-04T15:25:36.5025813Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule__native_batch_norm_legit_cuda_float32 XFAIL [0.0156s] [ 13%] 2025-12-04T15:25:36.5027099Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule__segment_reduce_lengths_cuda_float32 XFAIL [0.2117s] [ 14%] 2025-12-04T15:25:36.5028398Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule__segment_reduce_offsets_cuda_float32 XFAIL [0.2152s] [ 14%] 2025-12-04T15:25:36.5029754Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule__softmax_backward_data_cuda_float32 SKIPPED [0.0003s] (Skipped!) [ 14%] 2025-12-04T15:25:36.5031113Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule__unsafe_masked_index_cuda_float32 PASSED [0.4700s] [ 14%] 2025-12-04T15:25:36.5032444Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule__unsafe_masked_index_put_accumulate_cuda_float32 PASSED [0.7911s] [ 14%] 2025-12-04T15:25:36.5033810Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule__upsample_bilinear2d_aa_cuda_float32 PASSED [0.0302s] [ 14%] 2025-12-04T15:25:36.5035026Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_abs_cuda_float32 PASSED [0.3822s] [ 14%] 2025-12-04T15:25:36.5036151Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_acos_cuda_float32 PASSED [0.4201s] [ 14%] 2025-12-04T15:25:36.5037266Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_acosh_cuda_float32 PASSED [0.4166s] [ 14%] 2025-12-04T15:25:36.5038387Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_add_cuda_float32 PASSED [0.9055s] [ 14%] 2025-12-04T15:25:36.5039510Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_addbmm_cuda_float32 PASSED [0.0979s] [ 14%] 2025-12-04T15:25:36.5040654Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_addcdiv_cuda_float32 PASSED [0.5925s] [ 14%] 2025-12-04T15:25:36.5041790Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_addcmul_cuda_float32 PASSED [0.6072s] [ 14%] 2025-12-04T15:25:36.5042934Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_addmm_cuda_float32 PASSED [0.0841s] [ 14%] 2025-12-04T15:25:36.5044113Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_addmm_decomposed_cuda_float32 PASSED [0.0842s] [ 14%] 2025-12-04T15:25:36.5045294Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_addmv_cuda_float32 PASSED [0.0842s] [ 14%] 2025-12-04T15:25:36.5046401Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_addr_cuda_float32 PASSED [0.3123s] [ 14%] 2025-12-04T15:25:36.5047561Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_alias_copy_cuda_float32 PASSED [0.0132s] [ 14%] 2025-12-04T15:25:36.5048705Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_all_cuda_float32 PASSED [0.0663s] [ 14%] 2025-12-04T15:25:36.5049867Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_allclose_cuda_float32 XFAIL [0.0081s] [ 14%] 2025-12-04T15:25:36.5050989Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_amax_cuda_float32 PASSED [0.0857s] [ 14%] 2025-12-04T15:25:36.5052148Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_amin_cuda_float32 PASSED [0.0858s] [ 14%] 2025-12-04T15:25:36.5053278Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_aminmax_cuda_float32 PASSED [0.2636s] [ 15%] 2025-12-04T15:25:36.5054417Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_angle_cuda_float32 PASSED [0.1166s] [ 15%] 2025-12-04T15:25:36.5055582Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_any_cuda_float32 PASSED [0.0663s] [ 15%] 2025-12-04T15:25:36.5056707Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_arange_cuda_float32 PASSED [0.0286s] [ 15%] 2025-12-04T15:25:36.5057917Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_argmax_cuda_float32 PASSED [0.0450s] [ 15%] 2025-12-04T15:25:36.5059051Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_argmin_cuda_float32 PASSED [0.2529s] [ 15%] 2025-12-04T15:25:36.5060185Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_argsort_cuda_float32 PASSED [0.1245s] [ 15%] 2025-12-04T15:25:36.5061332Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_argwhere_cuda_float32 XFAIL [0.0083s] [ 15%] 2025-12-04T15:25:36.5062511Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_as_strided_copy_cuda_float32 XFAIL [0.2090s] [ 15%] 2025-12-04T15:25:36.5063698Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_as_strided_cuda_float32 XFAIL [0.2054s] [ 15%] 2025-12-04T15:25:36.5064913Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_as_strided_partial_views_cuda_float32 XFAIL [0.2090s] [ 15%] 2025-12-04T15:25:36.5066182Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_as_strided_scatter_cuda_float32 XFAIL [0.2133s] [ 15%] 2025-12-04T15:25:36.5067361Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_asin_cuda_float32 PASSED [0.5794s] [ 15%] 2025-12-04T15:25:36.5068482Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_asinh_cuda_float32 PASSED [0.5958s] [ 15%] 2025-12-04T15:25:36.5069594Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_atan2_cuda_float32 PASSED [1.5256s] [ 15%] 2025-12-04T15:25:36.5070723Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_atan_cuda_float32 PASSED [0.3847s] [ 15%] 2025-12-04T15:25:36.5071844Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_atanh_cuda_float32 PASSED [0.3882s] [ 15%] 2025-12-04T15:25:36.5073000Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_atleast_1d_cuda_float32 PASSED [0.0215s] [ 15%] 2025-12-04T15:25:36.5074161Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_atleast_2d_cuda_float32 PASSED [0.0213s] [ 15%] 2025-12-04T15:25:36.5075345Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_atleast_3d_cuda_float32 PASSED [0.0213s] [ 15%] 2025-12-04T15:25:36.5076502Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_baddbmm_cuda_float32 PASSED [0.0877s] [ 15%] 2025-12-04T15:25:36.5077650Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bernoulli_cuda_float32 XFAIL [0.0194s] [ 16%] 2025-12-04T15:25:36.5078795Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bfloat16_cuda_float32 XFAIL [0.2385s] [ 16%] 2025-12-04T15:25:36.5080072Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bfloat16_functorch_no_channels_last_cuda_float32 PASSED [0.2209s] [ 16%] 2025-12-04T15:25:36.5081391Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bincount_cuda_int64 XFAIL [0.0083s] [ 16%] 2025-12-04T15:25:36.5082540Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bitwise_and_cuda_int64 PASSED [0.8482s] [ 16%] 2025-12-04T15:25:36.5083764Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bitwise_left_shift_cuda_int64 PASSED [0.6401s] [ 16%] 2025-12-04T15:25:36.5084969Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bitwise_not_cuda_int64 PASSED [0.1575s] [ 16%] 2025-12-04T15:25:36.5086134Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bitwise_or_cuda_int64 PASSED [0.6470s] [ 16%] 2025-12-04T15:25:36.5087400Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bitwise_right_shift_cuda_int64 PASSED [0.6418s] [ 16%] 2025-12-04T15:25:36.5088598Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bitwise_xor_cuda_int64 PASSED [0.6412s] [ 16%] 2025-12-04T15:25:36.5089773Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_block_diag_cuda_float32 PASSED [0.0540s] [ 16%] 2025-12-04T15:25:36.5090920Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bmm_cuda_float32 PASSED [0.0132s] [ 16%] 2025-12-04T15:25:36.5092025Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bool_cuda_float32 XFAIL [0.0202s] [ 16%] 2025-12-04T15:25:36.5093259Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bool_functorch_no_channels_last_cuda_float32 PASSED [0.2759s] [ 16%] 2025-12-04T15:25:36.5094570Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_broadcast_shapes_cuda_float32 PASSED [0.0100s] [ 16%] 2025-12-04T15:25:36.5095809Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_broadcast_tensors_cuda_float32 PASSED [0.2551s] [ 16%] 2025-12-04T15:25:36.5097260Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_broadcast_to_cuda_float32 PASSED [0.0287s] [ 16%] 2025-12-04T15:25:36.5098448Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_bucketize_cuda_float32 PASSED [0.2424s] [ 16%] 2025-12-04T15:25:36.5099571Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_byte_cuda_float32 XFAIL [0.0202s] [ 16%] 2025-12-04T15:25:36.5100813Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_byte_functorch_no_channels_last_cuda_float32 PASSED [0.2279s] [ 16%] 2025-12-04T15:25:36.5102117Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cartesian_prod_cuda_float32 PASSED [0.2321s] [ 16%] 2025-12-04T15:25:36.5103285Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cat_cuda_float32 PASSED [0.2625s] [ 17%] 2025-12-04T15:25:36.5104536Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cauchy_cuda_float32 SKIPPED [0.0003s] (Test expects tensor input) [ 17%] 2025-12-04T15:25:36.5105813Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cdist_cuda_float32 PASSED [1.5087s] [ 17%] 2025-12-04T15:25:36.5106947Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cdouble_cuda_float32 XFAIL [0.0211s] [ 17%] 2025-12-04T15:25:36.5108076Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ceil_cuda_float32 PASSED [0.4137s] [ 17%] 2025-12-04T15:25:36.5109189Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cfloat_cuda_float32 XFAIL [0.0227s] [ 17%] 2025-12-04T15:25:36.5110316Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_chalf_cuda_float32 XFAIL [0.2263s] [ 17%] 2025-12-04T15:25:36.5111423Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_char_cuda_float32 XFAIL [0.2324s] [ 17%] 2025-12-04T15:25:36.5112665Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_char_functorch_no_channels_last_cuda_float32 PASSED [0.2283s] [ 17%] 2025-12-04T15:25:36.5114140Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cholesky_cuda_float32 SKIPPED [0.0016s] (no MAGMA library detected) [ 17%] 2025-12-04T15:25:36.5115680Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cholesky_inverse_cuda_float32 SKIPPED [0.0017s] (no MAGMA library detected) [ 17%] 2025-12-04T15:25:36.5117191Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cholesky_solve_cuda_float32 SKIPPED [0.0014s] (no MAGMA library detected) [ 17%] 2025-12-04T15:25:36.5118581Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_chunk_cuda_float32 PASSED [0.1274s] [ 17%] 2025-12-04T15:25:36.5119698Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_clamp_cuda_float32 PASSED [0.5201s] [ 17%] 2025-12-04T15:25:36.5120833Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_clamp_max_cuda_float32 XFAIL [0.0090s] [ 17%] 2025-12-04T15:25:36.5121991Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_clamp_min_cuda_float32 XFAIL [0.2146s] [ 17%] 2025-12-04T15:25:36.5123129Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_clone_cuda_float32 XFAIL [0.1885s] [ 17%] 2025-12-04T15:25:36.5124272Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_column_stack_cuda_float32 PASSED [0.4153s] [ 17%] 2025-12-04T15:25:36.5125476Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_combinations_cuda_float32 PASSED [0.1000s] [ 17%] 2025-12-04T15:25:36.5126659Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_complex_cuda_float32 PASSED [0.6577s] [ 17%] 2025-12-04T15:25:36.5127798Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_conj_cuda_float32 PASSED [0.1213s] [ 17%] 2025-12-04T15:25:36.5128958Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_conj_physical_cuda_float32 PASSED [0.1112s] [ 17%] 2025-12-04T15:25:36.5130177Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_constant_pad_nd_cuda_float32 PASSED [0.1762s] [ 18%] 2025-12-04T15:25:36.5131378Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_contiguous_cuda_float32 XFAIL [0.1805s] [ 18%] 2025-12-04T15:25:36.5132548Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_copysign_cuda_float32 PASSED [1.1152s] [ 18%] 2025-12-04T15:25:36.5133699Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_corrcoef_cuda_float32 PASSED [0.0275s] [ 18%] 2025-12-04T15:25:36.5134837Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cos_cuda_float32 PASSED [0.2182s] [ 18%] 2025-12-04T15:25:36.5135957Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cosh_cuda_float32 PASSED [0.2295s] [ 18%] 2025-12-04T15:25:36.5137184Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_count_nonzero_cuda_float32 XFAIL [0.0082s] [ 18%] 2025-12-04T15:25:36.5138330Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cov_cuda_float32 PASSED [0.2625s] [ 18%] 2025-12-04T15:25:36.5139461Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cross_cuda_float32 PASSED [0.0240s] [ 18%] 2025-12-04T15:25:36.5140602Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cummax_cuda_float32 PASSED [0.0200s] [ 18%] 2025-12-04T15:25:36.5141745Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cummin_cuda_float32 PASSED [0.0199s] [ 18%] 2025-12-04T15:25:36.5142881Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cumprod_cuda_float32 PASSED [0.0369s] [ 18%] 2025-12-04T15:25:36.5144020Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cumsum_cuda_float32 PASSED [0.2398s] [ 18%] 2025-12-04T15:25:36.5145279Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_cumulative_trapezoid_cuda_float32 PASSED [0.0591s] [ 18%] 2025-12-04T15:25:36.5146531Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_deg2rad_cuda_float32 PASSED [0.2144s] [ 18%] 2025-12-04T15:25:36.5147650Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_diag_cuda_float32 PASSED [0.0632s] [ 18%] 2025-12-04T15:25:36.5148800Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_diag_embed_cuda_float32 PASSED [0.1083s] [ 18%] 2025-12-04T15:25:36.5149955Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_diagflat_cuda_float32 XFAIL [0.0080s] [ 18%] 2025-12-04T15:25:36.5151192Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_diagonal_copy_cuda_float32 PASSED [0.0963s] [ 18%] 2025-12-04T15:25:36.5152360Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_diagonal_cuda_float32 PASSED [0.0924s] [ 18%] 2025-12-04T15:25:36.5153560Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_diagonal_scatter_cuda_float32 PASSED [0.0985s] [ 18%] 2025-12-04T15:25:36.5154739Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_diff_cuda_float32 PASSED [0.5445s] [ 19%] 2025-12-04T15:25:36.5155870Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_digamma_cuda_float32 PASSED [0.5657s] [ 19%] 2025-12-04T15:25:36.5156991Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_dist_cuda_float32 PASSED [0.3108s] [ 19%] 2025-12-04T15:25:36.5158182Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_div_floor_rounding_cuda_float32 PASSED [1.6671s] [ 19%] 2025-12-04T15:25:36.5159436Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_div_no_rounding_mode_cuda_float32 PASSED [1.6514s] [ 19%] 2025-12-04T15:25:36.5160702Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_div_trunc_rounding_cuda_float32 PASSED [1.6748s] [ 19%] 2025-12-04T15:25:36.5161868Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_dot_cuda_float32 PASSED [0.0272s] [ 19%] 2025-12-04T15:25:36.5162990Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_double_cuda_float32 XFAIL [0.0202s] [ 19%] 2025-12-04T15:25:36.5164245Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_double_functorch_no_channels_last_cuda_float32 PASSED [0.2520s] [ 19%] 2025-12-04T15:25:36.5165520Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_dsplit_cuda_float32 PASSED [0.0240s] [ 19%] 2025-12-04T15:25:36.5166666Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_dstack_cuda_float32 PASSED [0.2170s] [ 19%] 2025-12-04T15:25:36.5167789Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_einsum_cuda_float32 PASSED [0.0117s] [ 19%] 2025-12-04T15:25:36.5168927Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_empty_cuda_float32 PASSED [0.0080s] [ 19%] 2025-12-04T15:25:36.5170078Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_empty_like_cuda_float32 PASSED [0.0744s] [ 19%] 2025-12-04T15:25:36.5171276Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_empty_permuted_cuda_float32 PASSED [0.0275s] [ 19%] 2025-12-04T15:25:36.5172468Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_empty_strided_cuda_float32 PASSED [0.0076s] [ 19%] 2025-12-04T15:25:36.5173616Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_eq_cuda_float32 XFAIL [0.2676s] [ 19%] 2025-12-04T15:25:36.5174724Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_equal_cuda_float32 XFAIL [0.2257s] [ 19%] 2025-12-04T15:25:36.5175834Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_erf_cuda_float32 PASSED [0.4989s] [ 19%] 2025-12-04T15:25:36.5177074Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_erfc_cuda_float32 PASSED [0.9285s] [ 19%] 2025-12-04T15:25:36.5178246Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_erfinv_cuda_float32 PASSED [0.4680s] [ 19%] 2025-12-04T15:25:36.5179379Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_exp2_cuda_float32 PASSED [0.3853s] [ 20%] 2025-12-04T15:25:36.5180498Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_exp_cuda_float32 PASSED [0.2258s] [ 20%] 2025-12-04T15:25:36.5181625Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_expand_as_cuda_float32 PASSED [0.0224s] [ 20%] 2025-12-04T15:25:36.5182875Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_expand_copy_cuda_float32 PASSED [0.0365s] [ 20%] 2025-12-04T15:25:36.5184044Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_expand_cuda_float32 PASSED [0.0348s] [ 20%] 2025-12-04T15:25:36.5185178Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_expm1_cuda_float32 PASSED [0.2907s] [ 20%] 2025-12-04T15:25:36.5186323Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_exponential_cuda_float32 XFAIL [0.0148s] [ 20%] 2025-12-04T15:25:36.5187465Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_eye_cuda_float32 PASSED [0.2732s] [ 20%] 2025-12-04T15:25:36.5188599Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_fft2_cuda_float32 PASSED [0.2535s] [ 20%] 2025-12-04T15:25:36.5189752Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_fft_cuda_float32 PASSED [0.0487s] [ 20%] 2025-12-04T15:25:36.5190886Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_fftn_cuda_float32 PASSED [0.0481s] [ 20%] 2025-12-04T15:25:36.5192057Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_fftshift_cuda_float32 PASSED [0.0220s] [ 20%] 2025-12-04T15:25:36.5193234Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_hfft2_cuda_float32 PASSED [0.0484s] [ 20%] 2025-12-04T15:25:36.5194399Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_hfft_cuda_float32 PASSED [0.0436s] [ 20%] 2025-12-04T15:25:36.5195543Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_hfftn_cuda_float32 PASSED [0.0443s] [ 20%] 2025-12-04T15:25:36.5196841Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_ifft2_cuda_float32 PASSED [0.2390s] [ 20%] 2025-12-04T15:25:36.5197997Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_ifft_cuda_float32 PASSED [0.0425s] [ 20%] 2025-12-04T15:25:36.5199156Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_ifftn_cuda_float32 PASSED [0.2476s] [ 20%] 2025-12-04T15:25:36.5200330Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_ifftshift_cuda_float32 PASSED [0.0222s] [ 20%] 2025-12-04T15:25:36.5201509Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_ihfft2_cuda_float32 XFAIL [0.0133s] [ 20%] 2025-12-04T15:25:36.5202666Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_ihfft_cuda_float32 PASSED [0.0488s] [ 20%] 2025-12-04T15:25:36.5203822Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_ihfftn_cuda_float32 XFAIL [0.0127s] [ 20%] 2025-12-04T15:25:36.5204974Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_irfft2_cuda_float32 PASSED [0.4618s] [ 21%] 2025-12-04T15:25:36.5206144Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_irfft_cuda_float32 PASSED [0.0399s] [ 21%] 2025-12-04T15:25:36.5207308Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_irfftn_cuda_float32 PASSED [0.2467s] [ 21%] 2025-12-04T15:25:36.5208558Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_rfft2_cuda_float32 PASSED [0.0460s] [ 21%] 2025-12-04T15:25:36.5209748Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_rfft_cuda_float32 PASSED [0.0385s] [ 21%] 2025-12-04T15:25:36.5210906Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fft_rfftn_cuda_float32 PASSED [0.2461s] [ 21%] 2025-12-04T15:25:36.5212038Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fill_cuda_float32 XFAIL [0.0086s] [ 21%] 2025-12-04T15:25:36.5213164Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_flatten_cuda_float32 PASSED [0.3419s] [ 21%] 2025-12-04T15:25:36.5214367Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_flip_cuda_float32 PASSED [0.0333s] [ 21%] 2025-12-04T15:25:36.5215493Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fliplr_cuda_float32 PASSED [0.0158s] [ 21%] 2025-12-04T15:25:36.5216633Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_flipud_cuda_float32 PASSED [0.0160s] [ 21%] 2025-12-04T15:25:36.5217843Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_float_cuda_float32 XFAIL [0.0193s] [ 21%] 2025-12-04T15:25:36.5219082Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_float_functorch_no_channels_last_cuda_float32 PASSED [0.2260s] [ 21%] 2025-12-04T15:25:36.5220376Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_float_power_cuda_float32 PASSED [0.8166s] [ 21%] 2025-12-04T15:25:36.5221535Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_floor_cuda_float32 PASSED [0.2058s] [ 21%] 2025-12-04T15:25:36.5222694Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_floor_divide_cuda_float32 PASSED [0.6224s] [ 21%] 2025-12-04T15:25:36.5223844Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fmax_cuda_float32 PASSED [0.5795s] [ 21%] 2025-12-04T15:25:36.5224969Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fmin_cuda_float32 PASSED [0.5833s] [ 21%] 2025-12-04T15:25:36.5226082Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_fmod_cuda_float32 PASSED [0.6182s] [ 21%] 2025-12-04T15:25:36.5227205Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_frac_cuda_float32 PASSED [0.2069s] [ 21%] 2025-12-04T15:25:36.5228312Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_frexp_cuda_float32 PASSED [0.1809s] [ 21%] 2025-12-04T15:25:36.5229441Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_full_cuda_float32 PASSED [0.0077s] [ 22%] 2025-12-04T15:25:36.5230579Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_full_like_cuda_float32 PASSED [0.3004s] [ 22%] 2025-12-04T15:25:36.5231728Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_gather_cuda_float32 PASSED [0.0699s] [ 22%] 2025-12-04T15:25:36.5232829Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_gcd_cuda_int64 PASSED [0.6007s] [ 22%] 2025-12-04T15:25:36.5233939Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ge_cuda_float32 XFAIL [0.4653s] [ 22%] 2025-12-04T15:25:36.5235206Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_geometric_cuda_float32 SKIPPED [0.0003s] (Test expects tensor input) [ 22%] 2025-12-04T15:25:36.5236501Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_geqrf_cuda_float32 PASSED [0.3825s] [ 22%] 2025-12-04T15:25:36.5237645Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_gradient_cuda_float32 PASSED [0.3360s] [ 22%] 2025-12-04T15:25:36.5238840Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_grid_sampler_2d_cuda_float32 PASSED [0.1215s] [ 22%] 2025-12-04T15:25:36.5240170Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_grid_sampler_3d_cuda_float32 SKIPPED [0.0003s] (Skipped!) [ 22%] 2025-12-04T15:25:36.5241422Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_gt_cuda_float32 XFAIL [0.4680s] [ 22%] 2025-12-04T15:25:36.5242511Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_half_cuda_float32 XFAIL [0.2471s] [ 22%] 2025-12-04T15:25:36.5243769Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_half_functorch_no_channels_last_cuda_float32 PASSED [0.2431s] [ 22%] 2025-12-04T15:25:36.5245175Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_hash_tensor_cuda_float32 SKIPPED [0.0003s] (Skipped!) [ 22%] 2025-12-04T15:25:36.5246417Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_heaviside_cuda_float32 PASSED [0.5913s] [ 22%] 2025-12-04T15:25:36.5247549Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_histc_cuda_float32 XFAIL [0.0086s] [ 22%] 2025-12-04T15:25:36.5248683Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_hsplit_cuda_float32 PASSED [0.2561s] [ 22%] 2025-12-04T15:25:36.5249824Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_hstack_cuda_float32 PASSED [0.2405s] [ 22%] 2025-12-04T15:25:36.5250959Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_hypot_cuda_float32 PASSED [0.6263s] [ 22%] 2025-12-04T15:25:36.5252057Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_i0_cuda_float32 PASSED [0.5888s] [ 22%] 2025-12-04T15:25:36.5253182Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_igamma_cuda_float32 PASSED [1.3963s] [ 22%] 2025-12-04T15:25:36.5254324Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_igammac_cuda_float32 PASSED [1.6209s] [ 23%] 2025-12-04T15:25:36.5255474Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_imag_cuda_complex64 PASSED [0.1684s] [ 23%] 2025-12-04T15:25:36.5256607Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_add_cuda_float32 PASSED [0.2034s] [ 23%] 2025-12-04T15:25:36.5257856Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_copy_cuda_float32 PASSED [0.0407s] [ 23%] 2025-12-04T15:25:36.5259035Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_fill_cuda_float32 PASSED [0.1054s] [ 23%] 2025-12-04T15:25:36.5260202Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_put_cuda_float32 XFAIL [0.0312s] [ 23%] 2025-12-04T15:25:36.5261406Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_put_functorch_cuda_float32 PASSED [0.4225s] [ 23%] 2025-12-04T15:25:36.5262653Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_reduce_amax_cuda_float32 XFAIL [0.0088s] [ 23%] 2025-12-04T15:25:36.5263889Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_reduce_amin_cuda_float32 XFAIL [0.2194s] [ 23%] 2025-12-04T15:25:36.5265115Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_reduce_mean_cuda_float32 XFAIL [0.2170s] [ 23%] 2025-12-04T15:25:36.5266329Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_reduce_prod_cuda_float32 XFAIL [0.2239s] [ 23%] 2025-12-04T15:25:36.5267546Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_index_select_cuda_float32 PASSED [0.4546s] [ 23%] 2025-12-04T15:25:36.5268717Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_inner_cuda_float32 PASSED [0.0192s] [ 23%] 2025-12-04T15:25:36.5269833Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_int_cuda_float32 XFAIL [0.0200s] [ 23%] 2025-12-04T15:25:36.5271048Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_int_functorch_no_channels_last_cuda_float32 PASSED [0.0186s] [ 23%] 2025-12-04T15:25:36.5272348Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_isclose_cuda_float32 XFAIL [0.0228s] [ 23%] 2025-12-04T15:25:36.5273531Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_isfinite_cuda_float32 PASSED [0.3501s] [ 23%] 2025-12-04T15:25:36.5274668Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_isin_cuda_float32 XFAIL [0.0086s] [ 23%] 2025-12-04T15:25:36.5275780Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_isinf_cuda_float32 PASSED [0.3398s] [ 23%] 2025-12-04T15:25:36.5276966Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_isnan_cuda_float32 PASSED [0.1147s] [ 23%] 2025-12-04T15:25:36.5278111Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_isneginf_cuda_float32 PASSED [0.1153s] [ 23%] 2025-12-04T15:25:36.5279279Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_isposinf_cuda_float32 PASSED [0.1153s] [ 23%] 2025-12-04T15:25:36.5280427Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_isreal_cuda_float32 PASSED [0.1264s] [ 24%] 2025-12-04T15:25:36.5281559Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_istft_cuda_complex64 XFAIL [0.0098s] [ 24%] 2025-12-04T15:25:36.5282682Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_item_cuda_float32 XFAIL [0.0244s] [ 24%] 2025-12-04T15:25:36.5283901Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_jiterator_2inputs_2outputs_cuda_float32 XFAIL [0.2481s] [ 24%] 2025-12-04T15:25:36.5285263Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_jiterator_4inputs_with_extra_args_cuda_float32 XFAIL [0.2847s] [ 24%] 2025-12-04T15:25:36.5286556Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_jiterator_binary_cuda_float32 XFAIL [0.2802s] [ 24%] 2025-12-04T15:25:36.5287850Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_jiterator_binary_return_by_ref_cuda_float32 XFAIL [0.0186s] [ 24%] 2025-12-04T15:25:36.5289147Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_jiterator_unary_cuda_float32 XFAIL [0.2727s] [ 24%] 2025-12-04T15:25:36.5290321Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_kron_cuda_float32 PASSED [0.2197s] [ 24%] 2025-12-04T15:25:36.5291453Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_kthvalue_cuda_float32 PASSED [0.0629s] [ 24%] 2025-12-04T15:25:36.5292584Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_lcm_cuda_int64 PASSED [0.5952s] [ 24%] 2025-12-04T15:25:36.5293700Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ldexp_cuda_float32 PASSED [0.6372s] [ 24%] 2025-12-04T15:25:36.5294809Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_le_cuda_float32 XFAIL [0.4605s] [ 24%] 2025-12-04T15:25:36.5295905Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_lerp_cuda_float32 PASSED [0.3450s] [ 24%] 2025-12-04T15:25:36.5297256Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_lgamma_cuda_float32 PASSED [0.3193s] [ 24%] 2025-12-04T15:25:36.5298438Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_cholesky_cuda_float32 PASSED [0.0583s] [ 24%] 2025-12-04T15:25:36.5299686Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_cholesky_ex_cuda_float32 PASSED [0.0708s] [ 24%] 2025-12-04T15:25:36.5300898Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_cond_cuda_float32 PASSED [0.0255s] [ 24%] 2025-12-04T15:25:36.5302095Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_cross_cuda_float32 PASSED [0.0371s] [ 24%] 2025-12-04T15:25:36.5303361Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_det_cuda_float32 PASSED [0.0710s] [ 24%] 2025-12-04T15:25:36.5304573Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_diagonal_cuda_float32 PASSED [0.0671s] [ 24%] 2025-12-04T15:25:36.5305939Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_eig_cuda_float32 SKIPPED [0.0017s] (no MAGMA library detected) [ 25%] 2025-12-04T15:25:36.5307324Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_eigh_cuda_float32 SKIPPED [0.0003s] (Skipped!) [ 25%] 2025-12-04T15:25:36.5308724Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_eigvals_cuda_float32 SKIPPED [0.0015s] (no MAGMA library detected) [ 25%] 2025-12-04T15:25:36.5310310Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_eigvalsh_cuda_float32 SKIPPED [0.0018s] (no MAGMA library detected) [ 25%] 2025-12-04T15:25:36.5311730Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_householder_product_cuda_float32 PASSED [0.1165s] [ 25%] 2025-12-04T15:25:36.5312992Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_inv_cuda_float32 PASSED [0.0598s] [ 25%] 2025-12-04T15:25:36.5314185Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_inv_ex_cuda_float32 PASSED [0.0433s] [ 25%] 2025-12-04T15:25:36.5315417Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_ldl_factor_cuda_float32 PASSED [0.0282s] [ 25%] 2025-12-04T15:25:36.5316669Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_ldl_factor_ex_cuda_float32 PASSED [0.0303s] [ 25%] 2025-12-04T15:25:36.5317977Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_ldl_solve_cuda_float32 SKIPPED [0.0003s] (Skipped!) [ 25%] 2025-12-04T15:25:36.5319380Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_lstsq_cuda_float32 SKIPPED [0.0019s] (no MAGMA library detected) [ 25%] 2025-12-04T15:25:36.5320910Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_lstsq_grad_oriented_cuda_float32 SKIPPED [0.0015s] (no MAGMA library detected) [ 25%] 2025-12-04T15:25:36.5322285Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_lu_cuda_float32 XFAIL [0.0091s] [ 25%] 2025-12-04T15:25:36.5323472Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_lu_factor_cuda_float32 PASSED [0.4681s] [ 25%] 2025-12-04T15:25:36.5324727Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_lu_factor_ex_cuda_float32 PASSED [0.2269s] [ 25%] 2025-12-04T15:25:36.5325967Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_lu_solve_cuda_float32 PASSED [1.0682s] [ 25%] 2025-12-04T15:25:36.5327208Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_matrix_norm_cuda_float32 PASSED [0.2693s] [ 25%] 2025-12-04T15:25:36.5328463Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_matrix_power_cuda_float32 PASSED [0.1306s] [ 25%] 2025-12-04T15:25:36.5329732Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_matrix_rank_cuda_float32 PASSED [0.3536s] [ 25%] 2025-12-04T15:25:36.5331044Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_matrix_rank_hermitian_cuda_float32 PASSED [0.0392s] [ 25%] 2025-12-04T15:25:36.5332335Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_multi_dot_cuda_float32 PASSED [0.2190s] [ 25%] 2025-12-04T15:25:36.5333537Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_norm_cuda_float32 PASSED [0.3888s] [ 26%] 2025-12-04T15:25:36.5334825Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_norm_subgradients_at_zero_cuda_float32 PASSED [0.2889s] [ 26%] 2025-12-04T15:25:36.5336147Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_pinv_cuda_float32 PASSED [0.1210s] [ 26%] 2025-12-04T15:25:36.5337707Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_pinv_hermitian_cuda_float32 SKIPPED [0.0017s] (no MAGMA library detected) [ 26%] 2025-12-04T15:25:36.5368100Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_pinv_singular_cuda_float32 SKIPPED [0.0019s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 26%] 2025-12-04T15:25:36.5369660Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_qr_cuda_float32 PASSED [0.1676s] [ 26%] 2025-12-04T15:25:36.5371013Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_slogdet_cuda_float32 PASSED [0.0544s] [ 26%] 2025-12-04T15:25:36.5372235Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_solve_cuda_float32 PASSED [0.1691s] [ 26%] 2025-12-04T15:25:36.5373467Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_solve_ex_cuda_float32 PASSED [0.2113s] [ 26%] 2025-12-04T15:25:36.5374755Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_solve_triangular_cuda_float32 PASSED [0.7662s] [ 26%] 2025-12-04T15:25:36.5376001Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_svd_cuda_float32 PASSED [1.1721s] [ 26%] 2025-12-04T15:25:36.5377292Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_svdvals_cuda_float32 PASSED [0.1228s] [ 26%] 2025-12-04T15:25:36.5378542Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_tensorinv_cuda_float32 PASSED [0.0179s] [ 26%] 2025-12-04T15:25:36.5379802Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_tensorsolve_cuda_float32 XFAIL [0.0090s] [ 26%] 2025-12-04T15:25:36.5381031Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_vander_cuda_float32 PASSED [0.2518s] [ 26%] 2025-12-04T15:25:36.5382258Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_vecdot_cuda_float32 PASSED [0.2438s] [ 26%] 2025-12-04T15:25:36.5383512Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linalg_vector_norm_cuda_float32 PASSED [0.5362s] [ 26%] 2025-12-04T15:25:36.5384743Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linspace_cuda_float32 PASSED [0.2422s] [ 26%] 2025-12-04T15:25:36.5389549Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_linspace_tensor_overload_cuda_float32 PASSED [0.7391s] [ 26%] 2025-12-04T15:25:36.5390841Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_log10_cuda_float32 PASSED [0.2252s] [ 26%] 2025-12-04T15:25:36.5391974Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_log1p_cuda_float32 PASSED [0.2878s] [ 26%] 2025-12-04T15:25:36.5393093Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_log2_cuda_float32 PASSED [0.2263s] [ 26%] 2025-12-04T15:25:36.5394198Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_log_cuda_float32 PASSED [0.2258s] [ 27%] 2025-12-04T15:25:36.5395464Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_log_normal_cuda_float32 SKIPPED [0.0003s] (Test expects tensor input) [ 27%] 2025-12-04T15:25:36.5396946Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_log_softmax_cuda_float32 PASSED [0.0588s] [ 27%] 2025-12-04T15:25:36.5398181Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_log_softmax_with_dtype_cuda_float32 PASSED [0.0612s] [ 27%] 2025-12-04T15:25:36.5399399Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logaddexp2_cuda_float32 PASSED [0.0120s] [ 27%] 2025-12-04T15:25:36.5400705Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logaddexp_cuda_float32 PASSED [0.5813s] [ 27%] 2025-12-04T15:25:36.5401902Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logcumsumexp_cuda_float32 PASSED [0.0315s] [ 27%] 2025-12-04T15:25:36.5403257Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logdet_cuda_float32 SKIPPED [0.0016s] (no MAGMA library detected) [ 27%] 2025-12-04T15:25:36.5404544Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logical_and_cuda_float32 PASSED [0.5885s] [ 27%] 2025-12-04T15:25:36.5405715Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logical_not_cuda_float32 PASSED [0.1265s] [ 27%] 2025-12-04T15:25:36.5406972Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logical_or_cuda_float32 PASSED [0.5871s] [ 27%] 2025-12-04T15:25:36.5408139Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logical_xor_cuda_float32 PASSED [0.5888s] [ 27%] 2025-12-04T15:25:36.5409281Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logit_cuda_float32 PASSED [0.3374s] [ 27%] 2025-12-04T15:25:36.5410424Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logspace_cuda_float32 PASSED [0.1573s] [ 27%] 2025-12-04T15:25:36.5411665Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logspace_tensor_overload_cuda_float32 PASSED [5.1527s] [ 27%] 2025-12-04T15:25:36.5412897Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_logsumexp_cuda_float32 PASSED [0.1298s] [ 27%] 2025-12-04T15:25:36.5414044Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_long_cuda_float32 XFAIL [0.0199s] [ 27%] 2025-12-04T15:25:36.5415292Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_long_functorch_no_channels_last_cuda_float32 PASSED [0.0179s] [ 27%] 2025-12-04T15:25:36.5416528Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_lt_cuda_float32 XFAIL [0.4586s] [ 27%] 2025-12-04T15:25:36.5417730Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_lu_cuda_float32 PASSED [0.4570s] [ 27%] 2025-12-04T15:25:36.5418860Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_lu_solve_cuda_float32 XFAIL [0.0096s] [ 27%] 2025-12-04T15:25:36.5420017Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_lu_unpack_cuda_float32 PASSED [0.6374s] [ 28%] 2025-12-04T15:25:36.5421150Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mH_cuda_float32 PASSED [0.0328s] [ 28%] 2025-12-04T15:25:36.5422244Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mT_cuda_float32 PASSED [0.2398s] [ 28%] 2025-12-04T15:25:36.5423389Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_amax_cuda_float32 PASSED [0.3821s] [ 28%] 2025-12-04T15:25:36.5424583Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_amin_cuda_float32 PASSED [0.3795s] [ 28%] 2025-12-04T15:25:36.5425780Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_argmax_cuda_float32 PASSED [0.2377s] [ 28%] 2025-12-04T15:25:36.5426983Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_argmin_cuda_float32 PASSED [0.2388s] [ 28%] 2025-12-04T15:25:36.5428191Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_cumprod_cuda_float32 PASSED [0.0768s] [ 28%] 2025-12-04T15:25:36.5429397Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_cumsum_cuda_float32 PASSED [0.0761s] [ 28%] 2025-12-04T15:25:36.5430587Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_fill_cuda_float32 XFAIL [0.0352s] [ 28%] 2025-12-04T15:25:36.5431873Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_fill_functorch_Scalar_only_cuda_float32 PASSED [0.2556s] [ 28%] 2025-12-04T15:25:36.5433255Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_log_softmax_cuda_float32 PASSED [0.0998s] [ 28%] 2025-12-04T15:25:36.5434535Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_logaddexp_cuda_float32 PASSED [0.1085s] [ 28%] 2025-12-04T15:25:36.5435785Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_logsumexp_cuda_float32 PASSED [0.4388s] [ 28%] 2025-12-04T15:25:36.5436994Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_mean_cuda_float32 PASSED [0.4294s] [ 28%] 2025-12-04T15:25:36.5438262Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_median_cuda_float32 PASSED [0.0768s] [ 28%] 2025-12-04T15:25:36.5439466Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_norm_cuda_float32 PASSED [1.8998s] [ 28%] 2025-12-04T15:25:36.5440685Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_normalize_cuda_float32 PASSED [0.1377s] [ 28%] 2025-12-04T15:25:36.5441887Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_prod_cuda_float32 PASSED [0.3828s] [ 28%] 2025-12-04T15:25:36.5443088Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_scatter_cuda_float32 XFAIL [0.0177s] [ 28%] 2025-12-04T15:25:36.5444286Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_select_cuda_float32 XFAIL [0.2212s] [ 28%] 2025-12-04T15:25:36.5445491Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_softmax_cuda_float32 PASSED [0.2974s] [ 29%] 2025-12-04T15:25:36.5446698Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_softmin_cuda_float32 PASSED [0.1024s] [ 29%] 2025-12-04T15:25:36.5447901Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_std_cuda_float32 PASSED [0.4160s] [ 29%] 2025-12-04T15:25:36.5449078Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_sum_cuda_float32 PASSED [0.3752s] [ 29%] 2025-12-04T15:25:36.5450259Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_masked_var_cuda_float32 PASSED [0.4064s] [ 29%] 2025-12-04T15:25:36.5451426Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_matmul_cuda_float32 PASSED [0.1784s] [ 29%] 2025-12-04T15:25:36.5452572Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_matrix_exp_cuda_float32 PASSED [0.0421s] [ 29%] 2025-12-04T15:25:36.5453759Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_max_binary_cuda_float32 PASSED [1.0581s] [ 29%] 2025-12-04T15:25:36.5455040Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_max_pool2d_with_indices_backward_cuda_float32 PASSED [5.5018s] [ 29%] 2025-12-04T15:25:36.5456378Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_max_reduction_no_dim_cuda_float32 PASSED [0.0127s] [ 29%] 2025-12-04T15:25:36.5457735Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_max_reduction_with_dim_cuda_float32 PASSED [0.0238s] [ 29%] 2025-12-04T15:25:36.5458962Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_maximum_cuda_float32 PASSED [0.5840s] [ 29%] 2025-12-04T15:25:36.5460097Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mean_cuda_float32 PASSED [0.0722s] [ 29%] 2025-12-04T15:25:36.5461221Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_median_cuda_float32 PASSED [0.0601s] [ 29%] 2025-12-04T15:25:36.5462439Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_meshgrid_list_of_tensors_cuda_float32 PASSED [0.0116s] [ 29%] 2025-12-04T15:25:36.5463762Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_meshgrid_variadic_tensors_cuda_float32 PASSED [0.1709s] [ 29%] 2025-12-04T15:25:36.5465055Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_min_binary_cuda_float32 PASSED [1.0653s] [ 29%] 2025-12-04T15:25:36.5466318Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_min_reduction_no_dim_cuda_float32 PASSED [0.0127s] [ 29%] 2025-12-04T15:25:36.5467584Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_min_reduction_with_dim_cuda_float32 PASSED [0.0236s] [ 29%] 2025-12-04T15:25:36.5468809Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_minimum_cuda_float32 PASSED [0.5852s] [ 29%] 2025-12-04T15:25:36.5470021Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mm_cuda_float32 PASSED [0.0230s] [ 29%] 2025-12-04T15:25:36.5471134Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mode_cuda_float32 PASSED [0.0417s] [ 29%] 2025-12-04T15:25:36.5472251Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_movedim_cuda_float32 XFAIL [0.0525s] [ 30%] 2025-12-04T15:25:36.5473384Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_msort_cuda_float32 PASSED [0.2556s] [ 30%] 2025-12-04T15:25:36.5474509Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mul_cuda_float32 PASSED [1.6194s] [ 30%] 2025-12-04T15:25:36.5475662Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_multinomial_cuda_float32 XFAIL [0.0369s] [ 30%] 2025-12-04T15:25:36.5476790Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mv_cuda_float32 PASSED [0.2522s] [ 30%] 2025-12-04T15:25:36.5477997Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mvlgamma_mvlgamma_p_1_cuda_float32 PASSED [0.5177s] [ 30%] 2025-12-04T15:25:36.5479293Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mvlgamma_mvlgamma_p_3_cuda_float32 PASSED [0.5445s] [ 30%] 2025-12-04T15:25:36.5480593Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_mvlgamma_mvlgamma_p_5_cuda_float32 PASSED [0.5779s] [ 30%] 2025-12-04T15:25:36.5481807Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nan_to_num_cuda_float32 PASSED [0.2275s] [ 30%] 2025-12-04T15:25:36.5482972Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nanmean_cuda_float32 PASSED [0.1301s] [ 30%] 2025-12-04T15:25:36.5484130Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nanmedian_cuda_float32 PASSED [0.0592s] [ 30%] 2025-12-04T15:25:36.5485310Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nanquantile_cuda_float32 XFAIL [0.0090s] [ 30%] 2025-12-04T15:25:36.5486463Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nansum_cuda_float32 PASSED [0.3409s] [ 30%] 2025-12-04T15:25:36.5487718Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_narrow_copy_cuda_float32 PASSED [0.0983s] [ 30%] 2025-12-04T15:25:36.5489356Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_narrow_cuda_float32 XFAIL [0.0360s] [ 30%] 2025-12-04T15:25:36.5490948Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_native_batch_norm_cuda_float32 XFAIL [0.2217s] [ 30%] 2025-12-04T15:25:36.5492212Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_native_dropout_backward_cuda_float32 PASSED [0.2643s] [ 30%] 2025-12-04T15:25:36.5493504Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_native_layer_norm_cuda_float32 PASSED [0.2550s] [ 30%] 2025-12-04T15:25:36.5494685Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ne_cuda_float32 XFAIL [0.4611s] [ 30%] 2025-12-04T15:25:36.5495794Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_neg_cuda_float32 PASSED [0.4151s] [ 30%] 2025-12-04T15:25:36.5497159Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_new_empty_cuda_float32 PASSED [0.0253s] [ 30%] 2025-12-04T15:25:36.5498566Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_new_empty_strided_cuda_float32 SKIPPED [0.0003s] (Skipped!) [ 31%] 2025-12-04T15:25:36.5499880Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_new_full_cuda_float32 PASSED [0.0309s] [ 31%] 2025-12-04T15:25:36.5501048Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_new_ones_cuda_float32 PASSED [0.0299s] [ 31%] 2025-12-04T15:25:36.5502197Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_new_zeros_cuda_float32 PASSED [0.0306s] [ 31%] 2025-12-04T15:25:36.5503460Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nextafter_cuda_float32 PASSED [0.5831s] [ 31%] 2025-12-04T15:25:36.5504752Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_adaptive_avg_pool1d_cuda_float32 PASSED [0.0321s] [ 31%] 2025-12-04T15:25:36.5506162Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_adaptive_avg_pool2d_cuda_float32 PASSED [0.0486s] [ 31%] 2025-12-04T15:25:36.5507566Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_adaptive_avg_pool3d_cuda_float32 PASSED [0.2811s] [ 31%] 2025-12-04T15:25:36.5508954Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_adaptive_max_pool1d_cuda_float32 PASSED [0.0454s] [ 31%] 2025-12-04T15:25:36.5510355Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_adaptive_max_pool2d_cuda_float32 PASSED [0.1126s] [ 31%] 2025-12-04T15:25:36.5511760Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_adaptive_max_pool3d_cuda_float32 PASSED [0.0990s] [ 31%] 2025-12-04T15:25:36.5513134Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_alpha_dropout_cuda_float32 XFAIL [0.0378s] [ 31%] 2025-12-04T15:25:36.5514447Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_avg_pool1d_cuda_float32 PASSED [0.2546s] [ 31%] 2025-12-04T15:25:36.5515766Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_avg_pool2d_cuda_float32 PASSED [0.0464s] [ 31%] 2025-12-04T15:25:36.5517077Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_avg_pool3d_cuda_float32 PASSED [0.0496s] [ 31%] 2025-12-04T15:25:36.5518395Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_batch_norm_cuda_float32 PASSED [0.4612s] [ 31%] 2025-12-04T15:25:36.5519763Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_batch_norm_without_cudnn_cuda_float32 PASSED [0.4564s] [ 31%] 2025-12-04T15:25:36.5521137Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_bilinear_cuda_float32 XFAIL [0.0091s] [ 31%] 2025-12-04T15:25:36.5522493Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_binary_cross_entropy_cuda_float32 PASSED [0.2840s] [ 31%] 2025-12-04T15:25:36.5523960Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_binary_cross_entropy_with_logits_cuda_float32 PASSED [0.2062s] [ 31%] 2025-12-04T15:25:36.5525334Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_celu_cuda_float32 PASSED [0.2301s] [ 31%] 2025-12-04T15:25:36.5526649Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_channel_shuffle_cuda_float32 PASSED [0.0162s] [ 32%] 2025-12-04T15:25:36.5527977Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv1d_cuda_float32 PASSED [0.2725s] [ 32%] 2025-12-04T15:25:36.5529249Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_cuda_float32 PASSED [0.7723s] [ 32%] 2025-12-04T15:25:36.5530600Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_no_bias_cuda_float32 PASSED [0.0175s] [ 32%] 2025-12-04T15:25:36.5532056Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_stride_depthwise_with_bias_cuda_float32 PASSED [0.0244s] [ 32%] 2025-12-04T15:25:36.5533576Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_stride_groups_with_bias_cuda_float32 PASSED [0.0258s] [ 32%] 2025-12-04T15:25:36.5535041Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_stride_no_bias_cuda_float32 PASSED [0.0170s] [ 32%] 2025-12-04T15:25:36.5536556Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_stride_padding_no_bias_cuda_float32 PASSED [0.0176s] [ 32%] 2025-12-04T15:25:36.5538100Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_stride_padding_with_bias_cuda_float32 PASSED [0.0215s] [ 32%] 2025-12-04T15:25:36.5539579Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_stride_with_bias_cuda_float32 PASSED [0.0201s] [ 32%] 2025-12-04T15:25:36.5541099Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_strided_padding_dilation_no_bias_cuda_float32 PASSED [0.2557s] [ 32%] 2025-12-04T15:25:36.5542705Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_strided_padding_dilation_with_bias_cuda_float32 PASSED [0.2434s] [ 32%] 2025-12-04T15:25:36.5544208Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv2d_with_bias_cuda_float32 PASSED [0.2372s] [ 32%] 2025-12-04T15:25:36.5545524Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv3d_cuda_float32 PASSED [0.4627s] [ 32%] 2025-12-04T15:25:36.5546846Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv_transpose1d_cuda_float32 PASSED [0.2211s] [ 32%] 2025-12-04T15:25:36.5548220Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv_transpose2d_cuda_float32 PASSED [0.2633s] [ 32%] 2025-12-04T15:25:36.5549601Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_conv_transpose3d_cuda_float32 PASSED [0.2485s] [ 32%] 2025-12-04T15:25:36.5550989Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_cosine_embedding_loss_cuda_float32 PASSED [0.1194s] [ 32%] 2025-12-04T15:25:36.5552406Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_cosine_similarity_cuda_float32 PASSED [0.0748s] [ 32%] 2025-12-04T15:25:36.5553776Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_cross_entropy_cuda_float32 PASSED [0.1685s] [ 32%] 2025-12-04T15:25:36.5555092Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_ctc_loss_cuda_float32 XFAIL [0.0095s] [ 32%] 2025-12-04T15:25:36.5556383Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_dropout2d_cuda_float32 XFAIL [0.2534s] [ 32%] 2025-12-04T15:25:36.5557663Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_dropout3d_cuda_float32 XFAIL [0.2531s] [ 33%] 2025-12-04T15:25:36.5558942Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_dropout_cuda_float32 XFAIL [0.2508s] [ 33%] 2025-12-04T15:25:36.5560205Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_elu_cuda_float32 PASSED [0.4441s] [ 33%] 2025-12-04T15:25:36.5561501Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_embedding_bag_cuda_float32 XFAIL [0.0085s] [ 33%] 2025-12-04T15:25:36.5562848Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_embedding_cuda_float32 XFAIL [0.2547s] [ 33%] 2025-12-04T15:25:36.5564205Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_embedding_functorch_cuda_float32 PASSED [0.2478s] [ 33%] 2025-12-04T15:25:36.5565700Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_feature_alpha_dropout_with_train_cuda_float32 XFAIL [0.0237s] [ 33%] 2025-12-04T15:25:36.5567240Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_feature_alpha_dropout_without_train_cuda_float32 PASSED [0.1433s] [ 33%] 2025-12-04T15:25:36.5568773Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_fractional_max_pool2d_cuda_float32 XFAIL [0.0099s] [ 33%] 2025-12-04T15:25:36.5570193Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_fractional_max_pool3d_cuda_float32 XFAIL [0.2279s] [ 33%] 2025-12-04T15:25:36.5571594Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_gaussian_nll_loss_cuda_float32 XFAIL [0.2431s] [ 33%] 2025-12-04T15:25:36.5572916Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_gelu_cuda_float32 PASSED [0.4769s] [ 33%] 2025-12-04T15:25:36.5574164Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_glu_cuda_float32 PASSED [0.1433s] [ 33%] 2025-12-04T15:25:36.5575434Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_grid_sample_cuda_float32 PASSED [0.2859s] [ 33%] 2025-12-04T15:25:36.5576760Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_group_norm_cuda_float32 PASSED [0.2199s] [ 33%] 2025-12-04T15:25:36.5578164Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_hardshrink_cuda_float32 PASSED [0.1436s] [ 33%] 2025-12-04T15:25:36.5579497Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_hardsigmoid_cuda_float32 PASSED [0.2255s] [ 33%] 2025-12-04T15:25:36.5580809Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_hardswish_cuda_float32 PASSED [0.0184s] [ 33%] 2025-12-04T15:25:36.5582119Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_hardtanh_cuda_float32 PASSED [0.1440s] [ 33%] 2025-12-04T15:25:36.5583482Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_hinge_embedding_loss_cuda_float32 PASSED [0.1558s] [ 33%] 2025-12-04T15:25:36.5584854Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_huber_loss_cuda_float32 PASSED [0.0733s] [ 33%] 2025-12-04T15:25:36.5586223Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_instance_norm_cuda_float32 PASSED [0.0118s] [ 34%] 2025-12-04T15:25:36.5587593Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_interpolate_area_cuda_float32 PASSED [0.0820s] [ 34%] 2025-12-04T15:25:36.5588998Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_interpolate_bicubic_cuda_float32 PASSED [0.1069s] [ 34%] 2025-12-04T15:25:36.5590423Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_interpolate_bilinear_cuda_float32 PASSED [0.1040s] [ 34%] 2025-12-04T15:25:36.5591819Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_interpolate_linear_cuda_float32 PASSED [0.0650s] [ 34%] 2025-12-04T15:25:36.5593255Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_interpolate_nearest-exact_cuda_float32 XFAIL [0.0080s] [ 34%] 2025-12-04T15:25:36.5594691Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_interpolate_nearest_cuda_float32 PASSED [0.0796s] [ 34%] 2025-12-04T15:25:36.5596363Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_interpolate_trilinear_cuda_float32 PASSED [0.0876s] [ 34%] 2025-12-04T15:25:36.5597762Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_kl_div_cuda_float32 XFAIL [0.0085s] [ 34%] 2025-12-04T15:25:36.5599028Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_l1_loss_cuda_float32 PASSED [0.2766s] [ 34%] 2025-12-04T15:25:36.5600331Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_layer_norm_cuda_float32 PASSED [0.1305s] [ 34%] 2025-12-04T15:25:36.5601744Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_leaky_relu_cuda_float32 PASSED [0.0567s] [ 34%] 2025-12-04T15:25:36.5603031Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_linear_cuda_float32 PASSED [0.1715s] [ 34%] 2025-12-04T15:25:36.5604363Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_local_response_norm_cuda_float32 PASSED [0.0520s] [ 34%] 2025-12-04T15:25:36.5605729Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_logsigmoid_cuda_float32 PASSED [0.1283s] [ 34%] 2025-12-04T15:25:36.5607092Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_margin_ranking_loss_cuda_float32 XFAIL [0.0167s] [ 34%] 2025-12-04T15:25:36.5608448Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_max_pool1d_cuda_float32 PASSED [3.6066s] [ 34%] 2025-12-04T15:25:36.5609752Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_max_pool2d_cuda_float32 PASSED [5.4914s] [ 34%] 2025-12-04T15:25:36.5611064Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_max_pool3d_cuda_float32 PASSED [2.2633s] [ 34%] 2025-12-04T15:25:36.5612385Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_max_unpool1d_cuda_float32 XFAIL [0.0089s] [ 34%] 2025-12-04T15:25:36.5613735Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_max_unpool1d_grad_cuda_float32 XFAIL [0.2211s] [ 34%] 2025-12-04T15:25:36.5615078Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_max_unpool2d_cuda_float32 XFAIL [0.2152s] [ 35%] 2025-12-04T15:25:36.5616406Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_max_unpool2d_grad_cuda_float32 XFAIL [0.2194s] [ 35%] 2025-12-04T15:25:36.5617815Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_max_unpool3d_cuda_float32 XFAIL [0.2173s] [ 35%] 2025-12-04T15:25:36.5619162Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_max_unpool3d_grad_cuda_float32 XFAIL [0.2231s] [ 35%] 2025-12-04T15:25:36.5620479Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_mish_cuda_float32 PASSED [0.4366s] [ 35%] 2025-12-04T15:25:36.5621749Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_mse_loss_cuda_float32 PASSED [0.0421s] [ 35%] 2025-12-04T15:25:36.5623098Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_mse_loss_functorch_cuda_float32 PASSED [0.0612s] [ 35%] 2025-12-04T15:25:36.5624525Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_multi_head_attention_forward_cuda_float32 XFAIL [3.4450s] [ 35%] 2025-12-04T15:25:36.5625956Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_multi_margin_loss_cuda_float32 XFAIL [0.2514s] [ 35%] 2025-12-04T15:25:36.5627340Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_multilabel_margin_loss_cuda_float32 XFAIL [0.2346s] [ 35%] 2025-12-04T15:25:36.5628844Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_multilabel_soft_margin_loss_cuda_float32 PASSED [0.2493s] [ 35%] 2025-12-04T15:25:36.5630269Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_nll_loss_cuda_float32 PASSED [0.5092s] [ 35%] 2025-12-04T15:25:36.5631575Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_normalize_cuda_float32 PASSED [0.0358s] [ 35%] 2025-12-04T15:25:36.5632865Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_one_hot_cuda_int64 XFAIL [0.0080s] [ 35%] 2025-12-04T15:25:36.5634213Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_pad_circular_cuda_float32 PASSED [0.0344s] [ 35%] 2025-12-04T15:25:36.5635552Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_pad_constant_cuda_float32 PASSED [0.1764s] [ 35%] 2025-12-04T15:25:36.5636888Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_pad_reflect_cuda_float32 PASSED [0.0461s] [ 35%] 2025-12-04T15:25:36.5638230Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_pad_replicate_cuda_float32 PASSED [0.0452s] [ 35%] 2025-12-04T15:25:36.5639615Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_pad_replicate_negative_cuda_float32 PASSED [0.0272s] [ 35%] 2025-12-04T15:25:36.5641031Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_pairwise_distance_cuda_float32 PASSED [0.0493s] [ 35%] 2025-12-04T15:25:36.5642371Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_pdist_cuda_float32 XFAIL [0.0082s] [ 35%] 2025-12-04T15:25:36.5643673Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_pixel_shuffle_cuda_float32 PASSED [0.2349s] [ 35%] 2025-12-04T15:25:36.5645023Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_pixel_unshuffle_cuda_float32 PASSED [0.0202s] [ 36%] 2025-12-04T15:25:36.5646387Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_poisson_nll_loss_cuda_float32 PASSED [0.5417s] [ 36%] 2025-12-04T15:25:36.5647718Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_prelu_cuda_float32 PASSED [0.5113s] [ 36%] 2025-12-04T15:25:36.5648983Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_relu6_cuda_float32 PASSED [0.1275s] [ 36%] 2025-12-04T15:25:36.5650235Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_relu_cuda_float32 PASSED [0.1330s] [ 36%] 2025-12-04T15:25:36.5651507Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_rms_norm_cuda_float32 PASSED [0.0860s] [ 36%] 2025-12-04T15:25:36.5652786Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_rrelu_cuda_float32 XFAIL [0.0148s] [ 36%] 2025-12-04T15:25:36.5654148Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_scaled_dot_product_attention_cuda_float32 XFAIL [0.3076s] [ 36%] 2025-12-04T15:25:36.5655516Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_selu_cuda_float32 PASSED [0.4388s] [ 36%] 2025-12-04T15:25:36.5656872Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_silu_complex_cuda_complex64 PASSED [0.3013s] [ 36%] 2025-12-04T15:25:36.5658180Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_silu_cuda_float32 PASSED [0.2263s] [ 36%] 2025-12-04T15:25:36.5659495Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_smooth_l1_loss_cuda_float32 PASSED [0.0549s] [ 36%] 2025-12-04T15:25:36.5660887Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_soft_margin_loss_cuda_float32 XFAIL [0.0156s] [ 36%] 2025-12-04T15:25:36.5662196Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_softmin_cuda_float32 PASSED [0.2528s] [ 36%] 2025-12-04T15:25:36.5663560Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_softmin_with_dtype_cuda_float32 PASSED [0.0286s] [ 36%] 2025-12-04T15:25:36.5664905Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_softplus_cuda_float32 PASSED [0.1266s] [ 36%] 2025-12-04T15:25:36.5666220Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_softshrink_cuda_float32 PASSED [0.1413s] [ 36%] 2025-12-04T15:25:36.5667585Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_softsign_cuda_float32 PASSED [0.1377s] [ 36%] 2025-12-04T15:25:36.5668897Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_tanhshrink_cuda_float32 PASSED [0.1307s] [ 36%] 2025-12-04T15:25:36.5670205Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_threshold_cuda_float32 PASSED [0.2518s] [ 36%] 2025-12-04T15:25:36.5671552Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_triplet_margin_loss_cuda_float32 XFAIL [0.0388s] [ 36%] 2025-12-04T15:25:36.5672997Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_triplet_margin_with_distance_loss_cuda_float32 PASSED [0.3825s] [ 37%] 2025-12-04T15:25:36.5674400Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_unfold_cuda_float32 PASSED [0.5601s] [ 37%] 2025-12-04T15:25:36.5675734Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_upsample_bilinear_cuda_float32 PASSED [0.0292s] [ 37%] 2025-12-04T15:25:36.5677123Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nn_functional_upsample_nearest_cuda_float32 PASSED [0.0455s] [ 37%] 2025-12-04T15:25:36.5678389Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nonzero_cuda_float32 XFAIL [0.0089s] [ 37%] 2025-12-04T15:25:36.5679646Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_nonzero_static_cuda_float32 SKIPPED [0.2132s] (Only runs on cpu) [ 37%] 2025-12-04T15:25:36.5680909Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_norm_cuda_float32 PASSED [0.1389s] [ 37%] 2025-12-04T15:25:36.5682049Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_norm_fro_cuda_float32 PASSED [0.0162s] [ 37%] 2025-12-04T15:25:36.5683207Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_norm_inf_cuda_float32 PASSED [0.2270s] [ 37%] 2025-12-04T15:25:36.5684350Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_norm_nuc_cuda_float32 PASSED [0.0201s] [ 37%] 2025-12-04T15:25:36.5685491Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_normal_cuda_float32 XFAIL [0.0099s] [ 37%] 2025-12-04T15:25:36.5686803Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_normal_in_place_cuda_float32 SKIPPED [0.0002s] (Test expects tensor input) [ 37%] 2025-12-04T15:25:36.5688173Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_normal_number_mean_cuda_float32 XFAIL [0.2246s] [ 37%] 2025-12-04T15:25:36.5689349Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ones_cuda_float32 PASSED [0.2252s] [ 37%] 2025-12-04T15:25:36.5690493Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ones_like_cuda_float32 PASSED [0.0293s] [ 37%] 2025-12-04T15:25:36.5691868Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ops_aten__new_zeros_with_same_feature_meta_functorchonly_cuda_float32 PASSED [0.0543s] [ 37%] 2025-12-04T15:25:36.5693376Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ops_aten_index_put_functorch_cuda_float32 PASSED [0.5952s] [ 37%] 2025-12-04T15:25:36.5694597Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ormqr_cuda_float32 XFAIL [0.0183s] [ 37%] 2025-12-04T15:25:36.5695748Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_outer_cuda_float32 PASSED [0.2362s] [ 37%] 2025-12-04T15:25:36.5697119Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_pca_lowrank_cuda_float32 XFAIL [0.0129s] [ 37%] 2025-12-04T15:25:36.5698312Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_permute_copy_cuda_float32 XFAIL [0.2264s] [ 37%] 2025-12-04T15:25:36.5699594Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_permute_cuda_float32 PASSED [0.4109s] [ 38%] 2025-12-04T15:25:36.5700757Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_pinverse_cuda_float32 PASSED [0.0464s] [ 38%] 2025-12-04T15:25:36.5701906Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_polar_cuda_float32 PASSED [0.6346s] [ 38%] 2025-12-04T15:25:36.5703128Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_polygamma_polygamma_n_0_cuda_float32 PASSED [0.2161s] [ 38%] 2025-12-04T15:25:36.5704434Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_polygamma_polygamma_n_1_cuda_float32 PASSED [2.4636s] [ 38%] 2025-12-04T15:25:36.5705747Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_polygamma_polygamma_n_2_cuda_float32 PASSED [0.3530s] [ 38%] 2025-12-04T15:25:36.5707064Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_polygamma_polygamma_n_3_cuda_float32 PASSED [0.2713s] [ 38%] 2025-12-04T15:25:36.5708377Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_polygamma_polygamma_n_4_cuda_float32 PASSED [0.2535s] [ 38%] 2025-12-04T15:25:36.5709602Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_positive_cuda_float32 PASSED [0.1098s] [ 38%] 2025-12-04T15:25:36.5710729Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_pow_cuda_float32 PASSED [0.6130s] [ 38%] 2025-12-04T15:25:36.5711840Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_prod_cuda_float32 PASSED [0.1231s] [ 38%] 2025-12-04T15:25:36.5712948Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_put_cuda_float32 XFAIL [0.0083s] [ 38%] 2025-12-04T15:25:36.5714028Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_qr_cuda_float32 PASSED [0.1659s] [ 38%] 2025-12-04T15:25:36.5715159Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_quantile_cuda_float32 XFAIL [0.0088s] [ 38%] 2025-12-04T15:25:36.5716303Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_rad2deg_cuda_float32 PASSED [0.4447s] [ 38%] 2025-12-04T15:25:36.5717450Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_rand_like_cuda_float32 XFAIL [0.0096s] [ 38%] 2025-12-04T15:25:36.5718729Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_randint_cuda_float32 SKIPPED [0.0002s] (Test expects tensor input) [ 38%] 2025-12-04T15:25:36.5720030Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_randint_like_cuda_float32 XFAIL [0.2192s] [ 38%] 2025-12-04T15:25:36.5721324Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_randn_cuda_float32 SKIPPED [0.0003s] (Test expects tensor input) [ 38%] 2025-12-04T15:25:36.5722612Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_randn_like_cuda_float32 XFAIL [0.2156s] [ 38%] 2025-12-04T15:25:36.5723753Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_ravel_cuda_float32 PASSED [0.2338s] [ 38%] 2025-12-04T15:25:36.5724860Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_real_cuda_float32 PASSED [0.1203s] [ 38%] 2025-12-04T15:25:36.5726061Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_reciprocal_cuda_float32 PASSED [0.2263s] [ 39%] 2025-12-04T15:25:36.5727285Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_remainder_cuda_float32 PASSED [0.6260s] [ 39%] 2025-12-04T15:25:36.5728431Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_renorm_cuda_float32 XFAIL [0.0127s] [ 39%] 2025-12-04T15:25:36.5729549Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_repeat_cuda_float32 PASSED [0.3528s] [ 39%] 2025-12-04T15:25:36.5730835Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_repeat_interleave_cuda_float32 PASSED [0.0232s] [ 39%] 2025-12-04T15:25:36.5732052Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_reshape_as_cuda_float32 PASSED [0.1596s] [ 39%] 2025-12-04T15:25:36.5733224Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_reshape_cuda_float32 PASSED [0.1479s] [ 39%] 2025-12-04T15:25:36.5734348Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_resize__cuda_float32 XFAIL [0.0115s] [ 39%] 2025-12-04T15:25:36.5735498Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_resize_as__cuda_float32 XFAIL [0.2253s] [ 39%] 2025-12-04T15:25:36.5736682Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_resolve_conj_cuda_float32 PASSED [0.2302s] [ 39%] 2025-12-04T15:25:36.5737947Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_resolve_neg_cuda_float32 PASSED [0.0121s] [ 39%] 2025-12-04T15:25:36.5739092Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_roll_cuda_float32 PASSED [0.0678s] [ 39%] 2025-12-04T15:25:36.5740216Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_rot90_cuda_float32 PASSED [0.1293s] [ 39%] 2025-12-04T15:25:36.5741342Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_round_cuda_float32 PASSED [0.2943s] [ 39%] 2025-12-04T15:25:36.5742529Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_round_decimals_0_cuda_float32 PASSED [0.3218s] [ 39%] 2025-12-04T15:25:36.5743752Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_round_decimals_3_cuda_float32 PASSED [0.3221s] [ 39%] 2025-12-04T15:25:36.5745012Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_round_decimals_neg_3_cuda_float32 PASSED [0.3196s] [ 39%] 2025-12-04T15:25:36.5746214Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_rsqrt_cuda_float32 PASSED [0.2267s] [ 39%] 2025-12-04T15:25:36.5747342Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_rsub_cuda_float32 PASSED [0.6356s] [ 39%] 2025-12-04T15:25:36.5748492Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_scalar_tensor_cuda_float32 PASSED [0.0072s] [ 39%] 2025-12-04T15:25:36.5749696Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_scatter_add_cuda_float32 PASSED [0.1648s] [ 39%] 2025-12-04T15:25:36.5750875Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_scatter_cuda_float32 PASSED [0.3129s] [ 40%] 2025-12-04T15:25:36.5752079Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_scatter_reduce_amax_cuda_float32 PASSED [0.4613s] [ 40%] 2025-12-04T15:25:36.5753334Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_scatter_reduce_amin_cuda_float32 PASSED [0.4610s] [ 40%] 2025-12-04T15:25:36.5754614Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_scatter_reduce_mean_cuda_float32 PASSED [0.5382s] [ 40%] 2025-12-04T15:25:36.5755886Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_scatter_reduce_prod_cuda_float32 PASSED [0.4596s] [ 40%] 2025-12-04T15:25:36.5757195Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_scatter_reduce_sum_cuda_float32 PASSED [0.4571s] [ 40%] 2025-12-04T15:25:36.5758408Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_searchsorted_cuda_float32 XFAIL [1.4206s] [ 40%] 2025-12-04T15:25:36.5759613Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_select_cuda_float32 PASSED [0.2641s] [ 40%] 2025-12-04T15:25:36.5760795Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_select_scatter_cuda_float32 PASSED [0.2486s] [ 40%] 2025-12-04T15:25:36.5761963Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sgn_cuda_float32 PASSED [0.2066s] [ 40%] 2025-12-04T15:25:36.5763125Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_short_cuda_float32 XFAIL [0.0199s] [ 40%] 2025-12-04T15:25:36.5764378Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_short_functorch_no_channels_last_cuda_float32 PASSED [0.2288s] [ 40%] 2025-12-04T15:25:36.5765654Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sigmoid_cuda_float32 PASSED [0.4084s] [ 40%] 2025-12-04T15:25:36.5766789Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sign_cuda_float32 PASSED [0.2073s] [ 40%] 2025-12-04T15:25:36.5768042Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_bartlett_cuda_float32 SKIPPED [0.0003s] (Skipped!) [ 40%] 2025-12-04T15:25:36.5769456Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_blackman_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 40%] 2025-12-04T15:25:36.5770866Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_cosine_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 40%] 2025-12-04T15:25:36.5772292Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_exponential_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 40%] 2025-12-04T15:25:36.5773723Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_gaussian_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 40%] 2025-12-04T15:25:36.5775163Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_general_cosine_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 40%] 2025-12-04T15:25:36.5776649Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_general_hamming_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 40%] 2025-12-04T15:25:36.5778169Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_hamming_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 41%] 2025-12-04T15:25:36.5779559Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_hann_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 41%] 2025-12-04T15:25:36.5780926Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_kaiser_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 41%] 2025-12-04T15:25:36.5782319Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signal_windows_nuttall_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 41%] 2025-12-04T15:25:36.5783601Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_signbit_cuda_float32 PASSED [0.1144s] [ 41%] 2025-12-04T15:25:36.5784736Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sin_cuda_float32 PASSED [0.1946s] [ 41%] 2025-12-04T15:25:36.5785832Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sinc_cuda_float32 PASSED [0.3040s] [ 41%] 2025-12-04T15:25:36.5786951Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sinh_cuda_float32 PASSED [0.2069s] [ 41%] 2025-12-04T15:25:36.5788067Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_slice_cuda_float32 PASSED [0.0187s] [ 41%] 2025-12-04T15:25:36.5789279Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_slice_scatter_cuda_float32 PASSED [0.0654s] [ 41%] 2025-12-04T15:25:36.5790446Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_softmax_cuda_float32 PASSED [0.0587s] [ 41%] 2025-12-04T15:25:36.5791680Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_softmax_with_dtype_cuda_float32 PASSED [0.0619s] [ 41%] 2025-12-04T15:25:36.5792869Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sort_cuda_float32 PASSED [0.1641s] [ 41%] 2025-12-04T15:25:36.5794132Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sparse_mm_reduce_cuda_float32 SKIPPED [0.0017s] (Only runs on cpu) [ 41%] 2025-12-04T15:25:36.5795523Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sparse_sampled_addmm_cuda_float32 XFAIL [0.0086s] [ 41%] 2025-12-04T15:25:36.5796913Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_airy_ai_cuda_float32 XFAIL [0.2464s] [ 41%] 2025-12-04T15:25:36.5798148Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_bessel_j0_cuda_float32 PASSED [2.1683s] [ 41%] 2025-12-04T15:25:36.5799404Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_bessel_j1_cuda_float32 PASSED [0.2633s] [ 41%] 2025-12-04T15:25:36.5800638Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_bessel_y0_cuda_float32 PASSED [2.7439s] [ 41%] 2025-12-04T15:25:36.5801887Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_bessel_y1_cuda_float32 PASSED [1.2832s] [ 41%] 2025-12-04T15:25:36.5803203Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_chebyshev_polynomial_t_cuda_float32 XFAIL [0.0138s] [ 41%] 2025-12-04T15:25:36.5804584Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_chebyshev_polynomial_u_cuda_float32 XFAIL [0.2443s] [ 41%] 2025-12-04T15:25:36.5805957Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_chebyshev_polynomial_v_cuda_float32 XFAIL [0.6314s] [ 42%] 2025-12-04T15:25:36.5807320Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_chebyshev_polynomial_w_cuda_float32 XFAIL [0.2183s] [ 42%] 2025-12-04T15:25:36.5808613Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_entr_cuda_float32 PASSED [0.4292s] [ 42%] 2025-12-04T15:25:36.5809819Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_erfcx_cuda_float32 PASSED [6.0289s] [ 42%] 2025-12-04T15:25:36.5811098Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_hermite_polynomial_h_cuda_float32 XFAIL [0.0134s] [ 42%] 2025-12-04T15:25:36.5812437Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_hermite_polynomial_he_cuda_float32 XFAIL [0.2356s] [ 42%] 2025-12-04T15:25:36.5813715Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_i0e_cuda_float32 PASSED [0.3375s] [ 42%] 2025-12-04T15:25:36.5814900Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_i1_cuda_float32 PASSED [0.5105s] [ 42%] 2025-12-04T15:25:36.5816079Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_i1e_cuda_float32 PASSED [0.1230s] [ 42%] 2025-12-04T15:25:36.5817398Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_laguerre_polynomial_l_cuda_float32 XFAIL [0.1086s] [ 42%] 2025-12-04T15:25:36.5818762Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_legendre_polynomial_p_cuda_float32 XFAIL [0.3319s] [ 42%] 2025-12-04T15:25:36.5820057Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_log_ndtr_cuda_float32 XFAIL [0.2170s] [ 42%] 2025-12-04T15:25:36.5821334Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_modified_bessel_i0_cuda_float32 PASSED [0.7321s] [ 42%] 2025-12-04T15:25:36.5822735Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_modified_bessel_i1_cuda_float32 PASSED [0.3807s] [ 42%] 2025-12-04T15:25:36.5824121Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_modified_bessel_k0_cuda_float32 PASSED [0.6599s] [ 42%] 2025-12-04T15:25:36.5825461Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_modified_bessel_k1_cuda_float32 PASSED [0.5845s] [ 42%] 2025-12-04T15:25:36.5826733Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_ndtr_cuda_float32 PASSED [0.1374s] [ 42%] 2025-12-04T15:25:36.5828009Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_ndtri_cuda_float32 PASSED [0.4464s] [ 42%] 2025-12-04T15:25:36.5829347Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_polygamma_special_polygamma_n_0_cuda_float32 PASSED [0.2091s] [ 42%] 2025-12-04T15:25:36.5830785Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_scaled_modified_bessel_k0_cuda_float32 PASSED [1.5188s] [ 42%] 2025-12-04T15:25:36.5832188Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_scaled_modified_bessel_k1_cuda_float32 PASSED [0.5869s] [ 42%] 2025-12-04T15:25:36.5833614Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_shifted_chebyshev_polynomial_t_cuda_float32 XFAIL [0.0135s] [ 43%] 2025-12-04T15:25:36.5835050Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_shifted_chebyshev_polynomial_u_cuda_float32 XFAIL [0.2586s] [ 43%] 2025-12-04T15:25:36.5836500Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_shifted_chebyshev_polynomial_v_cuda_float32 XFAIL [0.6467s] [ 43%] 2025-12-04T15:25:36.5837948Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_shifted_chebyshev_polynomial_w_cuda_float32 XFAIL [0.6430s] [ 43%] 2025-12-04T15:25:36.5839348Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_spherical_bessel_j0_cuda_float32 PASSED [0.6071s] [ 43%] 2025-12-04T15:25:36.5840625Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_xlog1py_cuda_float32 PASSED [0.6193s] [ 43%] 2025-12-04T15:25:36.5841833Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_special_zeta_cuda_float32 PASSED [3.1421s] [ 43%] 2025-12-04T15:25:36.5842994Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_split_cuda_float32 PASSED [0.0163s] [ 43%] 2025-12-04T15:25:36.5844173Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_split_list_args_cuda_float32 PASSED [0.0238s] [ 43%] 2025-12-04T15:25:36.5845406Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_split_with_sizes_copy_cuda_float32 PASSED [0.0296s] [ 43%] 2025-12-04T15:25:36.5846662Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_split_with_sizes_cuda_float32 PASSED [0.0281s] [ 43%] 2025-12-04T15:25:36.5847835Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sqrt_cuda_float32 PASSED [0.2088s] [ 43%] 2025-12-04T15:25:36.5848966Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_square_cuda_float32 PASSED [0.1269s] [ 43%] 2025-12-04T15:25:36.5850117Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_squeeze_copy_cuda_float32 XFAIL [0.0083s] [ 43%] 2025-12-04T15:25:36.5851288Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_squeeze_cuda_float32 PASSED [0.0506s] [ 43%] 2025-12-04T15:25:36.5852479Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_squeeze_multiple_cuda_float32 PASSED [0.0411s] [ 43%] 2025-12-04T15:25:36.5853661Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_stack_cuda_float32 PASSED [0.0099s] [ 43%] 2025-12-04T15:25:36.5854803Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_std_cuda_float32 PASSED [0.0491s] [ 43%] 2025-12-04T15:25:36.5855988Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_std_mean_cuda_float32 PASSED [0.0667s] [ 43%] 2025-12-04T15:25:36.5857245Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_std_mean_unbiased_cuda_float32 PASSED [0.0152s] [ 43%] 2025-12-04T15:25:36.5858464Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_std_unbiased_cuda_float32 PASSED [0.0127s] [ 43%] 2025-12-04T15:25:36.5859670Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_stft_cuda_float32 XFAIL [0.0093s] [ 44%] 2025-12-04T15:25:36.5860778Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sub_cuda_float32 PASSED [1.4228s] [ 44%] 2025-12-04T15:25:36.5861893Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sum_cuda_float32 PASSED [0.0657s] [ 44%] 2025-12-04T15:25:36.5863056Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_sum_to_size_cuda_float32 PASSED [0.0643s] [ 44%] 2025-12-04T15:25:36.5864187Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_svd_cuda_float32 PASSED [1.1825s] [ 44%] 2025-12-04T15:25:36.5865336Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_svd_lowrank_cuda_float32 XFAIL [0.0124s] [ 44%] 2025-12-04T15:25:36.5866490Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_t_copy_cuda_float32 XFAIL [0.2376s] [ 44%] 2025-12-04T15:25:36.5867608Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_t_cuda_float32 PASSED [0.4591s] [ 44%] 2025-12-04T15:25:36.5868748Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_take_along_dim_cuda_float32 PASSED [0.0391s] [ 44%] 2025-12-04T15:25:36.5869912Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_take_cuda_float32 XFAIL [0.0185s] [ 44%] 2025-12-04T15:25:36.5871023Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_tan_cuda_float32 PASSED [0.2095s] [ 44%] 2025-12-04T15:25:36.5872145Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_tanh_cuda_float32 PASSED [0.2930s] [ 44%] 2025-12-04T15:25:36.5873285Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_tensor_split_cuda_float32 XFAIL [0.0082s] [ 44%] 2025-12-04T15:25:36.5874466Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_tensordot_cuda_float32 PASSED [0.0272s] [ 44%] 2025-12-04T15:25:36.5875619Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_tile_cuda_float32 PASSED [0.1703s] [ 44%] 2025-12-04T15:25:36.5876790Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_to_cuda_float32 SKIPPED [0.0003s] (Skipped!) [ 44%] 2025-12-04T15:25:36.5877964Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_to_sparse_cuda_float32 XFAIL [0.0087s] [ 44%] 2025-12-04T15:25:36.5879107Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_topk_cuda_float32 PASSED [0.0689s] [ 44%] 2025-12-04T15:25:36.5880371Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_torch__scaled_mm_cuda_float8_e4m3fn SKIPPED [0.0003s] (Skipped!) [ 44%] 2025-12-04T15:25:36.5881772Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_torch__scaled_mm_v2_cuda_float8_e4m3fn SKIPPED [0.0002s] (Skipped!) [ 44%] 2025-12-04T15:25:36.5883228Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_torch_ops_aten__efficient_attention_forward_cuda_float32 XFAIL [0.0107s] [ 44%] 2025-12-04T15:25:36.5884916Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_torch_ops_aten__flash_attention_forward_cuda_float16 SKIPPED [0.2322s] (This platform doesn't support Flash Attention) [ 44%] 2025-12-04T15:25:36.5886615Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_torch_ops_aten__safe_softmax_default_cuda_float32 PASSED [0.2532s] [ 45%] 2025-12-04T15:25:36.5887921Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_trace_cuda_float32 PASSED [0.2304s] [ 45%] 2025-12-04T15:25:36.5889096Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_transpose_copy_cuda_float32 XFAIL [0.0086s] [ 45%] 2025-12-04T15:25:36.5890279Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_transpose_cuda_float32 PASSED [0.3476s] [ 45%] 2025-12-04T15:25:36.5891509Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_trapezoid_cuda_float32 PASSED [0.0576s] [ 45%] 2025-12-04T15:25:36.5892656Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_trapz_cuda_float32 PASSED [0.0576s] [ 45%] 2025-12-04T15:25:36.5893982Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_triangular_solve_cuda_float32 SKIPPED [0.0016s] (no MAGMA library detected) [ 45%] 2025-12-04T15:25:36.5895289Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_tril_cuda_float32 XFAIL [0.0082s] [ 45%] 2025-12-04T15:25:36.5896592Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_tril_indices_cuda_int64 PASSED [0.2241s] [ 45%] 2025-12-04T15:25:36.5897801Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_triu_cuda_float32 XFAIL [0.0082s] [ 45%] 2025-12-04T15:25:36.5898947Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_triu_indices_cuda_int64 PASSED [0.2256s] [ 45%] 2025-12-04T15:25:36.5900123Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_true_divide_cuda_float32 PASSED [0.8779s] [ 45%] 2025-12-04T15:25:36.5901285Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_trunc_cuda_float32 PASSED [0.1972s] [ 45%] 2025-12-04T15:25:36.5902443Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unbind_copy_cuda_float32 XFAIL [0.0150s] [ 45%] 2025-12-04T15:25:36.5903609Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unbind_cuda_float32 PASSED [0.2771s] [ 45%] 2025-12-04T15:25:36.5904754Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unflatten_cuda_float32 PASSED [0.0374s] [ 45%] 2025-12-04T15:25:36.5905931Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unfold_copy_cuda_float32 PASSED [0.0776s] [ 45%] 2025-12-04T15:25:36.5907094Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unfold_cuda_float32 PASSED [0.0741s] [ 45%] 2025-12-04T15:25:36.5908226Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_uniform_cuda_float32 XFAIL [0.0148s] [ 45%] 2025-12-04T15:25:36.5909405Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unique_consecutive_cuda_float32 XFAIL [0.2233s] [ 45%] 2025-12-04T15:25:36.5910599Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unique_cuda_float32 XFAIL [0.2212s] [ 45%] 2025-12-04T15:25:36.5911748Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unravel_index_cuda_int64 PASSED [0.2899s] [ 46%] 2025-12-04T15:25:36.5912944Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unsafe_chunk_cuda_float32 PASSED [0.1226s] [ 46%] 2025-12-04T15:25:36.5914134Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unsafe_split_cuda_float32 PASSED [0.0151s] [ 46%] 2025-12-04T15:25:36.5915339Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unsqueeze_copy_cuda_float32 XFAIL [0.0083s] [ 46%] 2025-12-04T15:25:36.5916527Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_unsqueeze_cuda_float32 PASSED [0.2756s] [ 46%] 2025-12-04T15:25:36.5917753Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_var_cuda_float32 PASSED [0.0493s] [ 46%] 2025-12-04T15:25:36.5918875Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_var_mean_cuda_float32 PASSED [0.2816s] [ 46%] 2025-12-04T15:25:36.5920120Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_var_mean_unbiased_cuda_float32 PASSED [0.0158s] [ 46%] 2025-12-04T15:25:36.5921347Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_var_unbiased_cuda_float32 PASSED [0.0125s] [ 46%] 2025-12-04T15:25:36.5922509Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_vdot_cuda_float32 PASSED [0.2408s] [ 46%] 2025-12-04T15:25:36.5923737Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_view_as_complex_cuda_float32 XFAIL [0.0098s] [ 46%] 2025-12-04T15:25:36.5924920Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_view_as_cuda_float32 PASSED [0.3735s] [ 46%] 2025-12-04T15:25:36.5926102Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_view_as_real_cuda_complex64 PASSED [0.0123s] [ 46%] 2025-12-04T15:25:36.5927290Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_view_copy_cuda_float32 PASSED [0.0514s] [ 46%] 2025-12-04T15:25:36.5928421Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_view_cuda_float32 PASSED [0.1483s] [ 46%] 2025-12-04T15:25:36.5929549Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_vsplit_cuda_float32 PASSED [0.0264s] [ 46%] 2025-12-04T15:25:36.5930692Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_vstack_cuda_float32 PASSED [0.2332s] [ 46%] 2025-12-04T15:25:36.5931828Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_where_cuda_float32 PASSED [0.1863s] [ 46%] 2025-12-04T15:25:36.5932948Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_xlogy_cuda_float32 PASSED [1.3438s] [ 46%] 2025-12-04T15:25:36.5934070Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_zero__cuda_float32 PASSED [0.0244s] [ 46%] 2025-12-04T15:25:36.5935196Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_zeros_cuda_float32 PASSED [0.0069s] [ 46%] 2025-12-04T15:25:36.5936349Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_op_has_batch_rule_zeros_like_cuda_float32 PASSED [0.0284s] [ 47%] 2025-12-04T15:25:36.5937550Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_searchsorted_bucketize_cuda PASSED [0.0058s] [ 47%] 2025-12-04T15:25:36.5938558Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_slogdet_cuda PASSED [0.0050s] [ 47%] 2025-12-04T15:25:36.5939508Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_sum_scalar_cuda PASSED [0.0033s] [ 47%] 2025-12-04T15:25:36.5940627Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_torch_return_types_returns_cuda ('RERUN', {'yellow': True}) [0.0029s] [ 47%] 2025-12-04T15:25:36.5941902Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_torch_return_types_returns_cuda ('RERUN', {'yellow': True}) [0.2203s] [ 47%] 2025-12-04T15:25:36.5943097Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_torch_return_types_returns_cuda FAILED [0.2215s] [ 47%] 2025-12-04T15:25:36.5943726Z 2025-12-04T15:25:36.5943870Z ==================================== RERUNS ==================================== 2025-12-04T15:25:36.5944477Z _______ TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda _______ 2025-12-04T15:25:36.5945045Z Traceback (most recent call last): 2025-12-04T15:25:36.5945769Z File "/var/lib/jenkins/workspace/test/functorch/test_vmap.py", line 5059, in test_torch_return_types_returns 2025-12-04T15:25:36.5946659Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.5947518Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/apis.py", line 208, in wrapped 2025-12-04T15:25:36.5948188Z return vmap_impl( 2025-12-04T15:25:36.5948868Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py", line 283, in vmap_impl 2025-12-04T15:25:36.5949556Z return _flat_vmap( 2025-12-04T15:25:36.5950197Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py", line 433, in _flat_vmap 2025-12-04T15:25:36.5950946Z batched_outputs = func(*batched_inputs, **kwargs) 2025-12-04T15:25:36.5952014Z RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T15:25:36.5952973Z 2025-12-04T15:25:36.5953202Z To execute this test, run the following from the base repo dir: 2025-12-04T15:25:36.5954135Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/functorch/test_vmap.py TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda 2025-12-04T15:25:36.5954876Z 2025-12-04T15:25:36.5955149Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T15:25:36.5955393Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T15:25:36.5957345Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5048: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.5957609Z isinstance(vmap(torch.min, (0, None))(t, 0), torch.return_types.min) 2025-12-04T15:25:36.5959524Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5051: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.5959769Z isinstance(vmap(torch.max, (0, None))(t, 0), torch.return_types.max) 2025-12-04T15:25:36.5961703Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5055: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.5961932Z vmap(torch.topk, (0, None, None))(t, 1, 0), torch.return_types.topk 2025-12-04T15:25:36.5963856Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5059: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.5964133Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.5964463Z _______ TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda _______ 2025-12-04T15:25:36.5964586Z Traceback (most recent call last): 2025-12-04T15:25:36.5965059Z File "/var/lib/jenkins/workspace/test/functorch/test_vmap.py", line 5059, in test_torch_return_types_returns 2025-12-04T15:25:36.5965352Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.5965780Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/apis.py", line 208, in wrapped 2025-12-04T15:25:36.5965899Z return vmap_impl( 2025-12-04T15:25:36.5966393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py", line 283, in vmap_impl 2025-12-04T15:25:36.5966496Z return _flat_vmap( 2025-12-04T15:25:36.5966975Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py", line 433, in _flat_vmap 2025-12-04T15:25:36.5967145Z batched_outputs = func(*batched_inputs, **kwargs) 2025-12-04T15:25:36.5967920Z RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T15:25:36.5967928Z 2025-12-04T15:25:36.5968244Z To execute this test, run the following from the base repo dir: 2025-12-04T15:25:36.5968837Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/functorch/test_vmap.py TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda 2025-12-04T15:25:36.5968843Z 2025-12-04T15:25:36.5969124Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T15:25:36.5969350Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T15:25:36.5971293Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5048: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.5971542Z isinstance(vmap(torch.min, (0, None))(t, 0), torch.return_types.min) 2025-12-04T15:25:36.5973482Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5051: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.5973726Z isinstance(vmap(torch.max, (0, None))(t, 0), torch.return_types.max) 2025-12-04T15:25:36.5975647Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5055: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.5975889Z vmap(torch.topk, (0, None, None))(t, 1, 0), torch.return_types.topk 2025-12-04T15:25:36.5977873Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5059: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.5978168Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.5978316Z =================================== FAILURES =================================== 2025-12-04T15:25:36.5978648Z _______ TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda _______ 2025-12-04T15:25:36.5978773Z Traceback (most recent call last): 2025-12-04T15:25:36.5979245Z File "/var/lib/jenkins/workspace/test/functorch/test_vmap.py", line 5059, in test_torch_return_types_returns 2025-12-04T15:25:36.5979542Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.5979976Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/apis.py", line 208, in wrapped 2025-12-04T15:25:36.5980096Z return vmap_impl( 2025-12-04T15:25:36.5980577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py", line 283, in vmap_impl 2025-12-04T15:25:36.5980684Z return _flat_vmap( 2025-12-04T15:25:36.5981167Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py", line 433, in _flat_vmap 2025-12-04T15:25:36.5981339Z batched_outputs = func(*batched_inputs, **kwargs) 2025-12-04T15:25:36.5982103Z RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T15:25:36.5982180Z 2025-12-04T15:25:36.5982398Z To execute this test, run the following from the base repo dir: 2025-12-04T15:25:36.5982991Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/functorch/test_vmap.py TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda 2025-12-04T15:25:36.5982997Z 2025-12-04T15:25:36.5983278Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T15:25:36.5983502Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T15:25:36.5985446Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5048: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.5985695Z isinstance(vmap(torch.min, (0, None))(t, 0), torch.return_types.min) 2025-12-04T15:25:36.5987623Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5051: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.5987865Z isinstance(vmap(torch.max, (0, None))(t, 0), torch.return_types.max) 2025-12-04T15:25:36.5989783Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5055: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.5990022Z vmap(torch.topk, (0, None, None))(t, 1, 0), torch.return_types.topk 2025-12-04T15:25:36.5991933Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5059: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.5992222Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.5992895Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-2355f4f75c6e0706.xml - 2025-12-04T15:25:36.5993082Z =========================== short test summary info ============================ 2025-12-04T15:25:36.5994389Z FAILED [0.2215s] functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_torch_return_types_returns_cuda - RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T15:25:36.5994395Z 2025-12-04T15:25:36.5994624Z To execute this test, run the following from the base repo dir: 2025-12-04T15:25:36.5995245Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/functorch/test_vmap.py TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda 2025-12-04T15:25:36.5995250Z 2025-12-04T15:25:36.5995546Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T15:25:36.5995738Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T15:25:36.5996143Z = 1 failed, 807 passed, 45 skipped, 154 xfailed, 2 rerun in 215.67s (0:03:35) == 2025-12-04T15:25:36.5996243Z Got exit code 1 2025-12-04T15:25:36.5996635Z Retrying single test... 2025-12-04T15:25:36.5997137Z Test results will be stored in test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-db01917ff3ce307b.xml 2025-12-04T15:25:36.5997313Z ============================= test session starts ============================== 2025-12-04T15:25:36.5997661Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T15:25:36.5997772Z cachedir: .pytest_cache 2025-12-04T15:25:36.5998301Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T15:25:36.5998427Z rootdir: /var/lib/jenkins/workspace 2025-12-04T15:25:36.5998535Z configfile: pytest.ini 2025-12-04T15:25:36.5999146Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T15:25:36.5999373Z collecting ... collected 2134 items / 2133 deselected / 1 selected 2025-12-04T15:25:36.6000073Z stepcurrent: skipping 1006 already run items. Running only test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_torch_return_types_returns_cuda 2025-12-04T15:25:36.6000191Z Running 1 items in this shard 2025-12-04T15:25:36.6000196Z 2025-12-04T15:25:36.6001191Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_torch_return_types_returns_cuda [W1204 15:21:11.594673555 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T15:25:36.6001211Z 2025-12-04T15:25:36.6001344Z ('RERUN', {'yellow': True}) [15.4444s] [100%] 2025-12-04T15:25:36.6002331Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_torch_return_types_returns_cuda [W1204 15:21:27.260630785 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T15:25:36.6002337Z 2025-12-04T15:25:36.6002481Z ('RERUN', {'yellow': True}) [0.2235s] [100%] 2025-12-04T15:25:36.6003461Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_torch_return_types_returns_cuda [W1204 15:21:27.464869284 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T15:25:36.6003466Z 2025-12-04T15:25:36.6003582Z FAILED [0.2020s] [100%] 2025-12-04T15:25:36.6003587Z 2025-12-04T15:25:36.6003731Z ==================================== RERUNS ==================================== 2025-12-04T15:25:36.6004047Z _______ TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda _______ 2025-12-04T15:25:36.6004183Z Traceback (most recent call last): 2025-12-04T15:25:36.6004661Z File "/var/lib/jenkins/workspace/test/functorch/test_vmap.py", line 5059, in test_torch_return_types_returns 2025-12-04T15:25:36.6004948Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.6005377Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/apis.py", line 208, in wrapped 2025-12-04T15:25:36.6005487Z return vmap_impl( 2025-12-04T15:25:36.6005936Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py", line 283, in vmap_impl 2025-12-04T15:25:36.6006039Z return _flat_vmap( 2025-12-04T15:25:36.6006478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py", line 433, in _flat_vmap 2025-12-04T15:25:36.6006705Z batched_outputs = func(*batched_inputs, **kwargs) 2025-12-04T15:25:36.6007470Z RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T15:25:36.6008286Z Exception raised from apply_magma_eig at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebra.cpp:2024 (most recent call first): 2025-12-04T15:25:36.6008397Z C++ CapturedTraceback: 2025-12-04T15:25:36.6009707Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T15:25:36.6010281Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T15:25:36.6010620Z #6 c10::detail::torchCheckFail(char const*, char const*, unsigned int, char const*) from ??:0 2025-12-04T15:25:36.6011439Z #7 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool)::{lambda()#1}::operator()() const::{lambda()#1}::operator()() const [clone .constprop.0] from BatchLinearAlgebra.cpp:0 2025-12-04T15:25:36.6011839Z #8 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T15:25:36.6012288Z #9 at::native::linalg_eig_out_info(at::Tensor const&, at::Tensor&, at::Tensor&, at::Tensor&, bool) from BatchLinearAlgebra.cpp:0 2025-12-04T15:25:36.6012564Z #10 at::native::linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T15:25:36.6013173Z #11 at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6015582Z #12 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&, at::Tensor&, at::Tensor&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&, at::Tensor&, at::Tensor&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6015868Z #13 at::_ops::linalg_eig_out::call(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T15:25:36.6016049Z #14 at::native::linalg_eig(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6018140Z #15 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA__linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6018453Z #16 at::_ops::linalg_eig::redispatch(c10::DispatchKeySet, at::Tensor const&) from ??:0 2025-12-04T15:25:36.6018996Z #17 torch::autograd::VariableType::(anonymous namespace)::linalg_eig(c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T15:25:36.6021366Z #18 c10::impl::make_boxed_from_unboxed_functor (c10::DispatchKeySet, at::Tensor const&), &torch::autograd::VariableType::(anonymous namespace)::linalg_eig>, std::tuple, c10::guts::typelist::typelist >, false>::call(c10::OperatorKernel*, c10::OperatorHandle const&, c10::DispatchKeySet, std::vector >*) from VariableType_2.cpp:0 2025-12-04T15:25:36.6022034Z #19 c10::Dispatcher::callBoxed(c10::OperatorHandle const&, std::vector >*) const [clone .isra.0] from VmapInterpreter.cpp:0 2025-12-04T15:25:36.6022672Z #20 at::functorch::Interpreter::sendToNextInterpreter(c10::OperatorHandle const&, std::vector >*, bool) from :0 2025-12-04T15:25:36.6023267Z #21 at::functorch::dynamicLayerBack(c10::OperatorHandle const&, std::vector >*, bool) from DynamicLayer.cpp:0 2025-12-04T15:25:36.6024108Z #22 c10::impl::BoxedKernelWrapper (at::Tensor const&), void>::call(c10::BoxedKernel const&, c10::OperatorHandle const&, c10::DispatchKeySet, at::Tensor const&) from :0 2025-12-04T15:25:36.6024299Z #23 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6026386Z #24 std::tuple at::functorch::linalg_eig_generated_plumbing, at::Tensor, std::optional > (*)(at::Tensor const&, std::optional), &at::functorch::(anonymous namespace)::LinalgCheckMatrixUnaryRuleHelper<&at::functorch::func_string_linalg_eig, std::tuple (*)(at::Tensor const&), &at::_ops::linalg_eig::call, c10::guts::typelist::typelist >::apply_two>(at::Tensor const&) from BatchRulesLinearAlgebra.cpp:0 2025-12-04T15:25:36.6027989Z #25 c10::impl::make_boxed_from_unboxed_functor (*)(at::Tensor const&), std::tuple, c10::guts::typelist::typelist >, false>::call(c10::OperatorKernel*, c10::OperatorHandle const&, c10::DispatchKeySet, std::vector >*) from :0 2025-12-04T15:25:36.6028629Z #26 c10::Dispatcher::callBoxed(c10::OperatorHandle const&, std::vector >*) const [clone .isra.0] from VmapInterpreter.cpp:0 2025-12-04T15:25:36.6029153Z #27 at::functorch::Interpreter::process(c10::OperatorHandle const&, std::vector >*) from :0 2025-12-04T15:25:36.6029780Z #28 at::functorch::dynamicLayerFrontFallback(c10::OperatorHandle const&, std::vector >*) from DynamicLayer.cpp:0 2025-12-04T15:25:36.6030569Z #29 c10::impl::BoxedKernelWrapper (at::Tensor const&), void>::call(c10::BoxedKernel const&, c10::OperatorHandle const&, c10::DispatchKeySet, at::Tensor const&) from :0 2025-12-04T15:25:36.6030749Z #30 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6031153Z #31 torch::autograd::THPVariable_linalg_eig(_object*, _object*, _object*) from python_linalg_functions.cpp:0 2025-12-04T15:25:36.6031489Z #32 cfunction_call from /usr/local/src/conda/python-3.10.14/Objects/methodobject.c:543 2025-12-04T15:25:36.6031762Z #33 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T15:25:36.6032033Z #34 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5917 2025-12-04T15:25:36.6032411Z #35 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6032672Z #36 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6033059Z #37 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6033317Z #38 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6033688Z #39 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6034110Z #40 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6034514Z #41 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6034819Z #42 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T15:25:36.6035103Z #43 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6035474Z #44 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6035742Z #45 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6036110Z #46 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6036439Z #47 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6036808Z #48 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6037061Z #49 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6037443Z #50 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6037855Z #51 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6038244Z #52 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6038648Z #53 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6039017Z #54 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6039439Z #55 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6039810Z #56 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6040219Z #57 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6040603Z #58 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6041006Z #59 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6041389Z #60 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6041678Z #61 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T15:25:36.6041933Z #62 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6042316Z #63 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6042667Z #64 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6042983Z #65 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6043281Z #66 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6043582Z #67 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6044001Z #68 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6044373Z #69 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6044790Z #70 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6045161Z #71 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6045421Z #72 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6045800Z #73 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6046206Z #74 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6046616Z #75 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6047068Z #76 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6047439Z #77 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6047798Z #78 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6048101Z #79 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6048454Z #80 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6048737Z #81 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T15:25:36.6048992Z #82 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6049380Z #83 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6049786Z #84 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6050161Z #85 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6050581Z #86 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6050953Z #87 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6051212Z #88 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6051593Z #89 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6051997Z #90 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6052382Z #91 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6052786Z #92 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6053160Z #93 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6053428Z #94 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6053798Z #95 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6054215Z #96 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6054591Z #97 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6054992Z #98 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6055377Z #99 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6055736Z #100 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6056061Z #101 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6056358Z #102 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6056670Z #103 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6057185Z #104 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6057572Z #105 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6057838Z #106 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6058233Z #107 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6058703Z #108 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6059096Z #109 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6059537Z #110 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6059918Z #111 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6060290Z #112 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6060683Z #113 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6060998Z #114 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6061306Z #115 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6061725Z #116 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6062122Z #117 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6062540Z #118 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6062933Z #119 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6063343Z #120 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6063727Z #121 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6064007Z #122 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6064385Z #123 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6064800Z #124 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6065192Z #125 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6065608Z #126 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6066000Z #127 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6066359Z #128 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6066675Z #129 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6066988Z #130 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6067296Z #131 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6067731Z #132 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6068109Z #133 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6068526Z #134 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6068920Z #135 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6069337Z #136 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6069729Z #137 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6070024Z #138 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T15:25:36.6070334Z #139 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T15:25:36.6070619Z #140 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T15:25:36.6070941Z #141 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T15:25:36.6071294Z #142 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T15:25:36.6071670Z #143 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T15:25:36.6071963Z #144 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T15:25:36.6072248Z #145 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T15:25:36.6072576Z #146 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T15:25:36.6072774Z #147 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T15:25:36.6072889Z #148 _start from ??:0 2025-12-04T15:25:36.6073009Z #149 from ??:0 2025-12-04T15:25:36.6073015Z 2025-12-04T15:25:36.6073020Z 2025-12-04T15:25:36.6073255Z To execute this test, run the following from the base repo dir: 2025-12-04T15:25:36.6073852Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/functorch/test_vmap.py TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda 2025-12-04T15:25:36.6073859Z 2025-12-04T15:25:36.6074125Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T15:25:36.6074361Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T15:25:36.6076290Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5048: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6076550Z isinstance(vmap(torch.min, (0, None))(t, 0), torch.return_types.min) 2025-12-04T15:25:36.6078466Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5051: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6078719Z isinstance(vmap(torch.max, (0, None))(t, 0), torch.return_types.max) 2025-12-04T15:25:36.6080644Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5055: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6080885Z vmap(torch.topk, (0, None, None))(t, 1, 0), torch.return_types.topk 2025-12-04T15:25:36.6082795Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5059: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6083069Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.6084232Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py:433: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T15:25:36.6084395Z batched_outputs = func(*batched_inputs, **kwargs) 2025-12-04T15:25:36.6084766Z _______ TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda _______ 2025-12-04T15:25:36.6084889Z Traceback (most recent call last): 2025-12-04T15:25:36.6085395Z File "/var/lib/jenkins/workspace/test/functorch/test_vmap.py", line 5059, in test_torch_return_types_returns 2025-12-04T15:25:36.6085686Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.6086118Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/apis.py", line 208, in wrapped 2025-12-04T15:25:36.6086237Z return vmap_impl( 2025-12-04T15:25:36.6086725Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py", line 283, in vmap_impl 2025-12-04T15:25:36.6086827Z return _flat_vmap( 2025-12-04T15:25:36.6087276Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py", line 433, in _flat_vmap 2025-12-04T15:25:36.6087447Z batched_outputs = func(*batched_inputs, **kwargs) 2025-12-04T15:25:36.6088208Z RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T15:25:36.6088990Z Exception raised from apply_magma_eig at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebra.cpp:2024 (most recent call first): 2025-12-04T15:25:36.6089103Z C++ CapturedTraceback: 2025-12-04T15:25:36.6090431Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T15:25:36.6090918Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T15:25:36.6091268Z #6 c10::detail::torchCheckFail(char const*, char const*, unsigned int, char const*) from ??:0 2025-12-04T15:25:36.6092071Z #7 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool)::{lambda()#1}::operator()() const::{lambda()#1}::operator()() const [clone .constprop.0] from BatchLinearAlgebra.cpp:0 2025-12-04T15:25:36.6092473Z #8 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T15:25:36.6092920Z #9 at::native::linalg_eig_out_info(at::Tensor const&, at::Tensor&, at::Tensor&, at::Tensor&, bool) from BatchLinearAlgebra.cpp:0 2025-12-04T15:25:36.6093197Z #10 at::native::linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T15:25:36.6093822Z #11 at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6096895Z #12 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&, at::Tensor&, at::Tensor&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&, at::Tensor&, at::Tensor&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6097205Z #13 at::_ops::linalg_eig_out::call(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T15:25:36.6097376Z #14 at::native::linalg_eig(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6099491Z #15 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA__linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6099824Z #16 at::_ops::linalg_eig::redispatch(c10::DispatchKeySet, at::Tensor const&) from ??:0 2025-12-04T15:25:36.6100371Z #17 torch::autograd::VariableType::(anonymous namespace)::linalg_eig(c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T15:25:36.6102706Z #18 c10::impl::make_boxed_from_unboxed_functor (c10::DispatchKeySet, at::Tensor const&), &torch::autograd::VariableType::(anonymous namespace)::linalg_eig>, std::tuple, c10::guts::typelist::typelist >, false>::call(c10::OperatorKernel*, c10::OperatorHandle const&, c10::DispatchKeySet, std::vector >*) from VariableType_2.cpp:0 2025-12-04T15:25:36.6103425Z #19 c10::Dispatcher::callBoxed(c10::OperatorHandle const&, std::vector >*) const [clone .isra.0] from VmapInterpreter.cpp:0 2025-12-04T15:25:36.6104061Z #20 at::functorch::Interpreter::sendToNextInterpreter(c10::OperatorHandle const&, std::vector >*, bool) from :0 2025-12-04T15:25:36.6104650Z #21 at::functorch::dynamicLayerBack(c10::OperatorHandle const&, std::vector >*, bool) from DynamicLayer.cpp:0 2025-12-04T15:25:36.6105445Z #22 c10::impl::BoxedKernelWrapper (at::Tensor const&), void>::call(c10::BoxedKernel const&, c10::OperatorHandle const&, c10::DispatchKeySet, at::Tensor const&) from :0 2025-12-04T15:25:36.6105623Z #23 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6107741Z #24 std::tuple at::functorch::linalg_eig_generated_plumbing, at::Tensor, std::optional > (*)(at::Tensor const&, std::optional), &at::functorch::(anonymous namespace)::LinalgCheckMatrixUnaryRuleHelper<&at::functorch::func_string_linalg_eig, std::tuple (*)(at::Tensor const&), &at::_ops::linalg_eig::call, c10::guts::typelist::typelist >::apply_two>(at::Tensor const&) from BatchRulesLinearAlgebra.cpp:0 2025-12-04T15:25:36.6109341Z #25 c10::impl::make_boxed_from_unboxed_functor (*)(at::Tensor const&), std::tuple, c10::guts::typelist::typelist >, false>::call(c10::OperatorKernel*, c10::OperatorHandle const&, c10::DispatchKeySet, std::vector >*) from :0 2025-12-04T15:25:36.6109999Z #26 c10::Dispatcher::callBoxed(c10::OperatorHandle const&, std::vector >*) const [clone .isra.0] from VmapInterpreter.cpp:0 2025-12-04T15:25:36.6110508Z #27 at::functorch::Interpreter::process(c10::OperatorHandle const&, std::vector >*) from :0 2025-12-04T15:25:36.6111135Z #28 at::functorch::dynamicLayerFrontFallback(c10::OperatorHandle const&, std::vector >*) from DynamicLayer.cpp:0 2025-12-04T15:25:36.6111924Z #29 c10::impl::BoxedKernelWrapper (at::Tensor const&), void>::call(c10::BoxedKernel const&, c10::OperatorHandle const&, c10::DispatchKeySet, at::Tensor const&) from :0 2025-12-04T15:25:36.6112105Z #30 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6112522Z #31 torch::autograd::THPVariable_linalg_eig(_object*, _object*, _object*) from python_linalg_functions.cpp:0 2025-12-04T15:25:36.6112847Z #32 cfunction_call from /usr/local/src/conda/python-3.10.14/Objects/methodobject.c:543 2025-12-04T15:25:36.6113153Z #33 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T15:25:36.6113428Z #34 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5917 2025-12-04T15:25:36.6113852Z #35 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6114124Z #36 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6114498Z #37 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6114752Z #38 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6115195Z #39 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6115603Z #40 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6115986Z #41 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6116281Z #42 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T15:25:36.6116534Z #43 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6116917Z #44 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6117170Z #45 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6117545Z #46 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6117813Z #47 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6118185Z #48 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6118449Z #49 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6118818Z #50 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6119230Z #51 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6119612Z #52 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6120020Z #53 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6120401Z #54 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6120806Z #55 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6121178Z #56 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6121595Z #57 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6121965Z #58 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6122367Z #59 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6122757Z #60 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6123049Z #61 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T15:25:36.6123361Z #62 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6123730Z #63 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6124081Z #64 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6124398Z #65 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6124692Z #66 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6125038Z #67 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6125442Z #68 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6125840Z #69 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6126258Z #70 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6126626Z #71 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6126953Z #72 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6127322Z #73 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6127729Z #74 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6128110Z #75 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6128515Z #76 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6128884Z #77 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6129240Z #78 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6129543Z #79 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6129850Z #80 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6130116Z #81 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T15:25:36.6130374Z #82 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6130757Z #83 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6131165Z #84 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6131551Z #85 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6131958Z #86 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6132330Z #87 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6132601Z #88 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6132973Z #89 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6133389Z #90 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6133760Z #91 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6134167Z #92 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6134548Z #93 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6134805Z #94 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6135175Z #95 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6135594Z #96 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6135969Z #97 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6136386Z #98 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6136755Z #99 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6137240Z #100 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6137570Z #101 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6137899Z #102 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6138221Z #103 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6138637Z #104 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6139083Z #105 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6139358Z #106 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6139736Z #107 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6140170Z #108 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6140550Z #109 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6140965Z #110 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6141355Z #111 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6141709Z #112 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6142020Z #113 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6142332Z #114 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6142637Z #115 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6143065Z #116 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6143447Z #117 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6143865Z #118 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6144254Z #119 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6144669Z #120 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6145061Z #121 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6145329Z #122 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6145707Z #123 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6146133Z #124 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6146514Z #125 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6146945Z #126 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6147325Z #127 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6147683Z #128 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6148008Z #129 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6148313Z #130 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6148619Z #131 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6149049Z #132 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6149465Z #133 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6149895Z #134 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6150306Z #135 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6150718Z #136 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6151113Z #137 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6151463Z #138 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T15:25:36.6151785Z #139 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T15:25:36.6152054Z #140 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T15:25:36.6152344Z #141 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T15:25:36.6152712Z #142 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T15:25:36.6153041Z #143 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T15:25:36.6153337Z #144 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T15:25:36.6153628Z #145 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T15:25:36.6153896Z #146 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T15:25:36.6154107Z #147 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T15:25:36.6154207Z #148 _start from ??:0 2025-12-04T15:25:36.6154330Z #149 from ??:0 2025-12-04T15:25:36.6154337Z 2025-12-04T15:25:36.6154343Z 2025-12-04T15:25:36.6154579Z To execute this test, run the following from the base repo dir: 2025-12-04T15:25:36.6155184Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/functorch/test_vmap.py TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda 2025-12-04T15:25:36.6155190Z 2025-12-04T15:25:36.6155473Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T15:25:36.6155699Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T15:25:36.6157640Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5048: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6157902Z isinstance(vmap(torch.min, (0, None))(t, 0), torch.return_types.min) 2025-12-04T15:25:36.6159823Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5051: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6160080Z isinstance(vmap(torch.max, (0, None))(t, 0), torch.return_types.max) 2025-12-04T15:25:36.6161991Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5055: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6162236Z vmap(torch.topk, (0, None, None))(t, 1, 0), torch.return_types.topk 2025-12-04T15:25:36.6164244Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5059: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6164531Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.6165739Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py:433: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T15:25:36.6165921Z batched_outputs = func(*batched_inputs, **kwargs) 2025-12-04T15:25:36.6166141Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T15:25:36.6168058Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5048: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6168317Z isinstance(vmap(torch.min, (0, None))(t, 0), torch.return_types.min) 2025-12-04T15:25:36.6170240Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5051: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6170493Z isinstance(vmap(torch.max, (0, None))(t, 0), torch.return_types.max) 2025-12-04T15:25:36.6172402Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5055: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6172642Z vmap(torch.topk, (0, None, None))(t, 1, 0), torch.return_types.topk 2025-12-04T15:25:36.6174552Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5059: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6174837Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.6174983Z =================================== FAILURES =================================== 2025-12-04T15:25:36.6175298Z _______ TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda _______ 2025-12-04T15:25:36.6175435Z Traceback (most recent call last): 2025-12-04T15:25:36.6175906Z File "/var/lib/jenkins/workspace/test/functorch/test_vmap.py", line 5059, in test_torch_return_types_returns 2025-12-04T15:25:36.6176199Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.6176633Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/apis.py", line 208, in wrapped 2025-12-04T15:25:36.6176737Z return vmap_impl( 2025-12-04T15:25:36.6177286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py", line 283, in vmap_impl 2025-12-04T15:25:36.6177394Z return _flat_vmap( 2025-12-04T15:25:36.6177863Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py", line 433, in _flat_vmap 2025-12-04T15:25:36.6178046Z batched_outputs = func(*batched_inputs, **kwargs) 2025-12-04T15:25:36.6178807Z RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T15:25:36.6179654Z Exception raised from apply_magma_eig at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebra.cpp:2024 (most recent call first): 2025-12-04T15:25:36.6179764Z C++ CapturedTraceback: 2025-12-04T15:25:36.6181072Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T15:25:36.6181571Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T15:25:36.6181908Z #6 c10::detail::torchCheckFail(char const*, char const*, unsigned int, char const*) from ??:0 2025-12-04T15:25:36.6182722Z #7 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool)::{lambda()#1}::operator()() const::{lambda()#1}::operator()() const [clone .constprop.0] from BatchLinearAlgebra.cpp:0 2025-12-04T15:25:36.6183128Z #8 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T15:25:36.6183580Z #9 at::native::linalg_eig_out_info(at::Tensor const&, at::Tensor&, at::Tensor&, at::Tensor&, bool) from BatchLinearAlgebra.cpp:0 2025-12-04T15:25:36.6183852Z #10 at::native::linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T15:25:36.6184461Z #11 at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6186868Z #12 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&, at::Tensor&, at::Tensor&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&, at::Tensor&, at::Tensor&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6187154Z #13 at::_ops::linalg_eig_out::call(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T15:25:36.6187333Z #14 at::native::linalg_eig(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6189351Z #15 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA__linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6189659Z #16 at::_ops::linalg_eig::redispatch(c10::DispatchKeySet, at::Tensor const&) from ??:0 2025-12-04T15:25:36.6190205Z #17 torch::autograd::VariableType::(anonymous namespace)::linalg_eig(c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T15:25:36.6192570Z #18 c10::impl::make_boxed_from_unboxed_functor (c10::DispatchKeySet, at::Tensor const&), &torch::autograd::VariableType::(anonymous namespace)::linalg_eig>, std::tuple, c10::guts::typelist::typelist >, false>::call(c10::OperatorKernel*, c10::OperatorHandle const&, c10::DispatchKeySet, std::vector >*) from VariableType_2.cpp:0 2025-12-04T15:25:36.6193212Z #19 c10::Dispatcher::callBoxed(c10::OperatorHandle const&, std::vector >*) const [clone .isra.0] from VmapInterpreter.cpp:0 2025-12-04T15:25:36.6193905Z #20 at::functorch::Interpreter::sendToNextInterpreter(c10::OperatorHandle const&, std::vector >*, bool) from :0 2025-12-04T15:25:36.6194500Z #21 at::functorch::dynamicLayerBack(c10::OperatorHandle const&, std::vector >*, bool) from DynamicLayer.cpp:0 2025-12-04T15:25:36.6195280Z #22 c10::impl::BoxedKernelWrapper (at::Tensor const&), void>::call(c10::BoxedKernel const&, c10::OperatorHandle const&, c10::DispatchKeySet, at::Tensor const&) from :0 2025-12-04T15:25:36.6195471Z #23 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6197793Z #24 std::tuple at::functorch::linalg_eig_generated_plumbing, at::Tensor, std::optional > (*)(at::Tensor const&, std::optional), &at::functorch::(anonymous namespace)::LinalgCheckMatrixUnaryRuleHelper<&at::functorch::func_string_linalg_eig, std::tuple (*)(at::Tensor const&), &at::_ops::linalg_eig::call, c10::guts::typelist::typelist >::apply_two>(at::Tensor const&) from BatchRulesLinearAlgebra.cpp:0 2025-12-04T15:25:36.6199400Z #25 c10::impl::make_boxed_from_unboxed_functor (*)(at::Tensor const&), std::tuple, c10::guts::typelist::typelist >, false>::call(c10::OperatorKernel*, c10::OperatorHandle const&, c10::DispatchKeySet, std::vector >*) from :0 2025-12-04T15:25:36.6200046Z #26 c10::Dispatcher::callBoxed(c10::OperatorHandle const&, std::vector >*) const [clone .isra.0] from VmapInterpreter.cpp:0 2025-12-04T15:25:36.6200573Z #27 at::functorch::Interpreter::process(c10::OperatorHandle const&, std::vector >*) from :0 2025-12-04T15:25:36.6201210Z #28 at::functorch::dynamicLayerFrontFallback(c10::OperatorHandle const&, std::vector >*) from DynamicLayer.cpp:0 2025-12-04T15:25:36.6201996Z #29 c10::impl::BoxedKernelWrapper (at::Tensor const&), void>::call(c10::BoxedKernel const&, c10::OperatorHandle const&, c10::DispatchKeySet, at::Tensor const&) from :0 2025-12-04T15:25:36.6202178Z #30 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6202585Z #31 torch::autograd::THPVariable_linalg_eig(_object*, _object*, _object*) from python_linalg_functions.cpp:0 2025-12-04T15:25:36.6202927Z #32 cfunction_call from /usr/local/src/conda/python-3.10.14/Objects/methodobject.c:543 2025-12-04T15:25:36.6203198Z #33 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T15:25:36.6203470Z #34 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5917 2025-12-04T15:25:36.6203848Z #35 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6204113Z #36 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6204499Z #37 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6204752Z #38 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6205185Z #39 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6205652Z #40 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6206026Z #41 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6206334Z #42 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T15:25:36.6206590Z #43 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6207047Z #44 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6207312Z #45 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6207683Z #46 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6207954Z #47 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6208325Z #48 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6208581Z #49 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6208965Z #50 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6209372Z #51 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6209741Z #52 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6210166Z #53 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6210534Z #54 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6210956Z #55 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6211323Z #56 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6211728Z #57 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6212112Z #58 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6212517Z #59 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6212897Z #60 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6213190Z #61 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T15:25:36.6213463Z #62 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6213845Z #63 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6214194Z #64 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6214511Z #65 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6214805Z #66 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6215109Z #67 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6215526Z #68 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6215900Z #69 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6216305Z #70 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6216685Z #71 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6217051Z #72 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6217434Z #73 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6217870Z #74 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6218240Z #75 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6218658Z #76 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6219085Z #77 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6219445Z #78 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6219750Z #79 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6220046Z #80 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6220326Z #81 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T15:25:36.6220583Z #82 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6220967Z #83 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6221374Z #84 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6221744Z #85 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6222166Z #86 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6222537Z #87 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6222796Z #88 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6223182Z #89 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6223589Z #90 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6223976Z #91 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6224381Z #92 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6224751Z #93 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6225026Z #94 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6225394Z #95 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6225811Z #96 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6226181Z #97 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6226585Z #98 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6226972Z #99 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6227335Z #100 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6227658Z #101 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6227960Z #102 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6228265Z #103 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6228694Z #104 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6229111Z #105 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6229372Z #106 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6229794Z #107 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6230211Z #108 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6230601Z #109 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6231015Z #110 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6231454Z #111 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6231824Z #112 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6232135Z #113 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6232447Z #114 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6232755Z #115 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6233180Z #116 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6233576Z #117 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6233988Z #118 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6234386Z #119 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6234800Z #120 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6235179Z #121 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6235462Z #122 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6235846Z #123 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6236258Z #124 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6236650Z #125 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6237063Z #126 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6237458Z #127 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6237815Z #128 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6238125Z #129 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6238443Z #130 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6238752Z #131 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6239187Z #132 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6239567Z #133 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6239986Z #134 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6240384Z #135 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6240798Z #136 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6241193Z #137 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6241518Z #138 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T15:25:36.6241828Z #139 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T15:25:36.6242140Z #140 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T15:25:36.6242427Z #141 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T15:25:36.6242781Z #142 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T15:25:36.6243125Z #143 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T15:25:36.6243477Z #144 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T15:25:36.6243760Z #145 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T15:25:36.6244025Z #146 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T15:25:36.6244229Z #147 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T15:25:36.6244343Z #148 _start from ??:0 2025-12-04T15:25:36.6244465Z #149 from ??:0 2025-12-04T15:25:36.6244472Z 2025-12-04T15:25:36.6244477Z 2025-12-04T15:25:36.6244710Z To execute this test, run the following from the base repo dir: 2025-12-04T15:25:36.6245305Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/functorch/test_vmap.py TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda 2025-12-04T15:25:36.6245311Z 2025-12-04T15:25:36.6245580Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T15:25:36.6245817Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T15:25:36.6247750Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5048: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6248009Z isinstance(vmap(torch.min, (0, None))(t, 0), torch.return_types.min) 2025-12-04T15:25:36.6249933Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5051: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6250198Z isinstance(vmap(torch.max, (0, None))(t, 0), torch.return_types.max) 2025-12-04T15:25:36.6252108Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5055: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6252345Z vmap(torch.topk, (0, None, None))(t, 1, 0), torch.return_types.topk 2025-12-04T15:25:36.6254256Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5059: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6254534Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.6255720Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py:433: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T15:25:36.6255913Z batched_outputs = func(*batched_inputs, **kwargs) 2025-12-04T15:25:36.6256145Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T15:25:36.6258146Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5048: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6258483Z isinstance(vmap(torch.min, (0, None))(t, 0), torch.return_types.min) 2025-12-04T15:25:36.6260396Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5051: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6260651Z isinstance(vmap(torch.max, (0, None))(t, 0), torch.return_types.max) 2025-12-04T15:25:36.6262569Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5055: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6262798Z vmap(torch.topk, (0, None, None))(t, 1, 0), torch.return_types.topk 2025-12-04T15:25:36.6264725Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5059: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6265000Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.6265692Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-db01917ff3ce307b.xml - 2025-12-04T15:25:36.6265864Z =========================== short test summary info ============================ 2025-12-04T15:25:36.6267178Z FAILED [0.2020s] functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_torch_return_types_returns_cuda - RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T15:25:36.6267941Z Exception raised from apply_magma_eig at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebra.cpp:2024 (most recent call first): 2025-12-04T15:25:36.6268051Z C++ CapturedTraceback: 2025-12-04T15:25:36.6269380Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T15:25:36.6269871Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T15:25:36.6270253Z #6 c10::detail::torchCheckFail(char const*, char const*, unsigned int, char const*) from ??:0 2025-12-04T15:25:36.6271092Z #7 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool)::{lambda()#1}::operator()() const::{lambda()#1}::operator()() const [clone .constprop.0] from BatchLinearAlgebra.cpp:0 2025-12-04T15:25:36.6271505Z #8 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T15:25:36.6271944Z #9 at::native::linalg_eig_out_info(at::Tensor const&, at::Tensor&, at::Tensor&, at::Tensor&, bool) from BatchLinearAlgebra.cpp:0 2025-12-04T15:25:36.6272271Z #10 at::native::linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T15:25:36.6272895Z #11 at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6275303Z #12 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&, at::Tensor&, at::Tensor&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&, at::Tensor&, at::Tensor&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6275599Z #13 at::_ops::linalg_eig_out::call(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T15:25:36.6275770Z #14 at::native::linalg_eig(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6277820Z #15 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA__linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6278112Z #16 at::_ops::linalg_eig::redispatch(c10::DispatchKeySet, at::Tensor const&) from ??:0 2025-12-04T15:25:36.6278671Z #17 torch::autograd::VariableType::(anonymous namespace)::linalg_eig(c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T15:25:36.6280984Z #18 c10::impl::make_boxed_from_unboxed_functor (c10::DispatchKeySet, at::Tensor const&), &torch::autograd::VariableType::(anonymous namespace)::linalg_eig>, std::tuple, c10::guts::typelist::typelist >, false>::call(c10::OperatorKernel*, c10::OperatorHandle const&, c10::DispatchKeySet, std::vector >*) from VariableType_2.cpp:0 2025-12-04T15:25:36.6281642Z #19 c10::Dispatcher::callBoxed(c10::OperatorHandle const&, std::vector >*) const [clone .isra.0] from VmapInterpreter.cpp:0 2025-12-04T15:25:36.6282266Z #20 at::functorch::Interpreter::sendToNextInterpreter(c10::OperatorHandle const&, std::vector >*, bool) from :0 2025-12-04T15:25:36.6282860Z #21 at::functorch::dynamicLayerBack(c10::OperatorHandle const&, std::vector >*, bool) from DynamicLayer.cpp:0 2025-12-04T15:25:36.6283655Z #22 c10::impl::BoxedKernelWrapper (at::Tensor const&), void>::call(c10::BoxedKernel const&, c10::OperatorHandle const&, c10::DispatchKeySet, at::Tensor const&) from :0 2025-12-04T15:25:36.6283832Z #23 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6285986Z #24 std::tuple at::functorch::linalg_eig_generated_plumbing, at::Tensor, std::optional > (*)(at::Tensor const&, std::optional), &at::functorch::(anonymous namespace)::LinalgCheckMatrixUnaryRuleHelper<&at::functorch::func_string_linalg_eig, std::tuple (*)(at::Tensor const&), &at::_ops::linalg_eig::call, c10::guts::typelist::typelist >::apply_two>(at::Tensor const&) from BatchRulesLinearAlgebra.cpp:0 2025-12-04T15:25:36.6287579Z #25 c10::impl::make_boxed_from_unboxed_functor (*)(at::Tensor const&), std::tuple, c10::guts::typelist::typelist >, false>::call(c10::OperatorKernel*, c10::OperatorHandle const&, c10::DispatchKeySet, std::vector >*) from :0 2025-12-04T15:25:36.6288286Z #26 c10::Dispatcher::callBoxed(c10::OperatorHandle const&, std::vector >*) const [clone .isra.0] from VmapInterpreter.cpp:0 2025-12-04T15:25:36.6288793Z #27 at::functorch::Interpreter::process(c10::OperatorHandle const&, std::vector >*) from :0 2025-12-04T15:25:36.6289439Z #28 at::functorch::dynamicLayerFrontFallback(c10::OperatorHandle const&, std::vector >*) from DynamicLayer.cpp:0 2025-12-04T15:25:36.6290214Z #29 c10::impl::BoxedKernelWrapper (at::Tensor const&), void>::call(c10::BoxedKernel const&, c10::OperatorHandle const&, c10::DispatchKeySet, at::Tensor const&) from :0 2025-12-04T15:25:36.6290392Z #30 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6290808Z #31 torch::autograd::THPVariable_linalg_eig(_object*, _object*, _object*) from python_linalg_functions.cpp:0 2025-12-04T15:25:36.6291131Z #32 cfunction_call from /usr/local/src/conda/python-3.10.14/Objects/methodobject.c:543 2025-12-04T15:25:36.6291415Z #33 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T15:25:36.6291675Z #34 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5917 2025-12-04T15:25:36.6292052Z #35 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6292323Z #36 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6292697Z #37 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6292953Z #38 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6293339Z #39 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6293749Z #40 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6294131Z #41 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6294426Z #42 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T15:25:36.6294681Z #43 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6295065Z #44 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6295320Z #45 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6295703Z #46 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6295963Z #47 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6296558Z #48 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6296903Z #49 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6297341Z #50 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6297763Z #51 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6298182Z #52 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6298594Z #53 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6298980Z #54 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6299387Z #55 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6299837Z #56 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6300253Z #57 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6300626Z #58 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6301042Z #59 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6301413Z #60 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6301707Z #61 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T15:25:36.6301973Z #62 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6302343Z #63 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6302707Z #64 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6303012Z #65 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6303307Z #66 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6303631Z #67 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6304038Z #68 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6304409Z #69 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6304823Z #70 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6305190Z #71 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6305464Z #72 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6305832Z #73 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6306236Z #74 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6306616Z #75 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6307018Z #76 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6307401Z #77 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6307749Z #78 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6308051Z #79 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6308359Z #80 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6308626Z #81 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T15:25:36.6308894Z #82 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6309258Z #83 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6324106Z #84 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6324652Z #85 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6325071Z #86 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6325461Z #87 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6325723Z #88 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6326188Z #89 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6326597Z #90 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6326967Z #91 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6327387Z #92 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6327759Z #93 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6328035Z #94 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6328409Z #95 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6328814Z #96 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6329199Z #97 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6329604Z #98 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6329974Z #99 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6330352Z #100 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6330665Z #101 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6330980Z #102 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6331287Z #103 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6331703Z #104 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6332102Z #105 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6332367Z #106 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6332763Z #107 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6333182Z #108 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6333561Z #109 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6333991Z #110 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6334369Z #111 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6334744Z #112 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6335061Z #113 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6335361Z #114 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6335708Z #115 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6336120Z #116 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6336530Z #117 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6337078Z #118 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6337462Z #119 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6337892Z #120 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6338273Z #121 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6338601Z #122 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6338996Z #123 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6339409Z #124 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6339806Z #125 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6340222Z #126 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6340603Z #127 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6340978Z #128 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6341291Z #129 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6341612Z #130 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6341924Z #131 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6342341Z #132 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6342738Z #133 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6343151Z #134 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6343533Z #135 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6343963Z #136 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6344343Z #137 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6344655Z #138 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T15:25:36.6344963Z #139 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T15:25:36.6345232Z #140 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T15:25:36.6345537Z #141 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T15:25:36.6345889Z #142 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T15:25:36.6346234Z #143 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T15:25:36.6346528Z #144 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T15:25:36.6346799Z #145 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T15:25:36.6347083Z #146 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T15:25:36.6347288Z #147 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T15:25:36.6347390Z #148 _start from ??:0 2025-12-04T15:25:36.6347529Z #149 from ??:0 2025-12-04T15:25:36.6347537Z 2025-12-04T15:25:36.6347542Z 2025-12-04T15:25:36.6347765Z To execute this test, run the following from the base repo dir: 2025-12-04T15:25:36.6348418Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/functorch/test_vmap.py TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda 2025-12-04T15:25:36.6348426Z 2025-12-04T15:25:36.6348728Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T15:25:36.6348913Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T15:25:36.6349135Z ================= 1 failed, 2133 deselected, 2 rerun in 16.07s ================= 2025-12-04T15:25:36.6349236Z Got exit code 1 2025-12-04T15:25:36.6349360Z Retrying single test... 2025-12-04T15:25:36.6349925Z Test results will be stored in test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-12ecde9ddbb651c9.xml 2025-12-04T15:25:36.6350089Z ============================= test session starts ============================== 2025-12-04T15:25:36.6350460Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T15:25:36.6350569Z cachedir: .pytest_cache 2025-12-04T15:25:36.6351095Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T15:25:36.6351235Z rootdir: /var/lib/jenkins/workspace 2025-12-04T15:25:36.6351348Z configfile: pytest.ini 2025-12-04T15:25:36.6351958Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T15:25:36.6352187Z collecting ... collected 2134 items / 2133 deselected / 1 selected 2025-12-04T15:25:36.6352883Z stepcurrent: skipping 1006 already run items. Running only test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_torch_return_types_returns_cuda 2025-12-04T15:25:36.6353013Z Running 1 items in this shard 2025-12-04T15:25:36.6353018Z 2025-12-04T15:25:36.6354016Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_torch_return_types_returns_cuda [W1204 15:21:37.846620697 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T15:25:36.6354023Z 2025-12-04T15:25:36.6354168Z ('RERUN', {'yellow': True}) [15.3700s] [100%] 2025-12-04T15:25:36.6355153Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_torch_return_types_returns_cuda [W1204 15:21:52.434056650 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T15:25:36.6355158Z 2025-12-04T15:25:36.6355301Z ('RERUN', {'yellow': True}) [0.2191s] [100%] 2025-12-04T15:25:36.6356280Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_torch_return_types_returns_cuda [W1204 15:21:52.619702766 Module.cpp:201] symbolizing C++ stack trace for exception; if this hangs, rerun with TORCH_DISABLE_ADDR2LINE=1... 2025-12-04T15:25:36.6356290Z 2025-12-04T15:25:36.6356391Z FAILED [0.1835s] [100%] 2025-12-04T15:25:36.6356396Z 2025-12-04T15:25:36.6356550Z ==================================== RERUNS ==================================== 2025-12-04T15:25:36.6356869Z _______ TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda _______ 2025-12-04T15:25:36.6357004Z Traceback (most recent call last): 2025-12-04T15:25:36.6357481Z File "/var/lib/jenkins/workspace/test/functorch/test_vmap.py", line 5059, in test_torch_return_types_returns 2025-12-04T15:25:36.6357758Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.6358203Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/apis.py", line 208, in wrapped 2025-12-04T15:25:36.6358308Z return vmap_impl( 2025-12-04T15:25:36.6358747Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py", line 283, in vmap_impl 2025-12-04T15:25:36.6358863Z return _flat_vmap( 2025-12-04T15:25:36.6359299Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py", line 433, in _flat_vmap 2025-12-04T15:25:36.6359482Z batched_outputs = func(*batched_inputs, **kwargs) 2025-12-04T15:25:36.6360274Z RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T15:25:36.6361068Z Exception raised from apply_magma_eig at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebra.cpp:2024 (most recent call first): 2025-12-04T15:25:36.6361195Z C++ CapturedTraceback: 2025-12-04T15:25:36.6362502Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T15:25:36.6363066Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T15:25:36.6363404Z #6 c10::detail::torchCheckFail(char const*, char const*, unsigned int, char const*) from ??:0 2025-12-04T15:25:36.6364215Z #7 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool)::{lambda()#1}::operator()() const::{lambda()#1}::operator()() const [clone .constprop.0] from BatchLinearAlgebra.cpp:0 2025-12-04T15:25:36.6364630Z #8 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T15:25:36.6365067Z #9 at::native::linalg_eig_out_info(at::Tensor const&, at::Tensor&, at::Tensor&, at::Tensor&, bool) from BatchLinearAlgebra.cpp:0 2025-12-04T15:25:36.6365360Z #10 at::native::linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T15:25:36.6365966Z #11 at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6368360Z #12 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&, at::Tensor&, at::Tensor&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&, at::Tensor&, at::Tensor&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6368640Z #13 at::_ops::linalg_eig_out::call(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T15:25:36.6368825Z #14 at::native::linalg_eig(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6370844Z #15 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA__linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6371137Z #16 at::_ops::linalg_eig::redispatch(c10::DispatchKeySet, at::Tensor const&) from ??:0 2025-12-04T15:25:36.6371698Z #17 torch::autograd::VariableType::(anonymous namespace)::linalg_eig(c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T15:25:36.6374068Z #18 c10::impl::make_boxed_from_unboxed_functor (c10::DispatchKeySet, at::Tensor const&), &torch::autograd::VariableType::(anonymous namespace)::linalg_eig>, std::tuple, c10::guts::typelist::typelist >, false>::call(c10::OperatorKernel*, c10::OperatorHandle const&, c10::DispatchKeySet, std::vector >*) from VariableType_2.cpp:0 2025-12-04T15:25:36.6374754Z #19 c10::Dispatcher::callBoxed(c10::OperatorHandle const&, std::vector >*) const [clone .isra.0] from VmapInterpreter.cpp:0 2025-12-04T15:25:36.6375375Z #20 at::functorch::Interpreter::sendToNextInterpreter(c10::OperatorHandle const&, std::vector >*, bool) from :0 2025-12-04T15:25:36.6375983Z #21 at::functorch::dynamicLayerBack(c10::OperatorHandle const&, std::vector >*, bool) from DynamicLayer.cpp:0 2025-12-04T15:25:36.6376891Z #22 c10::impl::BoxedKernelWrapper (at::Tensor const&), void>::call(c10::BoxedKernel const&, c10::OperatorHandle const&, c10::DispatchKeySet, at::Tensor const&) from :0 2025-12-04T15:25:36.6377092Z #23 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6379196Z #24 std::tuple at::functorch::linalg_eig_generated_plumbing, at::Tensor, std::optional > (*)(at::Tensor const&, std::optional), &at::functorch::(anonymous namespace)::LinalgCheckMatrixUnaryRuleHelper<&at::functorch::func_string_linalg_eig, std::tuple (*)(at::Tensor const&), &at::_ops::linalg_eig::call, c10::guts::typelist::typelist >::apply_two>(at::Tensor const&) from BatchRulesLinearAlgebra.cpp:0 2025-12-04T15:25:36.6380795Z #25 c10::impl::make_boxed_from_unboxed_functor (*)(at::Tensor const&), std::tuple, c10::guts::typelist::typelist >, false>::call(c10::OperatorKernel*, c10::OperatorHandle const&, c10::DispatchKeySet, std::vector >*) from :0 2025-12-04T15:25:36.6381442Z #26 c10::Dispatcher::callBoxed(c10::OperatorHandle const&, std::vector >*) const [clone .isra.0] from VmapInterpreter.cpp:0 2025-12-04T15:25:36.6381972Z #27 at::functorch::Interpreter::process(c10::OperatorHandle const&, std::vector >*) from :0 2025-12-04T15:25:36.6382606Z #28 at::functorch::dynamicLayerFrontFallback(c10::OperatorHandle const&, std::vector >*) from DynamicLayer.cpp:0 2025-12-04T15:25:36.6383388Z #29 c10::impl::BoxedKernelWrapper (at::Tensor const&), void>::call(c10::BoxedKernel const&, c10::OperatorHandle const&, c10::DispatchKeySet, at::Tensor const&) from :0 2025-12-04T15:25:36.6383587Z #30 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6383993Z #31 torch::autograd::THPVariable_linalg_eig(_object*, _object*, _object*) from python_linalg_functions.cpp:0 2025-12-04T15:25:36.6384336Z #32 cfunction_call from /usr/local/src/conda/python-3.10.14/Objects/methodobject.c:543 2025-12-04T15:25:36.6384612Z #33 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T15:25:36.6384872Z #34 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5917 2025-12-04T15:25:36.6385271Z #35 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6385532Z #36 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6385908Z #37 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6386177Z #38 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6386552Z #39 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6386976Z #40 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6387348Z #41 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6387679Z #42 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T15:25:36.6387954Z #43 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6388356Z #44 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6388626Z #45 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6388996Z #46 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6389310Z #47 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6389694Z #48 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6389949Z #49 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6390322Z #50 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6390747Z #51 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6391119Z #52 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6391542Z #53 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6391913Z #54 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6392319Z #55 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6392712Z #56 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6393119Z #57 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6393509Z #58 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6393916Z #59 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6394292Z #60 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6394604Z #61 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T15:25:36.6394861Z #62 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6395246Z #63 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6395602Z #64 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6395912Z #65 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6396392Z #66 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6396699Z #67 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6397107Z #68 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6397497Z #69 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6397903Z #70 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6398292Z #71 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6398559Z #72 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6398929Z #73 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6399345Z #74 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6399786Z #75 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6400207Z #76 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6400617Z #77 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6400969Z #78 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6401284Z #79 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6401657Z #80 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6401936Z #81 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T15:25:36.6402196Z #82 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6402566Z #83 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6402994Z #84 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6403367Z #85 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6403772Z #86 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6404153Z #87 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6404411Z #88 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6404798Z #89 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6405204Z #90 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6405571Z #91 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6405995Z #92 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6406366Z #93 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6406637Z #94 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6407005Z #95 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6407409Z #96 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6407797Z #97 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6408201Z #98 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6408587Z #99 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6408951Z #100 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6409268Z #101 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6409586Z #102 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6409893Z #103 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6410308Z #104 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6410704Z #105 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6410970Z #106 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6411366Z #107 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6411815Z #108 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6412193Z #109 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6412670Z #110 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6413049Z #111 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6413424Z #112 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6413811Z #113 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6414111Z #114 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6414433Z #115 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6414846Z #116 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6415227Z #117 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6415660Z #118 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6416038Z #119 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6416462Z #120 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6416903Z #121 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6417181Z #122 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6417577Z #123 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6417990Z #124 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6418383Z #125 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6418799Z #126 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6419177Z #127 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6419551Z #128 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6419863Z #129 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6420179Z #130 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6420490Z #131 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6420904Z #132 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6421302Z #133 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6421717Z #134 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6422096Z #135 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6422527Z #136 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6422905Z #137 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6423220Z #138 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T15:25:36.6423529Z #139 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T15:25:36.6423800Z #140 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T15:25:36.6424140Z #141 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T15:25:36.6424493Z #142 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T15:25:36.6424862Z #143 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T15:25:36.6425158Z #144 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T15:25:36.6425430Z #145 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T15:25:36.6425712Z #146 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T15:25:36.6425972Z #147 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T15:25:36.6426082Z #148 _start from ??:0 2025-12-04T15:25:36.6426222Z #149 from ??:0 2025-12-04T15:25:36.6426228Z 2025-12-04T15:25:36.6426233Z 2025-12-04T15:25:36.6426452Z To execute this test, run the following from the base repo dir: 2025-12-04T15:25:36.6427057Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/functorch/test_vmap.py TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda 2025-12-04T15:25:36.6427062Z 2025-12-04T15:25:36.6427343Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T15:25:36.6427565Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T15:25:36.6429503Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5048: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6429755Z isinstance(vmap(torch.min, (0, None))(t, 0), torch.return_types.min) 2025-12-04T15:25:36.6431692Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5051: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6431937Z isinstance(vmap(torch.max, (0, None))(t, 0), torch.return_types.max) 2025-12-04T15:25:36.6433857Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5055: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6434085Z vmap(torch.topk, (0, None, None))(t, 1, 0), torch.return_types.topk 2025-12-04T15:25:36.6436003Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5059: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6436290Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.6437444Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py:433: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T15:25:36.6437621Z batched_outputs = func(*batched_inputs, **kwargs) 2025-12-04T15:25:36.6437974Z _______ TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda _______ 2025-12-04T15:25:36.6438111Z Traceback (most recent call last): 2025-12-04T15:25:36.6438585Z File "/var/lib/jenkins/workspace/test/functorch/test_vmap.py", line 5059, in test_torch_return_types_returns 2025-12-04T15:25:36.6438893Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.6439337Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/apis.py", line 208, in wrapped 2025-12-04T15:25:36.6439441Z return vmap_impl( 2025-12-04T15:25:36.6439875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py", line 283, in vmap_impl 2025-12-04T15:25:36.6440047Z return _flat_vmap( 2025-12-04T15:25:36.6440486Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py", line 433, in _flat_vmap 2025-12-04T15:25:36.6440667Z batched_outputs = func(*batched_inputs, **kwargs) 2025-12-04T15:25:36.6441424Z RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T15:25:36.6442180Z Exception raised from apply_magma_eig at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebra.cpp:2024 (most recent call first): 2025-12-04T15:25:36.6442303Z C++ CapturedTraceback: 2025-12-04T15:25:36.6443621Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T15:25:36.6444118Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T15:25:36.6444452Z #6 c10::detail::torchCheckFail(char const*, char const*, unsigned int, char const*) from ??:0 2025-12-04T15:25:36.6445258Z #7 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool)::{lambda()#1}::operator()() const::{lambda()#1}::operator()() const [clone .constprop.0] from BatchLinearAlgebra.cpp:0 2025-12-04T15:25:36.6445668Z #8 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T15:25:36.6446103Z #9 at::native::linalg_eig_out_info(at::Tensor const&, at::Tensor&, at::Tensor&, at::Tensor&, bool) from BatchLinearAlgebra.cpp:0 2025-12-04T15:25:36.6446392Z #10 at::native::linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T15:25:36.6447004Z #11 at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6449409Z #12 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&, at::Tensor&, at::Tensor&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&, at::Tensor&, at::Tensor&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6449691Z #13 at::_ops::linalg_eig_out::call(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T15:25:36.6449873Z #14 at::native::linalg_eig(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6451914Z #15 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA__linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6452233Z #16 at::_ops::linalg_eig::redispatch(c10::DispatchKeySet, at::Tensor const&) from ??:0 2025-12-04T15:25:36.6452792Z #17 torch::autograd::VariableType::(anonymous namespace)::linalg_eig(c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T15:25:36.6455083Z #18 c10::impl::make_boxed_from_unboxed_functor (c10::DispatchKeySet, at::Tensor const&), &torch::autograd::VariableType::(anonymous namespace)::linalg_eig>, std::tuple, c10::guts::typelist::typelist >, false>::call(c10::OperatorKernel*, c10::OperatorHandle const&, c10::DispatchKeySet, std::vector >*) from VariableType_2.cpp:0 2025-12-04T15:25:36.6455795Z #19 c10::Dispatcher::callBoxed(c10::OperatorHandle const&, std::vector >*) const [clone .isra.0] from VmapInterpreter.cpp:0 2025-12-04T15:25:36.6456420Z #20 at::functorch::Interpreter::sendToNextInterpreter(c10::OperatorHandle const&, std::vector >*, bool) from :0 2025-12-04T15:25:36.6457139Z #21 at::functorch::dynamicLayerBack(c10::OperatorHandle const&, std::vector >*, bool) from DynamicLayer.cpp:0 2025-12-04T15:25:36.6457927Z #22 c10::impl::BoxedKernelWrapper (at::Tensor const&), void>::call(c10::BoxedKernel const&, c10::OperatorHandle const&, c10::DispatchKeySet, at::Tensor const&) from :0 2025-12-04T15:25:36.6458119Z #23 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6460197Z #24 std::tuple at::functorch::linalg_eig_generated_plumbing, at::Tensor, std::optional > (*)(at::Tensor const&, std::optional), &at::functorch::(anonymous namespace)::LinalgCheckMatrixUnaryRuleHelper<&at::functorch::func_string_linalg_eig, std::tuple (*)(at::Tensor const&), &at::_ops::linalg_eig::call, c10::guts::typelist::typelist >::apply_two>(at::Tensor const&) from BatchRulesLinearAlgebra.cpp:0 2025-12-04T15:25:36.6461791Z #25 c10::impl::make_boxed_from_unboxed_functor (*)(at::Tensor const&), std::tuple, c10::guts::typelist::typelist >, false>::call(c10::OperatorKernel*, c10::OperatorHandle const&, c10::DispatchKeySet, std::vector >*) from :0 2025-12-04T15:25:36.6462431Z #26 c10::Dispatcher::callBoxed(c10::OperatorHandle const&, std::vector >*) const [clone .isra.0] from VmapInterpreter.cpp:0 2025-12-04T15:25:36.6462952Z #27 at::functorch::Interpreter::process(c10::OperatorHandle const&, std::vector >*) from :0 2025-12-04T15:25:36.6463578Z #28 at::functorch::dynamicLayerFrontFallback(c10::OperatorHandle const&, std::vector >*) from DynamicLayer.cpp:0 2025-12-04T15:25:36.6464354Z #29 c10::impl::BoxedKernelWrapper (at::Tensor const&), void>::call(c10::BoxedKernel const&, c10::OperatorHandle const&, c10::DispatchKeySet, at::Tensor const&) from :0 2025-12-04T15:25:36.6464547Z #30 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6464953Z #31 torch::autograd::THPVariable_linalg_eig(_object*, _object*, _object*) from python_linalg_functions.cpp:0 2025-12-04T15:25:36.6465287Z #32 cfunction_call from /usr/local/src/conda/python-3.10.14/Objects/methodobject.c:543 2025-12-04T15:25:36.6465561Z #33 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T15:25:36.6465861Z #34 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5917 2025-12-04T15:25:36.6466250Z #35 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6466537Z #36 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6466909Z #37 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6467176Z #38 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6467618Z #39 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6468038Z #40 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6468407Z #41 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6468701Z #42 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T15:25:36.6468970Z #43 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6469343Z #44 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6469608Z #45 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6469980Z #46 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6470237Z #47 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6470622Z #48 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6470879Z #49 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6471247Z #50 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6471671Z #51 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6472042Z #52 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6472468Z #53 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6472843Z #54 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6473245Z #55 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6473627Z #56 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6474033Z #57 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6474419Z #58 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6474828Z #59 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6475198Z #60 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6475506Z #61 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T15:25:36.6475760Z #62 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6476142Z #63 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6476495Z #64 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6476800Z #65 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6477105Z #66 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6477441Z #67 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6477850Z #68 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6478263Z #69 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6478671Z #70 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6479054Z #71 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6479309Z #72 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6479736Z #73 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6480153Z #74 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6480521Z #75 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6480942Z #76 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6481311Z #77 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6481662Z #78 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6481979Z #79 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6482270Z #80 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6482554Z #81 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T15:25:36.6482812Z #82 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6483181Z #83 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6483603Z #84 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6483970Z #85 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6484377Z #86 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6484759Z #87 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6485015Z #88 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6485409Z #89 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6485812Z #90 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6486179Z #91 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6486594Z #92 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6486963Z #93 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6487235Z #94 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6487601Z #95 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6488006Z #96 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6488385Z #97 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6488794Z #98 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6489177Z #99 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6489564Z #100 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6489876Z #101 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6490219Z #102 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6490528Z #103 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6490943Z #104 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6491336Z #105 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6491656Z #106 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6492044Z #107 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6492460Z #108 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6492840Z #109 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6493269Z #110 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6493646Z #111 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6494017Z #112 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6494326Z #113 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6494629Z #114 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6494950Z #115 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6495364Z #116 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6495760Z #117 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6496330Z #118 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6496712Z #119 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6497239Z #120 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6497616Z #121 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6497884Z #122 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6498276Z #123 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6498688Z #124 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6499082Z #125 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6499492Z #126 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6499872Z #127 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6500242Z #128 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6500553Z #129 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6500869Z #130 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6501171Z #131 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6501587Z #132 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6502042Z #133 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6502454Z #134 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6502884Z #135 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6503295Z #136 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6503670Z #137 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6504050Z #138 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T15:25:36.6504355Z #139 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T15:25:36.6504621Z #140 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T15:25:36.6504917Z #141 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T15:25:36.6505269Z #142 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T15:25:36.6505606Z #143 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T15:25:36.6505898Z #144 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T15:25:36.6506169Z #145 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T15:25:36.6506442Z #146 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T15:25:36.6506643Z #147 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T15:25:36.6506741Z #148 _start from ??:0 2025-12-04T15:25:36.6506874Z #149 from ??:0 2025-12-04T15:25:36.6506879Z 2025-12-04T15:25:36.6506884Z 2025-12-04T15:25:36.6507101Z To execute this test, run the following from the base repo dir: 2025-12-04T15:25:36.6507708Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/functorch/test_vmap.py TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda 2025-12-04T15:25:36.6507714Z 2025-12-04T15:25:36.6507981Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T15:25:36.6508217Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T15:25:36.6510151Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5048: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6510399Z isinstance(vmap(torch.min, (0, None))(t, 0), torch.return_types.min) 2025-12-04T15:25:36.6512327Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5051: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6512574Z isinstance(vmap(torch.max, (0, None))(t, 0), torch.return_types.max) 2025-12-04T15:25:36.6514495Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5055: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6514724Z vmap(torch.topk, (0, None, None))(t, 1, 0), torch.return_types.topk 2025-12-04T15:25:36.6516753Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5059: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6517028Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.6518191Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py:433: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T15:25:36.6518415Z batched_outputs = func(*batched_inputs, **kwargs) 2025-12-04T15:25:36.6518636Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T15:25:36.6520560Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5048: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6520802Z isinstance(vmap(torch.min, (0, None))(t, 0), torch.return_types.min) 2025-12-04T15:25:36.6522728Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5051: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6522974Z isinstance(vmap(torch.max, (0, None))(t, 0), torch.return_types.max) 2025-12-04T15:25:36.6524884Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5055: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6525103Z vmap(torch.topk, (0, None, None))(t, 1, 0), torch.return_types.topk 2025-12-04T15:25:36.6527026Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5059: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6527300Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.6527457Z =================================== FAILURES =================================== 2025-12-04T15:25:36.6527771Z _______ TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda _______ 2025-12-04T15:25:36.6527894Z Traceback (most recent call last): 2025-12-04T15:25:36.6528375Z File "/var/lib/jenkins/workspace/test/functorch/test_vmap.py", line 5059, in test_torch_return_types_returns 2025-12-04T15:25:36.6528654Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.6529087Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/apis.py", line 208, in wrapped 2025-12-04T15:25:36.6529203Z return vmap_impl( 2025-12-04T15:25:36.6529665Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py", line 283, in vmap_impl 2025-12-04T15:25:36.6529778Z return _flat_vmap( 2025-12-04T15:25:36.6530216Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py", line 433, in _flat_vmap 2025-12-04T15:25:36.6530412Z batched_outputs = func(*batched_inputs, **kwargs) 2025-12-04T15:25:36.6531182Z RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T15:25:36.6531944Z Exception raised from apply_magma_eig at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebra.cpp:2024 (most recent call first): 2025-12-04T15:25:36.6532120Z C++ CapturedTraceback: 2025-12-04T15:25:36.6533436Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T15:25:36.6533920Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T15:25:36.6534266Z #6 c10::detail::torchCheckFail(char const*, char const*, unsigned int, char const*) from ??:0 2025-12-04T15:25:36.6535074Z #7 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool)::{lambda()#1}::operator()() const::{lambda()#1}::operator()() const [clone .constprop.0] from BatchLinearAlgebra.cpp:0 2025-12-04T15:25:36.6535485Z #8 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T15:25:36.6535919Z #9 at::native::linalg_eig_out_info(at::Tensor const&, at::Tensor&, at::Tensor&, at::Tensor&, bool) from BatchLinearAlgebra.cpp:0 2025-12-04T15:25:36.6536192Z #10 at::native::linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T15:25:36.6536893Z #11 at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6539306Z #12 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&, at::Tensor&, at::Tensor&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&, at::Tensor&, at::Tensor&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6539601Z #13 at::_ops::linalg_eig_out::call(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T15:25:36.6539770Z #14 at::native::linalg_eig(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6541793Z #15 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA__linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6542087Z #16 at::_ops::linalg_eig::redispatch(c10::DispatchKeySet, at::Tensor const&) from ??:0 2025-12-04T15:25:36.6542646Z #17 torch::autograd::VariableType::(anonymous namespace)::linalg_eig(c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T15:25:36.6545013Z #18 c10::impl::make_boxed_from_unboxed_functor (c10::DispatchKeySet, at::Tensor const&), &torch::autograd::VariableType::(anonymous namespace)::linalg_eig>, std::tuple, c10::guts::typelist::typelist >, false>::call(c10::OperatorKernel*, c10::OperatorHandle const&, c10::DispatchKeySet, std::vector >*) from VariableType_2.cpp:0 2025-12-04T15:25:36.6545666Z #19 c10::Dispatcher::callBoxed(c10::OperatorHandle const&, std::vector >*) const [clone .isra.0] from VmapInterpreter.cpp:0 2025-12-04T15:25:36.6546341Z #20 at::functorch::Interpreter::sendToNextInterpreter(c10::OperatorHandle const&, std::vector >*, bool) from :0 2025-12-04T15:25:36.6546934Z #21 at::functorch::dynamicLayerBack(c10::OperatorHandle const&, std::vector >*, bool) from DynamicLayer.cpp:0 2025-12-04T15:25:36.6547722Z #22 c10::impl::BoxedKernelWrapper (at::Tensor const&), void>::call(c10::BoxedKernel const&, c10::OperatorHandle const&, c10::DispatchKeySet, at::Tensor const&) from :0 2025-12-04T15:25:36.6547895Z #23 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6550001Z #24 std::tuple at::functorch::linalg_eig_generated_plumbing, at::Tensor, std::optional > (*)(at::Tensor const&, std::optional), &at::functorch::(anonymous namespace)::LinalgCheckMatrixUnaryRuleHelper<&at::functorch::func_string_linalg_eig, std::tuple (*)(at::Tensor const&), &at::_ops::linalg_eig::call, c10::guts::typelist::typelist >::apply_two>(at::Tensor const&) from BatchRulesLinearAlgebra.cpp:0 2025-12-04T15:25:36.6551600Z #25 c10::impl::make_boxed_from_unboxed_functor (*)(at::Tensor const&), std::tuple, c10::guts::typelist::typelist >, false>::call(c10::OperatorKernel*, c10::OperatorHandle const&, c10::DispatchKeySet, std::vector >*) from :0 2025-12-04T15:25:36.6552248Z #26 c10::Dispatcher::callBoxed(c10::OperatorHandle const&, std::vector >*) const [clone .isra.0] from VmapInterpreter.cpp:0 2025-12-04T15:25:36.6552751Z #27 at::functorch::Interpreter::process(c10::OperatorHandle const&, std::vector >*) from :0 2025-12-04T15:25:36.6553390Z #28 at::functorch::dynamicLayerFrontFallback(c10::OperatorHandle const&, std::vector >*) from DynamicLayer.cpp:0 2025-12-04T15:25:36.6554172Z #29 c10::impl::BoxedKernelWrapper (at::Tensor const&), void>::call(c10::BoxedKernel const&, c10::OperatorHandle const&, c10::DispatchKeySet, at::Tensor const&) from :0 2025-12-04T15:25:36.6554360Z #30 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6554765Z #31 torch::autograd::THPVariable_linalg_eig(_object*, _object*, _object*) from python_linalg_functions.cpp:0 2025-12-04T15:25:36.6555092Z #32 cfunction_call from /usr/local/src/conda/python-3.10.14/Objects/methodobject.c:543 2025-12-04T15:25:36.6555372Z #33 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T15:25:36.6555631Z #34 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5917 2025-12-04T15:25:36.6556007Z #35 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6556278Z #36 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6556648Z #37 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6556911Z #38 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6557308Z #39 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6557718Z #40 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6558122Z #41 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6558413Z #42 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T15:25:36.6558674Z #43 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6559039Z #44 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6559349Z #45 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6559726Z #46 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6559979Z #47 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6560345Z #48 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6560605Z #49 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6560982Z #50 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6561397Z #51 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6561766Z #52 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6562178Z #53 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6562557Z #54 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6562961Z #55 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6563343Z #56 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6563744Z #57 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6564113Z #58 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6564524Z #59 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6564893Z #60 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6565195Z #61 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T15:25:36.6565450Z #62 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6565821Z #63 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6566185Z #64 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6566491Z #65 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6566782Z #66 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6567097Z #67 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6567504Z #68 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6567884Z #69 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6568293Z #70 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6568662Z #71 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6568927Z #72 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6569346Z #73 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6569790Z #74 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6570158Z #75 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6570564Z #76 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6570942Z #77 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6571359Z #78 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6571672Z #79 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6571964Z #80 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6572234Z #81 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T15:25:36.6572499Z #82 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6572870Z #83 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6573271Z #84 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6573649Z #85 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6574056Z #86 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6574437Z #87 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6574692Z #88 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6575060Z #89 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6575478Z #90 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6575849Z #91 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6576265Z #92 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6576631Z #93 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6576954Z #94 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6577339Z #95 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6577743Z #96 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6578110Z #97 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6578529Z #98 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6578899Z #99 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6579269Z #100 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6579582Z #101 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6579878Z #102 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6580206Z #103 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6580623Z #104 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6581011Z #105 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6581312Z #106 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6581690Z #107 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6582145Z #108 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6582527Z #109 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6582951Z #110 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6583385Z #111 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6583743Z #112 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6584094Z #113 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6584396Z #114 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6584701Z #115 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6585129Z #116 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6585507Z #117 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6585928Z #118 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6586308Z #119 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6586718Z #120 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6587106Z #121 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6587370Z #122 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6587756Z #123 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6588167Z #124 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6588541Z #125 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6588964Z #126 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6589343Z #127 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6589713Z #128 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6590020Z #129 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6590317Z #130 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6590639Z #131 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6591049Z #132 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6591423Z #133 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6591843Z #134 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6592222Z #135 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6592640Z #136 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6593011Z #137 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6593331Z #138 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T15:25:36.6593646Z #139 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T15:25:36.6593941Z #140 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T15:25:36.6594231Z #141 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T15:25:36.6594583Z #142 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T15:25:36.6594906Z #143 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T15:25:36.6595264Z #144 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T15:25:36.6595533Z #145 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T15:25:36.6595797Z #146 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T15:25:36.6596164Z #147 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T15:25:36.6596271Z #148 _start from ??:0 2025-12-04T15:25:36.6596402Z #149 from ??:0 2025-12-04T15:25:36.6596409Z 2025-12-04T15:25:36.6596413Z 2025-12-04T15:25:36.6596633Z To execute this test, run the following from the base repo dir: 2025-12-04T15:25:36.6597229Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/functorch/test_vmap.py TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda 2025-12-04T15:25:36.6597235Z 2025-12-04T15:25:36.6597512Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T15:25:36.6597741Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T15:25:36.6599675Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5048: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6599919Z isinstance(vmap(torch.min, (0, None))(t, 0), torch.return_types.min) 2025-12-04T15:25:36.6601848Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5051: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6602094Z isinstance(vmap(torch.max, (0, None))(t, 0), torch.return_types.max) 2025-12-04T15:25:36.6604011Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5055: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6604235Z vmap(torch.topk, (0, None, None))(t, 1, 0), torch.return_types.topk 2025-12-04T15:25:36.6606135Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5059: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6606420Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.6607637Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_functorch/vmap.py:433: UserWarning: Unsupported unwinding pattern: Address not in range (Triggered internally at /var/lib/jenkins/workspace/torch/csrc/profiler/unwind/unwind.cpp:219.) 2025-12-04T15:25:36.6607813Z batched_outputs = func(*batched_inputs, **kwargs) 2025-12-04T15:25:36.6608076Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T15:25:36.6610007Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5048: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6610343Z isinstance(vmap(torch.min, (0, None))(t, 0), torch.return_types.min) 2025-12-04T15:25:36.6612272Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5051: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6612513Z isinstance(vmap(torch.max, (0, None))(t, 0), torch.return_types.max) 2025-12-04T15:25:36.6614445Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5055: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6614682Z vmap(torch.topk, (0, None, None))(t, 1, 0), torch.return_types.topk 2025-12-04T15:25:36.6616592Z /var/lib/jenkins/workspace/test/functorch/test_vmap.py:5059: FutureWarning: We've integrated functorch into PyTorch. As the final step of the integration, `functorch.vmap` is deprecated as of PyTorch 2.0 and will be deleted in a future version of PyTorch >= 2.3. Please use `torch.vmap` instead; see the PyTorch 2.0 release notes and/or the `torch.func` migration guide for more details https://pytorch.org/docs/main/func.migrating.html 2025-12-04T15:25:36.6616939Z isinstance(vmap(torch.linalg.eig, (0))(t), torch.return_types.linalg_eig) 2025-12-04T15:25:36.6617629Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-12ecde9ddbb651c9.xml - 2025-12-04T15:25:36.6617813Z =========================== short test summary info ============================ 2025-12-04T15:25:36.6619117Z FAILED [0.1835s] functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_torch_return_types_returns_cuda - RuntimeError: Calling torch.linalg.eig with MAGMA requires compiling PyTorch with MAGMA. Either transfer the tensor to the CPU before calling torch.linalg.eig or use cuSolver. 2025-12-04T15:25:36.6619882Z Exception raised from apply_magma_eig at /var/lib/jenkins/workspace/aten/src/ATen/native/cuda/linalg/BatchLinearAlgebra.cpp:2024 (most recent call first): 2025-12-04T15:25:36.6619993Z C++ CapturedTraceback: 2025-12-04T15:25:36.6621301Z #4 std::_Function_handler, std::allocator > > const> (), c10::SetStackTraceFetcher(std::function, std::allocator > ()>)::{lambda()#1}>::_M_invoke(std::_Any_data const&) from Logging.cpp:0 2025-12-04T15:25:36.6621800Z #5 c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string, std::allocator >) from ??:0 2025-12-04T15:25:36.6622137Z #6 c10::detail::torchCheckFail(char const*, char const*, unsigned int, char const*) from ??:0 2025-12-04T15:25:36.6623026Z #7 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool)::{lambda()#1}::operator()() const::{lambda()#1}::operator()() const [clone .constprop.0] from BatchLinearAlgebra.cpp:0 2025-12-04T15:25:36.6623430Z #8 at::native::lazy_linalg::linalg_eig_magma(at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor const&, bool) from ??:0 2025-12-04T15:25:36.6623867Z #9 at::native::linalg_eig_out_info(at::Tensor const&, at::Tensor&, at::Tensor&, at::Tensor&, bool) from BatchLinearAlgebra.cpp:0 2025-12-04T15:25:36.6624205Z #10 at::native::linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T15:25:36.6624815Z #11 at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out(at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6627222Z #12 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&, at::Tensor&, at::Tensor&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA_out_linalg_eig_out>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&, at::Tensor&, at::Tensor&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&, at::Tensor&, at::Tensor&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6627503Z #13 at::_ops::linalg_eig_out::call(at::Tensor const&, at::Tensor&, at::Tensor&) from ??:0 2025-12-04T15:25:36.6627683Z #14 at::native::linalg_eig(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6629707Z #15 c10::impl::wrap_kernel_functor_unboxed_ (at::Tensor const&), &at::(anonymous namespace)::(anonymous namespace)::wrapper_CUDA__linalg_eig>, std::tuple, c10::guts::typelist::typelist >, std::tuple (at::Tensor const&)>::call(c10::OperatorKernel*, c10::DispatchKeySet, at::Tensor const&) from RegisterCUDA_0.cpp:0 2025-12-04T15:25:36.6630013Z #16 at::_ops::linalg_eig::redispatch(c10::DispatchKeySet, at::Tensor const&) from ??:0 2025-12-04T15:25:36.6630551Z #17 torch::autograd::VariableType::(anonymous namespace)::linalg_eig(c10::DispatchKeySet, at::Tensor const&) from VariableType_2.cpp:0 2025-12-04T15:25:36.6632861Z #18 c10::impl::make_boxed_from_unboxed_functor (c10::DispatchKeySet, at::Tensor const&), &torch::autograd::VariableType::(anonymous namespace)::linalg_eig>, std::tuple, c10::guts::typelist::typelist >, false>::call(c10::OperatorKernel*, c10::OperatorHandle const&, c10::DispatchKeySet, std::vector >*) from VariableType_2.cpp:0 2025-12-04T15:25:36.6633496Z #19 c10::Dispatcher::callBoxed(c10::OperatorHandle const&, std::vector >*) const [clone .isra.0] from VmapInterpreter.cpp:0 2025-12-04T15:25:36.6634117Z #20 at::functorch::Interpreter::sendToNextInterpreter(c10::OperatorHandle const&, std::vector >*, bool) from :0 2025-12-04T15:25:36.6634720Z #21 at::functorch::dynamicLayerBack(c10::OperatorHandle const&, std::vector >*, bool) from DynamicLayer.cpp:0 2025-12-04T15:25:36.6635495Z #22 c10::impl::BoxedKernelWrapper (at::Tensor const&), void>::call(c10::BoxedKernel const&, c10::OperatorHandle const&, c10::DispatchKeySet, at::Tensor const&) from :0 2025-12-04T15:25:36.6635682Z #23 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6637831Z #24 std::tuple at::functorch::linalg_eig_generated_plumbing, at::Tensor, std::optional > (*)(at::Tensor const&, std::optional), &at::functorch::(anonymous namespace)::LinalgCheckMatrixUnaryRuleHelper<&at::functorch::func_string_linalg_eig, std::tuple (*)(at::Tensor const&), &at::_ops::linalg_eig::call, c10::guts::typelist::typelist >::apply_two>(at::Tensor const&) from BatchRulesLinearAlgebra.cpp:0 2025-12-04T15:25:36.6639419Z #25 c10::impl::make_boxed_from_unboxed_functor (*)(at::Tensor const&), std::tuple, c10::guts::typelist::typelist >, false>::call(c10::OperatorKernel*, c10::OperatorHandle const&, c10::DispatchKeySet, std::vector >*) from :0 2025-12-04T15:25:36.6640105Z #26 c10::Dispatcher::callBoxed(c10::OperatorHandle const&, std::vector >*) const [clone .isra.0] from VmapInterpreter.cpp:0 2025-12-04T15:25:36.6640624Z #27 at::functorch::Interpreter::process(c10::OperatorHandle const&, std::vector >*) from :0 2025-12-04T15:25:36.6641252Z #28 at::functorch::dynamicLayerFrontFallback(c10::OperatorHandle const&, std::vector >*) from DynamicLayer.cpp:0 2025-12-04T15:25:36.6642037Z #29 c10::impl::BoxedKernelWrapper (at::Tensor const&), void>::call(c10::BoxedKernel const&, c10::OperatorHandle const&, c10::DispatchKeySet, at::Tensor const&) from :0 2025-12-04T15:25:36.6642217Z #30 at::_ops::linalg_eig::call(at::Tensor const&) from ??:0 2025-12-04T15:25:36.6642620Z #31 torch::autograd::THPVariable_linalg_eig(_object*, _object*, _object*) from python_linalg_functions.cpp:0 2025-12-04T15:25:36.6642951Z #32 cfunction_call from /usr/local/src/conda/python-3.10.14/Objects/methodobject.c:543 2025-12-04T15:25:36.6643225Z #33 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T15:25:36.6643484Z #34 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5917 2025-12-04T15:25:36.6643870Z #35 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6644126Z #36 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6644505Z #37 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6644761Z #38 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6645138Z #39 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6645553Z #40 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6645922Z #41 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6646224Z #42 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T15:25:36.6646479Z #43 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6646850Z #44 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6647112Z #45 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6647483Z #46 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6647738Z #47 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6648116Z #48 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6648368Z #49 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6648743Z #50 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6649178Z #51 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6649546Z #52 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6649987Z #53 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6650361Z #54 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6650773Z #55 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6651201Z #56 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6651607Z #57 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6651989Z #58 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6652393Z #59 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6652772Z #60 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6653068Z #61 PyVectorcall_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:267 2025-12-04T15:25:36.6653322Z #62 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6653701Z #63 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6654052Z #64 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6654356Z #65 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6654661Z #66 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6654970Z #67 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6655386Z #68 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6655757Z #69 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6656162Z #70 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6656543Z #71 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6656803Z #72 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6657256Z #73 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6657661Z #74 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6658031Z #75 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6658447Z #76 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6658820Z #77 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6659183Z #78 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6659486Z #79 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6659777Z #80 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6660058Z #81 _PyObject_Call from /usr/local/src/conda/python-3.10.14/Objects/call.c:305 2025-12-04T15:25:36.6660313Z #82 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6660682Z #83 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6661142Z #84 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6661514Z #85 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6661954Z #86 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6662324Z #87 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6662580Z #88 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6663036Z #89 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6663444Z #90 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6663823Z #91 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6664229Z #92 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6664598Z #93 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6664864Z #94 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6665243Z #95 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6665662Z #96 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6666038Z #97 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6666445Z #98 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6666829Z #99 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6667194Z #100 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6667505Z #101 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6667821Z #102 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6668128Z #103 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6668557Z #104 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6668941Z #105 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6669206Z #106 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6669598Z #107 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6670016Z #108 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6670412Z #109 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6670828Z #110 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6671205Z #111 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6671583Z #112 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6671894Z #113 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6672214Z #114 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6672525Z #115 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6672941Z #116 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6673367Z #117 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6673782Z #118 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6674187Z #119 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6674612Z #120 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6674988Z #121 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6675317Z #122 do_call_core from /usr/local/src/conda/python-3.10.14/Python/ceval.c:5945 2025-12-04T15:25:36.6675697Z #123 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6676110Z #124 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6676505Z #125 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6676913Z #126 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6677304Z #127 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6677662Z #128 _PyObject_FastCallDictTstate from /usr/local/src/conda/python-3.10.14/Objects/call.c:153 2025-12-04T15:25:36.6677975Z #129 _PyObject_Call_Prepend from /usr/local/src/conda/python-3.10.14/Objects/call.c:431 2025-12-04T15:25:36.6678292Z #130 slot_tp_call from /usr/local/src/conda/python-3.10.14/Objects/typeobject.c:7494 2025-12-04T15:25:36.6678596Z #131 _PyObject_MakeTpCall from /usr/local/src/conda/python-3.10.14/Objects/call.c:215 2025-12-04T15:25:36.6679009Z #132 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:112 2025-12-04T15:25:36.6679405Z #133 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6679817Z #134 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6680210Z #135 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6680624Z #136 _PyObject_VectorcallTstate from /usr/local/src/conda/python-3.10.14/Include/cpython/abstract.h:114 2025-12-04T15:25:36.6681000Z #137 _PyEval_EvalFrame from /usr/local/src/conda/python-3.10.14/Include/internal/pycore_ceval.h:46 2025-12-04T15:25:36.6681308Z #138 PyEval_EvalCode from /usr/local/src/conda/python-3.10.14/Python/ceval.c:1134 2025-12-04T15:25:36.6681614Z #139 run_eval_code_obj from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1291 2025-12-04T15:25:36.6681894Z #140 run_mod from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1312 2025-12-04T15:25:36.6682178Z #141 pyrun_file from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:1208 2025-12-04T15:25:36.6682531Z #142 _PyRun_SimpleFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:456 2025-12-04T15:25:36.6682869Z #143 _PyRun_AnyFileObject from /usr/local/src/conda/python-3.10.14/Python/pythonrun.c:90 2025-12-04T15:25:36.6683159Z #144 pymain_run_file_obj from /usr/local/src/conda/python-3.10.14/Modules/main.c:357 2025-12-04T15:25:36.6683439Z #145 Py_BytesMain from /usr/local/src/conda/python-3.10.14/Modules/main.c:1090 2025-12-04T15:25:36.6683704Z #146 __libc_start_call_main from ./csu/../sysdeps/nptl/libc_start_call_main.h:58 2025-12-04T15:25:36.6683906Z #147 __libc_start_main_impl from ./csu/../csu/libc-start.c:392 2025-12-04T15:25:36.6684017Z #148 _start from ??:0 2025-12-04T15:25:36.6684138Z #149 from ??:0 2025-12-04T15:25:36.6684144Z 2025-12-04T15:25:36.6684149Z 2025-12-04T15:25:36.6684369Z To execute this test, run the following from the base repo dir: 2025-12-04T15:25:36.6685016Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=1 python test/functorch/test_vmap.py TestVmapOperatorsOpInfoCUDA.test_torch_return_types_returns_cuda 2025-12-04T15:25:36.6685022Z 2025-12-04T15:25:36.6685291Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T15:25:36.6685517Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-12-04T15:25:36.6685721Z ================= 1 failed, 2133 deselected, 2 rerun in 15.97s ================= 2025-12-04T15:25:36.6685818Z Got exit code 1 2025-12-04T15:25:36.6686348Z FAILED CONSISTENTLY: test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_torch_return_types_returns_cuda 2025-12-04T15:25:36.6686814Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-12-04T15:25:36.6687317Z Test results will be stored in test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-94056324a12f0026.xml 2025-12-04T15:25:36.6687479Z ============================= test session starts ============================== 2025-12-04T15:25:36.6687830Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T15:25:36.6687950Z cachedir: .pytest_cache 2025-12-04T15:25:36.6688475Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T15:25:36.6688601Z rootdir: /var/lib/jenkins/workspace 2025-12-04T15:25:36.6688721Z configfile: pytest.ini 2025-12-04T15:25:36.6689309Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, anyio-4.12.0, typeguard-4.3.0 2025-12-04T15:25:36.6689562Z collecting ... collected 2134 items / 1007 deselected / 1127 selected 2025-12-04T15:25:36.6689782Z stepcurrent: skipping 1007 already run items. 2025-12-04T15:25:36.6689899Z Running 1127 items in this shard 2025-12-04T15:25:36.6689905Z 2025-12-04T15:25:36.6690366Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_escaped_error_cuda PASSED [0.0106s] [ 0%] 2025-12-04T15:25:36.6691005Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_CubeGenVmapAutogradFunction_cuda_float32 PASSED [0.0308s] [ 0%] 2025-12-04T15:25:36.6691718Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ForwardHasDefaultArgsAutogradFunction_cuda_float32 PASSED [0.2089s] [ 0%] 2025-12-04T15:25:36.6692191Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_H_cuda_float32 PASSED [0.0131s] [ 0%] 2025-12-04T15:25:36.6692818Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_MulGenVmapAutogradFunction_cuda_float32 PASSED [0.0196s] [ 0%] 2025-12-04T15:25:36.6693393Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyCatCustomOp_cuda_float32 PASSED [0.0072s] [ 0%] 2025-12-04T15:25:36.6694008Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyCubeAutogradFunction_cuda_float32 PASSED [0.0145s] [ 0%] 2025-12-04T15:25:36.6694584Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyCubeCustomOp_cuda_float32 PASSED [0.0135s] [ 0%] 2025-12-04T15:25:36.6695279Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyCubeNotComposableAutogradFunction_cuda_float32 XFAIL [0.0354s] [ 0%] 2025-12-04T15:25:36.6695948Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyExpMarkDirtyAutogradFunction_cuda_float32 PASSED [0.2162s] [ 0%] 2025-12-04T15:25:36.6696759Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyMulAutogradFunction_cuda_float32 PASSED [0.0191s] [ 0%] 2025-12-04T15:25:36.6697381Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyMulCustomOp_cuda_float32 PASSED [0.0154s] [ 1%] 2025-12-04T15:25:36.6697993Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyMulScalarCustomOp_cuda_float32 PASSED [0.0109s] [ 1%] 2025-12-04T15:25:36.6698615Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyNMSCustomOp_cuda_float32 XFAIL [0.0129s] [ 1%] 2025-12-04T15:25:36.6699232Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyNonzeroCustomOp_cuda_float32 XFAIL [0.2054s] [ 1%] 2025-12-04T15:25:36.6699868Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpySortAutogradFunction_cuda_float32 PASSED [0.2127s] [ 1%] 2025-12-04T15:25:36.6700434Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpySortCustomOp_cuda_float32 PASSED [0.0149s] [ 1%] 2025-12-04T15:25:36.6701124Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpySplitCopyCustomOp_cuda_float32 PASSED [0.0168s] [ 1%] 2025-12-04T15:25:36.6701765Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpySplitCopyWithIntCustomOp_cuda_float32 PASSED [0.0174s] [ 1%] 2025-12-04T15:25:36.6702385Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyTakeAutogradFunction_cuda_float32 PASSED [0.0309s] [ 1%] 2025-12-04T15:25:36.6702957Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyTakeCustomOp_cuda_float32 PASSED [0.0263s] [ 1%] 2025-12-04T15:25:36.6703552Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_NumpyViewCopyCustomOp_cuda_float32 PASSED [0.2034s] [ 1%] 2025-12-04T15:25:36.6704224Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ScaleGradGenVmapAutogradFunction_cuda_float32 PASSED [0.2071s] [ 2%] 2025-12-04T15:25:36.6704821Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_SelectAutogradFunction_cuda_float32 PASSED [0.2023s] [ 2%] 2025-12-04T15:25:36.6705474Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_SelectGenVmapAutogradFunction_cuda_float32 PASSED [0.0130s] [ 2%] 2025-12-04T15:25:36.6706103Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_SortGenVmapAutogradFunction_cuda_float32 PASSED [0.0170s] [ 2%] 2025-12-04T15:25:36.6706582Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_T_cuda_float32 PASSED [0.0134s] [ 2%] 2025-12-04T15:25:36.6707288Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ZeroGradientsGenVmapAutogradFunction_cuda_float32 PASSED [0.0213s] [ 2%] 2025-12-04T15:25:36.6707810Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___getitem___cuda_float32 XFAIL [0.0685s] [ 2%] 2025-12-04T15:25:36.6708394Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___getitem___functorch_cuda_float32 PASSED [0.4414s] [ 2%] 2025-12-04T15:25:36.6708907Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___radd___cuda_float32 PASSED [0.6003s] [ 2%] 2025-12-04T15:25:36.6709399Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___rand___cuda_int64 PASSED [0.4407s] [ 2%] 2025-12-04T15:25:36.6709917Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___rdiv___cuda_float32 PASSED [0.6287s] [ 2%] 2025-12-04T15:25:36.6710436Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___rmatmul___cuda_float32 PASSED [0.9334s] [ 3%] 2025-12-04T15:25:36.6710950Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___rmod___cuda_float32 PASSED [0.6047s] [ 3%] 2025-12-04T15:25:36.6711451Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___rmul___cuda_float32 PASSED [0.5983s] [ 3%] 2025-12-04T15:25:36.6711937Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___ror___cuda_int64 PASSED [0.4426s] [ 3%] 2025-12-04T15:25:36.6712457Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___rpow___cuda_float32 PASSED [0.6116s] [ 3%] 2025-12-04T15:25:36.6712947Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___rsub___cuda_float32 XFAIL [0.2435s] [ 3%] 2025-12-04T15:25:36.6713488Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive___rxor___cuda_int64 PASSED [0.6832s] [ 3%] 2025-12-04T15:25:36.6714107Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive__batch_norm_with_update_cuda_float32 XFAIL [0.2343s] [ 3%] 2025-12-04T15:25:36.6714626Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive__chunk_cat_cuda_float32 PASSED [0.2329s] [ 3%] 2025-12-04T15:25:36.6715217Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive__native_batch_norm_legit_cuda_float32 XFAIL [0.0154s] [ 3%] 2025-12-04T15:25:36.6715856Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive__segment_reduce_lengths_cuda_float32 PASSED [0.6202s] [ 3%] 2025-12-04T15:25:36.6716449Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive__segment_reduce_offsets_cuda_float32 PASSED [0.3320s] [ 3%] 2025-12-04T15:25:36.6717085Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive__softmax_backward_data_cuda_float32 SKIPPED [0.0003s] (Skipped!) [ 4%] 2025-12-04T15:25:36.6717649Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive__unsafe_masked_index_cuda_float32 PASSED [0.2542s] [ 4%] 2025-12-04T15:25:36.6718298Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive__unsafe_masked_index_put_accumulate_cuda_float32 PASSED [0.7847s] [ 4%] 2025-12-04T15:25:36.6718882Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive__upsample_bilinear2d_aa_cuda_float32 PASSED [0.0301s] [ 4%] 2025-12-04T15:25:36.6719374Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_abs_cuda_float32 PASSED [0.3819s] [ 4%] 2025-12-04T15:25:36.6719867Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_acos_cuda_float32 PASSED [0.4067s] [ 4%] 2025-12-04T15:25:36.6720361Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_acosh_cuda_float32 PASSED [0.4065s] [ 4%] 2025-12-04T15:25:36.6720859Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_add_cuda_float32 PASSED [0.8904s] [ 4%] 2025-12-04T15:25:36.6721356Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_addbmm_cuda_float32 PASSED [0.1250s] [ 4%] 2025-12-04T15:25:36.6721863Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_addcdiv_cuda_float32 XFAIL [0.5941s] [ 4%] 2025-12-04T15:25:36.6722354Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_addcmul_cuda_float32 XFAIL [0.7897s] [ 4%] 2025-12-04T15:25:36.6722843Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_addmm_cuda_float32 PASSED [0.3130s] [ 5%] 2025-12-04T15:25:36.6723408Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_addmm_decomposed_cuda_float32 PASSED [0.1126s] [ 5%] 2025-12-04T15:25:36.6723898Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_addmv_cuda_float32 PASSED [0.1024s] [ 5%] 2025-12-04T15:25:36.6724395Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_addr_cuda_float32 PASSED [0.1068s] [ 5%] 2025-12-04T15:25:36.6724916Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_alias_copy_cuda_float32 PASSED [0.0123s] [ 5%] 2025-12-04T15:25:36.6725395Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_all_cuda_float32 PASSED [0.0644s] [ 5%] 2025-12-04T15:25:36.6725904Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_allclose_cuda_float32 XFAIL [0.0084s] [ 5%] 2025-12-04T15:25:36.6726392Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_amax_cuda_float32 PASSED [0.2677s] [ 5%] 2025-12-04T15:25:36.6726887Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_amin_cuda_float32 PASSED [0.0830s] [ 5%] 2025-12-04T15:25:36.6727382Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_aminmax_cuda_float32 PASSED [0.0493s] [ 5%] 2025-12-04T15:25:36.6727907Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_angle_cuda_float32 PASSED [0.1137s] [ 5%] 2025-12-04T15:25:36.6728424Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_any_cuda_float32 PASSED [0.0650s] [ 6%] 2025-12-04T15:25:36.6728919Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_arange_cuda_float32 PASSED [0.0277s] [ 6%] 2025-12-04T15:25:36.6729424Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_argmax_cuda_float32 PASSED [0.0446s] [ 6%] 2025-12-04T15:25:36.6729916Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_argmin_cuda_float32 PASSED [0.0451s] [ 6%] 2025-12-04T15:25:36.6730472Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_argsort_cuda_float32 PASSED [0.1214s] [ 6%] 2025-12-04T15:25:36.6730989Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_argwhere_cuda_float32 PASSED [0.0277s] [ 6%] 2025-12-04T15:25:36.6731527Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_as_strided_copy_cuda_float32 XFAIL [0.0149s] [ 6%] 2025-12-04T15:25:36.6732050Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_as_strided_cuda_float32 XFAIL [0.2121s] [ 6%] 2025-12-04T15:25:36.6732629Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_as_strided_partial_views_cuda_float32 XFAIL [0.1990s] [ 6%] 2025-12-04T15:25:36.6733174Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_as_strided_scatter_cuda_float32 XFAIL [0.2018s] [ 6%] 2025-12-04T15:25:36.6733676Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_asin_cuda_float32 PASSED [0.5682s] [ 6%] 2025-12-04T15:25:36.6734166Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_asinh_cuda_float32 PASSED [0.5773s] [ 7%] 2025-12-04T15:25:36.6734669Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_atan2_cuda_float32 PASSED [1.4903s] [ 7%] 2025-12-04T15:25:36.6735156Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_atan_cuda_float32 PASSED [0.3795s] [ 7%] 2025-12-04T15:25:36.6735647Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_atanh_cuda_float32 PASSED [0.3820s] [ 7%] 2025-12-04T15:25:36.6736182Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_atleast_1d_cuda_float32 PASSED [0.0218s] [ 7%] 2025-12-04T15:25:36.6736704Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_atleast_2d_cuda_float32 PASSED [0.0212s] [ 7%] 2025-12-04T15:25:36.6737300Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_atleast_3d_cuda_float32 PASSED [0.0216s] [ 7%] 2025-12-04T15:25:36.6737805Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_baddbmm_cuda_float32 PASSED [0.1079s] [ 7%] 2025-12-04T15:25:36.6738315Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bernoulli_cuda_float32 XFAIL [0.0186s] [ 7%] 2025-12-04T15:25:36.6738821Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bfloat16_cuda_float32 XFAIL [0.2242s] [ 7%] 2025-12-04T15:25:36.6739469Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bfloat16_functorch_no_channels_last_cuda_float32 PASSED [0.2097s] [ 7%] 2025-12-04T15:25:36.6739975Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bincount_cuda_int64 PASSED [0.2630s] [ 7%] 2025-12-04T15:25:36.6740487Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bitwise_and_cuda_int64 PASSED [0.6249s] [ 8%] 2025-12-04T15:25:36.6741035Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bitwise_left_shift_cuda_int64 PASSED [0.6232s] [ 8%] 2025-12-04T15:25:36.6741559Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bitwise_not_cuda_int64 PASSED [0.1533s] [ 8%] 2025-12-04T15:25:36.6742103Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bitwise_or_cuda_int64 PASSED [0.6255s] [ 8%] 2025-12-04T15:25:36.6742676Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bitwise_right_shift_cuda_int64 PASSED [0.6269s] [ 8%] 2025-12-04T15:25:36.6743217Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bitwise_xor_cuda_int64 PASSED [0.6266s] [ 8%] 2025-12-04T15:25:36.6743732Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_block_diag_cuda_float32 PASSED [0.0530s] [ 8%] 2025-12-04T15:25:36.6744228Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bmm_cuda_float32 PASSED [0.0128s] [ 8%] 2025-12-04T15:25:36.6744764Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bool_cuda_float32 XFAIL [0.0196s] [ 8%] 2025-12-04T15:25:36.6745395Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bool_functorch_no_channels_last_cuda_float32 PASSED [0.2645s] [ 8%] 2025-12-04T15:25:36.6745944Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_broadcast_shapes_cuda_float32 PASSED [0.0095s] [ 8%] 2025-12-04T15:25:36.6746503Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_broadcast_tensors_cuda_float32 PASSED [0.2500s] [ 9%] 2025-12-04T15:25:36.6747047Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_broadcast_to_cuda_float32 PASSED [0.0284s] [ 9%] 2025-12-04T15:25:36.6747559Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_bucketize_cuda_float32 PASSED [0.2382s] [ 9%] 2025-12-04T15:25:36.6748050Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_byte_cuda_float32 XFAIL [0.0200s] [ 9%] 2025-12-04T15:25:36.6748673Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_byte_functorch_no_channels_last_cuda_float32 PASSED [0.2074s] [ 9%] 2025-12-04T15:25:36.6749216Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cartesian_prod_cuda_float32 PASSED [0.2180s] [ 9%] 2025-12-04T15:25:36.6749706Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cat_cuda_float32 PASSED [0.2474s] [ 9%] 2025-12-04T15:25:36.6750350Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cauchy_cuda_float32 SKIPPED [0.0003s] (Test expects tensor input) [ 9%] 2025-12-04T15:25:36.6750851Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cdist_cuda_float32 PASSED [1.4736s] [ 9%] 2025-12-04T15:25:36.6751346Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cdouble_cuda_float32 XFAIL [0.0205s] [ 9%] 2025-12-04T15:25:36.6751836Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ceil_cuda_float32 PASSED [0.3977s] [ 9%] 2025-12-04T15:25:36.6752335Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cfloat_cuda_float32 XFAIL [0.0221s] [ 10%] 2025-12-04T15:25:36.6752827Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_chalf_cuda_float32 PASSED [0.3528s] [ 10%] 2025-12-04T15:25:36.6753317Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_char_cuda_float32 XFAIL [0.0196s] [ 10%] 2025-12-04T15:25:36.6753943Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_char_functorch_no_channels_last_cuda_float32 PASSED [0.2119s] [ 10%] 2025-12-04T15:25:36.6754608Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cholesky_cuda_float32 SKIPPED [0.0016s] (no MAGMA library detected) [ 10%] 2025-12-04T15:25:36.6755308Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cholesky_inverse_cuda_float32 SKIPPED [0.0017s] (no MAGMA library detected) [ 10%] 2025-12-04T15:25:36.6755988Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cholesky_solve_cuda_float32 SKIPPED [0.0014s] (no MAGMA library detected) [ 10%] 2025-12-04T15:25:36.6756529Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_chunk_cuda_float32 PASSED [0.1227s] [ 10%] 2025-12-04T15:25:36.6757013Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_clamp_cuda_float32 XFAIL [0.2840s] [ 10%] 2025-12-04T15:25:36.6757558Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_clamp_max_cuda_float32 XFAIL [0.2057s] [ 10%] 2025-12-04T15:25:36.6758067Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_clamp_min_cuda_float32 XFAIL [0.0080s] [ 10%] 2025-12-04T15:25:36.6758546Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_clone_cuda_float32 XFAIL [0.1847s] [ 11%] 2025-12-04T15:25:36.6759147Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_column_stack_cuda_float32 PASSED [0.3688s] [ 11%] 2025-12-04T15:25:36.6759682Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_combinations_cuda_float32 PASSED [0.0992s] [ 11%] 2025-12-04T15:25:36.6760202Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_complex_cuda_float32 PASSED [0.6397s] [ 11%] 2025-12-04T15:25:36.6760689Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_conj_cuda_float32 PASSED [0.1183s] [ 11%] 2025-12-04T15:25:36.6761226Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_conj_physical_cuda_float32 PASSED [0.2015s] [ 11%] 2025-12-04T15:25:36.6761786Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_constant_pad_nd_cuda_float32 PASSED [0.1704s] [ 11%] 2025-12-04T15:25:36.6762302Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_contiguous_cuda_float32 XFAIL [0.1768s] [ 11%] 2025-12-04T15:25:36.6762831Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_copysign_cuda_float32 PASSED [1.0856s] [ 11%] 2025-12-04T15:25:36.6763340Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_corrcoef_cuda_float32 PASSED [0.0269s] [ 11%] 2025-12-04T15:25:36.6763824Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cos_cuda_float32 PASSED [0.2159s] [ 11%] 2025-12-04T15:25:36.6764326Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cosh_cuda_float32 PASSED [0.2282s] [ 11%] 2025-12-04T15:25:36.6764860Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_count_nonzero_cuda_float32 PASSED [0.0738s] [ 12%] 2025-12-04T15:25:36.6765359Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cov_cuda_float32 PASSED [0.2569s] [ 12%] 2025-12-04T15:25:36.6765851Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cross_cuda_float32 PASSED [0.0232s] [ 12%] 2025-12-04T15:25:36.6766351Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cummax_cuda_float32 PASSED [0.0193s] [ 12%] 2025-12-04T15:25:36.6766857Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cummin_cuda_float32 PASSED [0.0192s] [ 12%] 2025-12-04T15:25:36.6767357Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cumprod_cuda_float32 PASSED [0.0644s] [ 12%] 2025-12-04T15:25:36.6767865Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cumsum_cuda_float32 PASSED [0.2336s] [ 12%] 2025-12-04T15:25:36.6768448Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_cumulative_trapezoid_cuda_float32 PASSED [0.0583s] [ 12%] 2025-12-04T15:25:36.6768945Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_deg2rad_cuda_float32 PASSED [0.2101s] [ 12%] 2025-12-04T15:25:36.6769450Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_diag_cuda_float32 PASSED [0.0615s] [ 12%] 2025-12-04T15:25:36.6769968Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_diag_embed_cuda_float32 PASSED [0.1072s] [ 12%] 2025-12-04T15:25:36.6770490Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_diagflat_cuda_float32 PASSED [0.0253s] [ 13%] 2025-12-04T15:25:36.6771071Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_diagonal_copy_cuda_float32 PASSED [0.0941s] [ 13%] 2025-12-04T15:25:36.6771612Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_diagonal_cuda_float32 PASSED [0.0913s] [ 13%] 2025-12-04T15:25:36.6772174Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_diagonal_scatter_cuda_float32 PASSED [0.0965s] [ 13%] 2025-12-04T15:25:36.6772660Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_diff_cuda_float32 PASSED [0.5290s] [ 13%] 2025-12-04T15:25:36.6773227Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_digamma_cuda_float32 PASSED [0.5608s] [ 13%] 2025-12-04T15:25:36.6773710Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_dist_cuda_float32 PASSED [0.3132s] [ 13%] 2025-12-04T15:25:36.6774263Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_div_floor_rounding_cuda_float32 PASSED [1.6345s] [ 13%] 2025-12-04T15:25:36.6774843Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_div_no_rounding_mode_cuda_float32 PASSED [1.6118s] [ 13%] 2025-12-04T15:25:36.6775399Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_div_trunc_rounding_cuda_float32 PASSED [1.6331s] [ 13%] 2025-12-04T15:25:36.6775892Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_dot_cuda_float32 PASSED [0.0266s] [ 13%] 2025-12-04T15:25:36.6776381Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_double_cuda_float32 XFAIL [0.0198s] [ 14%] 2025-12-04T15:25:36.6777083Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_double_functorch_no_channels_last_cuda_float32 PASSED [0.2401s] [ 14%] 2025-12-04T15:25:36.6777593Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_dsplit_cuda_float32 PASSED [0.0234s] [ 14%] 2025-12-04T15:25:36.6778088Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_dstack_cuda_float32 PASSED [0.1983s] [ 14%] 2025-12-04T15:25:36.6778593Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_einsum_cuda_float32 PASSED [0.0113s] [ 14%] 2025-12-04T15:25:36.6779086Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_empty_cuda_float32 PASSED [0.0077s] [ 14%] 2025-12-04T15:25:36.6779604Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_empty_like_cuda_float32 PASSED [0.0723s] [ 14%] 2025-12-04T15:25:36.6780154Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_empty_permuted_cuda_float32 PASSED [0.0261s] [ 14%] 2025-12-04T15:25:36.6780687Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_empty_strided_cuda_float32 PASSED [0.0073s] [ 14%] 2025-12-04T15:25:36.6781175Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_eq_cuda_float32 XFAIL [0.2601s] [ 14%] 2025-12-04T15:25:36.6781663Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_equal_cuda_float32 XFAIL [0.1970s] [ 14%] 2025-12-04T15:25:36.6782140Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_erf_cuda_float32 PASSED [0.4799s] [ 14%] 2025-12-04T15:25:36.6782641Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_erfc_cuda_float32 PASSED [0.3194s] [ 15%] 2025-12-04T15:25:36.6783135Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_erfinv_cuda_float32 PASSED [0.2902s] [ 15%] 2025-12-04T15:25:36.6783630Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_exp2_cuda_float32 PASSED [0.3166s] [ 15%] 2025-12-04T15:25:36.6784116Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_exp_cuda_float32 PASSED [0.2232s] [ 15%] 2025-12-04T15:25:36.6784632Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_expand_as_cuda_float32 PASSED [0.0221s] [ 15%] 2025-12-04T15:25:36.6785209Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_expand_copy_cuda_float32 PASSED [0.0357s] [ 15%] 2025-12-04T15:25:36.6785708Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_expand_cuda_float32 PASSED [0.0343s] [ 15%] 2025-12-04T15:25:36.6786238Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_expm1_cuda_float32 PASSED [0.2857s] [ 15%] 2025-12-04T15:25:36.6786761Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_exponential_cuda_float32 XFAIL [0.0143s] [ 15%] 2025-12-04T15:25:36.6787243Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_eye_cuda_float32 PASSED [0.2565s] [ 15%] 2025-12-04T15:25:36.6787827Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_fft2_cuda_float32 PASSED [0.2437s] [ 15%] 2025-12-04T15:25:36.6788323Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_fft_cuda_float32 PASSED [0.0491s] [ 16%] 2025-12-04T15:25:36.6788845Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_fftn_cuda_float32 PASSED [0.0479s] [ 16%] 2025-12-04T15:25:36.6789374Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_fftshift_cuda_float32 PASSED [0.0216s] [ 16%] 2025-12-04T15:25:36.6789885Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_hfft2_cuda_float32 PASSED [0.0481s] [ 16%] 2025-12-04T15:25:36.6790402Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_hfft_cuda_float32 PASSED [0.0450s] [ 16%] 2025-12-04T15:25:36.6790912Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_hfftn_cuda_float32 PASSED [0.0456s] [ 16%] 2025-12-04T15:25:36.6791434Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_ifft2_cuda_float32 PASSED [0.2325s] [ 16%] 2025-12-04T15:25:36.6791938Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_ifft_cuda_float32 PASSED [0.0427s] [ 16%] 2025-12-04T15:25:36.6792447Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_ifftn_cuda_float32 PASSED [0.2290s] [ 16%] 2025-12-04T15:25:36.6792994Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_ifftshift_cuda_float32 PASSED [0.0221s] [ 16%] 2025-12-04T15:25:36.6793505Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_ihfft2_cuda_float32 PASSED [0.0469s] [ 16%] 2025-12-04T15:25:36.6794028Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_ihfft_cuda_float32 PASSED [0.0464s] [ 17%] 2025-12-04T15:25:36.6794545Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_ihfftn_cuda_float32 PASSED [0.0483s] [ 17%] 2025-12-04T15:25:36.6795056Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_irfft2_cuda_float32 PASSED [0.0442s] [ 17%] 2025-12-04T15:25:36.6795576Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_irfft_cuda_float32 PASSED [0.2248s] [ 17%] 2025-12-04T15:25:36.6796263Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_irfftn_cuda_float32 PASSED [0.0434s] [ 17%] 2025-12-04T15:25:36.6796786Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_rfft2_cuda_float32 PASSED [0.2228s] [ 17%] 2025-12-04T15:25:36.6797293Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_rfft_cuda_float32 PASSED [0.0387s] [ 17%] 2025-12-04T15:25:36.6797801Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fft_rfftn_cuda_float32 PASSED [0.2262s] [ 17%] 2025-12-04T15:25:36.6798307Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fill_cuda_float32 PASSED [0.2355s] [ 17%] 2025-12-04T15:25:36.6798808Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_flatten_cuda_float32 PASSED [0.1334s] [ 17%] 2025-12-04T15:25:36.6799374Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_flip_cuda_float32 PASSED [0.0328s] [ 17%] 2025-12-04T15:25:36.6799874Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fliplr_cuda_float32 PASSED [0.2138s] [ 18%] 2025-12-04T15:25:36.6800415Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_flipud_cuda_float32 PASSED [0.0163s] [ 18%] 2025-12-04T15:25:36.6800917Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_float_cuda_float32 XFAIL [0.0192s] [ 18%] 2025-12-04T15:25:36.6801548Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_float_functorch_no_channels_last_cuda_float32 PASSED [0.2168s] [ 18%] 2025-12-04T15:25:36.6802163Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_float_power_cuda_float32 PASSED [0.8094s] [ 18%] 2025-12-04T15:25:36.6802650Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_floor_cuda_float32 PASSED [0.2042s] [ 18%] 2025-12-04T15:25:36.6803177Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_floor_divide_cuda_float32 PASSED [0.8773s] [ 18%] 2025-12-04T15:25:36.6803678Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fmax_cuda_float32 PASSED [0.7939s] [ 18%] 2025-12-04T15:25:36.6804167Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fmin_cuda_float32 PASSED [0.5771s] [ 18%] 2025-12-04T15:25:36.6804665Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_fmod_cuda_float32 PASSED [0.8744s] [ 18%] 2025-12-04T15:25:36.6805149Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_frac_cuda_float32 PASSED [0.2047s] [ 18%] 2025-12-04T15:25:36.6805647Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_frexp_cuda_float32 PASSED [0.1799s] [ 18%] 2025-12-04T15:25:36.6806144Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_full_cuda_float32 PASSED [0.0074s] [ 19%] 2025-12-04T15:25:36.6806652Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_full_like_cuda_float32 PASSED [0.2509s] [ 19%] 2025-12-04T15:25:36.6807157Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_gather_cuda_float32 PASSED [0.0674s] [ 19%] 2025-12-04T15:25:36.6807631Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_gcd_cuda_int64 PASSED [0.5814s] [ 19%] 2025-12-04T15:25:36.6808100Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ge_cuda_float32 XFAIL [0.4551s] [ 19%] 2025-12-04T15:25:36.6808777Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_geometric_cuda_float32 SKIPPED [0.0003s] (Test expects tensor input) [ 19%] 2025-12-04T15:25:36.6809279Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_geqrf_cuda_float32 PASSED [0.5017s] [ 19%] 2025-12-04T15:25:36.6809798Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_gradient_cuda_float32 PASSED [0.3025s] [ 19%] 2025-12-04T15:25:36.6810341Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_grid_sampler_2d_cuda_float32 PASSED [0.1169s] [ 19%] 2025-12-04T15:25:36.6810939Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_grid_sampler_3d_cuda_float32 SKIPPED [0.0003s] (Skipped!) [ 19%] 2025-12-04T15:25:36.6811422Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_gt_cuda_float32 XFAIL [0.4498s] [ 19%] 2025-12-04T15:25:36.6811900Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_half_cuda_float32 XFAIL [0.2163s] [ 20%] 2025-12-04T15:25:36.6812533Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_half_functorch_no_channels_last_cuda_float32 PASSED [0.2048s] [ 20%] 2025-12-04T15:25:36.6813117Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_hash_tensor_cuda_float32 SKIPPED [0.0003s] (Skipped!) [ 20%] 2025-12-04T15:25:36.6813634Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_heaviside_cuda_float32 PASSED [0.7955s] [ 20%] 2025-12-04T15:25:36.6814165Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_histc_cuda_float32 PASSED [0.3348s] [ 20%] 2025-12-04T15:25:36.6814689Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_hsplit_cuda_float32 PASSED [0.2205s] [ 20%] 2025-12-04T15:25:36.6815194Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_hstack_cuda_float32 PASSED [0.1945s] [ 20%] 2025-12-04T15:25:36.6815680Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_hypot_cuda_float32 PASSED [0.7893s] [ 20%] 2025-12-04T15:25:36.6816225Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_i0_cuda_float32 PASSED [0.3048s] [ 20%] 2025-12-04T15:25:36.6816728Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_igamma_cuda_float32 PASSED [1.6131s] [ 20%] 2025-12-04T15:25:36.6817296Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_igammac_cuda_float32 PASSED [1.6430s] [ 20%] 2025-12-04T15:25:36.6817808Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_imag_cuda_complex64 PASSED [0.1666s] [ 21%] 2025-12-04T15:25:36.6818321Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_add_cuda_float32 PASSED [0.1984s] [ 21%] 2025-12-04T15:25:36.6818841Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_copy_cuda_float32 PASSED [0.0673s] [ 21%] 2025-12-04T15:25:36.6819368Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_fill_cuda_float32 PASSED [0.1045s] [ 21%] 2025-12-04T15:25:36.6819876Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_put_cuda_float32 XFAIL [0.0309s] [ 21%] 2025-12-04T15:25:36.6820452Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_put_functorch_cuda_float32 PASSED [0.4237s] [ 21%] 2025-12-04T15:25:36.6821004Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_reduce_amax_cuda_float32 PASSED [0.1861s] [ 21%] 2025-12-04T15:25:36.6821557Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_reduce_amin_cuda_float32 PASSED [0.1837s] [ 21%] 2025-12-04T15:25:36.6822126Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_reduce_mean_cuda_float32 PASSED [0.2230s] [ 21%] 2025-12-04T15:25:36.6822678Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_reduce_prod_cuda_float32 PASSED [0.1851s] [ 21%] 2025-12-04T15:25:36.6823214Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_index_select_cuda_float32 PASSED [0.2188s] [ 21%] 2025-12-04T15:25:36.6823706Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_inner_cuda_float32 PASSED [0.0186s] [ 22%] 2025-12-04T15:25:36.6824177Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_int_cuda_float32 XFAIL [0.0194s] [ 22%] 2025-12-04T15:25:36.6824807Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_int_functorch_no_channels_last_cuda_float32 PASSED [0.0176s] [ 22%] 2025-12-04T15:25:36.6825310Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_isclose_cuda_float32 PASSED [1.0423s] [ 22%] 2025-12-04T15:25:36.6825829Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_isfinite_cuda_float32 PASSED [0.1343s] [ 22%] 2025-12-04T15:25:36.6826318Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_isin_cuda_float32 PASSED [0.0269s] [ 22%] 2025-12-04T15:25:36.6826811Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_isinf_cuda_float32 PASSED [0.1160s] [ 22%] 2025-12-04T15:25:36.6827312Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_isnan_cuda_float32 PASSED [0.1118s] [ 22%] 2025-12-04T15:25:36.6827820Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_isneginf_cuda_float32 PASSED [0.1114s] [ 22%] 2025-12-04T15:25:36.6828378Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_isposinf_cuda_float32 PASSED [0.1110s] [ 22%] 2025-12-04T15:25:36.6828903Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_isreal_cuda_float32 PASSED [0.1222s] [ 22%] 2025-12-04T15:25:36.6829402Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_istft_cuda_complex64 PASSED [0.0791s] [ 22%] 2025-12-04T15:25:36.6829894Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_item_cuda_float32 XFAIL [0.0235s] [ 23%] 2025-12-04T15:25:36.6830564Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_jiterator_2inputs_2outputs_cuda_float32 XFAIL [0.2363s] [ 23%] 2025-12-04T15:25:36.6831203Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_jiterator_4inputs_with_extra_args_cuda_float32 XFAIL [0.2045s] [ 23%] 2025-12-04T15:25:36.6831750Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_jiterator_binary_cuda_float32 XFAIL [0.2156s] [ 23%] 2025-12-04T15:25:36.6832359Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_jiterator_binary_return_by_ref_cuda_float32 XFAIL [0.0175s] [ 23%] 2025-12-04T15:25:36.6832912Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_jiterator_unary_cuda_float32 XFAIL [0.2031s] [ 23%] 2025-12-04T15:25:36.6833400Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_kron_cuda_float32 PASSED [0.2052s] [ 23%] 2025-12-04T15:25:36.6833919Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_kthvalue_cuda_float32 PASSED [0.0618s] [ 23%] 2025-12-04T15:25:36.6834397Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_lcm_cuda_int64 PASSED [0.5813s] [ 23%] 2025-12-04T15:25:36.6834888Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ldexp_cuda_float32 PASSED [0.8740s] [ 23%] 2025-12-04T15:25:36.6835375Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_le_cuda_float32 XFAIL [0.4482s] [ 23%] 2025-12-04T15:25:36.6835858Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_lerp_cuda_float32 PASSED [0.3802s] [ 24%] 2025-12-04T15:25:36.6836367Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_lgamma_cuda_float32 PASSED [0.3120s] [ 24%] 2025-12-04T15:25:36.6836908Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_cholesky_cuda_float32 PASSED [0.0575s] [ 24%] 2025-12-04T15:25:36.6837466Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_cholesky_ex_cuda_float32 PASSED [0.0702s] [ 24%] 2025-12-04T15:25:36.6838003Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_cond_cuda_float32 PASSED [0.0272s] [ 24%] 2025-12-04T15:25:36.6838526Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_cross_cuda_float32 PASSED [0.0365s] [ 24%] 2025-12-04T15:25:36.6839058Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_det_cuda_float32 PASSED [0.0710s] [ 24%] 2025-12-04T15:25:36.6839600Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_diagonal_cuda_float32 PASSED [0.2711s] [ 24%] 2025-12-04T15:25:36.6840252Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_eig_cuda_float32 SKIPPED [0.0017s] (no MAGMA library detected) [ 24%] 2025-12-04T15:25:36.6840852Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_eigh_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 24%] 2025-12-04T15:25:36.6841536Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_eigvals_cuda_float32 SKIPPED [0.0016s] (no MAGMA library detected) [ 24%] 2025-12-04T15:25:36.6842232Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_eigvalsh_cuda_float32 SKIPPED [0.0014s] (no MAGMA library detected) [ 25%] 2025-12-04T15:25:36.6842869Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_householder_product_cuda_float32 PASSED [0.1136s] [ 25%] 2025-12-04T15:25:36.6843438Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_inv_cuda_float32 PASSED [0.0583s] [ 25%] 2025-12-04T15:25:36.6843970Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_inv_ex_cuda_float32 PASSED [0.0422s] [ 25%] 2025-12-04T15:25:36.6844519Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_ldl_factor_cuda_float32 PASSED [0.0278s] [ 25%] 2025-12-04T15:25:36.6845162Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_ldl_factor_ex_cuda_float32 PASSED [0.0295s] [ 25%] 2025-12-04T15:25:36.6845709Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_ldl_solve_cuda_float32 PASSED [0.2396s] [ 25%] 2025-12-04T15:25:36.6846389Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_lstsq_cuda_float32 SKIPPED [0.0017s] (no MAGMA library detected) [ 25%] 2025-12-04T15:25:36.6847124Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_lstsq_grad_oriented_cuda_float32 SKIPPED [0.0016s] (no MAGMA library detected) [ 25%] 2025-12-04T15:25:36.6847640Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_lu_cuda_float32 PASSED [0.2659s] [ 25%] 2025-12-04T15:25:36.6848206Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_lu_factor_cuda_float32 PASSED [0.2388s] [ 25%] 2025-12-04T15:25:36.6848765Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_lu_factor_ex_cuda_float32 PASSED [0.2225s] [ 25%] 2025-12-04T15:25:36.6849325Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_lu_solve_cuda_float32 PASSED [1.0419s] [ 26%] 2025-12-04T15:25:36.6849884Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_matrix_norm_cuda_float32 PASSED [0.2644s] [ 26%] 2025-12-04T15:25:36.6850452Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_matrix_power_cuda_float32 PASSED [0.1287s] [ 26%] 2025-12-04T15:25:36.6851031Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_matrix_rank_cuda_float32 PASSED [0.3507s] [ 26%] 2025-12-04T15:25:36.6851643Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_matrix_rank_hermitian_cuda_float32 PASSED [0.0387s] [ 26%] 2025-12-04T15:25:36.6852208Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_multi_dot_cuda_float32 PASSED [0.2130s] [ 26%] 2025-12-04T15:25:36.6852737Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_norm_cuda_float32 PASSED [0.3848s] [ 26%] 2025-12-04T15:25:36.6853362Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_norm_subgradients_at_zero_cuda_float32 PASSED [0.2824s] [ 26%] 2025-12-04T15:25:36.6853900Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_pinv_cuda_float32 PASSED [0.1186s] [ 26%] 2025-12-04T15:25:36.6854622Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_pinv_hermitian_cuda_float32 SKIPPED [0.0016s] (no MAGMA library detected) [ 26%] 2025-12-04T15:25:36.6855483Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_pinv_singular_cuda_float32 SKIPPED [0.0016s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 26%] 2025-12-04T15:25:36.6855996Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_qr_cuda_float32 PASSED [0.1638s] [ 27%] 2025-12-04T15:25:36.6856550Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_slogdet_cuda_float32 PASSED [0.0532s] [ 27%] 2025-12-04T15:25:36.6857143Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_solve_cuda_float32 PASSED [0.1657s] [ 27%] 2025-12-04T15:25:36.6857728Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_solve_ex_cuda_float32 PASSED [0.2074s] [ 27%] 2025-12-04T15:25:36.6858360Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_solve_triangular_cuda_float32 PASSED [0.7415s] [ 27%] 2025-12-04T15:25:36.6858880Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_svd_cuda_float32 PASSED [1.1403s] [ 27%] 2025-12-04T15:25:36.6859429Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_svdvals_cuda_float32 PASSED [0.1220s] [ 27%] 2025-12-04T15:25:36.6860038Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_tensorinv_cuda_float32 PASSED [0.0176s] [ 27%] 2025-12-04T15:25:36.6860601Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_tensorsolve_cuda_float32 PASSED [0.0477s] [ 27%] 2025-12-04T15:25:36.6861150Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_vander_cuda_float32 PASSED [0.0396s] [ 27%] 2025-12-04T15:25:36.6861681Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_vecdot_cuda_float32 PASSED [0.4329s] [ 27%] 2025-12-04T15:25:36.6862256Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linalg_vector_norm_cuda_float32 PASSED [0.5252s] [ 28%] 2025-12-04T15:25:36.6862763Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linspace_cuda_float32 PASSED [0.0361s] [ 28%] 2025-12-04T15:25:36.6863353Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_linspace_tensor_overload_cuda_float32 PASSED [0.7221s] [ 28%] 2025-12-04T15:25:36.6863860Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_log10_cuda_float32 PASSED [0.2228s] [ 28%] 2025-12-04T15:25:36.6864351Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_log1p_cuda_float32 PASSED [0.2825s] [ 28%] 2025-12-04T15:25:36.6864852Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_log2_cuda_float32 PASSED [0.2209s] [ 28%] 2025-12-04T15:25:36.6865331Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_log_cuda_float32 PASSED [0.2214s] [ 28%] 2025-12-04T15:25:36.6865990Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_log_normal_cuda_float32 SKIPPED [0.0003s] (Test expects tensor input) [ 28%] 2025-12-04T15:25:36.6866522Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_log_softmax_cuda_float32 PASSED [0.0576s] [ 28%] 2025-12-04T15:25:36.6867097Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_log_softmax_with_dtype_cuda_float32 PASSED [0.0601s] [ 28%] 2025-12-04T15:25:36.6867635Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logaddexp2_cuda_float32 PASSED [0.2209s] [ 28%] 2025-12-04T15:25:36.6868148Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logaddexp_cuda_float32 PASSED [0.5701s] [ 29%] 2025-12-04T15:25:36.6868689Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logcumsumexp_cuda_float32 PASSED [0.0308s] [ 29%] 2025-12-04T15:25:36.6869349Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logdet_cuda_float32 SKIPPED [0.0016s] (no MAGMA library detected) [ 29%] 2025-12-04T15:25:36.6869874Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logical_and_cuda_float32 PASSED [0.7846s] [ 29%] 2025-12-04T15:25:36.6870413Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logical_not_cuda_float32 PASSED [0.1238s] [ 29%] 2025-12-04T15:25:36.6870935Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logical_or_cuda_float32 PASSED [0.5734s] [ 29%] 2025-12-04T15:25:36.6871457Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logical_xor_cuda_float32 PASSED [0.5781s] [ 29%] 2025-12-04T15:25:36.6871991Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logit_cuda_float32 PASSED [0.3320s] [ 29%] 2025-12-04T15:25:36.6872500Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logspace_cuda_float32 PASSED [0.1504s] [ 29%] 2025-12-04T15:25:36.6873128Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logspace_tensor_overload_cuda_float32 PASSED [5.0321s] [ 29%] 2025-12-04T15:25:36.6873639Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_logsumexp_cuda_float32 PASSED [0.1267s] [ 29%] 2025-12-04T15:25:36.6874116Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_long_cuda_float32 XFAIL [0.0193s] [ 29%] 2025-12-04T15:25:36.6874822Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_long_functorch_no_channels_last_cuda_float32 PASSED [0.0174s] [ 30%] 2025-12-04T15:25:36.6875289Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_lt_cuda_float32 XFAIL [0.4493s] [ 30%] 2025-12-04T15:25:36.6875778Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_lu_cuda_float32 PASSED [0.4316s] [ 30%] 2025-12-04T15:25:36.6876285Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_lu_solve_cuda_float32 PASSED [0.3026s] [ 30%] 2025-12-04T15:25:36.6876794Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_lu_unpack_cuda_float32 PASSED [0.4196s] [ 30%] 2025-12-04T15:25:36.6877280Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mH_cuda_float32 PASSED [0.0310s] [ 30%] 2025-12-04T15:25:36.6877755Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mT_cuda_float32 PASSED [0.0202s] [ 30%] 2025-12-04T15:25:36.6878295Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_amax_cuda_float32 PASSED [0.3710s] [ 30%] 2025-12-04T15:25:36.6878817Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_amin_cuda_float32 PASSED [0.3741s] [ 30%] 2025-12-04T15:25:36.6879347Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_argmax_cuda_float32 PASSED [0.2386s] [ 30%] 2025-12-04T15:25:36.6879893Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_argmin_cuda_float32 PASSED [0.2327s] [ 30%] 2025-12-04T15:25:36.6880431Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_cumprod_cuda_float32 PASSED [0.0747s] [ 31%] 2025-12-04T15:25:36.6880977Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_cumsum_cuda_float32 PASSED [0.0748s] [ 31%] 2025-12-04T15:25:36.6881504Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_fill_cuda_float32 PASSED [0.1495s] [ 31%] 2025-12-04T15:25:36.6882134Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_fill_functorch_Scalar_only_cuda_float32 PASSED [0.2405s] [ 31%] 2025-12-04T15:25:36.6882706Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_log_softmax_cuda_float32 PASSED [0.0988s] [ 31%] 2025-12-04T15:25:36.6883260Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_logaddexp_cuda_float32 PASSED [0.1071s] [ 31%] 2025-12-04T15:25:36.6883823Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_logsumexp_cuda_float32 PASSED [0.4315s] [ 31%] 2025-12-04T15:25:36.6884348Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_mean_cuda_float32 PASSED [0.4278s] [ 31%] 2025-12-04T15:25:36.6884879Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_median_cuda_float32 PASSED [0.0768s] [ 31%] 2025-12-04T15:25:36.6885419Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_norm_cuda_float32 PASSED [1.8752s] [ 31%] 2025-12-04T15:25:36.6885967Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_normalize_cuda_float32 PASSED [0.1384s] [ 31%] 2025-12-04T15:25:36.6886533Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_prod_cuda_float32 PASSED [0.3881s] [ 32%] 2025-12-04T15:25:36.6887071Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_scatter_cuda_float32 PASSED [0.1049s] [ 32%] 2025-12-04T15:25:36.6887624Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_select_cuda_float32 XFAIL [0.0187s] [ 32%] 2025-12-04T15:25:36.6888174Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_softmax_cuda_float32 PASSED [0.2960s] [ 32%] 2025-12-04T15:25:36.6888714Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_softmin_cuda_float32 PASSED [0.1002s] [ 32%] 2025-12-04T15:25:36.6889303Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_std_cuda_float32 PASSED [0.4018s] [ 32%] 2025-12-04T15:25:36.6889821Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_sum_cuda_float32 PASSED [0.3683s] [ 32%] 2025-12-04T15:25:36.6890336Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_masked_var_cuda_float32 PASSED [0.3980s] [ 32%] 2025-12-04T15:25:36.6890847Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_matmul_cuda_float32 PASSED [0.1743s] [ 32%] 2025-12-04T15:25:36.6891363Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_matrix_exp_cuda_float32 PASSED [0.0418s] [ 32%] 2025-12-04T15:25:36.6891890Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_max_binary_cuda_float32 PASSED [1.0549s] [ 32%] 2025-12-04T15:25:36.6892517Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_max_pool2d_with_indices_backward_cuda_float32 PASSED [5.3303s] [ 33%] 2025-12-04T15:25:36.6893096Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_max_reduction_no_dim_cuda_float32 PASSED [0.0123s] [ 33%] 2025-12-04T15:25:36.6893675Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_max_reduction_with_dim_cuda_float32 PASSED [0.0229s] [ 33%] 2025-12-04T15:25:36.6894176Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_maximum_cuda_float32 PASSED [0.5663s] [ 33%] 2025-12-04T15:25:36.6894676Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mean_cuda_float32 PASSED [0.0714s] [ 33%] 2025-12-04T15:25:36.6895173Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_median_cuda_float32 PASSED [0.0598s] [ 33%] 2025-12-04T15:25:36.6895769Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_meshgrid_list_of_tensors_cuda_float32 PASSED [0.0114s] [ 33%] 2025-12-04T15:25:36.6896519Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_meshgrid_variadic_tensors_cuda_float32 PASSED [0.1708s] [ 33%] 2025-12-04T15:25:36.6897096Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_min_binary_cuda_float32 PASSED [1.0501s] [ 33%] 2025-12-04T15:25:36.6897687Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_min_reduction_no_dim_cuda_float32 PASSED [0.0124s] [ 33%] 2025-12-04T15:25:36.6898267Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_min_reduction_with_dim_cuda_float32 PASSED [0.0229s] [ 33%] 2025-12-04T15:25:36.6898784Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_minimum_cuda_float32 PASSED [0.5675s] [ 33%] 2025-12-04T15:25:36.6899261Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mm_cuda_float32 PASSED [0.0223s] [ 34%] 2025-12-04T15:25:36.6899754Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mode_cuda_float32 PASSED [0.0406s] [ 34%] 2025-12-04T15:25:36.6900266Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_movedim_cuda_float32 XFAIL [0.0508s] [ 34%] 2025-12-04T15:25:36.6900762Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_msort_cuda_float32 PASSED [0.2312s] [ 34%] 2025-12-04T15:25:36.6901327Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mul_cuda_float32 PASSED [1.5744s] [ 34%] 2025-12-04T15:25:36.6901897Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_multinomial_cuda_float32 XFAIL [0.0360s] [ 34%] 2025-12-04T15:25:36.6902374Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mv_cuda_float32 PASSED [0.2301s] [ 34%] 2025-12-04T15:25:36.6902971Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mvlgamma_mvlgamma_p_1_cuda_float32 PASSED [0.5028s] [ 34%] 2025-12-04T15:25:36.6903627Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mvlgamma_mvlgamma_p_3_cuda_float32 PASSED [0.5324s] [ 34%] 2025-12-04T15:25:36.6904217Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_mvlgamma_mvlgamma_p_5_cuda_float32 PASSED [0.5655s] [ 34%] 2025-12-04T15:25:36.6904737Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nan_to_num_cuda_float32 PASSED [0.2226s] [ 34%] 2025-12-04T15:25:36.6905238Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nanmean_cuda_float32 PASSED [0.1277s] [ 35%] 2025-12-04T15:25:36.6905773Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nanmedian_cuda_float32 PASSED [0.0579s] [ 35%] 2025-12-04T15:25:36.6906300Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nanquantile_cuda_float32 PASSED [0.5095s] [ 35%] 2025-12-04T15:25:36.6906807Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nansum_cuda_float32 PASSED [0.1080s] [ 35%] 2025-12-04T15:25:36.6907334Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_narrow_copy_cuda_float32 PASSED [0.0941s] [ 35%] 2025-12-04T15:25:36.6907822Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_narrow_cuda_float32 XFAIL [0.0350s] [ 35%] 2025-12-04T15:25:36.6908385Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_native_batch_norm_cuda_float32 XFAIL [0.2245s] [ 35%] 2025-12-04T15:25:36.6908970Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_native_dropout_backward_cuda_float32 PASSED [0.2466s] [ 35%] 2025-12-04T15:25:36.6909542Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_native_layer_norm_cuda_float32 PASSED [0.2452s] [ 35%] 2025-12-04T15:25:36.6910011Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ne_cuda_float32 XFAIL [0.4509s] [ 35%] 2025-12-04T15:25:36.6910491Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_neg_cuda_float32 PASSED [0.5955s] [ 35%] 2025-12-04T15:25:36.6911019Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_new_empty_cuda_float32 PASSED [0.0245s] [ 36%] 2025-12-04T15:25:36.6911628Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_new_empty_strided_cuda_float32 SKIPPED [0.0003s] (Skipped!) [ 36%] 2025-12-04T15:25:36.6912148Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_new_full_cuda_float32 PASSED [0.0299s] [ 36%] 2025-12-04T15:25:36.6912654Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_new_ones_cuda_float32 PASSED [0.2301s] [ 36%] 2025-12-04T15:25:36.6913161Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_new_zeros_cuda_float32 PASSED [0.0298s] [ 36%] 2025-12-04T15:25:36.6913689Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nextafter_cuda_float32 PASSED [0.7914s] [ 36%] 2025-12-04T15:25:36.6914327Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_adaptive_avg_pool1d_cuda_float32 PASSED [0.0308s] [ 36%] 2025-12-04T15:25:36.6914975Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_adaptive_avg_pool2d_cuda_float32 PASSED [0.0469s] [ 36%] 2025-12-04T15:25:36.6915639Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_adaptive_avg_pool3d_cuda_float32 PASSED [0.2543s] [ 36%] 2025-12-04T15:25:36.6916269Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_adaptive_max_pool1d_cuda_float32 PASSED [0.0438s] [ 36%] 2025-12-04T15:25:36.6916940Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_adaptive_max_pool2d_cuda_float32 PASSED [0.1089s] [ 36%] 2025-12-04T15:25:36.6917571Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_adaptive_max_pool3d_cuda_float32 PASSED [0.0963s] [ 37%] 2025-12-04T15:25:36.6918240Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_alpha_dropout_cuda_float32 XFAIL [0.0369s] [ 37%] 2025-12-04T15:25:36.6918829Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_avg_pool1d_cuda_float32 PASSED [0.2305s] [ 37%] 2025-12-04T15:25:36.6919417Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_avg_pool2d_cuda_float32 PASSED [0.0442s] [ 37%] 2025-12-04T15:25:36.6920012Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_avg_pool3d_cuda_float32 PASSED [0.0480s] [ 37%] 2025-12-04T15:25:36.6920605Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_batch_norm_cuda_float32 PASSED [0.4504s] [ 37%] 2025-12-04T15:25:36.6921268Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_batch_norm_without_cudnn_cuda_float32 PASSED [0.4460s] [ 37%] 2025-12-04T15:25:36.6921851Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_bilinear_cuda_float32 PASSED [0.6371s] [ 37%] 2025-12-04T15:25:36.6922497Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_binary_cross_entropy_cuda_float32 PASSED [0.0725s] [ 37%] 2025-12-04T15:25:36.6923198Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_binary_cross_entropy_with_logits_cuda_float32 PASSED [0.2029s] [ 37%] 2025-12-04T15:25:36.6923760Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_celu_cuda_float32 PASSED [0.2270s] [ 37%] 2025-12-04T15:25:36.6924392Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_channel_shuffle_cuda_float32 PASSED [0.0161s] [ 37%] 2025-12-04T15:25:36.6924962Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv1d_cuda_float32 PASSED [1.6210s] [ 38%] 2025-12-04T15:25:36.6925538Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_cuda_float32 PASSED [0.7792s] [ 38%] 2025-12-04T15:25:36.6926147Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_no_bias_cuda_float32 PASSED [0.0176s] [ 38%] 2025-12-04T15:25:36.6926847Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_stride_depthwise_with_bias_cuda_float32 PASSED [0.0244s] [ 38%] 2025-12-04T15:25:36.6927538Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_stride_groups_with_bias_cuda_float32 PASSED [0.0242s] [ 38%] 2025-12-04T15:25:36.6928175Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_stride_no_bias_cuda_float32 PASSED [0.2140s] [ 38%] 2025-12-04T15:25:36.6928859Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_stride_padding_no_bias_cuda_float32 PASSED [0.0183s] [ 38%] 2025-12-04T15:25:36.6929551Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_stride_padding_with_bias_cuda_float32 PASSED [0.0217s] [ 38%] 2025-12-04T15:25:36.6930201Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_stride_with_bias_cuda_float32 PASSED [0.0203s] [ 38%] 2025-12-04T15:25:36.6930976Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_strided_padding_dilation_no_bias_cuda_float32 PASSED [0.0258s] [ 38%] 2025-12-04T15:25:36.6931756Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_strided_padding_dilation_with_bias_cuda_float32 PASSED [0.0225s] [ 38%] 2025-12-04T15:25:36.6932389Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv2d_with_bias_cuda_float32 PASSED [0.0197s] [ 39%] 2025-12-04T15:25:36.6932957Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv3d_cuda_float32 PASSED [0.4636s] [ 39%] 2025-12-04T15:25:36.6933650Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv_transpose1d_cuda_float32 PASSED [0.2263s] [ 39%] 2025-12-04T15:25:36.6934270Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv_transpose2d_cuda_float32 PASSED [0.2694s] [ 39%] 2025-12-04T15:25:36.6934897Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_conv_transpose3d_cuda_float32 PASSED [0.2547s] [ 39%] 2025-12-04T15:25:36.6935554Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_cosine_embedding_loss_cuda_float32 PASSED [0.1242s] [ 39%] 2025-12-04T15:25:36.6936184Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_cosine_similarity_cuda_float32 PASSED [0.2825s] [ 39%] 2025-12-04T15:25:36.6936802Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_cross_entropy_cuda_float32 PASSED [0.1735s] [ 39%] 2025-12-04T15:25:36.6937451Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_ctc_loss_cuda_float32 PASSED [0.4156s] [ 39%] 2025-12-04T15:25:36.6938031Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_dropout2d_cuda_float32 XFAIL [0.0377s] [ 39%] 2025-12-04T15:25:36.6938622Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_dropout3d_cuda_float32 XFAIL [0.2374s] [ 39%] 2025-12-04T15:25:36.6939193Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_dropout_cuda_float32 XFAIL [0.2319s] [ 40%] 2025-12-04T15:25:36.6939764Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_elu_cuda_float32 PASSED [0.4269s] [ 40%] 2025-12-04T15:25:36.6940364Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_embedding_bag_cuda_float32 XFAIL [0.0734s] [ 40%] 2025-12-04T15:25:36.6940947Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_embedding_cuda_float32 XFAIL [0.0538s] [ 40%] 2025-12-04T15:25:36.6941594Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_embedding_functorch_cuda_float32 PASSED [0.2431s] [ 40%] 2025-12-04T15:25:36.6942284Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_feature_alpha_dropout_with_train_cuda_float32 XFAIL [0.0234s] [ 40%] 2025-12-04T15:25:36.6943004Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_feature_alpha_dropout_without_train_cuda_float32 PASSED [0.1428s] [ 40%] 2025-12-04T15:25:36.6943640Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_fractional_max_pool2d_cuda_float32 XFAIL [0.0096s] [ 40%] 2025-12-04T15:25:36.6944285Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_fractional_max_pool3d_cuda_float32 XFAIL [0.2062s] [ 40%] 2025-12-04T15:25:36.6944904Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_gaussian_nll_loss_cuda_float32 XFAIL [0.2153s] [ 40%] 2025-12-04T15:25:36.6945464Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_gelu_cuda_float32 PASSED [0.2318s] [ 40%] 2025-12-04T15:25:36.6946067Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_glu_cuda_float32 PASSED [0.1417s] [ 40%] 2025-12-04T15:25:36.6946692Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_grid_sample_cuda_float32 PASSED [0.2896s] [ 41%] 2025-12-04T15:25:36.6947301Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_group_norm_cuda_float32 PASSED [0.2184s] [ 41%] 2025-12-04T15:25:36.6947891Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_hardshrink_cuda_float32 PASSED [0.1414s] [ 41%] 2025-12-04T15:25:36.6948546Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_hardsigmoid_cuda_float32 PASSED [0.2244s] [ 41%] 2025-12-04T15:25:36.6949146Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_hardswish_cuda_float32 PASSED [0.0182s] [ 41%] 2025-12-04T15:25:36.6949726Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_hardtanh_cuda_float32 PASSED [0.1417s] [ 41%] 2025-12-04T15:25:36.6950376Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_hinge_embedding_loss_cuda_float32 PASSED [0.1550s] [ 41%] 2025-12-04T15:25:36.6950963Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_huber_loss_cuda_float32 PASSED [0.0725s] [ 41%] 2025-12-04T15:25:36.6951568Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_instance_norm_cuda_float32 PASSED [0.0117s] [ 41%] 2025-12-04T15:25:36.6952204Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_interpolate_area_cuda_float32 PASSED [0.0813s] [ 41%] 2025-12-04T15:25:36.6952841Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_interpolate_bicubic_cuda_float32 PASSED [0.1059s] [ 41%] 2025-12-04T15:25:36.6953497Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_interpolate_bilinear_cuda_float32 PASSED [0.1025s] [ 42%] 2025-12-04T15:25:36.6954133Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_interpolate_linear_cuda_float32 PASSED [0.0647s] [ 42%] 2025-12-04T15:25:36.6954801Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_interpolate_nearest-exact_cuda_float32 PASSED [0.0853s] [ 42%] 2025-12-04T15:25:36.6955453Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_interpolate_nearest_cuda_float32 PASSED [0.0791s] [ 42%] 2025-12-04T15:25:36.6956102Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_interpolate_trilinear_cuda_float32 PASSED [0.0875s] [ 42%] 2025-12-04T15:25:36.6956685Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_kl_div_cuda_float32 PASSED [0.1185s] [ 42%] 2025-12-04T15:25:36.6957262Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_l1_loss_cuda_float32 PASSED [0.2766s] [ 42%] 2025-12-04T15:25:36.6957864Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_layer_norm_cuda_float32 PASSED [0.1312s] [ 42%] 2025-12-04T15:25:36.6958457Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_leaky_relu_cuda_float32 PASSED [0.0563s] [ 42%] 2025-12-04T15:25:36.6959021Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_linear_cuda_float32 PASSED [0.1714s] [ 42%] 2025-12-04T15:25:36.6959670Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_local_response_norm_cuda_float32 PASSED [0.0520s] [ 42%] 2025-12-04T15:25:36.6960258Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_logsigmoid_cuda_float32 PASSED [0.1260s] [ 43%] 2025-12-04T15:25:36.6960941Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_margin_ranking_loss_cuda_float32 PASSED [0.4826s] [ 43%] 2025-12-04T15:25:36.6961528Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_max_pool1d_cuda_float32 PASSED [3.3933s] [ 43%] 2025-12-04T15:25:36.6962149Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_max_pool2d_cuda_float32 PASSED [5.4679s] [ 43%] 2025-12-04T15:25:36.6962750Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_max_pool3d_cuda_float32 PASSED [2.2572s] [ 43%] 2025-12-04T15:25:36.6963350Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_max_unpool1d_cuda_float32 PASSED [1.0783s] [ 43%] 2025-12-04T15:25:36.6964042Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_max_unpool1d_grad_cuda_float32 PASSED [0.1330s] [ 43%] 2025-12-04T15:25:36.6964641Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_max_unpool2d_cuda_float32 PASSED [2.0944s] [ 43%] 2025-12-04T15:25:36.6965265Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_max_unpool2d_grad_cuda_float32 PASSED [0.2446s] [ 43%] 2025-12-04T15:25:36.6965881Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_max_unpool3d_cuda_float32 PASSED [0.7350s] [ 43%] 2025-12-04T15:25:36.6966502Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_max_unpool3d_grad_cuda_float32 PASSED [0.1422s] [ 43%] 2025-12-04T15:25:36.6967075Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_mish_cuda_float32 PASSED [0.2266s] [ 44%] 2025-12-04T15:25:36.6967655Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_mse_loss_cuda_float32 PASSED [0.0428s] [ 44%] 2025-12-04T15:25:36.6968297Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_mse_loss_functorch_cuda_float32 PASSED [0.0614s] [ 44%] 2025-12-04T15:25:36.6968964Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_multi_head_attention_forward_cuda_float32 XFAIL [3.5022s] [ 44%] 2025-12-04T15:25:36.6969598Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_multi_margin_loss_cuda_float32 PASSED [1.5291s] [ 44%] 2025-12-04T15:25:36.6970265Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_multilabel_margin_loss_cuda_float32 PASSED [0.2152s] [ 44%] 2025-12-04T15:25:36.6970937Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_multilabel_soft_margin_loss_cuda_float32 PASSED [0.0414s] [ 44%] 2025-12-04T15:25:36.6971532Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_nll_loss_cuda_float32 PASSED [0.5134s] [ 44%] 2025-12-04T15:25:36.6972122Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_normalize_cuda_float32 PASSED [0.0350s] [ 44%] 2025-12-04T15:25:36.6972675Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_one_hot_cuda_int64 XFAIL [0.0078s] [ 44%] 2025-12-04T15:25:36.6973286Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_pad_circular_cuda_float32 PASSED [0.0342s] [ 44%] 2025-12-04T15:25:36.6973881Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_pad_constant_cuda_float32 PASSED [0.1744s] [ 44%] 2025-12-04T15:25:36.6974489Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_pad_reflect_cuda_float32 PASSED [0.0458s] [ 45%] 2025-12-04T15:25:36.6975101Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_pad_replicate_cuda_float32 PASSED [0.0447s] [ 45%] 2025-12-04T15:25:36.6975746Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_pad_replicate_negative_cuda_float32 PASSED [0.0267s] [ 45%] 2025-12-04T15:25:36.6976422Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_pairwise_distance_cuda_float32 PASSED [0.0488s] [ 45%] 2025-12-04T15:25:36.6977079Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_pdist_cuda_float32 PASSED [0.0348s] [ 45%] 2025-12-04T15:25:36.6977698Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_pixel_shuffle_cuda_float32 PASSED [0.0198s] [ 45%] 2025-12-04T15:25:36.6978308Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_pixel_unshuffle_cuda_float32 PASSED [0.0196s] [ 45%] 2025-12-04T15:25:36.6979054Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_poisson_nll_loss_cuda_float32 PASSED [0.5439s] [ 45%] 2025-12-04T15:25:36.6979632Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_prelu_cuda_float32 PASSED [0.5115s] [ 45%] 2025-12-04T15:25:36.6980198Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_relu6_cuda_float32 PASSED [0.1273s] [ 45%] 2025-12-04T15:25:36.6980769Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_relu_cuda_float32 PASSED [0.1328s] [ 45%] 2025-12-04T15:25:36.6981342Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_rms_norm_cuda_float32 PASSED [0.0872s] [ 46%] 2025-12-04T15:25:36.6981913Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_rrelu_cuda_float32 XFAIL [0.0151s] [ 46%] 2025-12-04T15:25:36.6982582Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_scaled_dot_product_attention_cuda_float32 XFAIL [0.3090s] [ 46%] 2025-12-04T15:25:36.6983137Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_selu_cuda_float32 PASSED [0.4268s] [ 46%] 2025-12-04T15:25:36.6983760Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_silu_complex_cuda_complex64 PASSED [0.2992s] [ 46%] 2025-12-04T15:25:36.6984317Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_silu_cuda_float32 PASSED [0.2252s] [ 46%] 2025-12-04T15:25:36.6984938Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_smooth_l1_loss_cuda_float32 PASSED [0.0550s] [ 46%] 2025-12-04T15:25:36.6985558Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_soft_margin_loss_cuda_float32 PASSED [0.2652s] [ 46%] 2025-12-04T15:25:36.6986134Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_softmin_cuda_float32 PASSED [0.0278s] [ 46%] 2025-12-04T15:25:36.6986771Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_softmin_with_dtype_cuda_float32 PASSED [0.0281s] [ 46%] 2025-12-04T15:25:36.6987354Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_softplus_cuda_float32 PASSED [0.1249s] [ 46%] 2025-12-04T15:25:36.6987964Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_softshrink_cuda_float32 PASSED [0.1407s] [ 47%] 2025-12-04T15:25:36.6988539Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_softsign_cuda_float32 PASSED [0.1367s] [ 47%] 2025-12-04T15:25:36.6989131Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_tanhshrink_cuda_float32 PASSED [0.1297s] [ 47%] 2025-12-04T15:25:36.6989730Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_threshold_cuda_float32 PASSED [0.2483s] [ 47%] 2025-12-04T15:25:36.6990364Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_triplet_margin_loss_cuda_float32 PASSED [0.1552s] [ 47%] 2025-12-04T15:25:36.6991123Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_triplet_margin_with_distance_loss_cuda_float32 PASSED [0.1505s] [ 47%] 2025-12-04T15:25:36.6991722Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_unfold_cuda_float32 PASSED [0.5598s] [ 47%] 2025-12-04T15:25:36.6992349Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_upsample_bilinear_cuda_float32 PASSED [0.0290s] [ 47%] 2025-12-04T15:25:36.6992982Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nn_functional_upsample_nearest_cuda_float32 PASSED [0.0447s] [ 47%] 2025-12-04T15:25:36.6993541Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nonzero_cuda_float32 XFAIL [0.0088s] [ 47%] 2025-12-04T15:25:36.6994188Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_nonzero_static_cuda_float32 SKIPPED [0.2220s] (Only runs on cpu) [ 47%] 2025-12-04T15:25:36.6994678Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_norm_cuda_float32 PASSED [0.1375s] [ 48%] 2025-12-04T15:25:36.6995201Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_norm_fro_cuda_float32 PASSED [0.0161s] [ 48%] 2025-12-04T15:25:36.6995710Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_norm_inf_cuda_float32 PASSED [0.2230s] [ 48%] 2025-12-04T15:25:36.6996367Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_norm_nuc_cuda_float32 PASSED [0.0216s] [ 48%] 2025-12-04T15:25:36.6996872Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_normal_cuda_float32 XFAIL [0.0100s] [ 48%] 2025-12-04T15:25:36.6997558Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_normal_in_place_cuda_float32 SKIPPED [0.0003s] (Test expects tensor input) [ 48%] 2025-12-04T15:25:36.6998125Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_normal_number_mean_cuda_float32 XFAIL [0.2023s] [ 48%] 2025-12-04T15:25:36.6998616Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ones_cuda_float32 PASSED [0.1927s] [ 48%] 2025-12-04T15:25:36.6999132Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ones_like_cuda_float32 PASSED [0.0282s] [ 48%] 2025-12-04T15:25:36.6999882Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ops_aten__new_zeros_with_same_feature_meta_functorchonly_cuda_float32 PASSED [0.0531s] [ 48%] 2025-12-04T15:25:36.7000482Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ops_aten_index_put_functorch_cuda_float32 PASSED [0.5955s] [ 48%] 2025-12-04T15:25:36.7000991Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ormqr_cuda_float32 PASSED [1.7833s] [ 48%] 2025-12-04T15:25:36.7001480Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_outer_cuda_float32 PASSED [0.0172s] [ 49%] 2025-12-04T15:25:36.7001997Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_pca_lowrank_cuda_float32 XFAIL [0.0128s] [ 49%] 2025-12-04T15:25:36.7002538Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_permute_copy_cuda_float32 PASSED [0.4229s] [ 49%] 2025-12-04T15:25:36.7003044Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_permute_cuda_float32 PASSED [0.1982s] [ 49%] 2025-12-04T15:25:36.7003568Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_pinverse_cuda_float32 PASSED [0.0463s] [ 49%] 2025-12-04T15:25:36.7004056Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_polar_cuda_float32 PASSED [0.6319s] [ 49%] 2025-12-04T15:25:36.7004647Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_polygamma_polygamma_n_0_cuda_float32 PASSED [0.4058s] [ 49%] 2025-12-04T15:25:36.7005247Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_polygamma_polygamma_n_1_cuda_float32 PASSED [0.6189s] [ 49%] 2025-12-04T15:25:36.7005902Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_polygamma_polygamma_n_2_cuda_float32 PASSED [0.5062s] [ 49%] 2025-12-04T15:25:36.7006551Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_polygamma_polygamma_n_3_cuda_float32 PASSED [0.5093s] [ 49%] 2025-12-04T15:25:36.7007143Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_polygamma_polygamma_n_4_cuda_float32 PASSED [0.4979s] [ 49%] 2025-12-04T15:25:36.7007654Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_positive_cuda_float32 PASSED [0.1084s] [ 50%] 2025-12-04T15:25:36.7008229Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_pow_cuda_float32 PASSED [0.8706s] [ 50%] 2025-12-04T15:25:36.7008718Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_prod_cuda_float32 PASSED [0.1232s] [ 50%] 2025-12-04T15:25:36.7009211Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_put_cuda_float32 PASSED [0.5878s] [ 50%] 2025-12-04T15:25:36.7009692Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_qr_cuda_float32 PASSED [0.1670s] [ 50%] 2025-12-04T15:25:36.7010200Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_quantile_cuda_float32 PASSED [0.5049s] [ 50%] 2025-12-04T15:25:36.7010714Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_rad2deg_cuda_float32 PASSED [0.2115s] [ 50%] 2025-12-04T15:25:36.7011223Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_rand_like_cuda_float32 XFAIL [0.0095s] [ 50%] 2025-12-04T15:25:36.7011885Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_randint_cuda_float32 SKIPPED [0.0002s] (Test expects tensor input) [ 50%] 2025-12-04T15:25:36.7012413Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_randint_like_cuda_float32 XFAIL [0.2493s] [ 50%] 2025-12-04T15:25:36.7013047Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_randn_cuda_float32 SKIPPED [0.0003s] (Test expects tensor input) [ 50%] 2025-12-04T15:25:36.7013574Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_randn_like_cuda_float32 XFAIL [0.1997s] [ 51%] 2025-12-04T15:25:36.7014068Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_ravel_cuda_float32 PASSED [0.2144s] [ 51%] 2025-12-04T15:25:36.7014569Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_real_cuda_float32 PASSED [0.1199s] [ 51%] 2025-12-04T15:25:36.7015092Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_reciprocal_cuda_float32 PASSED [0.2298s] [ 51%] 2025-12-04T15:25:36.7015611Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_remainder_cuda_float32 PASSED [0.8943s] [ 51%] 2025-12-04T15:25:36.7016117Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_renorm_cuda_float32 PASSED [0.0394s] [ 51%] 2025-12-04T15:25:36.7016616Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_repeat_cuda_float32 PASSED [0.1278s] [ 51%] 2025-12-04T15:25:36.7017240Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_repeat_interleave_cuda_float32 PASSED [0.0233s] [ 51%] 2025-12-04T15:25:36.7017766Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_reshape_as_cuda_float32 PASSED [0.1579s] [ 51%] 2025-12-04T15:25:36.7018269Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_reshape_cuda_float32 PASSED [0.1457s] [ 51%] 2025-12-04T15:25:36.7018772Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_resize__cuda_float32 XFAIL [0.0109s] [ 51%] 2025-12-04T15:25:36.7019283Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_resize_as__cuda_float32 XFAIL [0.2242s] [ 51%] 2025-12-04T15:25:36.7019825Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_resolve_conj_cuda_float32 PASSED [0.2003s] [ 52%] 2025-12-04T15:25:36.7020386Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_resolve_neg_cuda_float32 PASSED [0.0117s] [ 52%] 2025-12-04T15:25:36.7020877Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_roll_cuda_float32 PASSED [0.0656s] [ 52%] 2025-12-04T15:25:36.7021407Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_rot90_cuda_float32 PASSED [0.1272s] [ 52%] 2025-12-04T15:25:36.7021899Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_round_cuda_float32 PASSED [0.2880s] [ 52%] 2025-12-04T15:25:36.7022460Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_round_decimals_0_cuda_float32 PASSED [0.3202s] [ 52%] 2025-12-04T15:25:36.7023069Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_round_decimals_3_cuda_float32 PASSED [0.3197s] [ 52%] 2025-12-04T15:25:36.7023639Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_round_decimals_neg_3_cuda_float32 PASSED [0.3193s] [ 52%] 2025-12-04T15:25:36.7024143Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_rsqrt_cuda_float32 PASSED [0.2243s] [ 52%] 2025-12-04T15:25:36.7024630Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_rsub_cuda_float32 PASSED [0.6320s] [ 52%] 2025-12-04T15:25:36.7025175Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_scalar_tensor_cuda_float32 PASSED [0.0071s] [ 52%] 2025-12-04T15:25:36.7025699Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_scatter_add_cuda_float32 PASSED [0.1684s] [ 53%] 2025-12-04T15:25:36.7026201Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_scatter_cuda_float32 PASSED [0.5656s] [ 53%] 2025-12-04T15:25:36.7026776Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_scatter_reduce_amax_cuda_float32 PASSED [0.4627s] [ 53%] 2025-12-04T15:25:36.7027339Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_scatter_reduce_amin_cuda_float32 PASSED [0.4587s] [ 53%] 2025-12-04T15:25:36.7027916Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_scatter_reduce_mean_cuda_float32 PASSED [0.5398s] [ 53%] 2025-12-04T15:25:36.7028481Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_scatter_reduce_prod_cuda_float32 PASSED [0.4602s] [ 53%] 2025-12-04T15:25:36.7029038Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_scatter_reduce_sum_cuda_float32 PASSED [0.4621s] [ 53%] 2025-12-04T15:25:36.7029580Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_searchsorted_cuda_float32 PASSED [1.5371s] [ 53%] 2025-12-04T15:25:36.7030077Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_select_cuda_float32 PASSED [0.0242s] [ 53%] 2025-12-04T15:25:36.7030624Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_select_scatter_cuda_float32 PASSED [0.0382s] [ 53%] 2025-12-04T15:25:36.7031105Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sgn_cuda_float32 PASSED [0.2055s] [ 53%] 2025-12-04T15:25:36.7031587Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_short_cuda_float32 XFAIL [0.0197s] [ 54%] 2025-12-04T15:25:36.7032227Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_short_functorch_no_channels_last_cuda_float32 PASSED [0.0176s] [ 54%] 2025-12-04T15:25:36.7032728Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sigmoid_cuda_float32 PASSED [0.4052s] [ 54%] 2025-12-04T15:25:36.7033227Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sign_cuda_float32 PASSED [0.2053s] [ 54%] 2025-12-04T15:25:36.7033878Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_bartlett_cuda_float32 SKIPPED [0.0003s] (Skipped!) [ 54%] 2025-12-04T15:25:36.7034525Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_blackman_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 54%] 2025-12-04T15:25:36.7035195Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_cosine_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 54%] 2025-12-04T15:25:36.7035887Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_exponential_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 54%] 2025-12-04T15:25:36.7036545Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_gaussian_cuda_float32 SKIPPED [0.0006s] (Skipped!) [ 54%] 2025-12-04T15:25:36.7037210Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_general_cosine_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 54%] 2025-12-04T15:25:36.7037973Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_general_hamming_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 54%] 2025-12-04T15:25:36.7038612Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_hamming_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 55%] 2025-12-04T15:25:36.7039232Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_hann_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 55%] 2025-12-04T15:25:36.7039875Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_kaiser_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 55%] 2025-12-04T15:25:36.7040509Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signal_windows_nuttall_cuda_float32 SKIPPED [0.0002s] (Skipped!) [ 55%] 2025-12-04T15:25:36.7041027Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_signbit_cuda_float32 PASSED [0.1137s] [ 55%] 2025-12-04T15:25:36.7041508Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sin_cuda_float32 PASSED [0.1946s] [ 55%] 2025-12-04T15:25:36.7041999Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sinc_cuda_float32 PASSED [0.3034s] [ 55%] 2025-12-04T15:25:36.7042498Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sinh_cuda_float32 PASSED [0.2054s] [ 55%] 2025-12-04T15:25:36.7042991Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_slice_cuda_float32 PASSED [0.0185s] [ 55%] 2025-12-04T15:25:36.7043533Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_slice_scatter_cuda_float32 PASSED [0.0660s] [ 55%] 2025-12-04T15:25:36.7044036Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_softmax_cuda_float32 PASSED [0.0584s] [ 55%] 2025-12-04T15:25:36.7044601Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_softmax_with_dtype_cuda_float32 PASSED [0.0617s] [ 55%] 2025-12-04T15:25:36.7045097Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sort_cuda_float32 PASSED [0.1648s] [ 56%] 2025-12-04T15:25:36.7045739Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sparse_mm_reduce_cuda_float32 SKIPPED [0.0019s] (Only runs on cpu) [ 56%] 2025-12-04T15:25:36.7046319Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sparse_sampled_addmm_cuda_float32 XFAIL [0.0078s] [ 56%] 2025-12-04T15:25:36.7046866Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_airy_ai_cuda_float32 PASSED [2.3077s] [ 56%] 2025-12-04T15:25:36.7047417Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_bessel_j0_cuda_float32 PASSED [0.1310s] [ 56%] 2025-12-04T15:25:36.7047981Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_bessel_j1_cuda_float32 PASSED [0.1503s] [ 56%] 2025-12-04T15:25:36.7048536Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_bessel_y0_cuda_float32 PASSED [0.1374s] [ 56%] 2025-12-04T15:25:36.7049100Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_bessel_y1_cuda_float32 PASSED [0.1350s] [ 56%] 2025-12-04T15:25:36.7049756Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_chebyshev_polynomial_t_cuda_float32 PASSED [0.4311s] [ 56%] 2025-12-04T15:25:36.7050410Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_chebyshev_polynomial_u_cuda_float32 PASSED [1.2579s] [ 56%] 2025-12-04T15:25:36.7051051Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_chebyshev_polynomial_v_cuda_float32 PASSED [0.9985s] [ 56%] 2025-12-04T15:25:36.7051671Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_chebyshev_polynomial_w_cuda_float32 PASSED [0.9891s] [ 57%] 2025-12-04T15:25:36.7052275Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_entr_cuda_float32 PASSED [0.1231s] [ 57%] 2025-12-04T15:25:36.7052805Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_erfcx_cuda_float32 PASSED [0.3776s] [ 57%] 2025-12-04T15:25:36.7053434Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_hermite_polynomial_h_cuda_float32 PASSED [2.4743s] [ 57%] 2025-12-04T15:25:36.7054050Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_hermite_polynomial_he_cuda_float32 PASSED [0.2287s] [ 57%] 2025-12-04T15:25:36.7054575Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_i0e_cuda_float32 PASSED [0.1254s] [ 57%] 2025-12-04T15:25:36.7055110Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_i1_cuda_float32 PASSED [0.1251s] [ 57%] 2025-12-04T15:25:36.7055641Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_i1e_cuda_float32 PASSED [0.1235s] [ 57%] 2025-12-04T15:25:36.7056273Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_laguerre_polynomial_l_cuda_float32 PASSED [0.2017s] [ 57%] 2025-12-04T15:25:36.7056957Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_legendre_polynomial_p_cuda_float32 PASSED [0.2162s] [ 57%] 2025-12-04T15:25:36.7057512Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_log_ndtr_cuda_float32 PASSED [0.3895s] [ 57%] 2025-12-04T15:25:36.7058139Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_modified_bessel_i0_cuda_float32 PASSED [0.1292s] [ 58%] 2025-12-04T15:25:36.7058738Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_modified_bessel_i1_cuda_float32 PASSED [0.1304s] [ 58%] 2025-12-04T15:25:36.7059351Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_modified_bessel_k0_cuda_float32 PASSED [0.1592s] [ 58%] 2025-12-04T15:25:36.7059955Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_modified_bessel_k1_cuda_float32 PASSED [0.1335s] [ 58%] 2025-12-04T15:25:36.7060490Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_ndtr_cuda_float32 PASSED [0.1369s] [ 58%] 2025-12-04T15:25:36.7061043Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_ndtri_cuda_float32 PASSED [0.1301s] [ 58%] 2025-12-04T15:25:36.7061715Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_polygamma_special_polygamma_n_0_cuda_float32 PASSED [0.2102s] [ 58%] 2025-12-04T15:25:36.7062363Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_scaled_modified_bessel_k0_cuda_float32 PASSED [0.1319s] [ 58%] 2025-12-04T15:25:36.7063000Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_scaled_modified_bessel_k1_cuda_float32 PASSED [0.1326s] [ 58%] 2025-12-04T15:25:36.7063668Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_shifted_chebyshev_polynomial_t_cuda_float32 PASSED [0.4460s] [ 58%] 2025-12-04T15:25:36.7064347Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_shifted_chebyshev_polynomial_u_cuda_float32 PASSED [1.2911s] [ 58%] 2025-12-04T15:25:36.7065269Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_shifted_chebyshev_polynomial_v_cuda_float32 PASSED [1.0183s] [ 59%] 2025-12-04T15:25:36.7065986Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_shifted_chebyshev_polynomial_w_cuda_float32 PASSED [0.9872s] [ 59%] 2025-12-04T15:25:36.7066594Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_spherical_bessel_j0_cuda_float32 PASSED [0.1271s] [ 59%] 2025-12-04T15:25:36.7067155Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_xlog1py_cuda_float32 PASSED [0.6242s] [ 59%] 2025-12-04T15:25:36.7067748Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_special_zeta_cuda_float32 PASSED [0.9359s] [ 59%] 2025-12-04T15:25:36.7068239Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_split_cuda_float32 PASSED [0.0149s] [ 59%] 2025-12-04T15:25:36.7068797Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_split_list_args_cuda_float32 PASSED [0.0227s] [ 59%] 2025-12-04T15:25:36.7069371Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_split_with_sizes_copy_cuda_float32 PASSED [0.0288s] [ 59%] 2025-12-04T15:25:36.7069932Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_split_with_sizes_cuda_float32 PASSED [0.0276s] [ 59%] 2025-12-04T15:25:36.7070418Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sqrt_cuda_float32 PASSED [0.2067s] [ 59%] 2025-12-04T15:25:36.7070916Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_square_cuda_float32 PASSED [0.2345s] [ 59%] 2025-12-04T15:25:36.7071461Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_squeeze_copy_cuda_float32 PASSED [0.0333s] [ 59%] 2025-12-04T15:25:36.7071964Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_squeeze_cuda_float32 PASSED [0.0495s] [ 60%] 2025-12-04T15:25:36.7072533Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_squeeze_multiple_cuda_float32 PASSED [0.0410s] [ 60%] 2025-12-04T15:25:36.7073025Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_stack_cuda_float32 PASSED [0.0094s] [ 60%] 2025-12-04T15:25:36.7073506Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_std_cuda_float32 PASSED [0.0494s] [ 60%] 2025-12-04T15:25:36.7074028Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_std_mean_cuda_float32 PASSED [0.0672s] [ 60%] 2025-12-04T15:25:36.7074579Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_std_mean_unbiased_cuda_float32 PASSED [0.0150s] [ 60%] 2025-12-04T15:25:36.7075117Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_std_unbiased_cuda_float32 PASSED [0.0124s] [ 60%] 2025-12-04T15:25:36.7075609Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_stft_cuda_float32 PASSED [0.0496s] [ 60%] 2025-12-04T15:25:36.7076091Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sub_cuda_float32 PASSED [2.0503s] [ 60%] 2025-12-04T15:25:36.7076583Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sum_cuda_float32 PASSED [0.0668s] [ 60%] 2025-12-04T15:25:36.7077103Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_sum_to_size_cuda_float32 PASSED [0.0641s] [ 60%] 2025-12-04T15:25:36.7077597Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_svd_cuda_float32 PASSED [1.1510s] [ 61%] 2025-12-04T15:25:36.7078117Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_svd_lowrank_cuda_float32 XFAIL [0.0121s] [ 61%] 2025-12-04T15:25:36.7078610Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_t_copy_cuda_float32 PASSED [0.2282s] [ 61%] 2025-12-04T15:25:36.7079123Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_t_cuda_float32 PASSED [0.0261s] [ 61%] 2025-12-04T15:25:36.7079658Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_take_along_dim_cuda_float32 PASSED [0.2424s] [ 61%] 2025-12-04T15:25:36.7080183Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_take_cuda_float32 PASSED [0.0732s] [ 61%] 2025-12-04T15:25:36.7080663Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_tan_cuda_float32 PASSED [0.2115s] [ 61%] 2025-12-04T15:25:36.7081146Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_tanh_cuda_float32 PASSED [0.2961s] [ 61%] 2025-12-04T15:25:36.7081743Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_tensor_split_cuda_float32 XFAIL [0.0084s] [ 61%] 2025-12-04T15:25:36.7082256Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_tensordot_cuda_float32 PASSED [0.2403s] [ 61%] 2025-12-04T15:25:36.7082751Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_tile_cuda_float32 PASSED [0.1692s] [ 61%] 2025-12-04T15:25:36.7083225Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_to_cuda_float32 PASSED [0.0701s] [ 62%] 2025-12-04T15:25:36.7083732Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_to_sparse_cuda_float32 XFAIL [0.0095s] [ 62%] 2025-12-04T15:25:36.7084224Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_topk_cuda_float32 PASSED [0.0700s] [ 62%] 2025-12-04T15:25:36.7084851Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_torch__scaled_mm_cuda_float8_e4m3fn SKIPPED [0.0003s] (Skipped!) [ 62%] 2025-12-04T15:25:36.7085501Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_torch__scaled_mm_v2_cuda_float8_e4m3fn SKIPPED [0.0002s] (Skipped!) [ 62%] 2025-12-04T15:25:36.7086169Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_torch_ops_aten__efficient_attention_forward_cuda_float32 XFAIL [0.0106s] [ 62%] 2025-12-04T15:25:36.7087056Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_torch_ops_aten__flash_attention_forward_cuda_float16 SKIPPED [0.2088s] (This platform doesn't support Flash Attention) [ 62%] 2025-12-04T15:25:36.7087705Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_torch_ops_aten__safe_softmax_default_cuda_float32 PASSED [0.2427s] [ 62%] 2025-12-04T15:25:36.7088196Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_trace_cuda_float32 PASSED [0.2183s] [ 62%] 2025-12-04T15:25:36.7088750Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_transpose_copy_cuda_float32 PASSED [0.0347s] [ 62%] 2025-12-04T15:25:36.7089265Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_transpose_cuda_float32 PASSED [0.1291s] [ 62%] 2025-12-04T15:25:36.7089777Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_trapezoid_cuda_float32 PASSED [0.0593s] [ 62%] 2025-12-04T15:25:36.7090281Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_trapz_cuda_float32 PASSED [0.0582s] [ 63%] 2025-12-04T15:25:36.7090979Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_triangular_solve_cuda_float32 SKIPPED [0.0016s] (no MAGMA library detected) [ 63%] 2025-12-04T15:25:36.7091476Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_tril_cuda_float32 XFAIL [0.0080s] [ 63%] 2025-12-04T15:25:36.7091993Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_tril_indices_cuda_int64 PASSED [0.2146s] [ 63%] 2025-12-04T15:25:36.7092475Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_triu_cuda_float32 XFAIL [0.0079s] [ 63%] 2025-12-04T15:25:36.7093002Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_triu_indices_cuda_int64 PASSED [0.2139s] [ 63%] 2025-12-04T15:25:36.7093553Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_true_divide_cuda_float32 PASSED [0.8793s] [ 63%] 2025-12-04T15:25:36.7094056Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_trunc_cuda_float32 PASSED [0.3909s] [ 63%] 2025-12-04T15:25:36.7094602Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unbind_copy_cuda_float32 XFAIL [0.0149s] [ 63%] 2025-12-04T15:25:36.7095096Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unbind_cuda_float32 PASSED [0.2720s] [ 63%] 2025-12-04T15:25:36.7095619Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unflatten_cuda_float32 PASSED [0.0377s] [ 63%] 2025-12-04T15:25:36.7096398Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unfold_copy_cuda_float32 PASSED [0.0778s] [ 64%] 2025-12-04T15:25:36.7096969Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unfold_cuda_float32 PASSED [0.0744s] [ 64%] 2025-12-04T15:25:36.7097470Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_uniform_cuda_float32 XFAIL [0.0146s] [ 64%] 2025-12-04T15:25:36.7098022Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unique_consecutive_cuda_float32 XFAIL [0.2157s] [ 64%] 2025-12-04T15:25:36.7098531Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unique_cuda_float32 XFAIL [0.2142s] [ 64%] 2025-12-04T15:25:36.7099055Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unravel_index_cuda_int64 PASSED [0.2779s] [ 64%] 2025-12-04T15:25:36.7099594Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unsafe_chunk_cuda_float32 PASSED [0.1239s] [ 64%] 2025-12-04T15:25:36.7100125Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unsafe_split_cuda_float32 PASSED [0.0150s] [ 64%] 2025-12-04T15:25:36.7100664Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unsqueeze_copy_cuda_float32 PASSED [0.0370s] [ 64%] 2025-12-04T15:25:36.7101192Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_unsqueeze_cuda_float32 PASSED [0.0556s] [ 64%] 2025-12-04T15:25:36.7101670Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_var_cuda_float32 PASSED [0.2565s] [ 64%] 2025-12-04T15:25:36.7102192Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_var_mean_cuda_float32 PASSED [0.0685s] [ 65%] 2025-12-04T15:25:36.7102741Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_var_mean_unbiased_cuda_float32 PASSED [0.0151s] [ 65%] 2025-12-04T15:25:36.7103265Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_var_unbiased_cuda_float32 PASSED [0.0124s] [ 65%] 2025-12-04T15:25:36.7103773Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_vdot_cuda_float32 PASSED [0.2327s] [ 65%] 2025-12-04T15:25:36.7104302Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_view_as_complex_cuda_float32 XFAIL [0.0098s] [ 65%] 2025-12-04T15:25:36.7104814Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_view_as_cuda_float32 PASSED [0.3666s] [ 65%] 2025-12-04T15:25:36.7105348Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_view_as_real_cuda_complex64 PASSED [0.0123s] [ 65%] 2025-12-04T15:25:36.7105857Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_view_copy_cuda_float32 PASSED [0.0508s] [ 65%] 2025-12-04T15:25:36.7106357Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_view_cuda_float32 PASSED [0.1475s] [ 65%] 2025-12-04T15:25:36.7106854Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_vsplit_cuda_float32 PASSED [0.0264s] [ 65%] 2025-12-04T15:25:36.7107356Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_vstack_cuda_float32 PASSED [0.2220s] [ 65%] 2025-12-04T15:25:36.7107847Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_where_cuda_float32 PASSED [0.1880s] [ 66%] 2025-12-04T15:25:36.7108408Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_xlogy_cuda_float32 PASSED [1.6232s] [ 66%] 2025-12-04T15:25:36.7108957Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_zero__cuda_float32 PASSED [0.0245s] [ 66%] 2025-12-04T15:25:36.7109447Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_zeros_cuda_float32 PASSED [0.0066s] [ 66%] 2025-12-04T15:25:36.7109978Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_exhaustive_zeros_like_cuda_float32 PASSED [0.0282s] [ 66%] 2025-12-04T15:25:36.7110657Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_cholesky_cuda_float32 PASSED [0.2191s] [ 66%] 2025-12-04T15:25:36.7111270Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_cholesky_ex_cuda_float32 PASSED [0.0145s] [ 66%] 2025-12-04T15:25:36.7111866Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_cond_cuda_float32 PASSED [0.2140s] [ 66%] 2025-12-04T15:25:36.7112451Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_cross_cuda_float32 PASSED [0.0084s] [ 66%] 2025-12-04T15:25:36.7113047Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_det_cuda_float32 PASSED [0.2165s] [ 66%] 2025-12-04T15:25:36.7113641Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_diagonal_cuda_float32 XFAIL [0.0076s] [ 66%] 2025-12-04T15:25:36.7114373Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_eig_cuda_float32 SKIPPED [0.2034s] (no MAGMA library detected) [ 66%] 2025-12-04T15:25:36.7115090Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_eigh_cuda_float32 SKIPPED [0.0018s] (no MAGMA library detected) [ 67%] 2025-12-04T15:25:36.7115820Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_eigvals_cuda_float32 SKIPPED [0.0016s] (no MAGMA library detected) [ 67%] 2025-12-04T15:25:36.7116567Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_eigvalsh_cuda_float32 SKIPPED [0.0014s] (no MAGMA library detected) [ 67%] 2025-12-04T15:25:36.7117224Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_householder_product_cuda_float32 PASSED [0.2151s] [ 67%] 2025-12-04T15:25:36.7117817Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_inv_cuda_float32 PASSED [0.0129s] [ 67%] 2025-12-04T15:25:36.7118406Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_inv_ex_cuda_float32 PASSED [0.2159s] [ 67%] 2025-12-04T15:25:36.7119012Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_ldl_factor_cuda_float32 PASSED [0.0088s] [ 67%] 2025-12-04T15:25:36.7119644Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_ldl_factor_ex_cuda_float32 PASSED [0.2128s] [ 67%] 2025-12-04T15:25:36.7120299Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_ldl_solve_cuda_float32 SKIPPED [0.0003s] (Skipped!) [ 67%] 2025-12-04T15:25:36.7121033Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_lstsq_cuda_float32 SKIPPED [0.0016s] (no MAGMA library detected) [ 67%] 2025-12-04T15:25:36.7121817Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_lstsq_grad_oriented_cuda_float32 SKIPPED [0.0016s] (no MAGMA library detected) [ 67%] 2025-12-04T15:25:36.7122404Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_lu_cuda_float32 PASSED [0.0251s] [ 68%] 2025-12-04T15:25:36.7123035Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_lu_factor_cuda_float32 PASSED [0.2466s] [ 68%] 2025-12-04T15:25:36.7123651Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_lu_factor_ex_cuda_float32 PASSED [0.0249s] [ 68%] 2025-12-04T15:25:36.7124283Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_lu_solve_cuda_float32 PASSED [0.0656s] [ 68%] 2025-12-04T15:25:36.7124949Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_matrix_norm_cuda_float32 SKIPPED [0.0003s] (Skipped!) [ 68%] 2025-12-04T15:25:36.7125638Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_matrix_power_cuda_float32 PASSED [0.2164s] [ 68%] 2025-12-04T15:25:36.7126250Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_matrix_rank_cuda_float32 PASSED [0.0599s] [ 68%] 2025-12-04T15:25:36.7126908Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_matrix_rank_hermitian_cuda_float32 PASSED [0.0122s] [ 68%] 2025-12-04T15:25:36.7127578Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_multi_dot_cuda_float32 SKIPPED [0.0003s] (Skipped!) [ 68%] 2025-12-04T15:25:36.7128153Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_norm_cuda_float32 XFAIL [0.0224s] [ 68%] 2025-12-04T15:25:36.7128839Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_norm_subgradients_at_zero_cuda_float32 XFAIL [0.0202s] [ 68%] 2025-12-04T15:25:36.7129422Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_pinv_cuda_float32 PASSED [0.4209s] [ 69%] 2025-12-04T15:25:36.7130186Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_pinv_hermitian_cuda_float32 SKIPPED [0.0017s] (no MAGMA library detected) [ 69%] 2025-12-04T15:25:36.7131096Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_pinv_singular_cuda_float32 SKIPPED [0.0017s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 69%] 2025-12-04T15:25:36.7131672Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_qr_cuda_float32 PASSED [0.0308s] [ 69%] 2025-12-04T15:25:36.7132279Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_slogdet_cuda_float32 PASSED [0.2129s] [ 69%] 2025-12-04T15:25:36.7132867Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_solve_cuda_float32 PASSED [0.0248s] [ 69%] 2025-12-04T15:25:36.7133483Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_solve_ex_cuda_float32 PASSED [0.2206s] [ 69%] 2025-12-04T15:25:36.7134119Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_solve_triangular_cuda_float32 PASSED [0.0815s] [ 69%] 2025-12-04T15:25:36.7134693Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_svd_cuda_float32 PASSED [0.1588s] [ 69%] 2025-12-04T15:25:36.7135290Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_svdvals_cuda_float32 PASSED [0.0219s] [ 69%] 2025-12-04T15:25:36.7135889Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_tensorinv_cuda_float32 PASSED [0.0086s] [ 69%] 2025-12-04T15:25:36.7136512Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_tensorsolve_cuda_float32 PASSED [0.2089s] [ 70%] 2025-12-04T15:25:36.7137185Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_vander_cuda_float32 XFAIL [0.0099s] [ 70%] 2025-12-04T15:25:36.7137769Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_vecdot_cuda_float32 XFAIL [0.2139s] [ 70%] 2025-12-04T15:25:36.7138423Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_linalg_failure_1D_input_linalg_vector_norm_cuda_float32 XFAIL [0.0296s] [ 70%] 2025-12-04T15:25:36.7138953Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_multi_dot_failure_1D_input_cuda PASSED [0.1979s] [ 70%] 2025-12-04T15:25:36.7139446Z functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_vmap_with_anomaly_detection_cuda PASSED [0.0072s] [ 70%] 2025-12-04T15:25:36.7139810Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_add_cuda PASSED [0.0146s] [ 70%] 2025-12-04T15:25:36.7140322Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_binary_cross_entropy_cuda PASSED [0.0075s] [ 70%] 2025-12-04T15:25:36.7140728Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_diagonal_cuda PASSED [0.0061s] [ 70%] 2025-12-04T15:25:36.7141091Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_div_cuda PASSED [0.0239s] [ 70%] 2025-12-04T15:25:36.7141485Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_expand_cuda PASSED [0.0039s] [ 70%] 2025-12-04T15:25:36.7141858Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_index_cuda PASSED [0.0076s] [ 70%] 2025-12-04T15:25:36.7142298Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_inplace_manyview_cuda PASSED [0.0083s] [ 71%] 2025-12-04T15:25:36.7142724Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_inplace_view_cuda PASSED [0.0071s] [ 71%] 2025-12-04T15:25:36.7143103Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_lgamma_cuda PASSED [0.0062s] [ 71%] 2025-12-04T15:25:36.7143495Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_log1p_cuda PASSED [0.0067s] [ 71%] 2025-12-04T15:25:36.7143856Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_log_cuda PASSED [0.0060s] [ 71%] 2025-12-04T15:25:36.7144266Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_log_softmax_cuda PASSED [0.0064s] [ 71%] 2025-12-04T15:25:36.7144676Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_logsumexp_cuda PASSED [0.0075s] [ 71%] 2025-12-04T15:25:36.7145035Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_max_cuda PASSED [0.0044s] [ 71%] 2025-12-04T15:25:36.7145416Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_median_cuda PASSED [0.0044s] [ 71%] 2025-12-04T15:25:36.7145794Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_min_cuda PASSED [0.0043s] [ 71%] 2025-12-04T15:25:36.7146153Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_mul_cuda PASSED [0.0212s] [ 71%] 2025-12-04T15:25:36.7146552Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_permute_cuda PASSED [0.0036s] [ 72%] 2025-12-04T15:25:36.7147117Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_randomness_backend0_randomness_different_cuda PASSED [0.0034s] [ 72%] 2025-12-04T15:25:36.7147654Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_randomness_backend0_randomness_error_cuda PASSED [0.0032s] [ 72%] 2025-12-04T15:25:36.7148191Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_randomness_backend0_randomness_same_cuda PASSED [0.0033s] [ 72%] 2025-12-04T15:25:36.7148750Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_randomness_backend1_randomness_different_cuda PASSED [0.0028s] [ 72%] 2025-12-04T15:25:36.7149296Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_randomness_backend1_randomness_error_cuda PASSED [0.0023s] [ 72%] 2025-12-04T15:25:36.7149821Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_randomness_backend1_randomness_same_cuda PASSED [0.0024s] [ 72%] 2025-12-04T15:25:36.7150210Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_reshape_cuda PASSED [0.0036s] [ 72%] 2025-12-04T15:25:36.7150646Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_sdpa_backend0_cuda PASSED [0.0133s] [ 72%] 2025-12-04T15:25:36.7151065Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_sdpa_backend1_cuda PASSED [0.0071s] [ 72%] 2025-12-04T15:25:36.7151457Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_select_cuda PASSED [0.0083s] [ 72%] 2025-12-04T15:25:36.7151876Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_sigmoid_cuda PASSED [0.0063s] [ 73%] 2025-12-04T15:25:36.7152295Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_slice_cuda PASSED [0.0080s] [ 73%] 2025-12-04T15:25:36.7152678Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_stack_cuda PASSED [0.0043s] [ 73%] 2025-12-04T15:25:36.7153036Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_sub_cuda PASSED [0.0144s] [ 73%] 2025-12-04T15:25:36.7162407Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_threshold_cuda PASSED [0.0036s] [ 73%] 2025-12-04T15:25:36.7163010Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_trace_cuda PASSED [0.0051s] [ 73%] 2025-12-04T15:25:36.7163455Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_unrelated_output_cuda PASSED [0.0022s] [ 73%] 2025-12-04T15:25:36.7163973Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_unrelated_output_multiple_grad_cuda PASSED [0.0022s] [ 73%] 2025-12-04T15:25:36.7164394Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_vmap_fallback_check XFAIL [0.0023s] [ 73%] 2025-12-04T15:25:36.7164847Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_vmap_fallback_check_ok PASSED [0.0018s] [ 73%] 2025-12-04T15:25:36.7165224Z functorch/test_vmap.py::TestVmapBatchedGradientCUDA::test_where_cuda PASSED [0.0023s] [ 73%] 2025-12-04T15:25:36.7165815Z functorch/test_vmap.py::TestTransformFailureCUDA::test_fails_with_autograd_function_transform_grad_and_value_cuda PASSED [0.0019s] [ 74%] 2025-12-04T15:25:36.7166374Z functorch/test_vmap.py::TestTransformFailureCUDA::test_fails_with_autograd_function_transform_grad_cuda PASSED [0.0018s] [ 74%] 2025-12-04T15:25:36.7166927Z functorch/test_vmap.py::TestTransformFailureCUDA::test_fails_with_autograd_function_transform_jacfwd_cuda PASSED [0.1028s] [ 74%] 2025-12-04T15:25:36.7167492Z functorch/test_vmap.py::TestTransformFailureCUDA::test_fails_with_autograd_function_transform_jacrev_cuda PASSED [0.0022s] [ 74%] 2025-12-04T15:25:36.7168033Z functorch/test_vmap.py::TestTransformFailureCUDA::test_fails_with_autograd_function_transform_jvp_cuda PASSED [0.0020s] [ 74%] 2025-12-04T15:25:36.7168573Z functorch/test_vmap.py::TestTransformFailureCUDA::test_fails_with_autograd_function_transform_vjp_cuda PASSED [0.0018s] [ 74%] 2025-12-04T15:25:36.7169125Z functorch/test_vmap.py::TestTransformFailureCUDA::test_fails_with_autograd_function_transform_vmap_cuda PASSED [0.0018s] [ 74%] 2025-12-04T15:25:36.7169690Z functorch/test_vmap.py::TestRandomnessCUDA::test_alpha_dropout_randomness_different_batched_input_first_cuda PASSED [0.0027s] [ 74%] 2025-12-04T15:25:36.7170265Z functorch/test_vmap.py::TestRandomnessCUDA::test_alpha_dropout_randomness_different_batched_input_last_cuda PASSED [0.0031s] [ 74%] 2025-12-04T15:25:36.7170821Z functorch/test_vmap.py::TestRandomnessCUDA::test_alpha_dropout_randomness_different_batched_input_none_cuda PASSED [0.0024s] [ 74%] 2025-12-04T15:25:36.7171361Z functorch/test_vmap.py::TestRandomnessCUDA::test_alpha_dropout_randomness_error_batched_input_first_cuda PASSED [0.0021s] [ 74%] 2025-12-04T15:25:36.7171909Z functorch/test_vmap.py::TestRandomnessCUDA::test_alpha_dropout_randomness_error_batched_input_last_cuda PASSED [0.0019s] [ 74%] 2025-12-04T15:25:36.7172440Z functorch/test_vmap.py::TestRandomnessCUDA::test_alpha_dropout_randomness_error_batched_input_none_cuda PASSED [0.0019s] [ 75%] 2025-12-04T15:25:36.7172987Z functorch/test_vmap.py::TestRandomnessCUDA::test_alpha_dropout_randomness_same_batched_input_first_cuda PASSED [0.0020s] [ 75%] 2025-12-04T15:25:36.7173519Z functorch/test_vmap.py::TestRandomnessCUDA::test_alpha_dropout_randomness_same_batched_input_last_cuda PASSED [0.0025s] [ 75%] 2025-12-04T15:25:36.7174044Z functorch/test_vmap.py::TestRandomnessCUDA::test_alpha_dropout_randomness_same_batched_input_none_cuda PASSED [0.0020s] [ 75%] 2025-12-04T15:25:36.7174909Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_different_batched_input_first_batched_probability_first_cuda PASSED [0.0034s] [ 75%] 2025-12-04T15:25:36.7175749Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_different_batched_input_first_batched_probability_last_cuda PASSED [0.0033s] [ 75%] 2025-12-04T15:25:36.7176568Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_different_batched_input_first_batched_probability_none_cuda PASSED [0.0033s] [ 75%] 2025-12-04T15:25:36.7177472Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_different_batched_input_last_batched_probability_first_cuda PASSED [0.0038s] [ 75%] 2025-12-04T15:25:36.7178352Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_different_batched_input_last_batched_probability_last_cuda PASSED [0.0033s] [ 75%] 2025-12-04T15:25:36.7179148Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_different_batched_input_last_batched_probability_none_cuda PASSED [0.0033s] [ 75%] 2025-12-04T15:25:36.7179945Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_different_batched_input_none_batched_probability_first_cuda PASSED [0.0021s] [ 75%] 2025-12-04T15:25:36.7180750Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_different_batched_input_none_batched_probability_last_cuda PASSED [0.0019s] [ 76%] 2025-12-04T15:25:36.7181541Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_different_batched_input_none_batched_probability_none_cuda PASSED [0.0021s] [ 76%] 2025-12-04T15:25:36.7182345Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_error_batched_input_first_batched_probability_first_cuda PASSED [0.0024s] [ 76%] 2025-12-04T15:25:36.7183123Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_error_batched_input_first_batched_probability_last_cuda PASSED [0.0019s] [ 76%] 2025-12-04T15:25:36.7183920Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_error_batched_input_first_batched_probability_none_cuda PASSED [0.0019s] [ 76%] 2025-12-04T15:25:36.7184694Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_error_batched_input_last_batched_probability_first_cuda PASSED [0.0019s] [ 76%] 2025-12-04T15:25:36.7185459Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_error_batched_input_last_batched_probability_last_cuda PASSED [0.0019s] [ 76%] 2025-12-04T15:25:36.7186243Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_error_batched_input_last_batched_probability_none_cuda PASSED [0.0019s] [ 76%] 2025-12-04T15:25:36.7187018Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_error_batched_input_none_batched_probability_first_cuda PASSED [0.0023s] [ 76%] 2025-12-04T15:25:36.7187807Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_error_batched_input_none_batched_probability_last_cuda PASSED [0.0021s] [ 76%] 2025-12-04T15:25:36.7188570Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_error_batched_input_none_batched_probability_none_cuda PASSED [0.0019s] [ 76%] 2025-12-04T15:25:36.7189356Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_same_batched_input_first_batched_probability_first_cuda PASSED [0.0021s] [ 77%] 2025-12-04T15:25:36.7190118Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_same_batched_input_first_batched_probability_last_cuda PASSED [0.0019s] [ 77%] 2025-12-04T15:25:36.7190931Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_same_batched_input_first_batched_probability_none_cuda PASSED [0.0039s] [ 77%] 2025-12-04T15:25:36.7191734Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_same_batched_input_last_batched_probability_first_cuda PASSED [0.0023s] [ 77%] 2025-12-04T15:25:36.7192497Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_same_batched_input_last_batched_probability_last_cuda PASSED [0.0020s] [ 77%] 2025-12-04T15:25:36.7193332Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_same_batched_input_last_batched_probability_none_cuda PASSED [0.0039s] [ 77%] 2025-12-04T15:25:36.7194093Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_same_batched_input_none_batched_probability_first_cuda PASSED [0.0019s] [ 77%] 2025-12-04T15:25:36.7194872Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_same_batched_input_none_batched_probability_last_cuda PASSED [0.0019s] [ 77%] 2025-12-04T15:25:36.7195631Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_False_randomness_same_batched_input_none_batched_probability_none_cuda PASSED [0.0038s] [ 77%] 2025-12-04T15:25:36.7196618Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_different_batched_input_first_batched_probability_first_cuda PASSED [0.0033s] [ 77%] 2025-12-04T15:25:36.7197415Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_different_batched_input_first_batched_probability_last_cuda PASSED [0.0026s] [ 77%] 2025-12-04T15:25:36.7198208Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_different_batched_input_first_batched_probability_none_cuda PASSED [0.0026s] [ 77%] 2025-12-04T15:25:36.7199022Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_different_batched_input_last_batched_probability_first_cuda PASSED [0.0027s] [ 78%] 2025-12-04T15:25:36.7199809Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_different_batched_input_last_batched_probability_last_cuda PASSED [0.0027s] [ 78%] 2025-12-04T15:25:36.7200608Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_different_batched_input_last_batched_probability_none_cuda PASSED [0.0032s] [ 78%] 2025-12-04T15:25:36.7201403Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_different_batched_input_none_batched_probability_first_cuda PASSED [0.0019s] [ 78%] 2025-12-04T15:25:36.7202203Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_different_batched_input_none_batched_probability_last_cuda PASSED [0.0019s] [ 78%] 2025-12-04T15:25:36.7202993Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_different_batched_input_none_batched_probability_none_cuda PASSED [0.0019s] [ 78%] 2025-12-04T15:25:36.7203763Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_error_batched_input_first_batched_probability_first_cuda PASSED [0.0019s] [ 78%] 2025-12-04T15:25:36.7204543Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_error_batched_input_first_batched_probability_last_cuda PASSED [0.0019s] [ 78%] 2025-12-04T15:25:36.7205316Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_error_batched_input_first_batched_probability_none_cuda PASSED [0.0023s] [ 78%] 2025-12-04T15:25:36.7206160Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_error_batched_input_last_batched_probability_first_cuda PASSED [0.0019s] [ 78%] 2025-12-04T15:25:36.7206923Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_error_batched_input_last_batched_probability_last_cuda PASSED [0.0019s] [ 78%] 2025-12-04T15:25:36.7207744Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_error_batched_input_last_batched_probability_none_cuda PASSED [0.0022s] [ 79%] 2025-12-04T15:25:36.7208514Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_error_batched_input_none_batched_probability_first_cuda PASSED [0.0019s] [ 79%] 2025-12-04T15:25:36.7209349Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_error_batched_input_none_batched_probability_last_cuda PASSED [0.0019s] [ 79%] 2025-12-04T15:25:36.7210126Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_error_batched_input_none_batched_probability_none_cuda PASSED [0.0023s] [ 79%] 2025-12-04T15:25:36.7210896Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_same_batched_input_first_batched_probability_first_cuda PASSED [0.0019s] [ 79%] 2025-12-04T15:25:36.7211667Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_same_batched_input_first_batched_probability_last_cuda PASSED [0.0019s] [ 79%] 2025-12-04T15:25:36.7212424Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_same_batched_input_first_batched_probability_none_cuda PASSED [0.0032s] [ 79%] 2025-12-04T15:25:36.7213202Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_same_batched_input_last_batched_probability_first_cuda PASSED [0.0019s] [ 79%] 2025-12-04T15:25:36.7213959Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_same_batched_input_last_batched_probability_last_cuda PASSED [0.0019s] [ 79%] 2025-12-04T15:25:36.7214718Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_same_batched_input_last_batched_probability_none_cuda PASSED [0.0037s] [ 79%] 2025-12-04T15:25:36.7215489Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_same_batched_input_none_batched_probability_first_cuda PASSED [0.0019s] [ 79%] 2025-12-04T15:25:36.7216244Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_same_batched_input_none_batched_probability_last_cuda PASSED [0.0019s] [ 80%] 2025-12-04T15:25:36.7217081Z functorch/test_vmap.py::TestRandomnessCUDA::test_bernoulli_in_place_use_generator_True_randomness_same_batched_input_none_batched_probability_none_cuda PASSED [0.0031s] [ 80%] 2025-12-04T15:25:36.7217509Z functorch/test_vmap.py::TestRandomnessCUDA::test_chunk_vmap_in_dim_0_out_dim_0_cuda PASSED [0.0070s] [ 80%] 2025-12-04T15:25:36.7217936Z functorch/test_vmap.py::TestRandomnessCUDA::test_chunk_vmap_in_dim_0_out_dim_1_cuda PASSED [0.0070s] [ 80%] 2025-12-04T15:25:36.7218373Z functorch/test_vmap.py::TestRandomnessCUDA::test_chunk_vmap_in_dim_0_out_dim_2_cuda PASSED [0.0070s] [ 80%] 2025-12-04T15:25:36.7218789Z functorch/test_vmap.py::TestRandomnessCUDA::test_chunk_vmap_in_dim_1_out_dim_0_cuda PASSED [0.0071s] [ 80%] 2025-12-04T15:25:36.7219218Z functorch/test_vmap.py::TestRandomnessCUDA::test_chunk_vmap_in_dim_1_out_dim_1_cuda PASSED [0.0073s] [ 80%] 2025-12-04T15:25:36.7219633Z functorch/test_vmap.py::TestRandomnessCUDA::test_chunk_vmap_in_dim_1_out_dim_2_cuda PASSED [0.0073s] [ 80%] 2025-12-04T15:25:36.7220054Z functorch/test_vmap.py::TestRandomnessCUDA::test_chunk_vmap_in_dim_2_out_dim_0_cuda PASSED [0.0074s] [ 80%] 2025-12-04T15:25:36.7220486Z functorch/test_vmap.py::TestRandomnessCUDA::test_chunk_vmap_in_dim_2_out_dim_1_cuda PASSED [0.0080s] [ 80%] 2025-12-04T15:25:36.7220955Z functorch/test_vmap.py::TestRandomnessCUDA::test_chunk_vmap_in_dim_2_out_dim_2_cuda PASSED [0.0076s] [ 80%] 2025-12-04T15:25:36.7221511Z functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_randomness_different_batched_input_first_cuda PASSED [0.0025s] [ 81%] 2025-12-04T15:25:36.7222125Z functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_randomness_different_batched_input_last_cuda PASSED [0.0025s] [ 81%] 2025-12-04T15:25:36.7222659Z functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_randomness_different_batched_input_none_cuda PASSED [0.0025s] [ 81%] 2025-12-04T15:25:36.7223185Z functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_randomness_error_batched_input_first_cuda PASSED [0.0020s] [ 81%] 2025-12-04T15:25:36.7223755Z functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_randomness_error_batched_input_last_cuda PASSED [0.0019s] [ 81%] 2025-12-04T15:25:36.7224269Z functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_randomness_error_batched_input_none_cuda PASSED [0.0024s] [ 81%] 2025-12-04T15:25:36.7224771Z functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_randomness_same_batched_input_first_cuda PASSED [0.0021s] [ 81%] 2025-12-04T15:25:36.7225266Z functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_randomness_same_batched_input_last_cuda PASSED [0.0021s] [ 81%] 2025-12-04T15:25:36.7225777Z functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_randomness_same_batched_input_none_cuda PASSED [0.0021s] [ 81%] 2025-12-04T15:25:36.7226264Z functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_unbatched_randomness_different_cuda PASSED [0.0019s] [ 81%] 2025-12-04T15:25:36.7226737Z functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_unbatched_randomness_error_cuda PASSED [0.0019s] [ 81%] 2025-12-04T15:25:36.7227193Z functorch/test_vmap.py::TestRandomnessCUDA::test_dropout_unbatched_randomness_same_cuda PASSED [0.0025s] [ 81%] 2025-12-04T15:25:36.7227744Z functorch/test_vmap.py::TestRandomnessCUDA::test_factory_ops_randomness_different_use_generator_False_cuda PASSED [0.0097s] [ 82%] 2025-12-04T15:25:36.7228301Z functorch/test_vmap.py::TestRandomnessCUDA::test_factory_ops_randomness_different_use_generator_True_cuda PASSED [0.0064s] [ 82%] 2025-12-04T15:25:36.7228828Z functorch/test_vmap.py::TestRandomnessCUDA::test_factory_ops_randomness_error_use_generator_False_cuda PASSED [0.0018s] [ 82%] 2025-12-04T15:25:36.7229358Z functorch/test_vmap.py::TestRandomnessCUDA::test_factory_ops_randomness_error_use_generator_True_cuda PASSED [0.0018s] [ 82%] 2025-12-04T15:25:36.7229871Z functorch/test_vmap.py::TestRandomnessCUDA::test_factory_ops_randomness_same_use_generator_False_cuda PASSED [0.0107s] [ 82%] 2025-12-04T15:25:36.7230389Z functorch/test_vmap.py::TestRandomnessCUDA::test_factory_ops_randomness_same_use_generator_True_cuda PASSED [0.0074s] [ 82%] 2025-12-04T15:25:36.7231013Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_alpha_dropout_randomness_different_batched_input_first_cuda PASSED [0.0034s] [ 82%] 2025-12-04T15:25:36.7231611Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_alpha_dropout_randomness_different_batched_input_last_cuda PASSED [0.0027s] [ 82%] 2025-12-04T15:25:36.7232218Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_alpha_dropout_randomness_different_batched_input_none_cuda PASSED [0.0026s] [ 82%] 2025-12-04T15:25:36.7232799Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_alpha_dropout_randomness_error_batched_input_first_cuda PASSED [0.0019s] [ 82%] 2025-12-04T15:25:36.7233374Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_alpha_dropout_randomness_error_batched_input_last_cuda PASSED [0.0019s] [ 82%] 2025-12-04T15:25:36.7233960Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_alpha_dropout_randomness_error_batched_input_none_cuda PASSED [0.0019s] [ 83%] 2025-12-04T15:25:36.7234533Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_alpha_dropout_randomness_same_batched_input_first_cuda PASSED [0.0027s] [ 83%] 2025-12-04T15:25:36.7235140Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_alpha_dropout_randomness_same_batched_input_last_cuda PASSED [0.0022s] [ 83%] 2025-12-04T15:25:36.7235707Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_alpha_dropout_randomness_same_batched_input_none_cuda PASSED [0.0022s] [ 83%] 2025-12-04T15:25:36.7236350Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_different_batched_input_first_dim_2_cuda PASSED [0.0026s] [ 83%] 2025-12-04T15:25:36.7236957Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_different_batched_input_first_dim_3_cuda PASSED [0.0025s] [ 83%] 2025-12-04T15:25:36.7237554Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_different_batched_input_last_dim_2_cuda PASSED [0.0030s] [ 83%] 2025-12-04T15:25:36.7238216Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_different_batched_input_last_dim_3_cuda PASSED [0.0025s] [ 83%] 2025-12-04T15:25:36.7238806Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_different_batched_input_none_dim_2_cuda PASSED [0.0025s] [ 83%] 2025-12-04T15:25:36.7239414Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_different_batched_input_none_dim_3_cuda PASSED [0.0026s] [ 83%] 2025-12-04T15:25:36.7239987Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_error_batched_input_first_dim_2_cuda PASSED [0.0019s] [ 83%] 2025-12-04T15:25:36.7240560Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_error_batched_input_first_dim_3_cuda PASSED [0.0019s] [ 84%] 2025-12-04T15:25:36.7241136Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_error_batched_input_last_dim_2_cuda PASSED [0.0023s] [ 84%] 2025-12-04T15:25:36.7241706Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_error_batched_input_last_dim_3_cuda PASSED [0.0018s] [ 84%] 2025-12-04T15:25:36.7242287Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_error_batched_input_none_dim_2_cuda PASSED [0.0018s] [ 84%] 2025-12-04T15:25:36.7242856Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_error_batched_input_none_dim_3_cuda PASSED [0.0019s] [ 84%] 2025-12-04T15:25:36.7243429Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_same_batched_input_first_dim_2_cuda PASSED [0.0021s] [ 84%] 2025-12-04T15:25:36.7244003Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_same_batched_input_first_dim_3_cuda PASSED [0.0021s] [ 84%] 2025-12-04T15:25:36.7244565Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_same_batched_input_last_dim_2_cuda PASSED [0.0026s] [ 84%] 2025-12-04T15:25:36.7245151Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_same_batched_input_last_dim_3_cuda PASSED [0.0022s] [ 84%] 2025-12-04T15:25:36.7245715Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_same_batched_input_none_dim_2_cuda PASSED [0.0021s] [ 84%] 2025-12-04T15:25:36.7246275Z functorch/test_vmap.py::TestRandomnessCUDA::test_feature_dropout_randomness_same_batched_input_none_dim_3_cuda PASSED [0.0022s] [ 84%] 2025-12-04T15:25:36.7246668Z functorch/test_vmap.py::TestRandomnessCUDA::test_jacfwd_with_random_cuda PASSED [0.0032s] [ 85%] 2025-12-04T15:25:36.7247242Z functorch/test_vmap.py::TestRandomnessCUDA::test_like_functions_randomness_different_batched_input_first_cuda PASSED [0.0037s] [ 85%] 2025-12-04T15:25:36.7247819Z functorch/test_vmap.py::TestRandomnessCUDA::test_like_functions_randomness_different_batched_input_last_cuda PASSED [0.0032s] [ 85%] 2025-12-04T15:25:36.7248381Z functorch/test_vmap.py::TestRandomnessCUDA::test_like_functions_randomness_different_batched_input_none_cuda PASSED [0.0032s] [ 85%] 2025-12-04T15:25:36.7248929Z functorch/test_vmap.py::TestRandomnessCUDA::test_like_functions_randomness_error_batched_input_first_cuda PASSED [0.0018s] [ 85%] 2025-12-04T15:25:36.7249479Z functorch/test_vmap.py::TestRandomnessCUDA::test_like_functions_randomness_error_batched_input_last_cuda PASSED [0.0018s] [ 85%] 2025-12-04T15:25:36.7250043Z functorch/test_vmap.py::TestRandomnessCUDA::test_like_functions_randomness_error_batched_input_none_cuda PASSED [0.0018s] [ 85%] 2025-12-04T15:25:36.7250621Z functorch/test_vmap.py::TestRandomnessCUDA::test_like_functions_randomness_same_batched_input_first_cuda PASSED [0.0107s] [ 85%] 2025-12-04T15:25:36.7251155Z functorch/test_vmap.py::TestRandomnessCUDA::test_like_functions_randomness_same_batched_input_last_cuda PASSED [0.0100s] [ 85%] 2025-12-04T15:25:36.7251702Z functorch/test_vmap.py::TestRandomnessCUDA::test_like_functions_randomness_same_batched_input_none_cuda PASSED [0.0099s] [ 85%] 2025-12-04T15:25:36.7252503Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_different_batched_call_False_batched_input_first_cuda PASSED [0.0035s] [ 85%] 2025-12-04T15:25:36.7253240Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_different_batched_call_False_batched_input_last_cuda PASSED [0.0035s] [ 85%] 2025-12-04T15:25:36.7253991Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_different_batched_call_False_batched_input_none_cuda PASSED [0.0039s] [ 86%] 2025-12-04T15:25:36.7254729Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_different_batched_call_True_batched_input_first_cuda PASSED [0.0034s] [ 86%] 2025-12-04T15:25:36.7255468Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_different_batched_call_True_batched_input_last_cuda PASSED [0.0034s] [ 86%] 2025-12-04T15:25:36.7256202Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_different_batched_call_True_batched_input_none_cuda PASSED [0.0035s] [ 86%] 2025-12-04T15:25:36.7257005Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_error_batched_call_False_batched_input_first_cuda PASSED [0.0019s] [ 86%] 2025-12-04T15:25:36.7257738Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_error_batched_call_False_batched_input_last_cuda PASSED [0.0019s] [ 86%] 2025-12-04T15:25:36.7258446Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_error_batched_call_False_batched_input_none_cuda PASSED [0.0022s] [ 86%] 2025-12-04T15:25:36.7259175Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_error_batched_call_True_batched_input_first_cuda PASSED [0.0018s] [ 86%] 2025-12-04T15:25:36.7259886Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_error_batched_call_True_batched_input_last_cuda PASSED [0.0019s] [ 86%] 2025-12-04T15:25:36.7260609Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_error_batched_call_True_batched_input_none_cuda PASSED [0.0018s] [ 86%] 2025-12-04T15:25:36.7261318Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_same_batched_call_False_batched_input_first_cuda PASSED [0.0018s] [ 86%] 2025-12-04T15:25:36.7262027Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_same_batched_call_False_batched_input_last_cuda PASSED [0.0018s] [ 87%] 2025-12-04T15:25:36.7262747Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_same_batched_call_False_batched_input_none_cuda PASSED [0.0037s] [ 87%] 2025-12-04T15:25:36.7263453Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_same_batched_call_True_batched_input_first_cuda PASSED [0.0018s] [ 87%] 2025-12-04T15:25:36.7264172Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_same_batched_call_True_batched_input_last_cuda PASSED [0.0018s] [ 87%] 2025-12-04T15:25:36.7264912Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_False_randomness_same_batched_call_True_batched_input_none_cuda PASSED [0.0034s] [ 87%] 2025-12-04T15:25:36.7265694Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_different_batched_call_False_batched_input_first_cuda PASSED [0.0027s] [ 87%] 2025-12-04T15:25:36.7266430Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_different_batched_call_False_batched_input_last_cuda PASSED [0.0032s] [ 87%] 2025-12-04T15:25:36.7267153Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_different_batched_call_False_batched_input_none_cuda PASSED [0.0028s] [ 87%] 2025-12-04T15:25:36.7267974Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_different_batched_call_True_batched_input_first_cuda PASSED [0.0027s] [ 87%] 2025-12-04T15:25:36.7268700Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_different_batched_call_True_batched_input_last_cuda PASSED [0.0028s] [ 87%] 2025-12-04T15:25:36.7269442Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_different_batched_call_True_batched_input_none_cuda PASSED [0.0028s] [ 87%] 2025-12-04T15:25:36.7270157Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_error_batched_call_False_batched_input_first_cuda PASSED [0.0019s] [ 88%] 2025-12-04T15:25:36.7270865Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_error_batched_call_False_batched_input_last_cuda PASSED [0.0023s] [ 88%] 2025-12-04T15:25:36.7271582Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_error_batched_call_False_batched_input_none_cuda PASSED [0.0018s] [ 88%] 2025-12-04T15:25:36.7272288Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_error_batched_call_True_batched_input_first_cuda PASSED [0.0018s] [ 88%] 2025-12-04T15:25:36.7273004Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_error_batched_call_True_batched_input_last_cuda PASSED [0.0019s] [ 88%] 2025-12-04T15:25:36.7273702Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_error_batched_call_True_batched_input_none_cuda PASSED [0.0018s] [ 88%] 2025-12-04T15:25:36.7274419Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_same_batched_call_False_batched_input_first_cuda PASSED [0.0018s] [ 88%] 2025-12-04T15:25:36.7275124Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_same_batched_call_False_batched_input_last_cuda PASSED [0.0022s] [ 88%] 2025-12-04T15:25:36.7275822Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_same_batched_call_False_batched_input_none_cuda PASSED [0.0027s] [ 88%] 2025-12-04T15:25:36.7276534Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_same_batched_call_True_batched_input_first_cuda PASSED [0.0018s] [ 88%] 2025-12-04T15:25:36.7277229Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_same_batched_call_True_batched_input_last_cuda PASSED [0.0018s] [ 88%] 2025-12-04T15:25:36.7277932Z functorch/test_vmap.py::TestRandomnessCUDA::test_multinomial_use_generator_True_randomness_same_batched_call_True_batched_input_none_cuda PASSED [0.0027s] [ 88%] 2025-12-04T15:25:36.7278743Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_different_batched_input_first_batched_other_first_cuda PASSED [0.0053s] [ 89%] 2025-12-04T15:25:36.7279552Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_different_batched_input_first_batched_other_last_cuda PASSED [0.0056s] [ 89%] 2025-12-04T15:25:36.7280371Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_different_batched_input_first_batched_other_none_cuda PASSED [0.0051s] [ 89%] 2025-12-04T15:25:36.7281196Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_different_batched_input_last_batched_other_first_cuda PASSED [0.0051s] [ 89%] 2025-12-04T15:25:36.7281995Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_different_batched_input_last_batched_other_last_cuda PASSED [0.0057s] [ 89%] 2025-12-04T15:25:36.7282785Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_different_batched_input_last_batched_other_none_cuda PASSED [0.0052s] [ 89%] 2025-12-04T15:25:36.7283649Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_different_batched_input_none_batched_other_first_cuda PASSED [0.0052s] [ 89%] 2025-12-04T15:25:36.7284435Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_different_batched_input_none_batched_other_last_cuda PASSED [0.0057s] [ 89%] 2025-12-04T15:25:36.7285239Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_different_batched_input_none_batched_other_none_cuda PASSED [0.0052s] [ 89%] 2025-12-04T15:25:36.7286015Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_error_batched_input_first_batched_other_first_cuda PASSED [0.0019s] [ 89%] 2025-12-04T15:25:36.7286791Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_error_batched_input_first_batched_other_last_cuda PASSED [0.0019s] [ 89%] 2025-12-04T15:25:36.7287575Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_error_batched_input_first_batched_other_none_cuda PASSED [0.0019s] [ 90%] 2025-12-04T15:25:36.7288351Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_error_batched_input_last_batched_other_first_cuda PASSED [0.0019s] [ 90%] 2025-12-04T15:25:36.7289132Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_error_batched_input_last_batched_other_last_cuda PASSED [0.0023s] [ 90%] 2025-12-04T15:25:36.7289902Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_error_batched_input_last_batched_other_none_cuda PASSED [0.0019s] [ 90%] 2025-12-04T15:25:36.7290691Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_error_batched_input_none_batched_other_first_cuda PASSED [0.0018s] [ 90%] 2025-12-04T15:25:36.7291452Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_error_batched_input_none_batched_other_last_cuda PASSED [0.0018s] [ 90%] 2025-12-04T15:25:36.7292217Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_error_batched_input_none_batched_other_none_cuda PASSED [0.0018s] [ 90%] 2025-12-04T15:25:36.7293001Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_same_batched_input_first_batched_other_first_cuda PASSED [0.0018s] [ 90%] 2025-12-04T15:25:36.7293763Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_same_batched_input_first_batched_other_last_cuda PASSED [0.0023s] [ 90%] 2025-12-04T15:25:36.7294546Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_same_batched_input_first_batched_other_none_cuda PASSED [0.0018s] [ 90%] 2025-12-04T15:25:36.7295309Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_same_batched_input_last_batched_other_first_cuda PASSED [0.0018s] [ 90%] 2025-12-04T15:25:36.7296285Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_same_batched_input_last_batched_other_last_cuda PASSED [0.0018s] [ 91%] 2025-12-04T15:25:36.7297148Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_same_batched_input_last_batched_other_none_cuda PASSED [0.0019s] [ 91%] 2025-12-04T15:25:36.7297918Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_same_batched_input_none_batched_other_first_cuda PASSED [0.0019s] [ 91%] 2025-12-04T15:25:36.7298766Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_same_batched_input_none_batched_other_last_cuda PASSED [0.0027s] [ 91%] 2025-12-04T15:25:36.7299524Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_False_randomness_same_batched_input_none_batched_other_none_cuda PASSED [0.0063s] [ 91%] 2025-12-04T15:25:36.7300337Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_different_batched_input_first_batched_other_first_cuda PASSED [0.0039s] [ 91%] 2025-12-04T15:25:36.7301135Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_different_batched_input_first_batched_other_last_cuda PASSED [0.0040s] [ 91%] 2025-12-04T15:25:36.7301931Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_different_batched_input_first_batched_other_none_cuda PASSED [0.0043s] [ 91%] 2025-12-04T15:25:36.7302725Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_different_batched_input_last_batched_other_first_cuda PASSED [0.0039s] [ 91%] 2025-12-04T15:25:36.7303506Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_different_batched_input_last_batched_other_last_cuda PASSED [0.0040s] [ 91%] 2025-12-04T15:25:36.7304303Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_different_batched_input_last_batched_other_none_cuda PASSED [0.0042s] [ 91%] 2025-12-04T15:25:36.7305093Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_different_batched_input_none_batched_other_first_cuda PASSED [0.0038s] [ 92%] 2025-12-04T15:25:36.7305892Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_different_batched_input_none_batched_other_last_cuda PASSED [0.0040s] [ 92%] 2025-12-04T15:25:36.7306677Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_different_batched_input_none_batched_other_none_cuda PASSED [0.0042s] [ 92%] 2025-12-04T15:25:36.7307459Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_error_batched_input_first_batched_other_first_cuda PASSED [0.0019s] [ 92%] 2025-12-04T15:25:36.7308219Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_error_batched_input_first_batched_other_last_cuda PASSED [0.0019s] [ 92%] 2025-12-04T15:25:36.7308978Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_error_batched_input_first_batched_other_none_cuda PASSED [0.0019s] [ 92%] 2025-12-04T15:25:36.7309762Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_error_batched_input_last_batched_other_first_cuda PASSED [0.0019s] [ 92%] 2025-12-04T15:25:36.7310531Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_error_batched_input_last_batched_other_last_cuda PASSED [0.0019s] [ 92%] 2025-12-04T15:25:36.7311341Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_error_batched_input_last_batched_other_none_cuda PASSED [0.0024s] [ 92%] 2025-12-04T15:25:36.7312134Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_error_batched_input_none_batched_other_first_cuda PASSED [0.0019s] [ 92%] 2025-12-04T15:25:36.7312909Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_error_batched_input_none_batched_other_last_cuda PASSED [0.0019s] [ 92%] 2025-12-04T15:25:36.7313668Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_error_batched_input_none_batched_other_none_cuda PASSED [0.0019s] [ 92%] 2025-12-04T15:25:36.7314494Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_same_batched_input_first_batched_other_first_cuda PASSED [0.0019s] [ 93%] 2025-12-04T15:25:36.7315268Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_same_batched_input_first_batched_other_last_cuda PASSED [0.0019s] [ 93%] 2025-12-04T15:25:36.7316032Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_same_batched_input_first_batched_other_none_cuda PASSED [0.0023s] [ 93%] 2025-12-04T15:25:36.7316804Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_same_batched_input_last_batched_other_first_cuda PASSED [0.0019s] [ 93%] 2025-12-04T15:25:36.7317552Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_same_batched_input_last_batched_other_last_cuda PASSED [0.0019s] [ 93%] 2025-12-04T15:25:36.7318315Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_same_batched_input_last_batched_other_none_cuda PASSED [0.0018s] [ 93%] 2025-12-04T15:25:36.7319073Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_same_batched_input_none_batched_other_first_cuda PASSED [0.0019s] [ 93%] 2025-12-04T15:25:36.7319827Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_same_batched_input_none_batched_other_last_cuda PASSED [0.0019s] [ 93%] 2025-12-04T15:25:36.7320594Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_binary_out_of_place_use_generator_True_randomness_same_batched_input_none_batched_other_none_cuda PASSED [0.0054s] [ 93%] 2025-12-04T15:25:36.7321292Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_False_randomness_different_batched_input_first_cuda PASSED [0.0181s] [ 93%] 2025-12-04T15:25:36.7321989Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_False_randomness_different_batched_input_last_cuda PASSED [0.0186s] [ 93%] 2025-12-04T15:25:36.7322676Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_False_randomness_different_batched_input_none_cuda PASSED [0.0020s] [ 94%] 2025-12-04T15:25:36.7323358Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_False_randomness_error_batched_input_first_cuda PASSED [0.0019s] [ 94%] 2025-12-04T15:25:36.7324012Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_False_randomness_error_batched_input_last_cuda PASSED [0.0019s] [ 94%] 2025-12-04T15:25:36.7324674Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_False_randomness_error_batched_input_none_cuda PASSED [0.0019s] [ 94%] 2025-12-04T15:25:36.7325350Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_False_randomness_same_batched_input_first_cuda PASSED [0.0236s] [ 94%] 2025-12-04T15:25:36.7326003Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_False_randomness_same_batched_input_last_cuda PASSED [0.0237s] [ 94%] 2025-12-04T15:25:36.7326706Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_False_randomness_same_batched_input_none_cuda PASSED [0.0231s] [ 94%] 2025-12-04T15:25:36.7327446Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_True_randomness_different_batched_input_first_cuda PASSED [0.0115s] [ 94%] 2025-12-04T15:25:36.7328143Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_True_randomness_different_batched_input_last_cuda PASSED [0.0121s] [ 94%] 2025-12-04T15:25:36.7328821Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_True_randomness_different_batched_input_none_cuda PASSED [0.0020s] [ 94%] 2025-12-04T15:25:36.7329558Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_True_randomness_error_batched_input_first_cuda PASSED [0.0019s] [ 94%] 2025-12-04T15:25:36.7330234Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_True_randomness_error_batched_input_last_cuda PASSED [0.0020s] [ 95%] 2025-12-04T15:25:36.7330890Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_True_randomness_error_batched_input_none_cuda PASSED [0.0019s] [ 95%] 2025-12-04T15:25:36.7331539Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_True_randomness_same_batched_input_first_cuda PASSED [0.0172s] [ 95%] 2025-12-04T15:25:36.7332199Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_True_randomness_same_batched_input_last_cuda PASSED [0.0170s] [ 95%] 2025-12-04T15:25:36.7332850Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_inplace_use_generator_True_randomness_same_batched_input_none_cuda PASSED [0.0164s] [ 95%] 2025-12-04T15:25:36.7333561Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_False_randomness_different_batched_input_first_cuda PASSED [0.0144s] [ 95%] 2025-12-04T15:25:36.7334260Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_False_randomness_different_batched_input_last_cuda PASSED [0.0139s] [ 95%] 2025-12-04T15:25:36.7334951Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_False_randomness_different_batched_input_none_cuda PASSED [0.0137s] [ 95%] 2025-12-04T15:25:36.7335642Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_False_randomness_error_batched_input_first_cuda PASSED [0.0020s] [ 95%] 2025-12-04T15:25:36.7336315Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_False_randomness_error_batched_input_last_cuda PASSED [0.0020s] [ 95%] 2025-12-04T15:25:36.7337068Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_False_randomness_error_batched_input_none_cuda PASSED [0.0019s] [ 95%] 2025-12-04T15:25:36.7337742Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_False_randomness_same_batched_input_first_cuda PASSED [0.0019s] [ 96%] 2025-12-04T15:25:36.7338419Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_False_randomness_same_batched_input_last_cuda PASSED [0.0019s] [ 96%] 2025-12-04T15:25:36.7339084Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_False_randomness_same_batched_input_none_cuda PASSED [0.0173s] [ 96%] 2025-12-04T15:25:36.7339782Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_True_randomness_different_batched_input_first_cuda PASSED [0.0090s] [ 96%] 2025-12-04T15:25:36.7340487Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_True_randomness_different_batched_input_last_cuda PASSED [0.0089s] [ 96%] 2025-12-04T15:25:36.7341212Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_True_randomness_different_batched_input_none_cuda PASSED [0.0091s] [ 96%] 2025-12-04T15:25:36.7341897Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_True_randomness_error_batched_input_first_cuda PASSED [0.0019s] [ 96%] 2025-12-04T15:25:36.7342589Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_True_randomness_error_batched_input_last_cuda PASSED [0.0019s] [ 96%] 2025-12-04T15:25:36.7343266Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_True_randomness_error_batched_input_none_cuda PASSED [0.0019s] [ 96%] 2025-12-04T15:25:36.7343986Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_True_randomness_same_batched_input_first_cuda PASSED [0.0024s] [ 96%] 2025-12-04T15:25:36.7344642Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_True_randomness_same_batched_input_last_cuda PASSED [0.0019s] [ 96%] 2025-12-04T15:25:36.7345316Z functorch/test_vmap.py::TestRandomnessCUDA::test_random_unary_out_of_place_use_generator_True_randomness_same_batched_input_none_cuda PASSED [0.0126s] [ 96%] 2025-12-04T15:25:36.7345861Z functorch/test_vmap.py::TestRandomnessCUDA::test_randperm_randomness_different_use_generator_False_cuda PASSED [0.0038s] [ 97%] 2025-12-04T15:25:36.7346411Z functorch/test_vmap.py::TestRandomnessCUDA::test_randperm_randomness_different_use_generator_True_cuda PASSED [0.0036s] [ 97%] 2025-12-04T15:25:36.7346928Z functorch/test_vmap.py::TestRandomnessCUDA::test_randperm_randomness_error_use_generator_False_cuda PASSED [0.0023s] [ 97%] 2025-12-04T15:25:36.7347438Z functorch/test_vmap.py::TestRandomnessCUDA::test_randperm_randomness_error_use_generator_True_cuda PASSED [0.0027s] [ 97%] 2025-12-04T15:25:36.7347956Z functorch/test_vmap.py::TestRandomnessCUDA::test_randperm_randomness_same_use_generator_False_cuda PASSED [0.0031s] [ 97%] 2025-12-04T15:25:36.7348459Z functorch/test_vmap.py::TestRandomnessCUDA::test_randperm_randomness_same_use_generator_True_cuda PASSED [0.0030s] [ 97%] 2025-12-04T15:25:36.7348857Z functorch/test_vmap.py::TestRandomnessCUDA::test_unsupported_random_cuda PASSED [0.0024s] [ 97%] 2025-12-04T15:25:36.7349305Z functorch/test_vmap.py::TestRandomnessCUDA::test_vmap_chunksize_in_dim_0_out_dim_0_cuda PASSED [0.0072s] [ 97%] 2025-12-04T15:25:36.7349750Z functorch/test_vmap.py::TestRandomnessCUDA::test_vmap_chunksize_in_dim_0_out_dim_1_cuda PASSED [0.0071s] [ 97%] 2025-12-04T15:25:36.7350199Z functorch/test_vmap.py::TestRandomnessCUDA::test_vmap_chunksize_in_dim_0_out_dim_2_cuda PASSED [0.0073s] [ 97%] 2025-12-04T15:25:36.7350668Z functorch/test_vmap.py::TestRandomnessCUDA::test_vmap_chunksize_in_dim_1_out_dim_0_cuda PASSED [0.0073s] [ 97%] 2025-12-04T15:25:36.7351258Z functorch/test_vmap.py::TestRandomnessCUDA::test_vmap_chunksize_in_dim_1_out_dim_1_cuda PASSED [0.0076s] [ 98%] 2025-12-04T15:25:36.7351702Z functorch/test_vmap.py::TestRandomnessCUDA::test_vmap_chunksize_in_dim_1_out_dim_2_cuda PASSED [0.0076s] [ 98%] 2025-12-04T15:25:36.7352140Z functorch/test_vmap.py::TestRandomnessCUDA::test_vmap_chunksize_in_dim_2_out_dim_0_cuda PASSED [0.0075s] [ 98%] 2025-12-04T15:25:36.7352593Z functorch/test_vmap.py::TestRandomnessCUDA::test_vmap_chunksize_in_dim_2_out_dim_1_cuda PASSED [0.0075s] [ 98%] 2025-12-04T15:25:36.7353031Z functorch/test_vmap.py::TestRandomnessCUDA::test_vmap_chunksize_in_dim_2_out_dim_2_cuda PASSED [0.0075s] [ 98%] 2025-12-04T15:25:36.7353426Z functorch/test_vmap.py::TestVmapDeviceTypeCUDA::test__is_all_true_cuda PASSED [0.0037s] [ 98%] 2025-12-04T15:25:36.7353801Z functorch/test_vmap.py::TestVmapDeviceTypeCUDA::test__is_any_true_cuda PASSED [0.0028s] [ 98%] 2025-12-04T15:25:36.7354182Z functorch/test_vmap.py::TestVmapDeviceTypeCUDA::test_check_tensor_cuda PASSED [0.0157s] [ 98%] 2025-12-04T15:25:36.7354578Z functorch/test_vmap.py::TestVmapDeviceTypeCUDA::test_vmap_fallback_check XFAIL [0.0019s] [ 98%] 2025-12-04T15:25:36.7354985Z functorch/test_vmap.py::TestVmapDeviceTypeCUDA::test_vmap_fallback_check_ok PASSED [0.0019s] [ 98%] 2025-12-04T15:25:36.7355466Z functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_cat_batching_rule_cuda PASSED [0.0094s] [ 98%] 2025-12-04T15:25:36.7355905Z functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_fallback_binary_cuda PASSED [0.0044s] [ 99%] 2025-12-04T15:25:36.7356415Z functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_fallback_binary_nt_and_batched_dense_cuda PASSED [0.0042s] [ 99%] 2025-12-04T15:25:36.7356941Z functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_fallback_binary_nt_and_unbatched_dense_cuda PASSED [0.0041s] [ 99%] 2025-12-04T15:25:36.7357405Z functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_fallback_unary_cuda PASSED [0.0041s] [ 99%] 2025-12-04T15:25:36.7358022Z functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_fallback_with_nt_and_batched_dense_with_nonzero_bdim_raises_cuda PASSED [0.0023s] [ 99%] 2025-12-04T15:25:36.7358479Z functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_multilevel_vmap_raises_cuda PASSED [0.0029s] [ 99%] 2025-12-04T15:25:36.7358922Z functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_nt_acts_as_dense_in_vmap_cuda PASSED [0.0034s] [ 99%] 2025-12-04T15:25:36.7359403Z functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_nt_with_nonzero_in_dim_raises_cuda PASSED [0.0020s] [ 99%] 2025-12-04T15:25:36.7359873Z functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_nt_with_nonzero_out_dim_raises_cuda PASSED [0.0019s] [ 99%] 2025-12-04T15:25:36.7360252Z functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_shape_call_cuda XFAIL [0.0015s] [ 99%] 2025-12-04T15:25:36.7360666Z functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_vmap_fallback_check XFAIL [0.0018s] [ 99%] 2025-12-04T15:25:36.7361082Z functorch/test_vmap.py::TestVmapNestedTensorCUDA::test_vmap_fallback_check_ok PASSED [0.0025s] [100%] 2025-12-04T15:25:36.7361091Z 2025-12-04T15:25:36.7361788Z - generated xml file: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-94056324a12f0026.xml - 2025-12-04T15:25:36.7362050Z === 980 passed, 50 skipped, 1007 deselected, 97 xfailed in 212.41s (0:03:32) === 2025-12-04T15:25:36.7362676Z The following tests failed consistently: ['test/functorch/test_vmap.py::TestVmapOperatorsOpInfoCUDA::test_torch_return_types_returns_cuda'] 2025-12-04T15:25:36.7362696Z 2025-12-04T15:25:36.7363170Z FINISHED PRINTING LOG FILE of functorch/test_vmap 1/1 (test/test-reports/functorch.test_vmap_1.1_151d84486b0c8e0c_.log) 2025-12-04T15:25:36.7363176Z 2025-12-04T15:25:36.7363493Z Finished functorch/test_vmap 1/1 ... [2025-12-04 15:25:36.242317][22319.932705897], took 8.29min 2025-12-04T15:25:36.7364236Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-2355f4f75c6e0706.xml 2025-12-04T15:25:36.7364961Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-db01917ff3ce307b.xml 2025-12-04T15:25:36.7365692Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-12ecde9ddbb651c9.xml 2025-12-04T15:25:36.7366447Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-94056324a12f0026.xml 2025-12-04T15:25:37.9802565Z Uploading logs for 57119749282 to S3 2025-12-04T15:25:38.5149761Z Uploading artifacts took 1.95 seconds 2025-12-04T15:25:38.5150165Z functorch/test_vmap 1/1 failed! 2025-12-04T15:25:38.5154318Z Running dynamo/test_after_aot 1/1 ... [2025-12-04 15:25:38.515260][22322.20565049] 2025-12-04T15:25:38.5154882Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T15:25:38.5159312Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_after_aot.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 15:25:38.515706] 2025-12-04T15:25:55.3046349Z 2025-12-04T15:25:55.3048202Z dynamo/test_after_aot 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_after_aot_1.1_c57cf0a68fd0c5b4_.log 2025-12-04T15:25:55.3051167Z Running 2 items in this shard: test/dynamo/test_after_aot.py::TestAfterAot::test_dump_tensor, test/dynamo/test_after_aot.py::TestAfterAot::test_save_graph_repro 2025-12-04T15:25:55.3052739Z 2025-12-04T15:25:55.3053380Z Finished dynamo/test_after_aot 1/1 ... [2025-12-04 15:25:55.304425][22338.994819105], took 0.28min 2025-12-04T15:25:55.3448920Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_after_aot/dynamo.test_after_aot-9dfa8d2ed594e793.xml 2025-12-04T15:25:55.4258862Z Running inductor/test_snode_runtime 1/1 ... [2025-12-04 15:25:55.425595][22339.115987531] 2025-12-04T15:25:55.4259483Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T15:25:55.4262765Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_snode_runtime.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 15:25:55.426002] 2025-12-04T15:26:22.2303475Z 2025-12-04T15:26:22.2304679Z inductor/test_snode_runtime 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_snode_runtime_1.1_55d09bf84d516763_.log 2025-12-04T15:26:22.2314845Z Running 22 items in this shard: test/inductor/test_snode_runtime.py::UnsupportedTests::test_no_cuda, test/inductor/test_snode_runtime.py::UnsupportedTests::test_no_op, test/inductor/test_snode_runtime.py::ComputeBoundedTests::test_addmm, test/inductor/test_snode_runtime.py::ComputeBoundedTests::test_bmm, test/inductor/test_snode_runtime.py::ComputeBoundedTests::test_conv1d, test/inductor/test_snode_runtime.py::ComputeBoundedTests::test_conv2d, test/inductor/test_snode_runtime.py::ComputeBoundedTests::test_conv2d_transpose, test/inductor/test_snode_runtime.py::ComputeBoundedTests::test_conv3d, test/inductor/test_snode_runtime.py::ComputeBoundedTests::test_mm, test/inductor/test_snode_runtime.py::MemoryBoundedTests::test_dynamic, test/inductor/test_snode_runtime.py::MemoryBoundedTests::test_horizontal_reduction_pointwise, test/inductor/test_snode_runtime.py::MemoryBoundedTests::test_pointwise, test/inductor/test_snode_runtime.py::MemoryBoundedTests::test_relu, test/inductor/test_snode_runtime.py::TestCommAnalysis::test_all_gather_into_tensor, test/inductor/test_snode_runtime.py::TestCommAnalysis::test_all_gather_into_tensor_coalesced, test/inductor/test_snode_runtime.py::TestCommAnalysis::test_all_reduce, test/inductor/test_snode_runtime.py::TestCommAnalysis::test_all_reduce_coalesced, test/inductor/test_snode_runtime.py::TestCommAnalysis::test_legacy_all_gather_into_tensor_coalesced, test/inductor/test_snode_runtime.py::TestCommAnalysis::test_legacy_all_reduce, test/inductor/test_snode_runtime.py::TestCommAnalysis::test_legacy_all_reduce_coalesced, test/inductor/test_snode_runtime.py::TestCommAnalysis::test_reduce_scatter_tensor, test/inductor/test_snode_runtime.py::TestCommAnalysis::test_reduce_scatter_tensor_coalesced 2025-12-04T15:26:22.2323717Z 2025-12-04T15:26:22.2324086Z Finished inductor/test_snode_runtime 1/1 ... [2025-12-04 15:26:22.230152][22365.920544783], took 0.45min 2025-12-04T15:26:22.2700712Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_snode_runtime/inductor.test_snode_runtime-804cfaf2d58e1050.xml 2025-12-04T15:26:22.3535227Z Running inductor/test_minifier 1/1 ... [2025-12-04 15:26:22.353230][22366.04362026] 2025-12-04T15:26:22.3535839Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T15:26:22.3539273Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_minifier.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 15:26:22.353668] 2025-12-04T15:28:07.6304290Z 2025-12-04T15:28:07.6305420Z inductor/test_minifier 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_minifier_1.1_9cc263edd080c3f5_.log 2025-12-04T15:28:07.6311904Z Running 14 items in this shard: test/inductor/test_minifier.py::MinifierTests::test_accuracy_vs_strict_accuracy, test/inductor/test_minifier.py::MinifierTests::test_after_aot_cpu_accuracy_error, test/inductor/test_minifier.py::MinifierTests::test_after_aot_cpu_compile_error, test/inductor/test_minifier.py::MinifierTests::test_after_aot_gpu_accuracy_error, test/inductor/test_minifier.py::MinifierTests::test_after_aot_gpu_compile_error, test/inductor/test_minifier.py::MinifierTests::test_aoti_cpu_accuracy_error, test/inductor/test_minifier.py::MinifierTests::test_aoti_cpu_compile_error, test/inductor/test_minifier.py::MinifierTests::test_aoti_cpu_compile_error_unflatten, test/inductor/test_minifier.py::MinifierTests::test_aoti_gpu_accuracy_error, test/inductor/test_minifier.py::MinifierTests::test_aoti_gpu_compile_error, test/inductor/test_minifier.py::MinifierTests::test_aoti_gpu_compile_error_unflatten, test/inductor/test_minifier.py::MinifierTests::test_constant_in_graph, test/inductor/test_minifier.py::MinifierTests::test_offload_to_disk, test/inductor/test_minifier.py::MinifierTests::test_rmse_improves_over_atol 2025-12-04T15:28:07.6317771Z 2025-12-04T15:28:07.6318115Z Finished inductor/test_minifier 1/1 ... [2025-12-04 15:28:07.630581][22471.320972423], took 1.75min 2025-12-04T15:28:07.6711942Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_minifier/inductor.test_minifier-e48754fb8b6df415.xml 2025-12-04T15:28:07.7877311Z Running inductor/test_compiled_autograd 1/1 ... [2025-12-04 15:28:07.787360][22471.47775161] 2025-12-04T15:28:07.7877958Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T15:28:07.7881219Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_compiled_autograd.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 15:28:07.787822] 2025-12-04T15:47:47.1962414Z 2025-12-04T15:47:47.1965527Z inductor/test_compiled_autograd 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_compiled_autograd_1.1_bff035ff4424e9fc_.log 2025-12-04T15:47:47.2480470Z Running 861 items in this shard: test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_accumulate_grad_accuracy, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_accumulate_grad_polyfill_case_1_1, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_accumulate_grad_polyfill_case_1_2, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_accumulate_grad_polyfill_case_1_3, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_accumulate_grad_polyfill_case_1_5_1, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_accumulate_grad_polyfill_case_1_5_2, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_accumulate_grad_polyfill_case_2_1, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_accumulate_grad_polyfill_case_2_3_1, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_accumulate_grad_polyfill_case_2_3_2, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_accumulate_grad_polyfill_case_2_3_3, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_accumulate_grad_polyfill_case_3_1, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_accumulate_grad_polyfill_case_3_2, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_accumulate_without_zero, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_anomaly_mode_already_nan, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_anomaly_mode_backward, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_anomaly_mode_grad, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_aot_bwd_gm_runnable, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_autograd_cpp_node_basic_is_traceable_False, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_autograd_cpp_node_basic_is_traceable_True, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_autograd_cpp_node_data_dependent_is_traceable_False, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_autograd_cpp_node_data_dependent_is_traceable_True, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_autograd_cpp_node_id_is_traceable_False, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_autograd_cpp_node_id_is_traceable_True, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_autograd_cpp_node_non_traceable, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_autograd_cpp_node_saved_basic_is_traceable_False, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_autograd_cpp_node_saved_basic_is_traceable_True, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_autograd_cpp_node_saved_dynamic_is_traceable_False, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_autograd_cpp_node_saved_dynamic_is_traceable_True, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_autograd_cpp_node_saved_float_is_traceable_False, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_autograd_cpp_node_saved_float_is_traceable_True, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_autograd_cpp_node_saved_int_is_traceable_False, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_autograd_cpp_node_saved_int_is_traceable_True, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_backward_hook_relative_ordering_partial, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_basic, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_cache_hit, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_callback_graph_break_throws_error, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_checkpointing_sac, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_checkpointing_simple_reentrant_False, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_checkpointing_simple_reentrant_True, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_compile_api_api_compile_backend_aot_eager, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_compile_api_api_compile_backend_eager, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_compile_api_api_compile_backend_inductor, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_compile_api_api_optimize_backend_aot_eager, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_compile_api_api_optimize_backend_eager, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_compile_api_api_optimize_backend_inductor, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_compile_api_disable_api_compile_backend_aot_eager, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_compile_api_disable_api_compile_backend_eager, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_compile_api_disable_api_compile_backend_inductor, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_compile_api_disable_api_optimize_backend_aot_eager, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_compile_api_disable_api_optimize_backend_eager, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_compile_api_disable_api_optimize_backend_inductor, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_compiled_autograd_does_not_specialize_on_bw_symints, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_cpu_offloading, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_cudagraphs_cpu_division, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_cudagraphs_cpu_graph, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_cudagraphs_cpu_scalar_used_in_cpp_custom_op, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_cudagraphs_cpu_scalar_used_in_python_custom_op, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_cudagraphs_sdpa, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_custom_fn_bw_graph_break, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_custom_fn_compiled_fw_bw_graph_break, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_custom_fn_compiled_fw_graph_break, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_custom_fn_dynamically_defined_class, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_custom_fn_multiple_grads, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_custom_fn_non_variable_input, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_custom_fn_output_metadata, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_custom_fn_saved_attr, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_custom_fn_saved_multiple_tensors, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_custom_fn_saved_multiple_tensors_dedup, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_custom_fn_saved_shape_tensor, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_custom_fn_saved_tensors, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_custom_fn_with_same_graph, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_ddp_cpp_reducer_error, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_ddp_python_reducer, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_disk_offloading, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_dont_dce_side_effects, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_dynamic_shapes, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_dynamic_shapes_annotations, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_dynamic_shapes_eager_node, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_dynamic_shapes_from_forward, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_dynamo_boxed, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_dynamo_flaky_segfault, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_flex_attention, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_free_activation_memory, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_free_activation_memory_subclass, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_graph_break_custom_op, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_higher_order_gradients, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_hipify_not_loaded_with_import_cpp_extension, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_hipify_not_loaded_with_import_torch, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_implicit_add, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_inplace_grad_update, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_inputs_aliasing_bytecode_attr_mutations, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_inputs_aliasing_bytecode_stack_restore, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_issue106555, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_keep_graph_simple, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_keep_graph_usage_after_compiled, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_logging_tensor_flaky, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_logs, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_logs_aot_bwd_reuse, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_mismatch_fake_tensor_mode, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_mismatch_fake_tensor_mode_dynamic_shape, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_multiple_torch_compile, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_nested_compile, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_nested_context_manager, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_no_nested_compiled_autograd, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_no_output_nodes_all_leaves, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_no_output_nodes_different_leaves_will_recompile, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_no_output_nodes_some_leaves, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_optimize_assert_backend_aot_eager, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_optimize_assert_backend_eager, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_optimize_assert_backend_inductor, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_output_nodes_all_leaves, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_output_nodes_some_leaves, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_reorder_acc_grad, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_reorder_all_bwd_hooks, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_reorder_multi_post_hooks, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_reorder_multi_pre_hooks, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_reorder_multi_tensor_pre_hooks, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_reorder_post_hook1, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_reorder_post_hook2, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_reorder_post_hook3, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_reset, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_saved_tensor_unpack_hook_ordering, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_tensor_grad_hook1, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_tensor_grad_hook2, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_tensor_grad_hook3, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_tensor_subclass_basic, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_torch_compile, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_torch_compile_api_dynamic_shapes, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_torch_compile_graph_break, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_torch_compile_graph_break2, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_torch_compile_only_backward_call, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_torch_dispatch_mode, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_torch_function_mode, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_trace_auto_functionalized, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_trace_auto_functionalized_v2, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_trace_run_with_rng_state, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_verbose_logs_aot_dispatcher_nodes, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_verbose_logs_aot_dispatcher_nodes_hop, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_verbose_logs_aot_id, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_verbose_logs_cpp, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_verbose_logs_dynamic_shapes, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_verbose_logs_graph, test/inductor/test_compiled_autograd.py::TestCompiledAutograd::test_verbose_logs_snapshot, test/inductor/test_compiled_autograd.py::WrapTestClassTests::test_wrap_preserves_inheritance_and_super, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_access_saved_tensor_twice_without_recomputation_works, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_accumulate_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_accumulate_grad_posthooks_can_observe_tensor_prehook, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_accumulate_grad_posthooks_should_not_execute, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_accumulate_grad_tensor_reference, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_accumulate_grad_with_zero_numel_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_anomaly_assign_parent_cleanup, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_anomaly_detect_nan, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_anomaly_grad_warnings, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_anomaly_mode_no_check_nan, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_attribute_deletion, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_autograd_inplace_view_of_view, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_autograd_inplace_views_creation_meta, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_autograd_inplace_views_cross_dtype, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_autograd_multiple_views_python, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_autograd_node_isinstance, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_autograd_print_tensor, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_autograd_python_custom_function_inplace, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_autograd_simple_views_python, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_autograd_views_codegen, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_backward, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_backward_badcalls, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_backward_copy, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_backward_create_graph_warns, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_backward_hook_relative_ordering, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_backward_no_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_backward_to_node, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_backward_twice_retained_graph_with_saved_values, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_backward_twice_retained_graph_without_saved_values, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_backward_twice_with_saved_values, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_backward_twice_without_saved_values, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_backward_with_inputs, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_backward_with_nonleaf_inputs, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_backward_with_scalar_input, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_calculate_shape_util, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_callback_adds_callback, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_callback_propagates_errors_from_device_thread, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_cant_create_saved_tensors, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpoint_detects_non_determinism, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpoint_graph_execution_group, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpoint_sequential_warns_if_use_reentrant_not_passed_explcitly, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpoint_valid_reset_on_error, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpoint_warns_if_use_reentrant_not_passed_explcitly, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_non_reentrant_autocast_cpu, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_non_reentrant_autocast_gpu, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_arbitrary_input_output, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_correct_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_custom_function_works, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_dataparallel, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_detached_tensor_use_reentrant_False, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_detached_tensor_use_reentrant_True, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_input_requires_grad_False, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_input_requires_grad_True, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_memory_savings, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_parameter_used_in_an_out, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_saved_object_identity, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_checkpointing_without_reentrant_with_context_fn, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_copy_slices_graph_task_updates, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_create_graph_and_full_backward_hook_cycle, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_current_graph_task_execution_order, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_current_graph_task_id, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_current_node, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_autograd_ac_early_stop, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_autograd_no_early_free, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_autograd_repeated_grad_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_cycle, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_error, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_exception, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_forward_mode_forward_is_no_op, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_forward_mode_inplace_checks, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_forward_mode_non_differentiable, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_forward_mode_non_tensor_before_tensor_args, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_forward_mode_view_checks, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_forward_mode_wrong_formula, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_inplace_on_non_default_view, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_inplace_on_view_of_leaf, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_local_inplace, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_mark_dirty_not_differentiable, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_mark_output_view_of_intermediate, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_no_tensors, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_non_tensor_inputs_outputs, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_preserve_torch_function_when_return_as_is, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_return_view_in_nograd, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_save_for_forward, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_saved_tensors, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_saving_mutated_view_no_leak, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_setup_context_multi_input, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_setup_context_multi_output, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_setup_context_simple, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_custom_function_vmap_defaults, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_deep_reentrant, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_default_saved_tensors_hooks_double_backward, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_dep_nograd, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_dependent_backward, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_detach, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_detach_base, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_detach_then_inplace_raises_in_autograd, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_diagonal_expanded_v, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_dir, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_disabling_saved_tensor_hooks, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_disabling_saved_tensor_hooks_nested, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_dont_materialize_grads, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_duplicate_backward_root, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_enable_grad_decorator_no_paren, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_first_grad_fn_access_in_no_grad_mode, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_free_deep_graph, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_free_deep_graph_complicated, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_free_deep_graph_pyfunction, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_full_backward_hook_double_backward, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_function, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_function_returns_input, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_function_returns_undefined_tensor, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gc_in_destructor, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_get_data_and_hooks_from_raw_saved_variable, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_badcalls, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_batched_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_dtype, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_empty_inputs, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_fn_attr_bindings, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_fn_badcalls, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_fn_input_metadata, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_fn_prehooks, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_fn_prehooks_multiple_outputs, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_fn_prehooks_remove_hooks, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_materialize_grads, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_mode_class_decoration, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_mode_restored_reentrant, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_nonleaf, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_nonleaf_many_outputs, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_nonleaf_register_hook, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_thread_safety, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_to_node, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_to_node_inplace, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_to_node_materialize, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_to_node_multi, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_to_node_set, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_unreachable, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_grad_unreachable_discovery, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_backward_mul_by_grad_output, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_check_batched_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_check_forward_or_backward_only, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_check_no_differentiable_outputs, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_complex_non_complex_outputs, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_custom_error, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_default_device_placement_context, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_dense_and_sparse_inputs, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_forward_ad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_forward_ad_batched_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_forward_ad_respects_requires_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_forward_ad_runs_with_no_requires_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_get_analytical_jacobian, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_get_numerical_jacobian, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_input_layout0, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_input_layout1, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_input_layout2, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_input_layout3, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_input_layout4, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_jacobian_mismatch, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_multiple_mkldnn_inputs, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_nondeterministic, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_output_shape_or_dtype_depend_on_values, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_single_input, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_test_outputs, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_undefined_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_validates_input_mkldnn, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradcheck_validates_inputs, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradient_edge_graph_ownership, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_gradient_edge_output, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_graph_save_on_cpu, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_graph_save_on_cpu_cuda, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_hessian_vector, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_hook_closure_cycle_use_custom_function_False_use_tensor_hook_False, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_hook_closure_cycle_use_custom_function_False_use_tensor_hook_True, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_hook_closure_cycle_use_custom_function_True_use_tensor_hook_False, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_hook_closure_cycle_use_custom_function_True_use_tensor_hook_True, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_hook_edge_case_when_called_with_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_hook_none, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_hook_with_no_name, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_hooks, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_hooks_cpp, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_increment_version, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_index_backward_does_not_save_tensor, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_indexing, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_indexing_duplicates, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_inplace, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_inplace_not_requires_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_inplace_on_view_backward, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_inplace_on_view_leaf_errors, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_inplace_on_view_saved_output, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_inplace_on_view_weak_grad_fn, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_input_buffer_accum, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_integer_outputs, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_invalid_gradients, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_isolated_node, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_leaf_assignment, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_legacy_function_deprecation_exception, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_lobpcg, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_mark_non_differentiable, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_mark_non_differentiable_mixed, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_mark_non_differentiable_none, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_materialize_grads, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_multi_backward, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_multi_backward_no_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_multi_grad_all_hooks, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_multi_grad_any_hooks, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_multi_grad_hooks_invalid_mode, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_multiple_insert_removal_caching, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_named_tensor_for_complex_views, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_naughty_anomaly_access, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_naughty_autograd_function_attribute_access, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_naughty_autograd_function_stashing_ctx, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_nested_anomaly_detect_nan, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_nested_anomaly_printstack_cleanup, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_next_functions, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_no_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_no_grad_assignment, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_no_grad_copy, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_no_grad_copy_sparse, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_no_grad_input, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_no_grad_modifies_version, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_no_grad_python_function, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_no_requires_grad_inplace, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_no_unnecessary_save, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_no_unnecessary_unwrapping, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_node_ordering_when_none_returned, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_node_post_hook_registered_during_unpack_hook, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_not_implemented_fwad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_not_implemented_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_numpy_requires_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_once_differentiable, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_out_variant_raises_when_inputs_require_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_pack_hook_with_inplace_modification_should_fail, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_pickle, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_post_accumulate_grad_hook_e2e, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_post_accumulate_grad_hook_gets_cleaned_up, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_post_accumulate_grad_hook_multiple_hooks, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_post_accumulate_grad_hook_multiple_tensors, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_post_accumulate_grad_hook_on_non_leaf, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_post_accumulate_grad_hook_ordering, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_post_accumulate_grad_hook_returns_not_None, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_pow_zero_tensor_gradient, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_power_function, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_prehook_ordering, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_profiler, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_profiler_aggregation_fake, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_profiler_aggregation_lstm, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_profiler_aggregation_table, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_profiler_function_event_avg, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_profiler_propagation, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_profiler_seq_nr, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_profiler_shapes, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_profiler_unboxed_only, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_pynode_destruction_deadlock, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_record_function, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_record_function_callbacks, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_record_function_legacy, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_record_function_multithreaded, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_reentrant_child_error, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_reentrant_priority, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_reentrant_with_callbacks_both_depths, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_reentrant_with_callbacks_depth_0, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_reentrant_with_callbacks_depth_1, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_reentrant_with_leaf_variable_hook, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_reentrant_with_non_leaf_variable_hook, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_requires_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_requires_grad_, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_requires_grad_inplace, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_retain_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_retain_grad_cycle, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_retain_grad_inplace, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_retain_grad_inplace_over_view, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_retains_grad_can_always_observe_tensor_prehook, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_retains_grad_inplace_multiple_outputs, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_return_duplicate, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_return_duplicate_inplace, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_return_leaf, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_return_leaf_inplace, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_save_none_for_backward, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_save_on_cpu_and_checkpoint, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_save_output_nr, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_saved_tensor_hooks_custom_error_propagation, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_saved_tensor_hooks_custom_function_intermediates, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_saved_tensor_hooks_extra_enter_during_bw_no_leak, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_saved_tensor_hooks_extra_exit_during_bw_no_crash, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_saved_tensors_hook_version_counter_not_shared, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_saved_variable_packing_unpacking_did_not_save_original_with_default_hooks, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_saved_variable_packing_unpacking_did_not_save_original_with_hooks, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_saved_variable_packing_unpacking_saved_original_with_default_hooks, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_saved_variable_packing_unpacking_saved_original_with_hooks, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_saved_variable_saved_original_inplace_detach, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_saved_variable_version_counter, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_saved_variables_deprecated, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_saving_variable_to_disk, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_scalar_grad_mixed_device, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_select_expanded_v, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_select_sum, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_set_data_preserve_pyobj, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_set_data_self_requires_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_set_data_tensorimpl_type, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_set_grad_coroutines, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_set_grad_coroutines_benign_exceptions, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_set_grad_coroutines_critical_exceptions, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_set_grad_coroutines_exit, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_set_grad_enabled, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_set_grad_enabled_wraps, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_set_grad_generator_functions, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_set_grad_generator_functions_recursive, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_set_materialize_non_diff_grads, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_setitem, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_setitem_mask, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_setting_default_saved_variable_hooks_twice_should_not_fail, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_setting_default_saved_variable_hooks_twice_should_use_inner, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_setup_context_when_forward_has_default_args, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_shape, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_sharded_grad, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_simple_reentrant, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_slice_expanded_v, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_sparse_gather_both_scalar, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_sparse_gather_dim0, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_sparse_gather_dim1, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_sparse_gather_dim_neg, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_sparse_gather_ind_scalar, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_sparse_gather_x_scalar, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_sparse_mm_backward, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_tensor_grad_warnings, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_tensor_hooks_inplace, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_tensor_hooks_inplace_multiple_outputs, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_tensor_hooks_inplace_over_view, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_thread_shutdown, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_to_sparse_backward, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_too_many_grads, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_type_conversions, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_unpack_hooks_exec_count, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_unrelated_inputs, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_unsafe_set_version_counter, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_unused_grad_requires_grad_with_materialize, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_unused_output, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_var_mean_differentiable, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_variable_traverse, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_version_counter, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_view_func_replay, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_view_func_replay_with_modified_state, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_view_replay_enabled, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_volatile_deprecated, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_will_engine_execute_node, test/inductor/test_compiled_autograd.py::TestAutogradWithCompiledAutograd::test_wrapped_number_saved_tensors_hooks, test/inductor/test_compiled_autograd.py::TestNestedCheckpointWithCompiledAutograd::test_nested_checkpoint_early_stop_False, test/inductor/test_compiled_autograd.py::TestNestedCheckpointWithCompiledAutograd::test_nested_checkpoint_early_stop_True, test/inductor/test_compiled_autograd.py::TestNestedCheckpointWithCompiledAutograd::test_nested_checkpoint_kwargs_early_stop_False, test/inductor/test_compiled_autograd.py::TestNestedCheckpointWithCompiledAutograd::test_nested_checkpoint_kwargs_early_stop_True, test/inductor/test_compiled_autograd.py::TestNestedCheckpointWithCompiledAutograd::test_nested_checkpoint_non_tensor_inputs_and_outputs_early_stop_False, test/inductor/test_compiled_autograd.py::TestNestedCheckpointWithCompiledAutograd::test_nested_checkpoint_non_tensor_inputs_and_outputs_early_stop_True, test/inductor/test_compiled_autograd.py::TestNestedCheckpointWithCompiledAutograd::test_nested_checkpoint_reentrant_backwards_early_stop_False, test/inductor/test_compiled_autograd.py::TestNestedCheckpointWithCompiledAutograd::test_nested_checkpoint_reentrant_backwards_early_stop_True, test/inductor/test_compiled_autograd.py::TestNestedCheckpointWithCompiledAutograd::test_nested_checkpoint_same_graph_early_stop_False, test/inductor/test_compiled_autograd.py::TestNestedCheckpointWithCompiledAutograd::test_nested_checkpoint_same_graph_early_stop_True, test/inductor/test_compiled_autograd.py::TestNestedCheckpointWithCompiledAutograd::test_nested_checkpoint_set_early_stop, test/inductor/test_compiled_autograd.py::TestNestedCheckpointWithCompiledAutograd::test_nested_checkpoint_set_early_stop_no_recompution_needed, test/inductor/test_compiled_autograd.py::TestNestedCheckpointWithCompiledAutograd::test_nested_checkpoint_two_children_early_stop_False, test/inductor/test_compiled_autograd.py::TestNestedCheckpointWithCompiledAutograd::test_nested_checkpoint_two_children_early_stop_True, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_abstract_impl_on_existing_op, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_abstract_impl_on_existing_op_with_CompositeExplicitAutograd, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_abstract_impl_on_existing_op_with_CompositeImplicitAutograd, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_abstract_impl_on_existing_op_with_meta, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_autogen_aten_ops_are_pt2_compliant, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_autograd_function_backed_op, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_autograd_notimplemented, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_autograd_notimplemented_gradmode, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_dict_grad_for_nontensor, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_dict_invalid_keys, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_dict_requires_keys_for_input_optional_tensors, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_dict_requires_keys_for_input_tensors, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_grads_are_tensor_or_none, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_impl_on_existing_op, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_impl_on_existing_op_CompositeImplicitAutograd, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_impl_on_existing_op_incorrect_schema_mutable, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_impl_on_existing_op_incorrect_schema_no_output, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_impl_on_existing_op_incorrect_schema_views, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_impl_on_existing_op_with_key_key_Autograd, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_impl_on_existing_op_with_key_key_AutogradCPU, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_impl_on_existing_op_with_key_key_AutogradCUDA, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_output_differentiability_non_tensor, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_output_differentiability_numel, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_output_differentiability_tensorlist, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_output_differentiability_type, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_partially_registered, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_returns_dict, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_tensorlist_input_requires_list_grads, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_tensorlist_input_requires_list_grads_none_or_Tensor, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_backward_tensorlist_input_requires_list_grads_with_same_numel, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_basic_make_fx, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_builtin_aten_ops_are_pt2_compliant, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_builtin_torchscript_ops, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_data_dependent_basic, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_data_dependent_compile, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_data_dependent_fake_tracing, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_data_dependent_nms_dynamic_compile, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_define_and_impl, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_define_bad_schema, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_define_validation, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_define_with_tags_list, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_define_with_tags_single, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_define_with_tags_tuple, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_defined_in_python, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_duplicate_impl, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_functionalize_error, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_impl_abstract_overload, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_impl_cpu, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_impl_device_cpu, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_impl_device_cuda, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_impl_device_function, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_impl_device_invalid, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_impl_function, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_impl_invalid_devices, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_impl_meta, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_impl_multiple, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_impl_on_existing_op, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_impl_on_existing_op_with_cpu_registration_key_CPU, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_impl_on_existing_op_with_cpu_registration_key_CUDA, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_impl_on_existing_op_with_cpu_registration_key_CompositeExplicitAutograd, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_impl_on_existing_op_with_cpu_registration_key_CompositeImplicitAutograd, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_impl_separate, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_incorrect_schema_types, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_infer_schema_no_return, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_infer_schema_supported, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_infer_schema_unsupported, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_invalid_qualname, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_invalid_schemas, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_is_functional_schema, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_is_tensorlist_like_type, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_legacy_define, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_legacy_impl, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_lifetime, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_load_library, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_meta_for_data_dependent_shape_operation, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_name_must_match, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_new_data_dependent_symint, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_not_implemented_error, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_override_cea, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_override_fake, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_override_impl, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_override_meta, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_private_ctor, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_reserved_ns, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_resolve_packet, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_save_for_backward_inputs_are_namedtuple, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_schema_matches_signature, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_sequences, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_supported_param_types, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_supported_return_types_multi_return, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_supported_return_types_single_return, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_supported_schemas, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_symints, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_unsupported_param_types, test/inductor/test_compiled_autograd.py::TestCustomOpWithCompiledAutograd::test_unsupported_schemas, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_access_module_attr, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_allow_python_side_effects_utility, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_capture_constants, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_capture_global_num, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_capture_global_num_adds_guard, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_capture_input_num, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_capture_numpy_number, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_capture_tracked, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_capture_tracked_nested, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_capture_untracked_global, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_capture_untracked_global_nested, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_capture_untracked_nonlocal, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_capture_value_created_in_subgraph, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_concat_unbacked_shape_tensor, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_cond_branches_no_arguments, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_cond_branches_no_arguments_no_closure, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_cond_free_variable_in_both_branches, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_cond_graph_break_in_one_branch, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_cond_pytree_operands, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_cond_pytree_operands_with_non_tensor_leaves, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_cond_side_effect_in_one_branches, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_cond_source_fn_stack, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_cond_subgraph_name_is_valid, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_cond_with_constant_pred, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_cond_with_empty_operands, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_dynamic_shapes_over_vmap_batch_size, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_enum_arg, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_error_message_sane, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_fallback_on_graph_break_complicated, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_fallback_on_graph_break_simple, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_flat_list_output, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_fn_with_kwargs_in_torch_ops, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_freevars_as_inputs_to_wrap, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_grad_source_fn_stack, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_hints_wrapper, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_hints_wrapper_incorrect_type, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_hints_wrapper_no_hints, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_hints_wrapper_pytree_inputs, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_hooks, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_hopify_generic_wrap, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_inlined_functions, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_internal_nonlocal, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_lift_tensor_constant, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_lift_tensors_with_compound_expressions, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_lift_tensors_with_shared_symbols, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_make_closure, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_map_example_value_metadata_consistent_with_eager, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_map_graph_break, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_map_kwargs, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_map_lowers_to_graph, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_map_multi_return, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_map_pytree_return, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_map_side_effect, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_map_source_fn_stack, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_map_subgraph_name_is_valid, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_map_symint_input, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_modules, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_nested_tuple_output, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_nested_wrap, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_no_freevars, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_output_with_dict, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_register_mode, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_register_subclass, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_return_captured_var, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_return_captured_var_used_multiple_times, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_return_captured_vars, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_same_freevar_twice, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_del_existing_attr_global_module, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_del_existing_attr_global_obj, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_del_existing_attr_nonlocal_module, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_del_existing_attr_nonlocal_obj, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_in_body, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_local_list_append_no_graph_break, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_mutate_global_list, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_mutate_global_num, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_mutate_global_num_builtin, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_mutate_global_tensor, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_mutate_global_tensor_builtin, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_mutate_nonlocal_num, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_mutate_nonlocal_num_builtin, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_mutate_nonlocal_tensor, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_mutate_nonlocal_tensor_builtin, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_nested_nonlocal_list_append_graph_break, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_nonlocal_list_append_graph_break, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_set_existing_attr_global_module, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_set_existing_attr_global_obj, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_set_existing_attr_nonlocal_module, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_set_existing_attr_nonlocal_obj, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_set_new_attr_global_module, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_set_new_attr_global_obj, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_set_new_attr_nonlocal_module, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_side_effect_set_new_attr_nonlocal_obj, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_support_float_in_output, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_symint_in_slice, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_symint_input, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_tensor_and_unbacked_symbol_closure, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_tensor_to_list_closure, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_tensor_with_unbacked_shape_closure, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_unbacked_symbol_closure, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_vmap_multiply_scalar, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_vmap_source_fn_stack, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_wrap_all_kwarg, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_wrap_allow_local_assign_in_body_fn, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_wrap_inductor_compiled_regions_option, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_wrap_inductor_compiled_regions_with_backward, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_wrap_kwarg, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_wrap_kwarg_default, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_wrap_kwarg_default_else_branch, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_wrap_kwarg_default_if_branch, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_wrap_kwarg_int, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_wrap_kwarg_only, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_wrap_kwarg_recompile, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_wrap_pytree_args_nested, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_wrap_pytree_args_not_const_symint_tensor, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_wrap_pytree_args_with_symint_constant, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_wrap_pytree_kwargs, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_wrap_source_fn_stack, test/inductor/test_compiled_autograd.py::HigherOrderOpTestsWithCompiledAutograd::test_wrap_subgraph_name_is_valid, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_functional_call, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_functional_call_disable_inline_nn_module, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_functional_call_sequential_params_and_buffers, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_grad, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_grad_call_compiled_backward_fn, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_grad_call_torch_compile_fn, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_grad_capture_tensor, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_grad_closure_scalar, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_grad_fn_with_kwargs, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_grad_freevar_python_scalar, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_grad_freevar_tensor, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_grad_has_aux, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_grad_non_tensor_input, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_grad_over_grad, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_grad_pytree, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_grad_recompile, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_grad_two_tensor_all_grad_has_aux, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_grad_two_tensor_has_aux, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_grad_with_graph_break, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_grad_with_side_effect, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_hessian, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_hessian_argnums, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_jacfwd, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_jacfwd_has_aux, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_jacfwd_randomness, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_jacfwd_two_tensors_argnums, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_jacrev, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_jacrev_has_aux, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_jacrev_two_tensors_argnums, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_jvp_call_torch_compile_fn, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_jvp_freevar_python_scalar, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_jvp_freevar_tensor, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_jvp_has_aux, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_jvp_jvp, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_jvp_simple, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_jvp_two_tensors_disable_enable_disable_grad, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_jvp_two_tensors_disable_grad, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_jvp_two_tensors_has_aux, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_linearize_jvp_fn, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_teardown_resets_nested_graph_breaks, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vjp, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vjp_call_compiled_backward_fn, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vjp_has_aux, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vjp_multiple_outputs, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vjp_multiple_outputs_python_struct, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_call_compiled_backward_fn, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_call_torch_compile_fn, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_free_const, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_free_tensor, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_get_wrapped, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_kwargs, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_multiple_invocation_in_dims, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_multiple_invocation_out_dims, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_multiple_outputs, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_multiple_outputs_diff_dims, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_multiple_outputs_out_dims_tuple, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_new_tensor_implicit_via_op, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_new_tensor_in_body, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_new_tensor_unused_in_body, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_out_dims_None, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_over_vmap_captured, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_over_vmap_two_inputs, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_previous_illegal_op_no_graph_break, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_pytree_inputs, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_recompile, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_recompile_different_config, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_recompile_same_config, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_recompile_with_randomness, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_side_effects, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_side_effects_append_input, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_two_inputs, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_two_inputs_tuple_in_dims, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_with_conditional_graph_break, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_with_graph_break, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_with_graph_break_2, test/inductor/test_compiled_autograd.py::FuncTorchHigherOrderOpTestsWithCompiledAutograd::test_vmap_with_graph_break_lambda, test/inductor/test_compiled_autograd.py::ActivationCheckpointingTestsWithCompiledAutograd::test_cond_with_invalid_kwargs, test/inductor/test_compiled_autograd.py::ActivationCheckpointingTestsWithCompiledAutograd::test_cond_with_kwargs, test/inductor/test_compiled_autograd.py::ActivationCheckpointingTestsWithCompiledAutograd::test_cond_with_mismatched_output, test/inductor/test_compiled_autograd.py::ActivationCheckpointingTestsWithCompiledAutograd::test_dropout, test/inductor/test_compiled_autograd.py::ActivationCheckpointingTestsWithCompiledAutograd::test_dropout_inductor, test/inductor/test_compiled_autograd.py::ActivationCheckpointingTestsWithCompiledAutograd::test_fallback, test/inductor/test_compiled_autograd.py::ActivationCheckpointingTestsWithCompiledAutograd::test_flop_counter_for_cond, test/inductor/test_compiled_autograd.py::ActivationCheckpointingTestsWithCompiledAutograd::test_flop_counter_for_cond_unbalanced_branches, test/inductor/test_compiled_autograd.py::ActivationCheckpointingTestsWithCompiledAutograd::test_flop_counter_for_nested_cond, test/inductor/test_compiled_autograd.py::ActivationCheckpointingTestsWithCompiledAutograd::test_function, test/inductor/test_compiled_autograd.py::ActivationCheckpointingTestsWithCompiledAutograd::test_function_with_kwargs, test/inductor/test_compiled_autograd.py::ActivationCheckpointingTestsWithCompiledAutograd::test_module, test/inductor/test_compiled_autograd.py::ActivationCheckpointingTestsWithCompiledAutograd::test_non_aliasing_util, test/inductor/test_compiled_autograd.py::ActivationCheckpointingTestsWithCompiledAutograd::test_override_fallthrough_dispatch_key, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_device_mesh_compile, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dtensor_attribute_access_on_intermediate, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dtensor_basic, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dtensor_basic_export, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dtensor_constructor_w_dynamo_disable, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dtensor_constructor_w_graph_break, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dtensor_contiguous_dtensor_noncontiguous_local_as_tangent, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dtensor_different_gradient_placement, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dtensor_dont_recompile_on_same_placement_devicemesh, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dtensor_dynamic, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dtensor_dynamic_cat, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dtensor_dynamic_loss_parallel_log_softmax, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dtensor_dynamic_recompiles, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dtensor_dynamic_slice, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dtensor_dynamo_device_mesh_attrs, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dtensor_noncontiguous_output, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dtensor_partial_placement_graph_output, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dtensor_partial_placement_redistribute_unbalanced_correct_strides, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dtensor_requires_grad_recompile, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dynamo_dtensor, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dynamo_dtensor_from_local, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dynamo_dtensor_from_local_dynamic_shapes, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dynamo_dtensor_from_local_redistribute, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dynamo_dtensor_from_local_redistribute_async, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dynamo_dtensor_recompile, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dynamo_from_local_grad_placements_sequence_intermediate, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dynamo_from_local_grad_placements_sequence_intermediate_as_args, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dynamo_to_local_grad_placements_sequence, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dynamo_to_local_grad_placements_sequence_intermediate, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dynamo_to_local_kwargs, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_dynamo_to_local_kwargs_forward_hook, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_fakify_dtensor, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_get_local_rank_compile, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_graph_input_is_async, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_placement_compile, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_tp_compile_comm_reordering, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_tp_compile_comm_reordering_graph_partition, test/inductor/test_compiled_autograd.py::TestDTensorCompileWithCompiledAutograd::test_unwrap_async_collective_tensor_tangent, test/inductor/test_compiled_autograd.py::TestCompiledAutogradOpInfoCUDA::test_hops_in_bwd_auto_functionalize_simple_cuda_float32, test/inductor/test_compiled_autograd.py::TestCompiledAutogradOpInfoCUDA::test_hops_in_bwd_cond_simple_cuda_float32, test/inductor/test_compiled_autograd.py::TestCompiledAutogradOpInfoCUDA::test_hops_in_bwd_flex_attention_backward_simple_cuda_float32, test/inductor/test_compiled_autograd.py::TestCompiledAutogradOpInfoCUDA::test_hops_in_bwd_flex_attention_simple_cuda_float32, test/inductor/test_compiled_autograd.py::TestCompiledAutogradOpInfoCUDA::test_hops_in_bwd_invoke_quant_packed_simple_cuda_float32, test/inductor/test_compiled_autograd.py::TestCompiledAutogradOpInfoCUDA::test_hops_in_bwd_invoke_quant_simple_cuda_float32, test/inductor/test_compiled_autograd.py::TestCompiledAutogradOpInfoCUDA::test_hops_in_bwd_invoke_subgraph_simple_cuda_float32, test/inductor/test_compiled_autograd.py::TestCompiledAutogradOpInfoCUDA::test_hops_in_bwd_map_nested_cuda_float32, test/inductor/test_compiled_autograd.py::TestCompiledAutogradOpInfoCUDA::test_hops_in_bwd_map_simple_cuda_float32, test/inductor/test_compiled_autograd.py::TestCompiledAutogradOpInfoCUDA::test_hops_in_bwd_map_triple_nested_cuda_float32, test/inductor/test_compiled_autograd.py::TestCompiledAutogradOpInfoCUDA::test_hops_in_bwd_scan_simple_cuda_float32, test/inductor/test_compiled_autograd.py::TestCompiledAutogradOpInfoCUDA::test_hops_in_bwd_while_loop_simple_cuda_float32, test/inductor/test_compiled_autograd.py::TestCompiledAutogradOpInfoCUDA::test_hops_in_bwd_while_loop_stack_output_simple_cuda_float32 2025-12-04T15:47:47.2983825Z 2025-12-04T15:47:47.2984269Z Finished inductor/test_compiled_autograd 1/1 ... [2025-12-04 15:47:47.197640][23650.888028785], took 19.66min 2025-12-04T15:47:47.2985767Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_compiled_autograd/inductor.test_compiled_autograd-d055bdf97c26d8bd.xml 2025-12-04T15:47:48.6967263Z Uploading artifacts took 1.34 seconds 2025-12-04T15:47:48.6971162Z Running inductor/test_layout_optim 1/1 ... [2025-12-04 15:47:48.696915][23652.387307195] 2025-12-04T15:47:48.6971769Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T15:47:48.6976604Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_layout_optim.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 15:47:48.697382] 2025-12-04T15:47:58.6974574Z 2025-12-04T15:47:58.6975982Z inductor/test_layout_optim 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_layout_optim_1.1_12d2e0f1c8d2c698_.log 2025-12-04T15:47:58.6977012Z Running 0 items in this shard: 2025-12-04T15:47:58.6977246Z 2025-12-04T15:47:58.6977613Z Finished inductor/test_layout_optim 1/1 ... [2025-12-04 15:47:58.697264][23662.387656472], took 0.17min 2025-12-04T15:47:58.7386166Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_layout_optim/inductor.test_layout_optim-50eed008ca1555c3.xml 2025-12-04T15:47:58.8063459Z Running dynamo/test_unspec 1/1 ... [2025-12-04 15:47:58.806014][23662.496403875] 2025-12-04T15:47:58.8064034Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T15:47:58.8067114Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_unspec.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 15:47:58.806467] 2025-12-04T15:48:26.2611367Z 2025-12-04T15:48:26.2612397Z dynamo/test_unspec 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_unspec_1.1_9e5ffa0a28ec4624_.log 2025-12-04T15:48:26.2631044Z Running 51 items in this shard: test/dynamo/test_unspec.py::UnspecTests::test_argmin_coerces_symint_to_intlist_spec, test/dynamo/test_unspec.py::UnspecTests::test_bool_tensor_ctor, test/dynamo/test_unspec.py::UnspecTests::test_builtin_getitem, test/dynamo/test_unspec.py::UnspecTests::test_builtin_max_min, test/dynamo/test_unspec.py::UnspecTests::test_compiled_random_calls_are_random, test/dynamo/test_unspec.py::UnspecTests::test_conv1d_symint_padding, test/dynamo/test_unspec.py::UnspecTests::test_data_dependent_evaluate_expr_graph_break, test/dynamo/test_unspec.py::UnspecTests::test_defaults, test/dynamo/test_unspec.py::UnspecTests::test_exponential, test/dynamo/test_unspec.py::UnspecTests::test_feed_random_values_into_graph_only, test/dynamo/test_unspec.py::UnspecTests::test_isinstance_symint, test/dynamo/test_unspec.py::UnspecTests::test_item_max, test/dynamo/test_unspec.py::UnspecTests::test_mark_01_dynamic, test/dynamo/test_unspec.py::UnspecTests::test_mark_static_inside, test/dynamo/test_unspec.py::UnspecTests::test_mark_unbacked, test/dynamo/test_unspec.py::UnspecTests::test_mark_unbacked_channels_last, test/dynamo/test_unspec.py::UnspecTests::test_mark_unbacked_hint_consistency, test/dynamo/test_unspec.py::UnspecTests::test_multiple_consecutive_random_calls_before_graph, test/dynamo/test_unspec.py::UnspecTests::test_no_recompilations, test/dynamo/test_unspec.py::UnspecTests::test_no_recompilations_with_efficient_attention, test/dynamo/test_unspec.py::UnspecTests::test_no_recompiles, test/dynamo/test_unspec.py::UnspecTests::test_no_recompiles_prod_backward, test/dynamo/test_unspec.py::UnspecTests::test_numpy_correctness, test/dynamo/test_unspec.py::UnspecTests::test_propagate_dynamic_dim, test/dynamo/test_unspec.py::UnspecTests::test_prune_torch_check, test/dynamo/test_unspec.py::UnspecTests::test_random_call_with_while_loop, test/dynamo/test_unspec.py::UnspecTests::test_random_object, test/dynamo/test_unspec.py::UnspecTests::test_random_object_methods, test/dynamo/test_unspec.py::UnspecTests::test_random_object_overridden_methods, test/dynamo/test_unspec.py::UnspecTests::test_random_values_with_graph_break, test/dynamo/test_unspec.py::UnspecTests::test_rshift_dynamic, test/dynamo/test_unspec.py::UnspecTests::test_shape_graph_break, test/dynamo/test_unspec.py::UnspecTests::test_specializing_numpy_float_in_control_flow, test/dynamo/test_unspec.py::UnspecTests::test_split_aot_autograd, test/dynamo/test_unspec.py::UnspecTests::test_sum_dimlist_spec, test/dynamo/test_unspec.py::UnspecTests::test_sym_int_conversion, test/dynamo/test_unspec.py::UnspecTests::test_symbol_guard_limit_before_specialize, test/dynamo/test_unspec.py::UnspecTests::test_symfloat_no_replacement, test/dynamo/test_unspec.py::UnspecTests::test_symfloat_to_tensor, test/dynamo/test_unspec.py::UnspecTests::test_tensorfiy_python_scalars_1, test/dynamo/test_unspec.py::UnspecTests::test_tensorfiy_python_scalars_2, test/dynamo/test_unspec.py::UnspecTests::test_tensorfiy_python_scalars_3, test/dynamo/test_unspec.py::UnspecTests::test_to_tensor, test/dynamo/test_unspec.py::UnspecTests::test_unspec_float_input, test/dynamo/test_unspec.py::UnspecTests::test_unspec_float_input_f64, test/dynamo/test_unspec.py::UnspecTests::test_unspec_float_output, test/dynamo/test_unspec.py::UnspecTests::test_unspec_float_precision, test/dynamo/test_unspec.py::UnspecTests::test_unspec_roundtrip_float_input, test/dynamo/test_unspec.py::UnspecTests::test_unspecialized_float_multiply_precision, test/dynamo/test_unspec.py::UnspecTests::test_use_and_specialize, test/dynamo/test_unspec.py::UnspecTestsDeviceCUDA::test_builtin_functions_on_device_cuda 2025-12-04T15:48:26.2648741Z 2025-12-04T15:48:26.2649073Z Finished dynamo/test_unspec 1/1 ... [2025-12-04 15:48:26.260959][23689.951348957], took 0.46min 2025-12-04T15:48:26.3016445Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_unspec/dynamo.test_unspec-4bfba7972407dae5.xml 2025-12-04T15:48:26.3816212Z Running inductor/test_mmdecomp 1/1 ... [2025-12-04 15:48:26.381274][23690.071664709] 2025-12-04T15:48:26.3816873Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T15:48:26.3819650Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_mmdecomp.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 15:48:26.381720] 2025-12-04T15:48:47.5267074Z 2025-12-04T15:48:47.5267861Z inductor/test_mmdecomp 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_mmdecomp_1.1_8ac94a9aa95bcda4_.log 2025-12-04T15:48:47.5279421Z Running 24 items in this shard: test/inductor/test_mmdecomp.py::TestDecompCUDA::test_batched_mm_float32_bs_10_cuda_float32, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_batched_mm_float32_bs_1_cuda_float32, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_batched_mm_float32_bs_2_cuda_float32, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_batched_mm_float32_bs_4_cuda_float32, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_bmm_batch2_last_dim_size_is_one_cuda, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_dynamic_shape_mm_bfloat16_cuda_bfloat16, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_dynamic_shape_mm_float32_cuda_float32, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_simple_mm_bfloat16_cuda_bfloat16, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_simple_mm_float32_cuda_float32, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_some_batched_bfloat16_bs_10_cuda_bfloat16, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_some_batched_bfloat16_bs_1_cuda_bfloat16, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_some_batched_bfloat16_bs_2_cuda_bfloat16, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_some_batched_bfloat16_bs_4_cuda_bfloat16, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_some_batched_float32_bs_10_cuda_float32, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_some_batched_float32_bs_1_cuda_float32, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_some_batched_float32_bs_2_cuda_float32, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_some_batched_float32_bs_4_cuda_float32, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_some_batched_int32_bs_10_cuda_int32, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_some_batched_int32_bs_1_cuda_int32, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_some_batched_int32_bs_2_cuda_int32, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_some_batched_int32_bs_4_cuda_int32, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_some_bfloat16_cuda_bfloat16, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_some_float32_cuda_float32, test/inductor/test_mmdecomp.py::TestDecompCUDA::test_some_int32_cuda_int32 2025-12-04T15:48:47.5290250Z 2025-12-04T15:48:47.5290611Z Finished inductor/test_mmdecomp 1/1 ... [2025-12-04 15:48:47.526486][23711.216879674], took 0.35min 2025-12-04T15:48:47.5676938Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mmdecomp/inductor.test_mmdecomp-6ec8d10db2d708da.xml 2025-12-04T15:48:47.6558758Z Running dynamo/test_ctx_manager 1/1 ... [2025-12-04 15:48:47.655521][23711.345911905] 2025-12-04T15:48:47.6559350Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T15:48:47.6562250Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_ctx_manager.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 15:48:47.655942] 2025-12-04T15:49:10.4541556Z 2025-12-04T15:49:10.4544715Z dynamo/test_ctx_manager 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_ctx_manager_1.1_7cf597ef36c79024_.log 2025-12-04T15:49:10.4593926Z Running 104 items in this shard: test/dynamo/test_ctx_manager.py::CtxManagerTests::test_311_resume_block_keyerror, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_311_resume_block_keyerror2, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_autocast, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_autocast_arguments_binding, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_autocast_cpu, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_autocast_cpu_graph_break, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_autocast_cpu_graph_break_2, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_autocast_cpu_graph_break_inner_fn, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_autocast_decorator, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_autocast_device, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_autocast_float64, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_autocast_graph_break_method, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_autocast_sdpa, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_autograd_profiler, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_autograd_profiler_enabled, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_context_wrapping_grad_mode_decorator, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_context_wrapping_grad_mode_nested_function_decorator, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_context_wrapping_set_grad_enabled_nested_function, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_cuda_amp_autocast, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_cuda_device, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_cuda_event_across_graph_break, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_cuda_event_created_outside_of_graph, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_cuda_event_method, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_cuda_event_method_create_stream_outside_of_compile, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_cuda_event_reconstruct, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_cuda_stream_across_graph_break, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_cuda_stream_compared_with_constant, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_cuda_stream_compared_with_stream, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_cuda_stream_context_manager1, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_cuda_stream_context_manager2, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_cuda_stream_method, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_disable_saved_tensors_hooks, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_disable_saved_tensors_hooks_graph_break, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_disable_saved_tensors_hooks_prev_disabled, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_disable_saved_tensors_hooks_prev_disabled_nested, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_generic_context_manager_CustomizedCtxManager, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_generic_context_manager_customized_ctx_manager, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_generic_context_manager_with_graph_break_CustomizedCtxManager, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_generic_context_manager_with_graph_break_customized_ctx_manager, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_generic_ctx_manager_with_graph_break_CustomizedCtxManagerWithGraphBreak, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_generic_ctx_manager_with_graph_break_customized_ctx_manager_with_graph_break, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_grad_mode_guard, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_graph_break_inlining_autocast, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_graph_break_inlining_grad, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_inactive_context_graph_break_local, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_inactive_context_graph_break_local_nullctx, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_inactive_context_graph_break_local_nullctx2, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_inactive_context_graph_break_stack, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_inactive_context_graph_break_stack2, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_is_autocast_cpu_enabled, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_nested_generic_context_manager_CustomizedCtxManager, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_nested_generic_context_manager_customized_ctx_manager, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_nested_generic_context_manager_with_graph_break_CustomizedCtxManager, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_nested_generic_context_manager_with_graph_break_customized_ctx_manager, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_nested_grad_mode_graph_break, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_no_grad, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_return_context_manager, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_return_context_manager_with_graph_break, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_sdpa_kernel_ctx_manager1, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_sdpa_kernel_ctx_manager2, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_sdpa_kernel_ctx_manager3, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_sdpa_kernel_ctx_manager_as_decorator, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_sdpa_kernel_ctx_manager_kwargs, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_sdpa_kernel_ctx_manager_set_priority, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_store_attr_graph_break_key_error, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_torch_profiler, test/dynamo/test_ctx_manager.py::CtxManagerTests::test_torch_profiler_use_after_with_block, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_WITH_EXCEPT_START, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_advanced_contextmanager_as_argument, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_advanced_contextmanager_as_argument_error, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_change_parent_0, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_change_parent_1, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_change_parent_global_0, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_change_parent_global_1, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_change_parent_nonlocal_0, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_change_parent_nonlocal_1, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_contextlib_nullcontext, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_contextlib_suppress_name_stderr, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_contextlib_suppress_name_stdout, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_contextlib_suppress_name_suppress, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_contextmanager_as_argument, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_contextmanager_as_argument_only___enter__, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_contextmanager_as_argument_only___exit__, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_ctx_basic0, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_ctx_basic1, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_disable___enter__, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_disable___exit__, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_disable_ctx_manager, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_disable_trace_contextmanager, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_dynamo_disable_ctx, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_globals_change_in_other_file, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_graph_break_after___enter__, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_graph_break_and_disable___enter__, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_graph_break_before___enter__, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_graph_break_before___enter___and_disable___exit__, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_graph_break_before_and_after___enter__, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_graph_break_in_finally, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_graph_break_inside___enter__, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_graph_break_inside_ctx, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_graph_break_inside_ctx_1, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_graph_break_inside_ctx_2, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_graph_break_inside_ctx_with_side_effects, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_return_advanced_contextmanager, test/dynamo/test_ctx_manager.py::ContextlibContextManagerTests::test_return_new_contextmanager 2025-12-04T15:49:10.4641707Z 2025-12-04T15:49:10.4642054Z Finished dynamo/test_ctx_manager 1/1 ... [2025-12-04 15:49:10.454051][23734.144442398], took 0.38min 2025-12-04T15:49:10.4947868Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_ctx_manager/dynamo.test_ctx_manager-2753bd6e949ccdb3.xml 2025-12-04T15:49:10.5774906Z Running dynamo/test_exc 1/1 ... [2025-12-04 15:49:10.577169][23734.26756103] 2025-12-04T15:49:10.5775458Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T15:49:10.5778533Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_exc.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 15:49:10.577574] 2025-12-04T15:49:22.2582060Z 2025-12-04T15:49:22.2583070Z dynamo/test_exc 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_exc_1.1_d00a7ab51481eb46_.log 2025-12-04T15:49:22.2587127Z Running 10 items in this shard: test/dynamo/test_exc.py::ExcTests::test_backend_suppress_line, test/dynamo/test_exc.py::ExcTests::test_graph_break_log, test/dynamo/test_exc.py::ExcTests::test_graph_break_log_generic_jump, test/dynamo/test_exc.py::ExcTests::test_internal_error_no_suppress, test/dynamo/test_exc.py::ExcTests::test_internal_error_suppress_errors, test/dynamo/test_exc.py::ExcTests::test_not_implemented_error, test/dynamo/test_exc.py::ExcTests::test_trigger_bisect_on_error, test/dynamo/test_exc.py::ExcTests::test_trigger_on_error, test/dynamo/test_exc.py::ExcTests::test_unsupported_error, test/dynamo/test_exc.py::ExcTests::test_unsupported_real_stack 2025-12-04T15:49:22.2590396Z 2025-12-04T15:49:22.2590705Z Finished dynamo/test_exc 1/1 ... [2025-12-04 15:49:22.258029][23745.948418911], took 0.19min 2025-12-04T15:49:22.2989873Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_exc/dynamo.test_exc-e06b4c2ef4a1702a.xml 2025-12-04T15:49:22.3993323Z Running dynamo/test_misc 1/1 ... [2025-12-04 15:49:22.399006][23746.089398274] 2025-12-04T15:49:22.3993903Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T15:49:22.3996971Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_misc.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 15:49:22.399437] 2025-12-04T15:52:07.2592122Z 2025-12-04T15:52:07.2593201Z dynamo/test_misc 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_misc_1.1_cef59efd3c8e72ba_.log 2025-12-04T15:52:07.2812635Z Running 664 items in this shard: test/dynamo/test_misc.py::MiscTests::test_312_binary_slice_with_graph_break1, test/dynamo/test_misc.py::MiscTests::test_312_binary_slice_with_graph_break2, test/dynamo/test_misc.py::MiscTests::test_RAISE_VARARGS_0, test/dynamo/test_misc.py::MiscTests::test_T_tensor_attribute, test/dynamo/test_misc.py::MiscTests::test_add_sizes, test/dynamo/test_misc.py::MiscTests::test_add_to_set, test/dynamo/test_misc.py::MiscTests::test_anomaly_aot_autograd, test/dynamo/test_misc.py::MiscTests::test_any_all_symnode, test/dynamo/test_misc.py::MiscTests::test_aot_autograd_propagate_unbacked_symints_shape, test/dynamo/test_misc.py::MiscTests::test_arange_length_with_float32_dtype, test/dynamo/test_misc.py::MiscTests::test_argwhere_with_dynamic_shapes, test/dynamo/test_misc.py::MiscTests::test_assert, test/dynamo/test_misc.py::MiscTests::test_assert_size_stride, test/dynamo/test_misc.py::MiscTests::test_assigning_function_to_class_attribute, test/dynamo/test_misc.py::MiscTests::test_assigning_function_to_object_attribute, test/dynamo/test_misc.py::MiscTests::test_assume_32_bit_indexing, test/dynamo/test_misc.py::MiscTests::test_backend_match_guard, test/dynamo/test_misc.py::MiscTests::test_backend_match_guard_multi_threads, test/dynamo/test_misc.py::MiscTests::test_backward_deterministic_mode_mismatch_warning, test/dynamo/test_misc.py::MiscTests::test_boolarg, test/dynamo/test_misc.py::MiscTests::test_bound_shape_checks, test/dynamo/test_misc.py::MiscTests::test_build_tuple_unpack, test/dynamo/test_misc.py::MiscTests::test_builder_for_class_with_metaclass, test/dynamo/test_misc.py::MiscTests::test_builtin_abs, test/dynamo/test_misc.py::MiscTests::test_builtin_bool_on_symbool, test/dynamo/test_misc.py::MiscTests::test_builtin_bool_on_symfloat, test/dynamo/test_misc.py::MiscTests::test_builtin_bool_on_symint, test/dynamo/test_misc.py::MiscTests::test_builtin_complex, test/dynamo/test_misc.py::MiscTests::test_builtin_complex_args, test/dynamo/test_misc.py::MiscTests::test_builtin_isinstance, test/dynamo/test_misc.py::MiscTests::test_builtin_str_on_user_defined_function, test/dynamo/test_misc.py::MiscTests::test_builtin_subclasses_as_method_on_class_type, test/dynamo/test_misc.py::MiscTests::test_builtin_subclasses_as_method_on_var, test/dynamo/test_misc.py::MiscTests::test_call_parent_non_class_methods_from_child, test/dynamo/test_misc.py::MiscTests::test_callpacked, test/dynamo/test_misc.py::MiscTests::test_cannot_trace_mark_dynamic, test/dynamo/test_misc.py::MiscTests::test_cannot_trace_mark_dynamic_safe_unreached, test/dynamo/test_misc.py::MiscTests::test_cast, test/dynamo/test_misc.py::MiscTests::test_cat_unbacked, test/dynamo/test_misc.py::MiscTests::test_catch_watchings1, test/dynamo/test_misc.py::MiscTests::test_catch_watchings2, test/dynamo/test_misc.py::MiscTests::test_cell_captured_by_existing_func_but_not_root_frame, test/dynamo/test_misc.py::MiscTests::test_cell_output1, test/dynamo/test_misc.py::MiscTests::test_cell_output2, test/dynamo/test_misc.py::MiscTests::test_check_assert_error_at_runtime_when_predicate_false_and_message_has_closure, test/dynamo/test_misc.py::MiscTests::test_check_assert_error_at_runtime_when_predicate_true_and_message_has_closure, test/dynamo/test_misc.py::MiscTests::test_check_compiles_when_predicate_true_and_message_None, test/dynamo/test_misc.py::MiscTests::test_check_compiles_when_predicate_true_and_message_has_global, test/dynamo/test_misc.py::MiscTests::test_check_compiles_when_predicate_true_and_message_has_no_closure, test/dynamo/test_misc.py::MiscTests::test_check_compiles_when_predicate_true_constant_and_message_None, test/dynamo/test_misc.py::MiscTests::test_check_compiles_when_predicate_true_constant_and_message_has_no_closure, test/dynamo/test_misc.py::MiscTests::test_check_raises_at_runtime_when_predicate_false_and_message_None, test/dynamo/test_misc.py::MiscTests::test_check_raises_at_runtime_when_predicate_false_and_message_has_global, test/dynamo/test_misc.py::MiscTests::test_check_raises_at_runtime_when_predicate_false_and_message_has_no_closure, test/dynamo/test_misc.py::MiscTests::test_check_raises_at_runtime_when_predicate_false_constant_and_message_None, test/dynamo/test_misc.py::MiscTests::test_check_raises_at_runtime_when_predicate_false_constant_and_message_has_no_closure, test/dynamo/test_misc.py::MiscTests::test_check_simplification, test/dynamo/test_misc.py::MiscTests::test_class_binop, test/dynamo/test_misc.py::MiscTests::test_class_duner_flags, test/dynamo/test_misc.py::MiscTests::test_class_duner_mro, test/dynamo/test_misc.py::MiscTests::test_class_has_instancecheck_method, test/dynamo/test_misc.py::MiscTests::test_clone_sparse_input, test/dynamo/test_misc.py::MiscTests::test_closure_out_of_scope_cell, test/dynamo/test_misc.py::MiscTests::test_closure_out_of_scope_cell_with_cond, test/dynamo/test_misc.py::MiscTests::test_closure_out_of_scope_cell_with_mutation, test/dynamo/test_misc.py::MiscTests::test_closure_recompiles, test/dynamo/test_misc.py::MiscTests::test_closure_with_mutation_and_graph_break, test/dynamo/test_misc.py::MiscTests::test_closure_write_across_functions, test/dynamo/test_misc.py::MiscTests::test_compare_shapes_eq, test/dynamo/test_misc.py::MiscTests::test_compare_shapes_neq, test/dynamo/test_misc.py::MiscTests::test_compare_shapes_tuple_eq, test/dynamo/test_misc.py::MiscTests::test_compare_shapes_tuple_neq, test/dynamo/test_misc.py::MiscTests::test_compare_shapes_with_constant, test/dynamo/test_misc.py::MiscTests::test_compare_tensor_with_none, test/dynamo/test_misc.py::MiscTests::test_compilation_metrics_size_limit, test/dynamo/test_misc.py::MiscTests::test_compiled_class_graph_break, test/dynamo/test_misc.py::MiscTests::test_cond, test/dynamo/test_misc.py::MiscTests::test_cond_export, test/dynamo/test_misc.py::MiscTests::test_cond_export_single_arg, test/dynamo/test_misc.py::MiscTests::test_cond_nested, test/dynamo/test_misc.py::MiscTests::test_cond_runtime_assert_generation, test/dynamo/test_misc.py::MiscTests::test_cond_side_effects, test/dynamo/test_misc.py::MiscTests::test_cond_with_quantization, test/dynamo/test_misc.py::MiscTests::test_conditional_list_comp_in_context, test/dynamo/test_misc.py::MiscTests::test_config_getattr_default, test/dynamo/test_misc.py::MiscTests::test_config_obj, test/dynamo/test_misc.py::MiscTests::test_const_dict_variable_python_type, test/dynamo/test_misc.py::MiscTests::test_constant_getattr, test/dynamo/test_misc.py::MiscTests::test_constant_hasattr_returns_bool, test/dynamo/test_misc.py::MiscTests::test_cross_entropy_loss_fancy_ctor1, test/dynamo/test_misc.py::MiscTests::test_cross_entropy_loss_fancy_ctor2, test/dynamo/test_misc.py::MiscTests::test_cross_entropy_loss_simple_ctor, test/dynamo/test_misc.py::MiscTests::test_custom_dict, test/dynamo/test_misc.py::MiscTests::test_custom_module_free, test/dynamo/test_misc.py::MiscTests::test_data_access_in_inference_mode, test/dynamo/test_misc.py::MiscTests::test_data_ptr_graph_break_aten, test/dynamo/test_misc.py::MiscTests::test_data_ptr_graph_break_builtin, test/dynamo/test_misc.py::MiscTests::test_dataclass, test/dynamo/test_misc.py::MiscTests::test_dataclass_fields, test/dynamo/test_misc.py::MiscTests::test_dataclass_local_hasattr, test/dynamo/test_misc.py::MiscTests::test_default_args_device_dtype, test/dynamo/test_misc.py::MiscTests::test_default_dtype_change, test/dynamo/test_misc.py::MiscTests::test_defaultdict, test/dynamo/test_misc.py::MiscTests::test_deque_append_left, test/dynamo/test_misc.py::MiscTests::test_deque_input, test/dynamo/test_misc.py::MiscTests::test_derpy_nn_module_usage, test/dynamo/test_misc.py::MiscTests::test_descriptor, test/dynamo/test_misc.py::MiscTests::test_descriptor_side_effect, test/dynamo/test_misc.py::MiscTests::test_deterministic_algorithms_mutated, test/dynamo/test_misc.py::MiscTests::test_dictcomp, test/dynamo/test_misc.py::MiscTests::test_dim_order, test/dynamo/test_misc.py::MiscTests::test_disable_flag, test/dynamo/test_misc.py::MiscTests::test_dtypes_no_graphbreaks, test/dynamo/test_misc.py::MiscTests::test_dunder_methods, test/dynamo/test_misc.py::MiscTests::test_dunder_new_function_inlining, test/dynamo/test_misc.py::MiscTests::test_dunder_new_function_inlining1, test/dynamo/test_misc.py::MiscTests::test_dunder_new_function_inlining2, test/dynamo/test_misc.py::MiscTests::test_dunder_new_function_inlining3, test/dynamo/test_misc.py::MiscTests::test_dunder_new_function_inlining4, test/dynamo/test_misc.py::MiscTests::test_dunder_weakref, test/dynamo/test_misc.py::MiscTests::test_duplicate_graph_break_log, test/dynamo/test_misc.py::MiscTests::test_dynamic_one_hot, test/dynamo/test_misc.py::MiscTests::test_dynamic_shapes_as_strided, test/dynamo/test_misc.py::MiscTests::test_dynamic_sources_dynamic_override, test/dynamo/test_misc.py::MiscTests::test_dynamic_sources_dynamic_override_regex, test/dynamo/test_misc.py::MiscTests::test_dynamic_sources_force_parameter_static_shapes_and_property_static_shapes_override, test/dynamo/test_misc.py::MiscTests::test_dynamic_sources_graph_break, test/dynamo/test_misc.py::MiscTests::test_dynamic_sources_int, test/dynamo/test_misc.py::MiscTests::test_dynamic_sources_precedence_over_int_specialization, test/dynamo/test_misc.py::MiscTests::test_dynamic_sources_tensor, test/dynamo/test_misc.py::MiscTests::test_dynamo_cache_invalidate, test/dynamo/test_misc.py::MiscTests::test_dynamo_cache_move_to_front, test/dynamo/test_misc.py::MiscTests::test_dynamo_compiling_fake_tensor_to_vararg_int, test/dynamo/test_misc.py::MiscTests::test_dynamo_disabled_in_custom_op_kernels, test/dynamo/test_misc.py::MiscTests::test_dynamo_inside_custom_op, test/dynamo/test_misc.py::MiscTests::test_dynamo_min_operator_with_shape, test/dynamo/test_misc.py::MiscTests::test_dynamo_reset_clears_cache, test/dynamo/test_misc.py::MiscTests::test_empty_list, test/dynamo/test_misc.py::MiscTests::test_enum_as_dict_key, test/dynamo/test_misc.py::MiscTests::test_enum_as_dict_key_with_overloaded_str, test/dynamo/test_misc.py::MiscTests::test_enum_guards, test/dynamo/test_misc.py::MiscTests::test_enum_method, test/dynamo/test_misc.py::MiscTests::test_enum_no_graphbreaks, test/dynamo/test_misc.py::MiscTests::test_enum_subclass, test/dynamo/test_misc.py::MiscTests::test_error_on_nested_fx_trace, test/dynamo/test_misc.py::MiscTests::test_error_on_recompile, test/dynamo/test_misc.py::MiscTests::test_escaping_closure_var_with_backward_hook, test/dynamo/test_misc.py::MiscTests::test_escaping_closure_var_with_nonlocal_var, test/dynamo/test_misc.py::MiscTests::test_existing_func_that_creates_capturing_nested_func, test/dynamo/test_misc.py::MiscTests::test_fail_on_recompile_error_message, test/dynamo/test_misc.py::MiscTests::test_flat_name_to_original_fqn, test/dynamo/test_misc.py::MiscTests::test_float_speculation_log_divergence, test/dynamo/test_misc.py::MiscTests::test_fn_hasattr__name__1, test/dynamo/test_misc.py::MiscTests::test_fn_hasattr__name__2, test/dynamo/test_misc.py::MiscTests::test_fn_hasattr__name__3, test/dynamo/test_misc.py::MiscTests::test_fold, test/dynamo/test_misc.py::MiscTests::test_free_var_and_local_name_collision, test/dynamo/test_misc.py::MiscTests::test_frozen_dataclass_attr_access, test/dynamo/test_misc.py::MiscTests::test_frozen_dataclass_default_factory, test/dynamo/test_misc.py::MiscTests::test_frozen_dataclass_default_value, test/dynamo/test_misc.py::MiscTests::test_frozen_dataclass_hashable, test/dynamo/test_misc.py::MiscTests::test_frozen_dataclass_kw_only, test/dynamo/test_misc.py::MiscTests::test_frozen_dict, test/dynamo/test_misc.py::MiscTests::test_frozenset_of_non_literals, test/dynamo/test_misc.py::MiscTests::test_frozenset_torch_func_contains, test/dynamo/test_misc.py::MiscTests::test_fullgraph_capture, test/dynamo/test_misc.py::MiscTests::test_funcname_cache, test/dynamo/test_misc.py::MiscTests::test_function_annotation, test/dynamo/test_misc.py::MiscTests::test_function_generic_alias_annotation, test/dynamo/test_misc.py::MiscTests::test_generate_tensor_from_list_of_numpy_primitive_type, test/dynamo/test_misc.py::MiscTests::test_generate_trivial_abstract_impl, test/dynamo/test_misc.py::MiscTests::test_get_attr_function, test/dynamo/test_misc.py::MiscTests::test_get_cache_entry, test/dynamo/test_misc.py::MiscTests::test_get_custom_tensor_attribute, test/dynamo/test_misc.py::MiscTests::test_get_instruction_source_311, test/dynamo/test_misc.py::MiscTests::test_getattr_dict, test/dynamo/test_misc.py::MiscTests::test_getattrvariable_as_python_constant, test/dynamo/test_misc.py::MiscTests::test_getset_descriptor, test/dynamo/test_misc.py::MiscTests::test_global_state_guard_serialization, test/dynamo/test_misc.py::MiscTests::test_grad, test/dynamo/test_misc.py::MiscTests::test_grad_non_none, test/dynamo/test_misc.py::MiscTests::test_grad_none, test/dynamo/test_misc.py::MiscTests::test_grad_state_mutated, test/dynamo/test_misc.py::MiscTests::test_graph_break_compilation_metrics, test/dynamo/test_misc.py::MiscTests::test_graph_break_compilation_metrics_on_failure, test/dynamo/test_misc.py::MiscTests::test_graph_break_correctly_when_passing_numpy_ndarray_to_torch_function, test/dynamo/test_misc.py::MiscTests::test_guard_failure_fn, test/dynamo/test_misc.py::MiscTests::test_guard_failure_fn2, test/dynamo/test_misc.py::MiscTests::test_guard_failure_fn_shape_control, test/dynamo/test_misc.py::MiscTests::test_guard_failure_fn_tensor_iter, test/dynamo/test_misc.py::MiscTests::test_guard_filter_fn_by_id, test/dynamo/test_misc.py::MiscTests::test_guard_filter_fn_by_is_global, test/dynamo/test_misc.py::MiscTests::test_guard_filter_fn_by_name_and_value, test/dynamo/test_misc.py::MiscTests::test_guard_filter_globals, test/dynamo/test_misc.py::MiscTests::test_guard_filter_inbuilt_nn_modules, test/dynamo/test_misc.py::MiscTests::test_guard_filter_nn_modules, test/dynamo/test_misc.py::MiscTests::test_guard_filter_tensors, test/dynamo/test_misc.py::MiscTests::test_guard_function_builder_with_cse, test/dynamo/test_misc.py::MiscTests::test_guard_size_oblivious_backed, test/dynamo/test_misc.py::MiscTests::test_guard_string_escaped, test/dynamo/test_misc.py::MiscTests::test_guard_sym_node_fstring_when_used, test/dynamo/test_misc.py::MiscTests::test_guards_cse_pass_multiple, test/dynamo/test_misc.py::MiscTests::test_guards_cse_pass_single, test/dynamo/test_misc.py::MiscTests::test_guards_strip_function_call, test/dynamo/test_misc.py::MiscTests::test_hasattr_nn_module_guard, test/dynamo/test_misc.py::MiscTests::test_hash_getitem_slice, test/dynamo/test_misc.py::MiscTests::test_hash_hop, test/dynamo/test_misc.py::MiscTests::test_id_guarded_class, test/dynamo/test_misc.py::MiscTests::test_id_guarded_module, test/dynamo/test_misc.py::MiscTests::test_id_guarded_object, test/dynamo/test_misc.py::MiscTests::test_id_of_nn_module, test/dynamo/test_misc.py::MiscTests::test_id_tensor, test/dynamo/test_misc.py::MiscTests::test_if_cond_nn_mod1, test/dynamo/test_misc.py::MiscTests::test_if_cond_nn_mod2, test/dynamo/test_misc.py::MiscTests::test_if_cond_nn_mod3, test/dynamo/test_misc.py::MiscTests::test_if_cond_user_defined_object, test/dynamo/test_misc.py::MiscTests::test_if_cond_user_defined_object2, test/dynamo/test_misc.py::MiscTests::test_if_cond_user_defined_object3, test/dynamo/test_misc.py::MiscTests::test_infer_unbacked_size_gt_zero, test/dynamo/test_misc.py::MiscTests::test_inference_mode, test/dynamo/test_misc.py::MiscTests::test_inference_mode_param, test/dynamo/test_misc.py::MiscTests::test_inline_closure_not_loaded_by_parent, test/dynamo/test_misc.py::MiscTests::test_inline_closure_returned_by_another_function_and_captures, test/dynamo/test_misc.py::MiscTests::test_inline_dict_function, test/dynamo/test_misc.py::MiscTests::test_inline_dict_function_passed_as_arg, test/dynamo/test_misc.py::MiscTests::test_inline_dict_mutation, test/dynamo/test_misc.py::MiscTests::test_inline_func_jump_on_tensor_condition, test/dynamo/test_misc.py::MiscTests::test_inline_list_mutation, test/dynamo/test_misc.py::MiscTests::test_inline_local_dict_clear, test/dynamo/test_misc.py::MiscTests::test_inline_module_attr_dict_clear, test/dynamo/test_misc.py::MiscTests::test_inline_user_defined_dict_attr_clear, test/dynamo/test_misc.py::MiscTests::test_inplace, test/dynamo/test_misc.py::MiscTests::test_inplace_desugaring, test/dynamo/test_misc.py::MiscTests::test_inplace_param_update, test/dynamo/test_misc.py::MiscTests::test_inplace_view_on_graph_input, test/dynamo/test_misc.py::MiscTests::test_input_cell_mutation, test/dynamo/test_misc.py::MiscTests::test_inspect_signature_bind, test/dynamo/test_misc.py::MiscTests::test_inspect_signature_bind_non_user_function, test/dynamo/test_misc.py::MiscTests::test_inspect_signature_parameters, test/dynamo/test_misc.py::MiscTests::test_int_int_comparisons, test/dynamo/test_misc.py::MiscTests::test_int_list, test/dynamo/test_misc.py::MiscTests::test_int_neg, test/dynamo/test_misc.py::MiscTests::test_int_shape_binops, test/dynamo/test_misc.py::MiscTests::test_int_shape_comparisons, test/dynamo/test_misc.py::MiscTests::test_int_shape_inplace_binops, test/dynamo/test_misc.py::MiscTests::test_intermediary_tensor_grad_access, test/dynamo/test_misc.py::MiscTests::test_invalid_args_builtin, test/dynamo/test_misc.py::MiscTests::test_is_compiling, test/dynamo/test_misc.py::MiscTests::test_is_floating_point, test/dynamo/test_misc.py::MiscTests::test_is_floating_point2, test/dynamo/test_misc.py::MiscTests::test_is_tensor, test/dynamo/test_misc.py::MiscTests::test_is_tensor2, test/dynamo/test_misc.py::MiscTests::test_is_tensor_like, test/dynamo/test_misc.py::MiscTests::test_is_tensor_like2, test/dynamo/test_misc.py::MiscTests::test_item, test/dynamo/test_misc.py::MiscTests::test_item_changes, test/dynamo/test_misc.py::MiscTests::test_item_changes_new_shape, test/dynamo/test_misc.py::MiscTests::test_iter_set, test/dynamo/test_misc.py::MiscTests::test_iter_type, test/dynamo/test_misc.py::MiscTests::test_iterator_limit, test/dynamo/test_misc.py::MiscTests::test_itertools_accumulate_symint_default_sum, test/dynamo/test_misc.py::MiscTests::test_itertools_accumulate_tensors_builtins, test/dynamo/test_misc.py::MiscTests::test_itertools_accumulate_tensors_default_sum, test/dynamo/test_misc.py::MiscTests::test_itertools_accumulate_tensors_kwargs, test/dynamo/test_misc.py::MiscTests::test_itertools_accumulate_tensors_user_defined, test/dynamo/test_misc.py::MiscTests::test_itertools_groupby_pure_python_default_identify_func, test/dynamo/test_misc.py::MiscTests::test_itertools_groupby_pure_python_key_func, test/dynamo/test_misc.py::MiscTests::test_itertools_infinite_count, test/dynamo/test_misc.py::MiscTests::test_itertools_infinite_cycle, test/dynamo/test_misc.py::MiscTests::test_itertools_infinite_repeat, test/dynamo/test_misc.py::MiscTests::test_itertools_infinite_repeat_mutation, test/dynamo/test_misc.py::MiscTests::test_itertools_islice, test/dynamo/test_misc.py::MiscTests::test_itertools_islice_default_end, test/dynamo/test_misc.py::MiscTests::test_itertools_islice_default_step, test/dynamo/test_misc.py::MiscTests::test_itertools_repeat, test/dynamo/test_misc.py::MiscTests::test_itertools_tee, test/dynamo/test_misc.py::MiscTests::test_jacfwd_one_hot_dynamic_compile, test/dynamo/test_misc.py::MiscTests::test_large_reduction_list, test/dynamo/test_misc.py::MiscTests::test_linear_module_free, test/dynamo/test_misc.py::MiscTests::test_list_append_return_none, test/dynamo/test_misc.py::MiscTests::test_list_class, test/dynamo/test_misc.py::MiscTests::test_list_hasattr1, test/dynamo/test_misc.py::MiscTests::test_list_hasattr2, test/dynamo/test_misc.py::MiscTests::test_list_iadd_side_effect, test/dynamo/test_misc.py::MiscTests::test_list_iadd_with_shape, test/dynamo/test_misc.py::MiscTests::test_list_iterator_contains, test/dynamo/test_misc.py::MiscTests::test_list_mul, test/dynamo/test_misc.py::MiscTests::test_list_slice_mul, test/dynamo/test_misc.py::MiscTests::test_listcomp, test/dynamo/test_misc.py::MiscTests::test_load_fast_and_clear_graph_break, test/dynamo/test_misc.py::MiscTests::test_mandelbrot_numpy, test/dynamo/test_misc.py::MiscTests::test_map_side_effects, test/dynamo/test_misc.py::MiscTests::test_map_with_quantization, test/dynamo/test_misc.py::MiscTests::test_mark_dynamic_with_ranges, test/dynamo/test_misc.py::MiscTests::test_mark_static, test/dynamo/test_misc.py::MiscTests::test_mark_unbacked_strict, test/dynamo/test_misc.py::MiscTests::test_matmul1, test/dynamo/test_misc.py::MiscTests::test_min_max_over_iterable, test/dynamo/test_misc.py::MiscTests::test_module_complex_iter, test/dynamo/test_misc.py::MiscTests::test_module_deepcopy, test/dynamo/test_misc.py::MiscTests::test_module_not_callable, test/dynamo/test_misc.py::MiscTests::test_mro_type_tensor_no_source, test/dynamo/test_misc.py::MiscTests::test_multiple_inheritance, test/dynamo/test_misc.py::MiscTests::test_mutable_mapping_multiple_inheritance, test/dynamo/test_misc.py::MiscTests::test_named_parameters, test/dynamo/test_misc.py::MiscTests::test_namedtuple1, test/dynamo/test_misc.py::MiscTests::test_namedtuple2, test/dynamo/test_misc.py::MiscTests::test_namedtuple3, test/dynamo/test_misc.py::MiscTests::test_namedtuple_class, test/dynamo/test_misc.py::MiscTests::test_namedtuple_source_dynamic_attributes, test/dynamo/test_misc.py::MiscTests::test_namedtuple_sourceless_dynamic_attributes, test/dynamo/test_misc.py::MiscTests::test_namedtuple_with_custom_getitem, test/dynamo/test_misc.py::MiscTests::test_nan, test/dynamo/test_misc.py::MiscTests::test_ne_operator_with_custom_eq, test/dynamo/test_misc.py::MiscTests::test_ne_operator_with_custom_graphbreak_eq, test/dynamo/test_misc.py::MiscTests::test_ne_operator_with_custom_ne, test/dynamo/test_misc.py::MiscTests::test_nested_closure, test/dynamo/test_misc.py::MiscTests::test_nested_closure_mutation, test/dynamo/test_misc.py::MiscTests::test_nested_dataclass_reconstruct, test/dynamo/test_misc.py::MiscTests::test_nested_frozen_dataclass_hashable, test/dynamo/test_misc.py::MiscTests::test_nested_function_resuming_with_correct_globals, test/dynamo/test_misc.py::MiscTests::test_nested_optimize, test/dynamo/test_misc.py::MiscTests::test_nested_optimize_decorator, test/dynamo/test_misc.py::MiscTests::test_nested_optimize_run, test/dynamo/test_misc.py::MiscTests::test_nested_sequential_try, test/dynamo/test_misc.py::MiscTests::test_nested_sequential_try_with, test/dynamo/test_misc.py::MiscTests::test_nested_sequential_try_with_graph_break, test/dynamo/test_misc.py::MiscTests::test_nested_sequential_with, test/dynamo/test_misc.py::MiscTests::test_nested_wraps, test/dynamo/test_misc.py::MiscTests::test_nesteduserfunction_setattr, test/dynamo/test_misc.py::MiscTests::test_new_with_int_list, test/dynamo/test_misc.py::MiscTests::test_newly_constructed_tensor_attr_mutation, test/dynamo/test_misc.py::MiscTests::test_nn_functional_reduction, test/dynamo/test_misc.py::MiscTests::test_nn_module_getattr, test/dynamo/test_misc.py::MiscTests::test_nn_module_getattribute, test/dynamo/test_misc.py::MiscTests::test_nn_sequential_invocation, test/dynamo/test_misc.py::MiscTests::test_nn_sequential_invocation_reposition_indices, test/dynamo/test_misc.py::MiscTests::test_no_error_on_nested_fx_trace, test/dynamo/test_misc.py::MiscTests::test_no_guard_for_unused_sym_node_fstring, test/dynamo/test_misc.py::MiscTests::test_no_raise_guard_partial_constraint, test/dynamo/test_misc.py::MiscTests::test_no_raise_guard_partial_constraint_across_break, test/dynamo/test_misc.py::MiscTests::test_non_pt2_compliant_ops_graph_break, test/dynamo/test_misc.py::MiscTests::test_not_dynamic_scope, test/dynamo/test_misc.py::MiscTests::test_numel, test/dynamo/test_misc.py::MiscTests::test_numpy_array_of_arrays, test/dynamo/test_misc.py::MiscTests::test_numpy_as_global, test/dynamo/test_misc.py::MiscTests::test_numpy_fallback_on_eager, test/dynamo/test_misc.py::MiscTests::test_numpy_force, test/dynamo/test_misc.py::MiscTests::test_numpy_gt, test/dynamo/test_misc.py::MiscTests::test_numpy_int_constant, test/dynamo/test_misc.py::MiscTests::test_numpy_iter, test/dynamo/test_misc.py::MiscTests::test_numpy_min, test/dynamo/test_misc.py::MiscTests::test_numpy_ndarray_graph_break, test/dynamo/test_misc.py::MiscTests::test_numpy_ndarray_graph_break_with_multiple_outputs, test/dynamo/test_misc.py::MiscTests::test_numpy_ndarray_works_with_builtin_function, test/dynamo/test_misc.py::MiscTests::test_numpy_no_raise, test/dynamo/test_misc.py::MiscTests::test_numpy_non_torch_dtype, test/dynamo/test_misc.py::MiscTests::test_numpy_random_config_to_numpy, test/dynamo/test_misc.py::MiscTests::test_numpy_readonly, test/dynamo/test_misc.py::MiscTests::test_numpy_recompilation_scalar, test/dynamo/test_misc.py::MiscTests::test_numpy_size_attr, test/dynamo/test_misc.py::MiscTests::test_numpy_subdtype, test/dynamo/test_misc.py::MiscTests::test_numpy_take_along_axis, test/dynamo/test_misc.py::MiscTests::test_numpy_tolist, test/dynamo/test_misc.py::MiscTests::test_numpy_torch_operators, test/dynamo/test_misc.py::MiscTests::test_numpy_ufunc_out, test/dynamo/test_misc.py::MiscTests::test_numpy_ufunc_out_graph_break, test/dynamo/test_misc.py::MiscTests::test_numpy_unique_f16, test/dynamo/test_misc.py::MiscTests::test_numpy_variable_isinstance, test/dynamo/test_misc.py::MiscTests::test_numpy_with_builtin_type, test/dynamo/test_misc.py::MiscTests::test_object_classmethod, test/dynamo/test_misc.py::MiscTests::test_object_setattr, test/dynamo/test_misc.py::MiscTests::test_object_staticmethod, test/dynamo/test_misc.py::MiscTests::test_onnx_shape_as_tensor, test/dynamo/test_misc.py::MiscTests::test_optimize_on_module, test/dynamo/test_misc.py::MiscTests::test_ordered_dict_alias_reconstruct, test/dynamo/test_misc.py::MiscTests::test_ordered_dict_move_to_end, test/dynamo/test_misc.py::MiscTests::test_os_environ_get, test/dynamo/test_misc.py::MiscTests::test_os_environ_set_graph_break, test/dynamo/test_misc.py::MiscTests::test_out_variant_custom_op, test/dynamo/test_misc.py::MiscTests::test_out_variants_with_resizing_on_graph_inputs, test/dynamo/test_misc.py::MiscTests::test_out_variants_with_resizing_on_graph_inputs_with_dynamic, test/dynamo/test_misc.py::MiscTests::test_out_variants_with_resizing_on_graph_inputs_with_dynamic1, test/dynamo/test_misc.py::MiscTests::test_outside_linear_module_free, test/dynamo/test_misc.py::MiscTests::test_overridden_getattribute, test/dynamo/test_misc.py::MiscTests::test_packaging_version_parse, test/dynamo/test_misc.py::MiscTests::test_pair, test/dynamo/test_misc.py::MiscTests::test_param_shape_binops, test/dynamo/test_misc.py::MiscTests::test_parameter_free, test/dynamo/test_misc.py::MiscTests::test_patched_builtin_functions, test/dynamo/test_misc.py::MiscTests::test_pep0479_convert_stopiteration, test/dynamo/test_misc.py::MiscTests::test_precompile_entries, test/dynamo/test_misc.py::MiscTests::test_precompile_entry_hit, test/dynamo/test_misc.py::MiscTests::test_precompile_entry_miss, test/dynamo/test_misc.py::MiscTests::test_precompile_fail_on_recompile, test/dynamo/test_misc.py::MiscTests::test_proxy_frozen_dataclass, test/dynamo/test_misc.py::MiscTests::test_pt2_compliant_ops_are_allowed, test/dynamo/test_misc.py::MiscTests::test_pt2_compliant_overload, test/dynamo/test_misc.py::MiscTests::test_pure_python_accumulate, test/dynamo/test_misc.py::MiscTests::test_py_guards_mark_dynamic, test/dynamo/test_misc.py::MiscTests::test_python_slice, test/dynamo/test_misc.py::MiscTests::test_raise_guard_full_constraint, test/dynamo/test_misc.py::MiscTests::test_raise_guard_indirect_full_constraint, test/dynamo/test_misc.py::MiscTests::test_raise_guard_partial_constraint_across_break, test/dynamo/test_misc.py::MiscTests::test_raise_guard_partial_constraint_no_graph_break, test/dynamo/test_misc.py::MiscTests::test_raise_on_backend_error, test/dynamo/test_misc.py::MiscTests::test_raises, test/dynamo/test_misc.py::MiscTests::test_raises_importerror1, test/dynamo/test_misc.py::MiscTests::test_raises_importerror2, test/dynamo/test_misc.py::MiscTests::test_range___iter__, test/dynamo/test_misc.py::MiscTests::test_range_input, test/dynamo/test_misc.py::MiscTests::test_range_iter_guards, test/dynamo/test_misc.py::MiscTests::test_range_iter_side_effects, test/dynamo/test_misc.py::MiscTests::test_range_with_shape, test/dynamo/test_misc.py::MiscTests::test_real_imag_tensor_attribute, test/dynamo/test_misc.py::MiscTests::test_recompile_message_on_parameter, test/dynamo/test_misc.py::MiscTests::test_recompile_on_disable_1, test/dynamo/test_misc.py::MiscTests::test_recompile_on_disable_2, test/dynamo/test_misc.py::MiscTests::test_recompile_on_global_state_change, test/dynamo/test_misc.py::MiscTests::test_reconstruct_frozen_dataclass, test/dynamo/test_misc.py::MiscTests::test_reconstruct_set_across_graph_break, test/dynamo/test_misc.py::MiscTests::test_recursion_depth_guards, test/dynamo/test_misc.py::MiscTests::test_recursive_inline_list_mutation, test/dynamo/test_misc.py::MiscTests::test_recursive_tensor_attribute, test/dynamo/test_misc.py::MiscTests::test_release_input_memory, test/dynamo/test_misc.py::MiscTests::test_release_module_memory, test/dynamo/test_misc.py::MiscTests::test_release_scope_memory, test/dynamo/test_misc.py::MiscTests::test_remove_set, test/dynamo/test_misc.py::MiscTests::test_repeat_interleave_graphbreaks, test/dynamo/test_misc.py::MiscTests::test_replay_side_effects_config, test/dynamo/test_misc.py::MiscTests::test_replay_side_effects_input_mut, test/dynamo/test_misc.py::MiscTests::test_replay_side_effects_model_attr, test/dynamo/test_misc.py::MiscTests::test_repr, test/dynamo/test_misc.py::MiscTests::test_repro_graph_breaks_in__get_item_by_idx, test/dynamo/test_misc.py::MiscTests::test_restore_graphstate, test/dynamo/test_misc.py::MiscTests::test_return_dict_with_graph_break_and_update, test/dynamo/test_misc.py::MiscTests::test_return_nested_function, test/dynamo/test_misc.py::MiscTests::test_returning_func_with_captured_func_and_tensor, test/dynamo/test_misc.py::MiscTests::test_returning_nested_func_with_captured_tensor, test/dynamo/test_misc.py::MiscTests::test_running_func_with_captured_func_and_tensor, test/dynamo/test_misc.py::MiscTests::test_running_nested_func_with_captured_tensor, test/dynamo/test_misc.py::MiscTests::test_runtime_assert_replacement, test/dynamo/test_misc.py::MiscTests::test_sample_input, test/dynamo/test_misc.py::MiscTests::test_scalar_device_movement, test/dynamo/test_misc.py::MiscTests::test_scalar_tensor_is_equivalent_to_int_list_argument, test/dynamo/test_misc.py::MiscTests::test_scalar_tensor_is_equivalent_to_symint_argument, test/dynamo/test_misc.py::MiscTests::test_scalar_tensor_is_equivalent_to_symint_list_argument, test/dynamo/test_misc.py::MiscTests::test_sequential_module_free, test/dynamo/test_misc.py::MiscTests::test_set_aliasing_recompiles, test/dynamo/test_misc.py::MiscTests::test_set_custom_tensor_attribute, test/dynamo/test_misc.py::MiscTests::test_set_descriptor, test/dynamo/test_misc.py::MiscTests::test_set_discard, test/dynamo/test_misc.py::MiscTests::test_set_update, test/dynamo/test_misc.py::MiscTests::test_setattr_mutation1, test/dynamo/test_misc.py::MiscTests::test_setattr_mutation2, test/dynamo/test_misc.py::MiscTests::test_setattr_mutation3, test/dynamo/test_misc.py::MiscTests::test_shape_and_tuple_equality, test/dynamo/test_misc.py::MiscTests::test_shape_env_equal_constructor, test/dynamo/test_misc.py::MiscTests::test_shape_env_equal_create_symbolic_sizes_strides_storage_offset, test/dynamo/test_misc.py::MiscTests::test_shape_env_equal_empty, test/dynamo/test_misc.py::MiscTests::test_shape_env_equal_evaluate_expr_divisible, test/dynamo/test_misc.py::MiscTests::test_shape_env_equal_evaluate_expr_refinement, test/dynamo/test_misc.py::MiscTests::test_shape_env_equal_evaluate_expr_replacement, test/dynamo/test_misc.py::MiscTests::test_shape_env_equal_runtime_assert, test/dynamo/test_misc.py::MiscTests::test_shape_env_equal_unbacked, test/dynamo/test_misc.py::MiscTests::test_shape_env_no_recording, test/dynamo/test_misc.py::MiscTests::test_shape_env_recorded_function_fallback, test/dynamo/test_misc.py::MiscTests::test_shape_int_comparisons, test/dynamo/test_misc.py::MiscTests::test_shape_int_inplace_binops, test/dynamo/test_misc.py::MiscTests::test_shape_type, test/dynamo/test_misc.py::MiscTests::test_shape_unpack, test/dynamo/test_misc.py::MiscTests::test_side_effects_codegen_update_mutated, test/dynamo/test_misc.py::MiscTests::test_simple_set_usage, test/dynamo/test_misc.py::MiscTests::test_size_dim, test/dynamo/test_misc.py::MiscTests::test_size_input, test/dynamo/test_misc.py::MiscTests::test_slice_input, test/dynamo/test_misc.py::MiscTests::test_source_non_input_grad_access, test/dynamo/test_misc.py::MiscTests::test_sourceless_namedtuple, test/dynamo/test_misc.py::MiscTests::test_sparse_output_inductor_should_break, test/dynamo/test_misc.py::MiscTests::test_storage_return, test/dynamo/test_misc.py::MiscTests::test_str___iter__, test/dynamo/test_misc.py::MiscTests::test_str_format_assert1, test/dynamo/test_misc.py::MiscTests::test_str_format_assert2, test/dynamo/test_misc.py::MiscTests::test_str_format_return1, test/dynamo/test_misc.py::MiscTests::test_str_format_return2, test/dynamo/test_misc.py::MiscTests::test_stride_dim, test/dynamo/test_misc.py::MiscTests::test_structseq1, test/dynamo/test_misc.py::MiscTests::test_structseq2, test/dynamo/test_misc.py::MiscTests::test_super_after_graph_break, test/dynamo/test_misc.py::MiscTests::test_super_calling_with_metaclass, test/dynamo/test_misc.py::MiscTests::test_sym_and_terms, test/dynamo/test_misc.py::MiscTests::test_sym_constrain_range_on_replaced_unbacked_symbol, test/dynamo/test_misc.py::MiscTests::test_symint_as_device_kwarg_multi_gpu, test/dynamo/test_misc.py::MiscTests::test_symint_as_device_kwarg_non_strict_export, test/dynamo/test_misc.py::MiscTests::test_symint_copy_into_unbacked_slice, test/dynamo/test_misc.py::MiscTests::test_symint_fold_nontrivial_product_modulo, test/dynamo/test_misc.py::MiscTests::test_sys_modules, test/dynamo/test_misc.py::MiscTests::test_tagging_tensors_mix_used_unused_structure, test/dynamo/test_misc.py::MiscTests::test_tagging_tensors_simple, test/dynamo/test_misc.py::MiscTests::test_tensor__iter__, test/dynamo/test_misc.py::MiscTests::test_tensor_build_list_unpack, test/dynamo/test_misc.py::MiscTests::test_tensor_ctor_list_of_tensor, test/dynamo/test_misc.py::MiscTests::test_tensor_data, test/dynamo/test_misc.py::MiscTests::test_tensor_dict1, test/dynamo/test_misc.py::MiscTests::test_tensor_dict2, test/dynamo/test_misc.py::MiscTests::test_tensor_dict3, test/dynamo/test_misc.py::MiscTests::test_tensor_dot_grad_no_graph_break, test/dynamo/test_misc.py::MiscTests::test_tensor_dynamic_method, test/dynamo/test_misc.py::MiscTests::test_tensor_hasattr, test/dynamo/test_misc.py::MiscTests::test_tensor_interacts_with_numpy_ndarray, test/dynamo/test_misc.py::MiscTests::test_tensor_is_contiguous, test/dynamo/test_misc.py::MiscTests::test_tensor_item_capture, test/dynamo/test_misc.py::MiscTests::test_tensor_item_no_capture, test/dynamo/test_misc.py::MiscTests::test_tensor_iter, test/dynamo/test_misc.py::MiscTests::test_tensor_layout, test/dynamo/test_misc.py::MiscTests::test_tensor_setattr_getset_descriptor, test/dynamo/test_misc.py::MiscTests::test_tensor_types, test/dynamo/test_misc.py::MiscTests::test_thread_local_setattr, test/dynamo/test_misc.py::MiscTests::test_tolist, test/dynamo/test_misc.py::MiscTests::test_tolist_0d, test/dynamo/test_misc.py::MiscTests::test_tolist_1d, test/dynamo/test_misc.py::MiscTests::test_tolist_float, test/dynamo/test_misc.py::MiscTests::test_tolist_kd, test/dynamo/test_misc.py::MiscTests::test_tolist_kd_dynamic, test/dynamo/test_misc.py::MiscTests::test_tolist_scalar, test/dynamo/test_misc.py::MiscTests::test_top_package_import, test/dynamo/test_misc.py::MiscTests::test_torch_check, test/dynamo/test_misc.py::MiscTests::test_torch_check_nonnegative, test/dynamo/test_misc.py::MiscTests::test_torch_check_symbolic_shape_rel, test/dynamo/test_misc.py::MiscTests::test_torch_compile_ctx_on_forward_and_training_step, test/dynamo/test_misc.py::MiscTests::test_torch_distributions_lazy_property, test/dynamo/test_misc.py::MiscTests::test_torch_dtype_python_type, test/dynamo/test_misc.py::MiscTests::test_torch_dynamo_codegen_pow, test/dynamo/test_misc.py::MiscTests::test_torch_generator_set_state, test/dynamo/test_misc.py::MiscTests::test_torch_guards_stack_frame_register_inlining, test/dynamo/test_misc.py::MiscTests::test_torch_guards_stack_frame_register_inlining_deep, test/dynamo/test_misc.py::MiscTests::test_torch_nn_parameter_isinstance, test/dynamo/test_misc.py::MiscTests::test_torch_objects_as_keys, test/dynamo/test_misc.py::MiscTests::test_torch_package_working_with_trace, test/dynamo/test_misc.py::MiscTests::test_torch_seed, test/dynamo/test_misc.py::MiscTests::test_torch_size, test/dynamo/test_misc.py::MiscTests::test_torch_size_numel, test/dynamo/test_misc.py::MiscTests::test_torch_size_numel_dynamic, test/dynamo/test_misc.py::MiscTests::test_torch_variable_hasattr, test/dynamo/test_misc.py::MiscTests::test_trace_ndarray_frame, test/dynamo/test_misc.py::MiscTests::test_trace_ndarray_frame_2, test/dynamo/test_misc.py::MiscTests::test_tracing_nested_py_tree_mixed_all, test/dynamo/test_misc.py::MiscTests::test_tuple_class, test/dynamo/test_misc.py::MiscTests::test_tuple_from_tuple_iter, test/dynamo/test_misc.py::MiscTests::test_tuple_hasattr, test/dynamo/test_misc.py::MiscTests::test_tuple_iadd_with_shape, test/dynamo/test_misc.py::MiscTests::test_tuple_mul, test/dynamo/test_misc.py::MiscTests::test_tuple_mul_with_shape, test/dynamo/test_misc.py::MiscTests::test_tying_union_new_syntax, test/dynamo/test_misc.py::MiscTests::test_type_copy, test/dynamo/test_misc.py::MiscTests::test_typing_dict, test/dynamo/test_misc.py::MiscTests::test_typing_typevar, test/dynamo/test_misc.py::MiscTests::test_typing_union_and_optional, test/dynamo/test_misc.py::MiscTests::test_typing_union_new_syntax_reconstruct, test/dynamo/test_misc.py::MiscTests::test_typing_variable_isinstance, test/dynamo/test_misc.py::MiscTests::test_unbacked_2d_expand, test/dynamo/test_misc.py::MiscTests::test_unbacked_empty_tensor, test/dynamo/test_misc.py::MiscTests::test_unbacked_repeat_cat, test/dynamo/test_misc.py::MiscTests::test_unbacked_sources_scalar, test/dynamo/test_misc.py::MiscTests::test_unbacked_sources_tensor, test/dynamo/test_misc.py::MiscTests::test_unbacked_strict_mode, test/dynamo/test_misc.py::MiscTests::test_unbacked_symint_split, test/dynamo/test_misc.py::MiscTests::test_unhandled_exception_in_dynamo, test/dynamo/test_misc.py::MiscTests::test_unhandled_exception_in_dynamo2, test/dynamo/test_misc.py::MiscTests::test_unique_consecutive, test/dynamo/test_misc.py::MiscTests::test_unpack4, test/dynamo/test_misc.py::MiscTests::test_unpack5, test/dynamo/test_misc.py::MiscTests::test_unpack_tensor_shape_mismatch, test/dynamo/test_misc.py::MiscTests::test_update_locals_and_stack_uses_shared_cache, test/dynamo/test_misc.py::MiscTests::test_user_code_statically_known, test/dynamo/test_misc.py::MiscTests::test_user_defined_binop, test/dynamo/test_misc.py::MiscTests::test_user_defined_class_name, test/dynamo/test_misc.py::MiscTests::test_user_defined_class_python_type, test/dynamo/test_misc.py::MiscTests::test_user_defined_iter, test/dynamo/test_misc.py::MiscTests::test_user_defined_object_class_interaction, test/dynamo/test_misc.py::MiscTests::test_user_defined_setattr1, test/dynamo/test_misc.py::MiscTests::test_user_defined_setattr2, test/dynamo/test_misc.py::MiscTests::test_user_function_variable_supports_enum_argument, test/dynamo/test_misc.py::MiscTests::test_user_function_variable_supports_function_argument, test/dynamo/test_misc.py::MiscTests::test_user_function_variable_supports_type_abcmeta_argument, test/dynamo/test_misc.py::MiscTests::test_user_getattr1, test/dynamo/test_misc.py::MiscTests::test_user_getattr2, test/dynamo/test_misc.py::MiscTests::test_user_getattribute, test/dynamo/test_misc.py::MiscTests::test_user_property, test/dynamo/test_misc.py::MiscTests::test_usr_cls_classmethod, test/dynamo/test_misc.py::MiscTests::test_usr_cls_staticmethod, test/dynamo/test_misc.py::MiscTests::test_validate_outputs_unbacked, test/dynamo/test_misc.py::MiscTests::test_validate_outputs_unbacked_by_custom_op, test/dynamo/test_misc.py::MiscTests::test_variable_access_in_exception, test/dynamo/test_misc.py::MiscTests::test_variable_tracker_recursively_contains, test/dynamo/test_misc.py::MiscTests::test_version_ci, test/dynamo/test_misc.py::MiscTests::test_with_builtin_type, test/dynamo/test_misc.py::MiscTests::test_write_to_cells_with_name_shadowing, test/dynamo/test_misc.py::MiscTests::test_write_to_closures_in_inlining, test/dynamo/test_misc.py::MiscTests::test_writes_to_cells_across_frames1, test/dynamo/test_misc.py::MiscTests::test_writes_to_cells_across_frames2, test/dynamo/test_misc.py::MiscTests::test_yield_from, test/dynamo/test_misc.py::MiscTests::test_yield_from_in_a_loop, test/dynamo/test_misc.py::MiscTests::test_yield_from_user_stop_iteration, test/dynamo/test_misc.py::MiscTests::test_yield_gen_and_from, test/dynamo/test_misc.py::MiscTests::test_yield_send_to_subgenerator_graph_break, test/dynamo/test_misc.py::MiscTestsPyTree::test_pytree_register_constant_with_side_effect, test/dynamo/test_misc.py::MiscTestsPyTree::test_pytree_tree_flatten_unflatten_cxx, test/dynamo/test_misc.py::MiscTestsPyTree::test_pytree_tree_flatten_unflatten_native_optree, test/dynamo/test_misc.py::MiscTestsPyTree::test_pytree_tree_flatten_unflatten_python, test/dynamo/test_misc.py::MiscTestsPyTree::test_pytree_tree_leaves_cxx, test/dynamo/test_misc.py::MiscTestsPyTree::test_pytree_tree_leaves_native_optree, test/dynamo/test_misc.py::MiscTestsPyTree::test_pytree_tree_leaves_python, test/dynamo/test_misc.py::MiscTestsPyTree::test_pytree_tree_map_cxx, test/dynamo/test_misc.py::MiscTestsPyTree::test_pytree_tree_map_dict_order_cxx, test/dynamo/test_misc.py::MiscTestsPyTree::test_pytree_tree_map_dict_order_native_optree, test/dynamo/test_misc.py::MiscTestsPyTree::test_pytree_tree_map_dict_order_python, test/dynamo/test_misc.py::MiscTestsPyTree::test_pytree_tree_map_native_optree, test/dynamo/test_misc.py::MiscTestsPyTree::test_pytree_tree_map_only_cxx, test/dynamo/test_misc.py::MiscTestsPyTree::test_pytree_tree_map_only_native_optree, test/dynamo/test_misc.py::MiscTestsPyTree::test_pytree_tree_map_only_python, test/dynamo/test_misc.py::MiscTestsPyTree::test_pytree_tree_map_python, test/dynamo/test_misc.py::MiscTestsPyTree::test_tracing_nested_dicts_cxx, test/dynamo/test_misc.py::MiscTestsPyTree::test_tracing_nested_dicts_native_optree, test/dynamo/test_misc.py::MiscTestsPyTree::test_tracing_nested_dicts_python, test/dynamo/test_misc.py::MiscTestsPyTree::test_tracing_nested_mixed_all_cxx, test/dynamo/test_misc.py::MiscTestsPyTree::test_tracing_nested_mixed_all_native_optree, test/dynamo/test_misc.py::MiscTestsPyTree::test_tracing_nested_mixed_all_python, test/dynamo/test_misc.py::MiscTestsPyTree::test_tracing_nested_pytree_cxx, test/dynamo/test_misc.py::MiscTestsPyTree::test_tracing_nested_pytree_native_optree, test/dynamo/test_misc.py::MiscTestsPyTree::test_tracing_nested_pytree_python, test/dynamo/test_misc.py::MiscTestsPyTree::test_tracing_nested_tensor_subclass_cxx, test/dynamo/test_misc.py::MiscTestsPyTree::test_tracing_nested_tensor_subclass_native_optree, test/dynamo/test_misc.py::MiscTestsPyTree::test_tracing_nested_tensor_subclass_python, test/dynamo/test_misc.py::MiscTestsPyTree::test_tracing_nested_tuples_cxx, test/dynamo/test_misc.py::MiscTestsPyTree::test_tracing_nested_tuples_native_optree, test/dynamo/test_misc.py::MiscTestsPyTree::test_tracing_nested_tuples_python, test/dynamo/test_misc.py::MiscTestsPyTree::test_tracing_pytree_cxx, test/dynamo/test_misc.py::MiscTestsPyTree::test_tracing_pytree_native_optree, test/dynamo/test_misc.py::MiscTestsPyTree::test_tracing_pytree_python, test/dynamo/test_misc.py::TestTracer::test_jit_save, test/dynamo/test_misc.py::TestCustomFunction::test_autograd_function_with_matmul_folding_at_output, test/dynamo/test_misc.py::TestCustomFunction::test_retain_grad, test/dynamo/test_misc.py::MiscTestsDeviceCUDA::test_dynamic_fill_diagonal__cuda, test/dynamo/test_misc.py::MiscTestsDeviceCUDA::test_dynamic_float_scalar_tensor_coersion_cuda, test/dynamo/test_misc.py::MiscTestsDeviceCUDA::test_full_graph_capture_dynamic_output_shape_ops_cuda, test/dynamo/test_misc.py::MiscTestsDeviceCUDA::test_full_graph_capture_scalar_outputs_cuda, test/dynamo/test_misc.py::MiscTestsDeviceCUDA::test_get_device_cuda, test/dynamo/test_misc.py::MiscTestsDeviceCUDA::test_gpu_set_device_cuda, test/dynamo/test_misc.py::MiscTestsDeviceCUDA::test_interpolate_propagate_real_tensors_cuda, test/dynamo/test_misc.py::MiscTestsDeviceCUDA::test_legacy_cuda_tensor_cuda, test/dynamo/test_misc.py::MiscTestsDeviceCUDA::test_parsing_sdpa_cuda, test/dynamo/test_misc.py::MiscTestsDeviceCUDA::test_rand_cuda, test/dynamo/test_misc.py::MiscTestsDeviceCUDA::test_randint_no_graphbreak_cuda, test/dynamo/test_misc.py::MiscTestsDeviceCUDA::test_scalar_isin_decomposition_cuda, test/dynamo/test_misc.py::MiscTestsDeviceCUDA::test_symint_as_device_kwarg_cuda, test/dynamo/test_misc.py::MiscTestsDeviceCUDA::test_torch_cudnn_is_acceptable_bad_inputs_cuda, test/dynamo/test_misc.py::MiscTestsDeviceCUDA::test_torch_cudnn_is_acceptable_cuda, test/dynamo/test_misc.py::MiscTestsDeviceCUDA::test_torch_device_is_available_cuda, test/dynamo/test_misc.py::MiscTestsDeviceCUDA::test_torch_device_python_type_cuda, test/dynamo/test_misc.py::DynamoOpPromotionTests::test_symbool_guard_or_false, test/dynamo/test_misc.py::DynamoOpPromotionTests::test_symbool_tensor_mul, test/dynamo/test_misc.py::DynamoOpPromotionTests::test_symbool_tensor_mul_does_not_fail, test/dynamo/test_misc.py::DynamoOpPromotionTests::test_tensorify_track_item_symint 2025-12-04T15:52:07.3029522Z 2025-12-04T15:52:07.3029848Z Finished dynamo/test_misc 1/1 ... [2025-12-04 15:52:07.259926][23910.950316225], took 2.75min 2025-12-04T15:52:07.3031015Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_misc/dynamo.test_misc-43b95b63c3862211.xml 2025-12-04T15:52:07.3941380Z Running inductor/test_aot_inductor_arrayref 2/2 ... [2025-12-04 15:52:07.393845][23911.084237505] 2025-12-04T15:52:07.3942009Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T15:52:07.3944980Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_aot_inductor_arrayref.py', '--shard-id=2', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 15:52:07.394239] 2025-12-04T16:01:51.1023310Z 2025-12-04T16:01:51.1027225Z inductor/test_aot_inductor_arrayref 2/2 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_aot_inductor_arrayref_2.2_d8fa4d4f4ba9cf83_.log 2025-12-04T16:01:51.1155839Z Running 150 items in this shard: test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test__int_mm_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_1_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test__weight_int4pack_mm_m_32_n_64_q_group_32_num_groups_2_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test__weight_int4pack_mm_m_32_n_64_q_group_64_num_groups_1_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test__weight_int4pack_mm_with_scales_and_zeros_m_32_n_64_q_group_32_num_groups_1_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test__weight_int4pack_mm_with_scales_and_zeros_m_32_n_64_q_group_64_num_groups_2_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_add_complex_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_addmm_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_addmm_multiple_dynamic_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_aliased_buffer_reuse_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_aot_inductor_consts_cpp_build_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_aoti_debug_printer_cpp_kernel_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_aoti_runtime_asserts_backed_symint_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_aoti_user_defined_triton_kernel_profiling_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_assert_async_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_assert_tensor_meta_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_autotune_with_constant_folding_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_autotuning_args_reuse_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_bmm_multiple_dynamic_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_boolean_indexing_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_buffer_mutation_4_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_buffer_mutation_and_force_mmap_weights_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_cond_cpu_predicate_cuda_operands_max_autotune_False_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_cond_mismatched_branch_output_dynamic_False_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_cond_non_tensor_predicates_dynamic_True_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_cond_predicate_on_cpu_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_cond_simple_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_cond_symint_input_disable_one_pass_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_cond_unbacked_symint_closure_dynamic_False_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_cond_with_multiple_outputs_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_cond_with_outer_code_before_after_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_cond_with_parameters_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_cond_with_replace_view_ops_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_constant_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_constant_folding_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_constant_original_fqn_and_dtype_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_constant_type_propagation_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_conv_freezing_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_copy_non_blocking_is_pinned_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_dup_unbacked_sym_decl_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_dup_unbacked_sym_decl_with_refinement_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_duplicate_constant_folding_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_dynamic_cat_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_embedding_bag_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_empty_graph_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_extract_constants_map_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_fallback_kernel_with_symexpr_output_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_fft_c2c_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_fp8_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_fp8_view_of_param_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_freezing_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_fx_gm_return_tuple_validation_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_index_put_with_none_index_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_inf_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_input_codegen_with_sympy_expr_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_int_list_input_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_issue_140766_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_large_mmaped_weights_on_disk_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_large_weight_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_linear_freezing_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_load_package_multiple_gpus_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_masked_select_dynamic_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_misaligned_input_1_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_misaligned_input_2_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_misc_1_max_autotune_False_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_misc_1_max_autotune_True_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_multiple_output_alias_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_narrow_fallback_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_no_args_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_non_contiguous_output_alias_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_non_tensor_input_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_pad_fallback_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_pad_non_zero_memory_leak_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_proxy_executor_abs_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_proxy_executor_hann_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_proxy_executor_squeeze_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_pytree_inputs_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_quantized_linear_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_repeat_interleave_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_repeat_output_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_repeated_calling_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_return_constant_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_reuse_kernel_dynamic_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_rocm_triton_autotuning_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_run_with_grad_enabled_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_runtime_checks_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_runtime_checks_device_type_failed_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_scaled_dot_product_efficient_attention_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_shifted_constraint_ranges_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_simple_embed_kernel_binary_False_max_autotune_True_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_simple_multi_arch_embed_kernel_binary_False_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_simple_split_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_size_with_unbacked_add_and_mul_expr_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_size_with_unbacked_add_expr_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_small_constant_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_so_without_weight_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_stft_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_sym_expr_indexing_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_symbool_item_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_symfloat_item_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_sympy_cpp_printer_min_max_minmax1_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_torchvision_transforms_functional_tensor_resize_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_dynamic_launcher_grid_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_dynamic_launcher_grid_infer_from_tensor_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_dynamic_shape_with_div_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_equal_to_1_arg_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_grid_type_1_num_dims_1_dynamic_False_autotune_False_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_grid_type_1_num_dims_1_dynamic_False_autotune_True_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_grid_type_1_num_dims_1_dynamic_True_autotune_False_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_grid_type_1_num_dims_1_dynamic_True_autotune_True_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_grid_type_1_num_dims_2_dynamic_False_autotune_False_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_grid_type_1_num_dims_2_dynamic_False_autotune_True_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_grid_type_1_num_dims_2_dynamic_True_autotune_True_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_grid_type_2_num_dims_1_dynamic_False_autotune_True_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_grid_type_2_num_dims_2_dynamic_False_autotune_True_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_grid_type_2_num_dims_2_dynamic_True_autotune_True_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_grid_type_3_num_dims_1_dynamic_True_autotune_False_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_grid_type_3_num_dims_2_dynamic_False_autotune_True_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_multi_output_arg_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_on_device_tma_dynamic_False_tma_version_new_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_on_device_tma_dynamic_True_tma_version_old_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_reinterpret_view_mem_leak_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_sympy_fn_like_arg_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_tma_descriptor_1d_dynamic_True_tma_version_old_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_unbacked_symint_in_grid_dynamic_True_autotuning_True_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_weird_param_order_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_kernel_with_none_input_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_triton_next_power_of_2_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_unbacked_equals_input_size_runtime_assertion_mark_unbacked_False_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_unbacked_expr_replacements_shift_k_0_use_static_size_False_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_unbacked_expr_replacements_shift_k_1_use_static_size_True_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_unbacked_expr_replacements_shift_k_2_use_static_size_False_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_unbacked_expr_replacements_shift_k_3_use_static_size_False_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_update_constant_buffer_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_update_constant_buffer_simple_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_update_user_managed_buffer_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_upper_bound_i64_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_view_outputs_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_while_loop_nested_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_while_loop_simple_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_while_loop_with_conv_dynamic_False_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_while_loop_with_mixed_device_dynamic_False_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_while_loop_with_outer_code_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_while_loop_with_parameters_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_while_loop_with_unbacked_symint_closure_dynamic_False_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_with_cudagraphs_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_with_no_triton_profiler_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_with_offset_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_with_profiler_cpu_with_stack_allocation, test/inductor/test_aot_inductor_arrayref.py::AOTInductorTestABICompatibleCpuWithStackAllocation::test_zero_grid_with_backed_symbols_cpu_with_stack_allocation 2025-12-04T16:01:51.1282809Z 2025-12-04T16:01:51.1283232Z Finished inductor/test_aot_inductor_arrayref 2/2 ... [2025-12-04 16:01:51.102438][24494.792829468], took 9.73min 2025-12-04T16:01:51.1443957Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_aot_inductor_arrayref/inductor.test_aot_inductor_arrayref-d7936498b94cbe56.xml 2025-12-04T16:01:51.2118594Z Running inductor/test_halide 1/1 ... [2025-12-04 16:01:51.211536][24494.901927415] 2025-12-04T16:01:51.2119315Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:01:51.2122374Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_halide.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:01:51.211955] 2025-12-04T16:02:03.1781579Z 2025-12-04T16:02:03.1782578Z inductor/test_halide 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_halide_1.1_8d6772cbd633567a_.log 2025-12-04T16:02:03.1783373Z 2025-12-04T16:02:03.1783942Z Finished inductor/test_halide 1/1 ... [2025-12-04 16:02:03.177947][24506.868340663], took 0.20min 2025-12-04T16:02:03.2194209Z Running inductor/test_xpu_basic 1/1 ... [2025-12-04 16:02:03.219121][24506.909514262] 2025-12-04T16:02:03.2194790Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:02:03.2197760Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_xpu_basic.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:02:03.219529] 2025-12-04T16:02:15.2235180Z 2025-12-04T16:02:15.2236446Z inductor/test_xpu_basic 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_xpu_basic_1.1_8048a1f844b679fa_.log 2025-12-04T16:02:15.2237290Z 2025-12-04T16:02:15.2238258Z Finished inductor/test_xpu_basic 1/1 ... [2025-12-04 16:02:15.223278][24518.913670853], took 0.20min 2025-12-04T16:02:15.2648953Z Running inductor/test_provenance_tracing 1/1 ... [2025-12-04 16:02:15.264497][24518.954889392] 2025-12-04T16:02:15.2649606Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:02:15.2651511Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_provenance_tracing.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:02:15.264900] 2025-12-04T16:03:44.0615864Z 2025-12-04T16:03:44.0617076Z inductor/test_provenance_tracing 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_provenance_tracing_1.1_fa80dd40f7fc5a69_.log 2025-12-04T16:03:44.0627989Z Running 16 items in this shard: test/inductor/test_provenance_tracing.py::TestProvenanceTracingArtifact::test_triton_kernel_to_post_grad_tracing_combo_kernel, test/inductor/test_provenance_tracing.py::TestProvenanceTracingArtifact::test_triton_kernel_to_post_grad_tracing_cpu, test/inductor/test_provenance_tracing.py::TestProvenanceTracingArtifact::test_triton_kernel_to_post_grad_tracing_cuda, test/inductor/test_provenance_tracing.py::TestProvenanceTracingArtifact::test_triton_kernel_to_post_grad_tracing_extern_kernel, test/inductor/test_provenance_tracing.py::TestProvenanceTracingNodeMapping::test_create_node_mapping, test/inductor/test_provenance_tracing.py::TestProvenanceTracingNodeMeta::test_pattern_matcher_transfer_meta, test/inductor/test_provenance_tracing.py::TestProvenanceTracingStackTraces::test_cpu_extern_kernel, test/inductor/test_provenance_tracing.py::TestProvenanceTracingStackTraces::test_create_kernel_information_json_function, test/inductor/test_provenance_tracing.py::TestProvenanceTracingStackTraces::test_deferred_triton_kernels, test/inductor/test_provenance_tracing.py::TestProvenanceTracingStackTraces::test_kernel_information_generation, test/inductor/test_provenance_tracing.py::TestProvenanceTracingStackTraces::test_no_kernel_information_without_provenance_tracking, test/inductor/test_provenance_tracing.py::TestProvenanceTracingStackTraces::test_tlparse_kernel_stack_traces, test/inductor/test_provenance_tracing.py::TestProvenanceTracingKernelContextCpu::test_aoti_python_stack_traces_cpu, test/inductor/test_provenance_tracing.py::TestProvenanceTracingKernelContextCpu::test_jit_inductor_with_flag_cpu, test/inductor/test_provenance_tracing.py::TestProvenanceTracingKernelContextGpu::test_aoti_python_stack_traces_cuda, test/inductor/test_provenance_tracing.py::TestProvenanceTracingKernelContextGpu::test_jit_inductor_with_flag_cuda 2025-12-04T16:03:44.0637975Z 2025-12-04T16:03:44.0638403Z Finished inductor/test_provenance_tracing 1/1 ... [2025-12-04 16:03:44.061378][24607.751768776], took 1.48min 2025-12-04T16:03:44.1039609Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_provenance_tracing/inductor.test_provenance_tracing-e41a43dd8ef6019c.xml 2025-12-04T16:03:44.1947595Z Running dynamo/test_buffers_override 1/1 ... [2025-12-04 16:03:44.194340][24607.884731276] 2025-12-04T16:03:44.1948287Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:03:44.1950358Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_buffers_override.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:03:44.194797] 2025-12-04T16:03:59.8813813Z 2025-12-04T16:03:59.8815597Z dynamo/test_buffers_override 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_buffers_override_1.1_f8bb7c2759bc3846_.log 2025-12-04T16:03:59.8819090Z Running 2 items in this shard: test/dynamo/test_buffers_override.py::TestBuffersOverride::test_buffers_override, test/dynamo/test_buffers_override.py::TestBuffersOverride::test_named_buffers_override 2025-12-04T16:03:59.8821100Z 2025-12-04T16:03:59.8821821Z Finished dynamo/test_buffers_override 1/1 ... [2025-12-04 16:03:59.881133][24623.571526546], took 0.26min 2025-12-04T16:03:59.9232783Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_buffers_override/dynamo.test_buffers_override-01fc103efe1995d6.xml 2025-12-04T16:03:59.9967493Z Running inductor/test_inplacing_pass 1/1 ... [2025-12-04 16:03:59.996416][24623.686808216] 2025-12-04T16:03:59.9968506Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:03:59.9972911Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_inplacing_pass.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:03:59.996919] 2025-12-04T16:04:22.9471548Z 2025-12-04T16:04:22.9472780Z inductor/test_inplacing_pass 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_inplacing_pass_1.1_1204e772d28433be_.log 2025-12-04T16:04:22.9486349Z Running 22 items in this shard: test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_counters_functionalize_old, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_counters_functionalize_v2, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_dont_modify_input, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_dont_modify_live, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_dont_modify_view_of_live, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_generalized_scatter, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_lists_functionalize_v2, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_lists_old_functionalize, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_multi_output_intermediate, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_multiple_intermediate, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_multiple_mutations, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_partitioner_recomputes_factory_empty_like_sin_op, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_partitioner_recomputes_factory_empty_like_sin_triton, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_partitioner_recomputes_factory_ones_like_sin_op, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_partitioner_recomputes_factory_ones_like_sin_triton, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_should_modify_inner, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_should_modify_input, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_view_inplaced2_functionalize_v2, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_view_inplaced_functionalize_v2, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_views_not_inplaced2_functionalize_v2, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_views_not_inplaced3_functionalize_v2, test/inductor/test_inplacing_pass.py::TestReinplacingPassCorrectness::test_views_not_inplaced_functionalize_v2 2025-12-04T16:04:22.9499202Z 2025-12-04T16:04:22.9499597Z Finished inductor/test_inplacing_pass 1/1 ... [2025-12-04 16:04:22.946967][24646.63736136], took 0.38min 2025-12-04T16:04:22.9895245Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_inplacing_pass/inductor.test_inplacing_pass-d670df374fd1e952.xml 2025-12-04T16:04:23.0946611Z Running inductor/test_aot_inductor_custom_ops 1/1 ... [2025-12-04 16:04:23.094304][24646.784695272] 2025-12-04T16:04:23.0947281Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:04:23.0950260Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_aot_inductor_custom_ops.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:04:23.094780] 2025-12-04T16:07:47.8658267Z 2025-12-04T16:07:47.8659616Z inductor/test_aot_inductor_custom_ops 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_aot_inductor_custom_ops_1.1_663d4eb0571b35c8_.log 2025-12-04T16:07:47.8682482Z Running 35 items in this shard: test/inductor/test_aot_inductor_custom_ops.py::AOTInductorLoggingTest::test_shape_env_reuse, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_boxed_run_inputs_clearing_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_add_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_add_output_path_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_all_inputs_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_missing_arg_with_default_value_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_out_variant_without_return_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_return_list_of_single_tensor_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_return_single_tensor_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_square_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_with_concat_inputs_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_with_multiple_outputs_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_with_reinterpret_view_inputs_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_fn_with_int_output_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_fn_with_optional_tensor_nullopt_output_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_fn_with_optional_tensor_output_2_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_fn_with_optional_tensor_output_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_incorrect_custom_op_schema_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleGpu::test_boxed_run_inputs_clearing_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleGpu::test_custom_op_add_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleGpu::test_custom_op_add_output_path_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleGpu::test_custom_op_all_inputs_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleGpu::test_custom_op_missing_arg_with_default_value_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleGpu::test_custom_op_out_variant_without_return_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleGpu::test_custom_op_return_list_of_single_tensor_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleGpu::test_custom_op_return_single_tensor_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleGpu::test_custom_op_square_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleGpu::test_custom_op_with_concat_inputs_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleGpu::test_custom_op_with_multiple_outputs_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleGpu::test_custom_op_with_reinterpret_view_inputs_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleGpu::test_fn_with_int_output_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleGpu::test_fn_with_optional_tensor_nullopt_output_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleGpu::test_fn_with_optional_tensor_output_2_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleGpu::test_fn_with_optional_tensor_output_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleGpu::test_incorrect_custom_op_schema_cuda 2025-12-04T16:07:47.8704145Z 2025-12-04T16:07:47.8704562Z Finished inductor/test_aot_inductor_custom_ops 1/1 ... [2025-12-04 16:07:47.865845][24851.556232782], took 3.41min 2025-12-04T16:07:47.9081453Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_aot_inductor_custom_ops/inductor.test_aot_inductor_custom_ops-a0ad7eb32926bcd8.xml 2025-12-04T16:07:48.0020707Z Running inductor/test_split_cat_fx_passes 1/1 ... [2025-12-04 16:07:48.001755][24851.692146794] 2025-12-04T16:07:48.0021332Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:07:48.0024732Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_split_cat_fx_passes.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:07:48.002220] 2025-12-04T16:09:05.7858149Z 2025-12-04T16:09:05.7859257Z inductor/test_split_cat_fx_passes 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_split_cat_fx_passes_1.1_64f7c65379e18b89_.log 2025-12-04T16:09:05.7865528Z Running 11 items in this shard: test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_cat_normalization, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_config_flag_is_respected, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_consecutive_split_merge, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_numpy_compat_normalization, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_split_cat_merge, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_split_cat_merge_mutation, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_split_cat_new_patterns, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_split_normalization, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_split_squeeze, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_stack_normalization_axis_kwarg, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_unbind_stack 2025-12-04T16:09:05.7871027Z 2025-12-04T16:09:05.7871412Z Finished inductor/test_split_cat_fx_passes 1/1 ... [2025-12-04 16:09:05.785591][24929.47598565], took 1.30min 2025-12-04T16:09:05.8283079Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_split_cat_fx_passes/inductor.test_split_cat_fx_passes-4aa43e0fa2d59cfb.xml 2025-12-04T16:09:07.3434422Z Uploading artifacts took 1.44 seconds 2025-12-04T16:09:07.3438751Z Running inductor/test_profiler 1/1 ... [2025-12-04 16:09:07.343694][24931.034084816] 2025-12-04T16:09:07.3439329Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:09:07.3444127Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_profiler.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:09:07.344170] 2025-12-04T16:09:26.7437138Z 2025-12-04T16:09:26.7438158Z inductor/test_profiler 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_profiler_1.1_eb004b5dbdc54b3f_.log 2025-12-04T16:09:26.7442299Z Running 7 items in this shard: test/inductor/test_profiler.py::DynamoProfilerTests::test_cupti_lazy_reinit, test/inductor/test_profiler.py::DynamoProfilerTests::test_inductor_profiling_kernel_names_foreach, test/inductor/test_profiler.py::DynamoProfilerTests::test_inductor_profiling_kernel_names_pointwise, test/inductor/test_profiler.py::DynamoProfilerTests::test_inductor_profiling_kernel_names_template, test/inductor/test_profiler.py::DynamoProfilerTests::test_inductor_profiling_triton_hooks, test/inductor/test_profiler.py::DynamoProfilerTests::test_inductor_profiling_triton_launch, test/inductor/test_profiler.py::DynamoProfilerTests::test_pt2_triton_attributes 2025-12-04T16:09:26.7445733Z 2025-12-04T16:09:26.7446077Z Finished inductor/test_profiler 1/1 ... [2025-12-04 16:09:26.743518][24950.43391211], took 0.32min 2025-12-04T16:09:26.7862092Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_profiler/inductor.test_profiler-cdccc94d406caaec.xml 2025-12-04T16:09:26.8745184Z Running inductor/test_memory_planning 1/1 ... [2025-12-04 16:09:26.874237][24950.564627544] 2025-12-04T16:09:26.8745788Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:09:26.8749333Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_memory_planning.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:09:26.874671] 2025-12-04T16:09:54.7803217Z 2025-12-04T16:09:54.7804728Z inductor/test_memory_planning 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_memory_planning_1.1_7d8eac35fee71fb8_.log 2025-12-04T16:09:54.7807932Z Running 4 items in this shard: test/inductor/test_memory_planning.py::TestMemoryPlanning::test_aoti, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_cpp_wrapper, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_python_wrapper, test/inductor/test_memory_planning.py::TestMemoryPlanning::test_unbacked_symint 2025-12-04T16:09:54.7809674Z 2025-12-04T16:09:54.7810060Z Finished inductor/test_memory_planning 1/1 ... [2025-12-04 16:09:54.780129][24978.470523428], took 0.47min 2025-12-04T16:09:54.8226583Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_memory_planning/inductor.test_memory_planning-480d70b98c6f91e2.xml 2025-12-04T16:09:54.9129776Z Running inductor/test_mem_estimation 1/1 ... [2025-12-04 16:09:54.912561][24978.602953141] 2025-12-04T16:09:54.9130779Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:09:54.9132789Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_mem_estimation.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:09:54.912990] 2025-12-04T16:10:06.9445811Z 2025-12-04T16:10:06.9447311Z inductor/test_mem_estimation 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_mem_estimation_1.1_db7eba60d0c86675_.log 2025-12-04T16:10:06.9450380Z Running 4 items in this shard: test/inductor/test_mem_estimation.py::TestMemoryProfilingResNet::test_conv_network, test/inductor/test_mem_estimation.py::TestMemoryProfilingResNet::test_simple_linear_layers, test/inductor/test_mem_estimation.py::TestMemoryTracker::test_memory_tracker_different_scheduling, test/inductor/test_mem_estimation.py::TestMemoryTracker::test_memory_tracker_original_order 2025-12-04T16:10:06.9452589Z 2025-12-04T16:10:06.9452966Z Finished inductor/test_mem_estimation 1/1 ... [2025-12-04 16:10:06.944355][24990.634749159], took 0.20min 2025-12-04T16:10:06.9869453Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_mem_estimation/inductor.test_mem_estimation-f5f67f53fc542f3b.xml 2025-12-04T16:10:07.0591874Z Running dynamo/test_view 1/1 ... [2025-12-04 16:10:07.058827][24990.749218153] 2025-12-04T16:10:07.0592640Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:10:07.0595441Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_view.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:10:07.059243] 2025-12-04T16:10:15.2853547Z 2025-12-04T16:10:15.2854906Z dynamo/test_view 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_view_1.1_285b149edbf19f8b_.log 2025-12-04T16:10:15.2857727Z Running 6 items in this shard: test/dynamo/test_view.py::ViewTests::test_tensor_view_with_tensor_args, test/dynamo/test_view.py::ViewTests::test_tensor_view_with_tensor_shape_params, test/dynamo/test_view.py::ViewTests::test_torch_reshape_with_tensor_shape_params, test/dynamo/test_view.py::ViewTests::test_view_to_1d, test/dynamo/test_view.py::ViewTests::test_view_to_2d, test/dynamo/test_view.py::ViewTests::test_view_with_tensor_shape_params 2025-12-04T16:10:15.2859840Z 2025-12-04T16:10:15.2860145Z Finished dynamo/test_view 1/1 ... [2025-12-04 16:10:15.285111][24998.975505838], took 0.14min 2025-12-04T16:10:15.3276042Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_view/dynamo.test_view-5c939a16d5367651.xml 2025-12-04T16:10:15.4259833Z Running inductor/test_cutlass_evt 1/1 ... [2025-12-04 16:10:15.425638][24999.116029906] 2025-12-04T16:10:15.4260581Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:10:15.4263724Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_cutlass_evt.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:10:15.426058] 2025-12-04T16:10:25.4043332Z 2025-12-04T16:10:25.4044581Z inductor/test_cutlass_evt 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_cutlass_evt_1.1_5894d03fbbc62b34_.log 2025-12-04T16:10:25.4049559Z Running 8 items in this shard: test/inductor/test_cutlass_evt.py::TestCutlassEVT::test_evt_argument_codegen, test/inductor/test_cutlass_evt.py::TestCutlassEVT::test_evt_argument_codegen_return_accumulator, test/inductor/test_cutlass_evt.py::TestCutlassEVT::test_evt_codegen, test/inductor/test_cutlass_evt.py::TestCutlassEVT::test_example_tensor_creation, test/inductor/test_cutlass_evt.py::TestCutlassEVT::test_py_codegen, test/inductor/test_cutlass_evt.py::TestCutlassEVT::test_py_codegen_accumulator_return, test/inductor/test_cutlass_evt.py::TestCutlassEVT::test_py_codegen_broadcasting, test/inductor/test_cutlass_evt.py::TestCutlassEVT::test_py_codegen_disjoint_read_indexing 2025-12-04T16:10:25.4052953Z 2025-12-04T16:10:25.4053327Z Finished inductor/test_cutlass_evt 1/1 ... [2025-12-04 16:10:25.404156][25009.094546765], took 0.17min 2025-12-04T16:10:25.4468712Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_cutlass_evt/inductor.test_cutlass_evt-a51e1bccbd1b02fd.xml 2025-12-04T16:10:25.5192921Z Running dynamo/test_reconstruct 1/1 ... [2025-12-04 16:10:25.518922][25009.209314037] 2025-12-04T16:10:25.5193707Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:10:25.5196661Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_reconstruct.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:10:25.519347] 2025-12-04T16:10:39.1529214Z 2025-12-04T16:10:39.1530799Z dynamo/test_reconstruct 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_reconstruct_1.1_abcfd742020bd6ec_.log 2025-12-04T16:10:39.1538681Z Running 16 items in this shard: test/dynamo/test_reconstruct.py::ReconstructTest::test_ConstDict_clear_reconstruct, test/dynamo/test_reconstruct.py::ReconstructTest::test_ConstDict_del_reconstruct, test/dynamo/test_reconstruct.py::ReconstructTest::test_ConstDict_get_reconstruct, test/dynamo/test_reconstruct.py::ReconstructTest::test_ConstDict_optimize_reconstruct, test/dynamo/test_reconstruct.py::ReconstructTest::test_ConstDict_pop_reconstruct, test/dynamo/test_reconstruct.py::ReconstructTest::test_ConstDict_popitem_reconstruct, test/dynamo/test_reconstruct.py::ReconstructTest::test_ConstDict_popitem_reconstruct_graph_break, test/dynamo/test_reconstruct.py::ReconstructTest::test_create_dict_reconstruct, test/dynamo/test_reconstruct.py::ReconstructTest::test_functional_call_reconstruct, test/dynamo/test_reconstruct.py::ReconstructTest::test_functional_call_reconstruct_2, test/dynamo/test_reconstruct.py::ReconstructTest::test_graph_break_in_wrapped_nested_function, test/dynamo/test_reconstruct.py::ReconstructTest::test_graph_break_in_wrapped_skipped_function, test/dynamo/test_reconstruct.py::ReconstructTest::test_graph_break_in_wrapped_user_function, test/dynamo/test_reconstruct.py::ReconstructTest::test_graph_break_in_wrapped_user_method, test/dynamo/test_reconstruct.py::ReconstructTest::test_tma_experimental_reconstruct, test/dynamo/test_reconstruct.py::ReconstructTest::test_tma_stable_reconstruct 2025-12-04T16:10:39.1546039Z 2025-12-04T16:10:39.1546393Z Finished dynamo/test_reconstruct 1/1 ... [2025-12-04 16:10:39.152700][25022.843092674], took 0.23min 2025-12-04T16:10:39.1953671Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_reconstruct/dynamo.test_reconstruct-7d75ef23ae79ec60.xml 2025-12-04T16:10:39.2685243Z Running dynamo/test_aot_autograd 1/1 ... [2025-12-04 16:10:39.268141][25022.958531417] 2025-12-04T16:10:39.2685861Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:10:39.2688398Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_aot_autograd.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:10:39.268555] 2025-12-04T16:11:52.2381251Z 2025-12-04T16:11:52.2382404Z dynamo/test_aot_autograd 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_aot_autograd_1.1_a815d5749913f0e4_.log 2025-12-04T16:11:52.2408855Z Running 51 items in this shard: test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_LSTM, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_alias_inputs, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_aot_autograd_expand_mutation_backwards, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_aot_autograd_expand_mutation_error, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_aot_autograd_expand_mutation_functionalizes, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_aot_autograd_raises_invalid_leaf_set, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_aot_autograd_stride_reconstruction_on_zero_dim_dynamic_shaped_tensor, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_aot_export_joint_simple_repro, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_aot_grad_mode_mutation, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_aot_sequence_nr, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_dupe_via_dynamo_recompiles, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_dupe_via_dynamo_recompiles_many_args, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_dupe_via_dynamo_recompiles_many_args_param, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_dupe_via_dynamo_recompiles_many_args_param_non_tensor_arg, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_dupe_via_dynamo_recompiles_many_args_param_non_tensor_arg_list, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_arg_dupe_via_dynamo_recompiles_many_with_global, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_autograd_function_tangent_mutation, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_call_fn_with_non_const_inputs_aot_safe, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_call_fn_with_non_const_inputs_aot_unsafe, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_call_fn_with_non_const_inputs_aot_unsafe_control_flow, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_data_ptr_access_copy, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_data_ptr_access_fails_in_backward, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_data_ptr_access_fails_in_forward, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_different_inputs_overlapping_set_with_mutation, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_donated_buffer1, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_donated_buffer2, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_donated_buffer3, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_donated_buffer4, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_donated_buffer5, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_donated_buffer6, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_donated_buffer_with_retain_or_create_graph1, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_donated_buffer_with_retain_or_create_graph2, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_donated_buffer_with_retain_or_create_graph3, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_donated_buffer_with_retain_or_create_graph4, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_double_backward_errors, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_eager_sequence_nr, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_grad_inputs_alias_inputs, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_inputs_overlapping_with_mutation_recompile, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_inputs_overlapping_with_mutation_stress, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_joint_custom_pass, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_multiple_aot_autograd_calls_dupe_args, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_mutation, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_mutation1, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_negative_testing, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_negative_testing_mutation, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_nn_parameter_construction, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_no_storage_overlap_guards_no_aliasing, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_no_storage_overlap_guards_no_mutation, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_requires_grad_fake_via_dynamo_recompiles, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_split_with_sizes_aot_autograd_cleans_up_traceback_meta, test/dynamo/test_aot_autograd.py::AotAutogradFallbackTests::test_unbacked_activation_specialized_in_inductor 2025-12-04T16:11:52.2434471Z 2025-12-04T16:11:52.2434818Z Finished dynamo/test_aot_autograd 1/1 ... [2025-12-04 16:11:52.238014][25095.928406237], took 1.22min 2025-12-04T16:11:52.2816338Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_aot_autograd/dynamo.test_aot_autograd-89fcf170216dd13a.xml 2025-12-04T16:11:52.3893953Z Running export/test_cpp_serdes 1/1 ... [2025-12-04 16:11:52.389061][25096.079452043] 2025-12-04T16:11:52.3894576Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:11:52.3898020Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'export/test_cpp_serdes.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:11:52.389501] 2025-12-04T16:13:39.7743653Z 2025-12-04T16:13:39.7744768Z export/test_cpp_serdes 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_cpp_serdes_1.1_9da4c12e7e041500_.log 2025-12-04T16:13:39.7958244Z Running 431 items in this shard: test/export/test_cpp_serdes.py::CppSerdesTestExport::test__scaled_dot_product_flash_attention_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_additional_inputs_constants_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_allow_explicit_guards_as_runtime_asserts_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_annotate_on_assert_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_args_type_checked_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_aten_lift_fresh_copy_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_attention_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_attr_assignment_extra_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_automatic_constrain_size_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_automatic_dynamic_shapes_constant_relation_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_automatic_dynamic_shapes_linear_relation_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_automatic_dynamic_shapes_simple_equality_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_baddbmm_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_basic_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_basic_non_strict_fake_tensor_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_basic_non_strict_real_tensor_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_bincount_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_buffer_util_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_capture_subclass_constructor_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_capture_subclass_constructor_torch_ir_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_capture_subclass_wrong_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_ccode_python_mod_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_cdist_forward_compute_mode_zero_export_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_check_specialized_int_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_checks_to_constrain_range_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_cleanup_dynamic_markers_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_colin_unbacked_backed_vr_sub_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_colon_parameter_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_compiling_state_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_cond_access_identical_symint_closure_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_cond_branches_return_constant_int_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_cond_branches_return_same_int_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_cond_buffers_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_cond_contains_unbacked_no_escape_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_cond_int_closure_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_cond_unflatten_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_cond_with_module_stack_export_with_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_cond_with_module_stack_export_with_unflatten_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_constant_aliasing_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_constant_input_naming_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_constant_no_user_inp_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_constant_output_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_constant_output_dup_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_constant_requires_grad_const_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_constant_return_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_constant_tensor_mutation_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_constant_tensor_with_non_functional_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_constant_tensor_with_non_functional_nested_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_constrain_decomp_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_constrain_size_in_eager_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_constrain_size_with_constrain_value_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_constrain_size_with_various_cases_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_conv_dynamic_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_crop_like_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_cse_for_symint_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_custom_op_auto_functionalize_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_custom_op_auto_functionalize_pre_dispatch_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_custom_op_auto_warn_pre_dispatch_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_custom_op_preserve_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_custom_pytree_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_custom_tag_metadata_re_export_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_decomp_batch_norm_functional_predispatch_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_decomp_item_in_prim_after_decomposition_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_decomp_item_in_prim_before_decomposition_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_default_decomposition_core_cia_ops_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_derived_dim_1_2_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_derived_dim_basic_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_derived_dim_integer_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_derived_dim_nested_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_derived_dim_out_of_order_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_derived_dim_out_of_order_repeat_derived_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_derived_dim_out_of_order_simplified_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_derived_dim_out_of_order_simplified_repeat_non_derived_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_derived_dim_repeat_derived_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_detect_leak_nonstrict_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_detect_leak_nonstrict_with_stacktrace_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_detect_leak_strict_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_device_to_dynamic_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_device_to_gpu_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_device_to_mutation_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_device_to_mutation_float_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_device_to_static_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dim_1_2_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dim_auto_and_dim_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dim_dynamic_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dim_dynamic_divisibility_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dim_dynamic_specialization_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dim_hint_range_violations_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dim_hint_ranges_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_disable_forced_specializations_errors_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_disable_forced_specializations_ok_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_distributed_all_gather_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_distributed_all_gather_into_tensor_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_distributed_all_reduce_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_distributed_all_to_all_single_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_distributed_reduce_scatter_tensor_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dont_duck_size_for_auto_dynamic_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_double_lifted_constants_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_draft_export_checks_aliasing_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_draft_export_checks_mutation_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_draft_export_checks_mutation_list_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_draft_export_checks_mutation_with_nan_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_draft_export_fake_kernel_inference_errors_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_draft_export_infers_fake_kernel_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_duplicate_modules_with_non_persistent_buffers_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dynamic_lr_shift_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dynamic_shapes_bounds_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dynamic_shapes_builder_basic_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dynamic_shapes_builder_kwargs_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dynamic_shapes_builder_pytree_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dynamic_shapes_dataclass_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dynamic_shapes_inferred_basic_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dynamic_shapes_serdes_generic_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dynamic_shapes_serdes_user_errors_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dynamic_shapes_serdes_various_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dynamic_shapes_spec_with_pytree_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dynamic_shapes_wrapped_with_shape_guards_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_dynamic_sym_round_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_ends_of_bounds_oblivious_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_enum_str_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_error_does_not_reference_eager_fallback_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_error_when_passing_mutating_primitive_op_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_exception_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_expand_copy_export_handles_implicit_true_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_api_with_dynamic_shapes_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_as_backend_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_associative_scan_lifted_buffers_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_associative_scan_symbol_dim_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_associative_scan_symbol_scandim_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_aten_to_unflatten_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_aten_to_unflatten_subclass_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_aten_to_unflatten_subclass_pre_dispatch_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_cond_preserve_torch_fn_for_subgraphs_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_cond_symbool_pred_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_cond_warns_constant_pred_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_custom_decomp_table_basic_pop_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_custom_decomp_table_container_methods_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_custom_op_lib_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_custom_triton_kernel_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_custom_triton_kernel_mutable_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_cyclic_reference_leak_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_decomp_torture_case_1_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_decomp_torture_case_2_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_decomps_dynamic_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_decomps_simple_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_dynamo_config_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_for_training_run_decomp_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_for_training_with_container_type_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_for_training_with_dynamic_shapes_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_for_training_with_mutation_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_for_training_with_state_dict_hooks_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_func_with_default_kwargs_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_func_with_keyword_only_args_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_func_with_kwargs_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_func_with_pytree_kwargs_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_func_with_var_keyword_args_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_func_with_var_keyword_pytree_args_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_func_with_var_postional_args_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_function_schema_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_graph_with_no_inputs_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_input_mutation_bug_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_input_mutation_dynamic_shape_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_input_mutation_static_shape_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_leak_compile_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_linear_preserve_dynamic_shape_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_max_nonstrict_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_max_onnx_reported_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_method_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_mod_constraints_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_module_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_preserve_linear_at_aot_level_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_preserve_linear_but_not_custom_op_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_rnn_variants_with_warning_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_scan_pytree_output_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_script_module_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_statically_known_true_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_then_compile_tensor_ctor_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_with_autocast_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_with_fake_tensor_inputs_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_with_fake_tensor_inputs_on_cuda_devices_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_with_inline_constraints_complex_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_with_inline_constraints_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_with_set_grad_enabled_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_export_with_wrong_inputs_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_external_call_non_strict_real_tensor_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_fake_inputs_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_fake_weights_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_filter_traceback_frames_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_flex_attention_export_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_float_conversion_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_float_conversion_from_int_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_fqn_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_from_node_metadata_export_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_full_on_scalar_tensor_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_function_holding_tensor_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_hints_wrapper_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_hoo_inline_users_issue_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_if_functional_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_if_post_autograd_op_preserved_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_inductor_backend_inside_nonstrict_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_inline_script_class_method_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_inline_script_class_method_recursive_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_inline_script_function_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_inline_script_method_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_int_shape_specialization_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_intermediate_shape_comp_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_invalid_pytree_dynamo_graph_capture_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_is_exporting_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_is_nonzero_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_isnonzero_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_issue_113041_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_issue_157289_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_issue_161902_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_istft_op_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_keep_composite_ops_invalid_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_keep_composite_ops_linear_convd_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_keep_composite_ops_linear_convd_for_training_ir_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_kwarg_dynamic_shapes_diff_order_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_kwargs_reorder_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_layer_norm_unbacked_normalized_shape_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_layer_sharing_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_lazy_module_kwargs_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_lifted_constants_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_linear_conv_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_malformed_fqn_from_source_name_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_map_buffers_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_map_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_mask_nonzero_static_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_masked_select_dynamic_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_math_pow_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_mismatched_dynamic_shapes_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_mixed_input_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_module_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_module_dict_key_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_module_input_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_module_input_subclasses_parameterization_nested_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_module_list_slice_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_module_with_dict_container_inp_out_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_modules_access_for_deleted_submodule_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_more_multidimensional_slicing_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_multidimensional_slicing_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_multinomial_dynamic_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_multiple_definitions_same_name_dim_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_namedtuple_input_export_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_native_multi_attention_head_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_nested_dynamic_shapes_spec_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_nested_module_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_nested_module_fake_tensor_leak_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_nested_module_with_constant_buffer_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_nested_module_with_init_buffer_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_nested_module_with_parameter_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_nn_module_stack_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_nn_module_stack_shared_submodule_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_no_check_is_size_error_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_no_suggested_fixes_for_data_dependent_errors_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_no_tensor_computation_2_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_no_tensor_computation_3_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_no_tensor_computation_4_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_no_tensor_computation_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_non_arg_name_dynamic_shapes_api_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_non_arg_name_dynamic_shapes_api_with_container_type_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_non_arg_name_dynamic_shapes_api_with_kwarg_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_non_persistent_buffer_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_non_strict_dynamic_shapes_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_non_strict_dynamic_shapes_suggested_fixes_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_none_buffers_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_nonstrict_retrace_preserves_metadata_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_nonzero_2_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_nonzero_dynamic_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_not_registered_parameter_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_operator_aten_tensor_mode_variant_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_output_node_name_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_pad_sequence_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_param_util_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_partial_patched_forward_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_placeholder_naming_collisions_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_placeholder_naming_collisions_hoo_subgraphs_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_placeholder_naming_order_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_placeholder_naming_order_variadic_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_placeholder_update_preserving_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_predispatch_cond_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_predispatch_grad_wrappers_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_preserve_annotation_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_preserve_module_call_signature_unflatten_specialization_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_preserve_requires_grad_placeholders_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_preserve_shape_dynamism_for_unused_inputs_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_profiling_code_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_python_asserts_with_sym_int_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_pytree_register_data_class_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_pytree_register_nested_data_class_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_raise_user_error_when_guard_on_data_dependent_operation_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_range_constraints_with_replacement_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_real_tensor_alias_dtype_mismatch_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_real_tensor_bool_cast_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_real_tensor_errors_on_aliasing_custom_op_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_real_tensor_for_max_op_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_real_tensor_size_mismatch_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_redundant_assert_max_upper_bound_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_redundant_asserts_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_refine_dynamic_shapes_from_suggested_fixes_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_register_constant_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_repeat_interleave_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_replace_unbacked_with_very_large_upperbound_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_replaced_unbacked_bindings_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_reshape_view_helper_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_retracable_ep_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_retrace_pre_autograd_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_run_decomposition_supports_user_input_mutation_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_run_decompositions_keep_metadata_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_run_decompositions_keep_tensor_constant_metadata_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_runtime_assert_for_prim_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_runtime_assert_for_prm_str_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_runtime_assert_with_size_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_sdpa_gqa_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_sequential_slicing_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_set_example_inputs_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_set_grad_as_side_effect_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_set_grad_empty_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_set_grad_unflatten_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_setgrad_lifted_tensor_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_shared_submodule_nn_module_stack_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_simple_export_for_training_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_simple_unbacked_view_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_size_input_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_slice_nn_module_stack_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_solver_unsupported_sympy_function_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_specialize_derived_dim_roots_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_split_const_gm_with_lifted_constants_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_stack_trace_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_stack_trace_make_fx_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_state_primitives_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_state_shape_attribute_assignment_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_state_tensors_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_static_dim_constraints_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_subclass_context_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_subclass_nested_attr_access_complicated_metadata_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_subclass_nested_attr_access_const_metadata_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_subclass_nested_attr_access_const_metadata_not_top_level_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_subclass_nested_attr_access_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_subclass_nested_attr_access_submodule_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_subclasses_parameterization_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_subclasses_parameterization_nested_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_suggest_torch_checks_with_non_negative_check_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_suggest_torch_checks_with_regular_check_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_suggested_fixes_for_data_dependent_errors_basic_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_suggested_fixes_for_data_dependent_errors_puzzlers_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_suggested_fixes_new_roots_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_sym_float_operators_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_sym_or_sym_and_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_sym_sqrt_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_symbool_item_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_symfloat_item_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_symint_input_additional_inputs_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_symint_input_basic_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_symint_input_ranges_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_symint_input_shapes_collection_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_symint_input_specialization_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_symint_item_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_symint_output_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_symint_tensor_return_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_tag_ac_export_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_tensor_attribute_zero_args_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_tensor_constant_aten_to_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_tensor_constant_with_wrapped_method_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_to_module_with_mutated_buffer_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_to_module_with_mutated_buffer_multiple_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_to_module_with_mutated_buffer_multiple_update_sub_later_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_tolist_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_torch_check_eq_commutativity_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_torch_fn_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_trace_under_fake_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_train_eval_on_exported_preautograd_module_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_tril_dynamic_diagonal_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_triu_dynamic_diagonal_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unbacked_3d_matmul_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unbacked_bincount_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unbacked_bindings_for_divisible_u_symint_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unbacked_deferred_runtime_retrace_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unbacked_expand_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unbacked_infer_size_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unbacked_kth_value_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unbacked_linear_layer_norm_input_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unbacked_noncontig_lin_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unbacked_pad_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unbacked_scalar_constructor_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unbacked_slice_forward_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unbacked_slice_simple_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unbacked_stack_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unbacked_to_cond_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unbacked_to_cond_passthrough_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unbacked_unsqueeze_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_asserts_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_buffer_update_child2parent_swap_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_closure_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_isinstance_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_multiple_graphs_dispatch_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_multiple_graphs_preserve_signature_no_error_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_multiple_graphs_shared_submodule_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_multiple_graphs_state_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_no_unroll_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_placeholder_update_child2parent_swap_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_placeholder_update_grandchild2cousin_swap_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_random_dag_5_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_random_dag_6_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_random_dag_buf_8_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_random_dag_const_preserving_3_1_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_random_dag_const_preserving_3_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_random_dag_mutating_buf_4_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_random_dag_mutating_buf_6_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_random_dag_mutating_buf_9_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_random_dag_mutating_buf_preserving_10_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_random_dag_mutating_buf_preserving_4_1_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_random_dag_mutating_buf_preserving_4_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_random_dag_mutating_buf_preserving_5_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_random_dag_mutating_buf_preserving_7_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unflatten_random_dag_preserving_4_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unused_aliases_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_unused_constant_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_uplift_common_custom_meta_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_uplift_common_custom_meta_with_multiple_calls_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_use_embedding_twice_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_user_input_and_buffer_mutation_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_vmap_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_vmap_custom_autograd_function_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_vmap_to_assert_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_where_decomp_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_while_loop_assert_separation_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_while_loop_index_assertions_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_while_loop_simple_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_while_loop_tensor_constant_idx_cpp_serdes, test/export/test_cpp_serdes.py::CppSerdesTestExport::test_wrapper_module_cpp_serdes 2025-12-04T16:13:39.8162958Z 2025-12-04T16:13:39.8163331Z Finished export/test_cpp_serdes 1/1 ... [2025-12-04 16:13:39.775007][25203.465398401], took 1.79min 2025-12-04T16:13:39.8187479Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/export.test_cpp_serdes/export.test_cpp_serdes-1f0305d948167935.xml 2025-12-04T16:13:39.9561825Z Running inductor/test_block_analysis 1/1 ... [2025-12-04 16:13:39.955829][25203.64622074] 2025-12-04T16:13:39.9562461Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:13:39.9565231Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_block_analysis.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:13:39.956264] 2025-12-04T16:13:50.0858185Z 2025-12-04T16:13:50.0859532Z inductor/test_block_analysis 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_block_analysis_1.1_b3c20d306fe594c2_.log 2025-12-04T16:13:50.0865766Z Running 10 items in this shard: test/inductor/test_block_analysis.py::BlockAnalysisTest::test_affine_identity_stride_3_symbol2_expr2, test/inductor/test_block_analysis.py::BlockAnalysisTest::test_affine_identity_stride_4_symbol1_expr1, test/inductor/test_block_analysis.py::BlockAnalysisTest::test_affine_identity_stride_5_symbol0_expr0, test/inductor/test_block_analysis.py::BlockAnalysisTest::test_index_with_dynamic_shapes, test/inductor/test_block_analysis.py::BlockAnalysisTest::test_mod_div_identity_dims0_strides0_symbol0_expr0, test/inductor/test_block_analysis.py::BlockAnalysisTest::test_mod_div_identity_dims1_strides1_symbol1_expr1, test/inductor/test_block_analysis.py::BlockAnalysisTest::test_mod_div_identity_dims2_strides2_symbol2_expr2, test/inductor/test_block_analysis.py::BlockAnalysisTest::test_subexpr_identity_symbol0_expr0_subexpr0, test/inductor/test_block_analysis.py::BlockAnalysisTest::test_subexpr_identity_symbol1_expr1_subexpr1, test/inductor/test_block_analysis.py::BlockAnalysisTest::test_subexpr_identity_symbol2_expr2_subexpr2 2025-12-04T16:13:50.0871256Z 2025-12-04T16:13:50.0871639Z Finished inductor/test_block_analysis 1/1 ... [2025-12-04 16:13:50.085609][25213.776002183], took 0.17min 2025-12-04T16:13:50.1298896Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_block_analysis/inductor.test_block_analysis-53978d8777987ceb.xml 2025-12-04T16:13:50.2189547Z Running dynamo/test_subgraphs 1/1 ... [2025-12-04 16:13:50.218602][25213.90899299] 2025-12-04T16:13:50.2190124Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:13:50.2193300Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_subgraphs.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:13:50.219056] 2025-12-04T16:14:04.3040440Z 2025-12-04T16:14:04.3041809Z dynamo/test_subgraphs 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_subgraphs_1.1_e1155ecf654700c6_.log 2025-12-04T16:14:04.3057475Z Running 44 items in this shard: test/dynamo/test_subgraphs.py::SubGraphTests::test_capi_call1, test/dynamo/test_subgraphs.py::SubGraphTests::test_capi_call2, test/dynamo/test_subgraphs.py::SubGraphTests::test_capi_call3, test/dynamo/test_subgraphs.py::SubGraphTests::test_control_flow1, test/dynamo/test_subgraphs.py::SubGraphTests::test_control_flow2, test/dynamo/test_subgraphs.py::SubGraphTests::test_control_flow3, test/dynamo/test_subgraphs.py::SubGraphTests::test_control_flow4, test/dynamo/test_subgraphs.py::SubGraphTests::test_control_flow5, test/dynamo/test_subgraphs.py::SubGraphTests::test_dynamic_duck_size, test/dynamo/test_subgraphs.py::SubGraphTests::test_dynamic_getitem, test/dynamo/test_subgraphs.py::SubGraphTests::test_dynamic_kwarg, test/dynamo/test_subgraphs.py::SubGraphTests::test_dynamic_order_dependence, test/dynamo/test_subgraphs.py::SubGraphTests::test_dynamic_zero_inference, test/dynamo/test_subgraphs.py::SubGraphTests::test_enumerate_not_break_graph, test/dynamo/test_subgraphs.py::SubGraphTests::test_extended_args, test/dynamo/test_subgraphs.py::SubGraphTests::test_graph_break_on_item, test/dynamo/test_subgraphs.py::SubGraphTests::test_indirect_unsupported1, test/dynamo/test_subgraphs.py::SubGraphTests::test_indirect_unsupported2, test/dynamo/test_subgraphs.py::SubGraphTests::test_indirect_unsupported3, test/dynamo/test_subgraphs.py::SubGraphTests::test_multigraph, test/dynamo/test_subgraphs.py::SubGraphTests::test_no_graph_break_on_item, test/dynamo/test_subgraphs.py::SubGraphTests::test_pop_after_resume, test/dynamo/test_subgraphs.py::SubGraphTests::test_restore_range, test/dynamo/test_subgraphs.py::SubGraphTests::test_restore_range_iter, test/dynamo/test_subgraphs.py::SubGraphTests::test_restore_state, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume1, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume2, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume3, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume4, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume5, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume_freevars, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume_paths_join, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume_tuple_iterator, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume_with_no_grad1, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume_with_no_grad2, test/dynamo/test_subgraphs.py::SubGraphTests::test_resume_with_no_grad3, test/dynamo/test_subgraphs.py::SubGraphTests::test_stack_state1, test/dynamo/test_subgraphs.py::SubGraphTests::test_stack_state2, test/dynamo/test_subgraphs.py::SubGraphTests::test_start1, test/dynamo/test_subgraphs.py::SubGraphTests::test_start2, test/dynamo/test_subgraphs.py::SubGraphTests::test_start3, test/dynamo/test_subgraphs.py::SubGraphTests::test_start4, test/dynamo/test_subgraphs.py::SubGraphTests::test_tuple_iterator_mutate, test/dynamo/test_subgraphs.py::SubGraphTests::test_tuple_iterator_return 2025-12-04T16:14:04.3072250Z 2025-12-04T16:14:04.3072617Z Finished dynamo/test_subgraphs 1/1 ... [2025-12-04 16:14:04.303898][25227.994292863], took 0.23min 2025-12-04T16:14:04.3473305Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_subgraphs/dynamo.test_subgraphs-e09a516058cd0e01.xml 2025-12-04T16:14:04.4280376Z Running dynamo/test_pre_dispatch 1/1 ... [2025-12-04 16:14:04.427739][25228.118130062] 2025-12-04T16:14:04.4281144Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:14:04.4284341Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_pre_dispatch.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:14:04.428152] 2025-12-04T16:14:12.2537250Z 2025-12-04T16:14:12.2538683Z dynamo/test_pre_dispatch 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_pre_dispatch_1.1_99fd2ec956edaa5c_.log 2025-12-04T16:14:12.2540689Z Running 3 items in this shard: test/dynamo/test_pre_dispatch.py::PreDispatchTests::test_autocast_simple, test/dynamo/test_pre_dispatch.py::PreDispatchTests::test_enable_grad_and_no_grad, test/dynamo/test_pre_dispatch.py::PreDispatchTests::test_no_grad_simple 2025-12-04T16:14:12.2542028Z 2025-12-04T16:14:12.2542386Z Finished dynamo/test_pre_dispatch 1/1 ... [2025-12-04 16:14:12.253509][25235.94390355], took 0.13min 2025-12-04T16:14:12.2968003Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_pre_dispatch/dynamo.test_pre_dispatch-4a36f2c5e912fcfb.xml 2025-12-04T16:14:12.3873488Z Running inductor/test_custom_post_grad_passes 1/1 ... [2025-12-04 16:14:12.387031][25236.077422391] 2025-12-04T16:14:12.3874194Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:14:12.3877858Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_custom_post_grad_passes.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:14:12.387464] 2025-12-04T16:14:33.1837578Z 2025-12-04T16:14:33.1838998Z inductor/test_custom_post_grad_passes 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_custom_post_grad_passes_1.1_14ce4ccc6fc3bcc1_.log 2025-12-04T16:14:33.1843458Z Running 6 items in this shard: test/inductor/test_custom_post_grad_passes.py::TestPostGradCustomPrePostPass::test_custom_backend_pass, test/inductor/test_custom_post_grad_passes.py::TestPostGradCustomPrePostPass::test_custom_joint_pass_post, test/inductor/test_custom_post_grad_passes.py::TestPostGradCustomPrePostPass::test_custom_joint_pass_pre, test/inductor/test_custom_post_grad_passes.py::TestPostGradCustomPrePostPass::test_custom_post_pass, test/inductor/test_custom_post_grad_passes.py::TestPostGradCustomPrePostPass::test_custom_pre_grad_pass, test/inductor/test_custom_post_grad_passes.py::TestPostGradCustomPrePostPass::test_custom_pre_pass 2025-12-04T16:14:33.1846793Z 2025-12-04T16:14:33.1847220Z Finished inductor/test_custom_post_grad_passes 1/1 ... [2025-12-04 16:14:33.183534][25256.873927027], took 0.35min 2025-12-04T16:14:33.2269985Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_custom_post_grad_passes/inductor.test_custom_post_grad_passes-e88ba17552693664.xml 2025-12-04T16:14:33.3007873Z Running dynamo/test_fx_annotate 1/1 ... [2025-12-04 16:14:33.300458][25256.990848387] 2025-12-04T16:14:33.3008463Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:14:33.3011206Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_fx_annotate.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:14:33.300871] 2025-12-04T16:14:47.0345908Z 2025-12-04T16:14:47.0346934Z dynamo/test_fx_annotate 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_fx_annotate_1.1_a80c3def106c395c_.log 2025-12-04T16:14:47.0350042Z Running 6 items in this shard: test/dynamo/test_fx_annotate.py::AnnotateTests::test_ac_flex_attention, test/dynamo/test_fx_annotate.py::AnnotateTests::test_activation_checkpointing, test/dynamo/test_fx_annotate.py::AnnotateTests::test_activation_checkpointing_annotation_inside, test/dynamo/test_fx_annotate.py::AnnotateTests::test_annotations, test/dynamo/test_fx_annotate.py::AnnotateTests::test_as_decorator, test/dynamo/test_fx_annotate.py::AnnotateTests::test_graph_break 2025-12-04T16:14:47.0352825Z 2025-12-04T16:14:47.0353170Z Finished dynamo/test_fx_annotate 1/1 ... [2025-12-04 16:14:47.034345][25270.724738224], took 0.23min 2025-12-04T16:14:47.0781032Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_fx_annotate/dynamo.test_fx_annotate-ea1b8128153e57c8.xml 2025-12-04T16:14:47.1521805Z Running dynamo/test_pgo 1/1 ... [2025-12-04 16:14:47.151820][25270.842212034] 2025-12-04T16:14:47.1522328Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:14:47.1525278Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_pgo.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:14:47.152255] 2025-12-04T16:14:58.1821151Z 2025-12-04T16:14:58.1822193Z dynamo/test_pgo 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_pgo_1.1_c150ddd3d0d86d52_.log 2025-12-04T16:14:58.1826206Z Running 11 items in this shard: test/dynamo/test_pgo.py::PgoTest::test_basic, test/dynamo/test_pgo.py::PgoTest::test_different_file_paths_local_pgo, test/dynamo/test_pgo.py::PgoTest::test_distinct_compile_id, test/dynamo/test_pgo.py::PgoTest::test_njt, test/dynamo/test_pgo.py::PgoTest::test_no_empty_graph_allowlist, test/dynamo/test_pgo.py::PgoTest::test_pgo_dynamic_false, test/dynamo/test_pgo.py::PgoTest::test_pgo_dynamic_params, test/dynamo/test_pgo.py::PgoTest::test_remote_basic, test/dynamo/test_pgo.py::PgoTest::test_sticky_pgo_read_write, test/dynamo/test_pgo.py::PgoTest::test_whitelist_ints_floats, test/dynamo/test_pgo.py::PgoTest::test_whitelist_suggestion 2025-12-04T16:14:58.1829399Z 2025-12-04T16:14:58.1829700Z Finished dynamo/test_pgo 1/1 ... [2025-12-04 16:14:58.181892][25281.872287045], took 0.18min 2025-12-04T16:14:58.2259769Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_pgo/dynamo.test_pgo-b9bc1426bb4bf646.xml 2025-12-04T16:14:58.3072687Z Running export/test_export_opinfo 1/1 ... [2025-12-04 16:14:58.306954][25281.997345491] 2025-12-04T16:14:58.3073315Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:14:58.3076651Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'export/test_export_opinfo.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:14:58.307364] 2025-12-04T16:15:55.7530876Z 2025-12-04T16:15:55.7535040Z export/test_export_opinfo 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_export_opinfo_1.1_ca68eeb5bfbf3cab_.log 2025-12-04T16:15:55.7541284Z Running 9 items in this shard: test/export/test_export_opinfo.py::TestExportOnFakeCudaCUDA::test_fake_export___getitem___cuda_float32, test/export/test_export_opinfo.py::TestExportOnFakeCudaCUDA::test_fake_export_nn_functional_batch_norm_cuda_float32, test/export/test_export_opinfo.py::TestExportOnFakeCudaCUDA::test_fake_export_nn_functional_batch_norm_without_cudnn_cuda_float32, test/export/test_export_opinfo.py::TestExportOnFakeCudaCUDA::test_fake_export_nn_functional_conv2d_cuda_float32, test/export/test_export_opinfo.py::TestExportOnFakeCudaCUDA::test_fake_export_nn_functional_instance_norm_cuda_float32, test/export/test_export_opinfo.py::TestExportOnFakeCudaCUDA::test_fake_export_nn_functional_multi_margin_loss_cuda_float32, test/export/test_export_opinfo.py::TestExportOnFakeCudaCUDA::test_fake_export_nn_functional_scaled_dot_product_attention_cuda_float32, test/export/test_export_opinfo.py::TestExportOnFakeCudaCUDA::test_fake_export_nonzero_cuda_float32, test/export/test_export_opinfo.py::TestExportOnFakeCudaCUDA::test_preserve_original_behavior_cuda 2025-12-04T16:15:55.7547102Z 2025-12-04T16:15:55.7547470Z Finished export/test_export_opinfo 1/1 ... [2025-12-04 16:15:55.752852][25339.443245177], took 0.96min 2025-12-04T16:15:55.7981627Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/export.test_export_opinfo/export.test_export_opinfo-d8b8b1860a5b7382.xml 2025-12-04T16:15:55.8670361Z Running inductor/test_control_flow 2/4 ... [2025-12-04 16:15:55.866689][25339.557080706] 2025-12-04T16:15:55.8670986Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:15:55.8673912Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_control_flow.py', '--shard-id=2', '--num-shards=4', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:15:55.867117] 2025-12-04T16:29:39.9154475Z 2025-12-04T16:29:39.9155553Z inductor/test_control_flow 2/4 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_control_flow_2.4_11abfa8a57b42ba2_.log 2025-12-04T16:29:39.9342387Z Running 184 items in this shard: test/inductor/test_control_flow.py::CondTests::test_cond_control_flow_with_precomputed_size, test/inductor/test_control_flow.py::CondTests::test_cond_decompose_ops_in_subgraph_device_cuda, test/inductor/test_control_flow.py::CondTests::test_cond_functional_call_device_cpu_dynamic_True, test/inductor/test_control_flow.py::CondTests::test_cond_functional_call_device_cuda_dynamic_False, test/inductor/test_control_flow.py::CondTests::test_cond_functional_call_device_cuda_dynamic_True, test/inductor/test_control_flow.py::CondTests::test_cond_mismatched_branch_output_size_device_cpu_dynamic_False, test/inductor/test_control_flow.py::CondTests::test_cond_multiple_outputs_device_cpu_dynamic_False, test/inductor/test_control_flow.py::CondTests::test_cond_multiple_outputs_device_cpu_dynamic_True, test/inductor/test_control_flow.py::CondTests::test_cond_nested_control_flow_device_cpu_dynamic_False, test/inductor/test_control_flow.py::CondTests::test_cond_nested_control_flow_device_cuda_dynamic_True, test/inductor/test_control_flow.py::CondTests::test_cond_non_tensor_predicates_device_cpu_dynamic_False, test/inductor/test_control_flow.py::CondTests::test_cond_non_tensor_predicates_device_cpu_dynamic_True, test/inductor/test_control_flow.py::CondTests::test_cond_outer_code_before_after_device_cpu_dynamic_True, test/inductor/test_control_flow.py::CondTests::test_cond_select_with_input_idx_device_cuda_dynamic_True, test/inductor/test_control_flow.py::CondTests::test_cond_simple_control_flow_device_cpu_dynamic_False, test/inductor/test_control_flow.py::CondTests::test_cond_simple_control_flow_device_cuda_dynamic_True, test/inductor/test_control_flow.py::CondTests::test_cond_subgraphs_with_parameters_device_cuda_dynamic_False, test/inductor/test_control_flow.py::CondTests::test_cond_unbacked_symint_inner_device_cpu, test/inductor/test_control_flow.py::CondTests::test_cond_use_buffers_from_outer_scope, test/inductor/test_control_flow.py::CondTests::test_output_on_different_device, test/inductor/test_control_flow.py::WhileLoopTests::test_while_loop_nested_control_flow_device_cpu_dynamic_True_autograd_True, test/inductor/test_control_flow.py::WhileLoopTests::test_while_loop_nested_control_flow_device_cuda_dynamic_False_autograd_False, test/inductor/test_control_flow.py::WhileLoopTests::test_while_loop_nested_control_flow_device_cuda_dynamic_False_autograd_True, test/inductor/test_control_flow.py::WhileLoopTests::test_while_loop_nested_control_flow_device_cuda_dynamic_True_autograd_False, test/inductor/test_control_flow.py::WhileLoopTests::test_while_loop_stack_output_simple_device_cpu_dynamic_False, test/inductor/test_control_flow.py::WhileLoopTests::test_while_loop_with_conv_device_cuda_dynamic_False_autograd_False, test/inductor/test_control_flow.py::WhileLoopTests::test_while_loop_with_data_dependent_in_out_device_cuda_dynamic_True_autograd_True, test/inductor/test_control_flow.py::WhileLoopTests::test_while_loop_with_data_dependent_ops_device_cpu_dynamic_False_autograd_True, test/inductor/test_control_flow.py::WhileLoopTests::test_while_loop_with_outer_buffers_device_cpu_dynamic_False_autograd_False, test/inductor/test_control_flow.py::WhileLoopTests::test_while_loop_with_outer_code_device_cpu_dynamic_True_autograd_False, test/inductor/test_control_flow.py::WhileLoopTests::test_while_loop_with_parameters_device_cpu_dynamic_False_autograd_False, test/inductor/test_control_flow.py::WhileLoopTests::test_while_loop_with_parameters_device_cuda_dynamic_False_autograd_True, test/inductor/test_control_flow.py::WhileLoopTests::test_while_loop_with_pytree_inputs_device_cuda_dynamic_True_autograd_False, test/inductor/test_control_flow.py::WhileLoopTests::test_while_loop_with_sym_expr_cond_device_cuda_dynamic_True_autograd_True, test/inductor/test_control_flow.py::WhileLoopTests::test_while_loop_with_unbacked_symint_closure_device_cpu_dynamic_True_autograd_False, test/inductor/test_control_flow.py::WhileLoopTests::test_while_loop_with_unbacked_symint_closure_device_cuda_dynamic_False_autograd_False, test/inductor/test_control_flow.py::WhileLoopTests::test_while_loop_with_unbacked_symint_closure_device_cuda_dynamic_True_autograd_True, test/inductor/test_control_flow.py::WhileLoopTests::test_while_loop_zero_loop_device_cuda_dynamic_False, test/inductor/test_control_flow.py::AssociativeScanTests::test_associative_scan_CUDA_flip_combine_mode_generic_backend_inductor_device_cuda, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cpu_dynamic_False_reverse_False_dim_0_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cpu_dynamic_False_reverse_False_dim_0_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cpu_dynamic_False_reverse_False_dim_1_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cpu_dynamic_False_reverse_False_dim_3_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cpu_dynamic_False_reverse_True_dim_0_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cpu_dynamic_False_reverse_True_dim_3_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cpu_dynamic_True_reverse_False_dim_0_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cpu_dynamic_True_reverse_True_dim_0_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cpu_dynamic_True_reverse_True_dim_0_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cpu_dynamic_True_reverse_True_dim_0_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cpu_dynamic_True_reverse_True_dim_3_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cuda_dynamic_False_reverse_False_dim_0_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cuda_dynamic_False_reverse_False_dim_1_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cuda_dynamic_False_reverse_True_dim_0_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cuda_dynamic_False_reverse_True_dim_1_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cuda_dynamic_False_reverse_True_dim_3_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cuda_dynamic_True_reverse_False_dim_0_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cuda_dynamic_True_reverse_False_dim_1_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cuda_dynamic_True_reverse_False_dim_3_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cuda_dynamic_True_reverse_False_dim_3_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cuda_dynamic_True_reverse_True_dim_0_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cuda_dynamic_True_reverse_True_dim_0_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_cond_in_scan_device_cuda_dynamic_True_reverse_True_dim_1_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_chunked_ce_device_cpu_dynamic_False_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_chunked_ce_device_cuda_dynamic_True_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_compare_chunked_ce_with_no_scan_device_cpu_dynamic_False, test/inductor/test_control_flow.py::ScanTests::test_scan_compare_chunked_ce_with_no_scan_device_cuda_dynamic_True, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cpu_dynamic_False_reverse_False_dim_0_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cpu_dynamic_False_reverse_False_dim_1_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cpu_dynamic_False_reverse_False_dim_1_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cpu_dynamic_False_reverse_False_dim_3_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cpu_dynamic_False_reverse_True_dim_0_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cpu_dynamic_False_reverse_True_dim_3_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cpu_dynamic_False_reverse_True_dim_3_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cpu_dynamic_True_reverse_False_dim_0_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cpu_dynamic_True_reverse_False_dim_3_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cpu_dynamic_True_reverse_False_dim_3_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cpu_dynamic_True_reverse_True_dim_1_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cpu_dynamic_True_reverse_True_dim_1_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cpu_dynamic_True_reverse_True_dim_1_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cuda_dynamic_False_reverse_False_dim_0_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cuda_dynamic_False_reverse_False_dim_1_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cuda_dynamic_False_reverse_False_dim_3_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cuda_dynamic_False_reverse_True_dim_0_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cuda_dynamic_False_reverse_True_dim_0_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cuda_dynamic_False_reverse_True_dim_1_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cuda_dynamic_False_reverse_True_dim_3_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cuda_dynamic_False_reverse_True_dim_3_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cuda_dynamic_True_reverse_False_dim_0_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cuda_dynamic_True_reverse_False_dim_0_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cuda_dynamic_True_reverse_False_dim_3_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_conv_device_cuda_dynamic_True_reverse_True_dim_0_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_False_reverse_False_dim_1_pred_True_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_False_reverse_False_dim_1_pred_True_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_False_reverse_False_dim_3_pred_False_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_False_reverse_True_dim_0_pred_False_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_False_reverse_True_dim_0_pred_True_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_False_reverse_True_dim_1_pred_False_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_False_reverse_True_dim_1_pred_False_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_False_reverse_True_dim_1_pred_True_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_True_reverse_False_dim_0_pred_False_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_True_reverse_False_dim_0_pred_True_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_True_reverse_False_dim_1_pred_False_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_True_reverse_False_dim_1_pred_False_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_True_reverse_False_dim_1_pred_True_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_True_reverse_False_dim_3_pred_False_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_True_reverse_False_dim_3_pred_True_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_True_reverse_True_dim_0_pred_False_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_True_reverse_True_dim_0_pred_False_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_True_reverse_True_dim_0_pred_True_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_True_reverse_True_dim_1_pred_False_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cpu_dynamic_True_reverse_True_dim_3_pred_False_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_False_reverse_False_dim_0_pred_True_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_False_reverse_False_dim_1_pred_True_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_False_reverse_False_dim_3_pred_True_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_False_reverse_False_dim_3_pred_True_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_False_reverse_True_dim_0_pred_False_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_False_reverse_True_dim_0_pred_False_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_False_reverse_True_dim_0_pred_True_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_False_reverse_True_dim_1_pred_False_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_False_reverse_True_dim_1_pred_False_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_False_reverse_True_dim_1_pred_False_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_False_reverse_True_dim_1_pred_True_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_False_reverse_True_dim_3_pred_False_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_False_reverse_True_dim_3_pred_False_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_False_reverse_True_dim_3_pred_False_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_True_reverse_False_dim_0_pred_False_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_True_reverse_False_dim_1_pred_False_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_True_reverse_False_dim_1_pred_True_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_True_reverse_False_dim_3_pred_True_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_True_reverse_True_dim_0_pred_False_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_in_cond_device_cuda_dynamic_True_reverse_True_dim_0_pred_False_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cpu_dynamic_False_reverse_False_dim_0_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cpu_dynamic_False_reverse_False_dim_0_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cpu_dynamic_False_reverse_False_dim_1_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cpu_dynamic_False_reverse_False_dim_3_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cpu_dynamic_False_reverse_True_dim_0_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cpu_dynamic_False_reverse_True_dim_0_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cpu_dynamic_False_reverse_True_dim_1_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cpu_dynamic_False_reverse_True_dim_1_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cpu_dynamic_True_reverse_False_dim_0_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cpu_dynamic_True_reverse_False_dim_0_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cpu_dynamic_True_reverse_False_dim_1_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cpu_dynamic_True_reverse_False_dim_1_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cpu_dynamic_True_reverse_False_dim_3_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cpu_dynamic_True_reverse_False_dim_3_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cpu_dynamic_True_reverse_True_dim_0_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cuda_dynamic_False_reverse_False_dim_0_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cuda_dynamic_False_reverse_False_dim_0_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cuda_dynamic_False_reverse_False_dim_1_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cuda_dynamic_False_reverse_False_dim_1_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cuda_dynamic_False_reverse_False_dim_1_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cuda_dynamic_False_reverse_False_dim_3_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cuda_dynamic_False_reverse_True_dim_0_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cuda_dynamic_False_reverse_True_dim_1_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cuda_dynamic_False_reverse_True_dim_1_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cuda_dynamic_True_reverse_False_dim_0_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cuda_dynamic_True_reverse_False_dim_0_scan_length_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cuda_dynamic_True_reverse_False_dim_0_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cuda_dynamic_True_reverse_False_dim_1_scan_length_5_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cuda_dynamic_True_reverse_False_dim_3_scan_length_5_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_nn_modules_device_cuda_dynamic_True_reverse_True_dim_1_scan_length_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_pytree_in_out_device_cpu_dynamic_False_reverse_False_dim_0_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_pytree_in_out_device_cpu_dynamic_False_reverse_False_dim_2_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_pytree_in_out_device_cpu_dynamic_False_reverse_False_dim_2_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_pytree_in_out_device_cpu_dynamic_False_reverse_True_dim_0_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_pytree_in_out_device_cpu_dynamic_False_reverse_True_dim_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_pytree_in_out_device_cpu_dynamic_True_reverse_False_dim_2_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_pytree_in_out_device_cpu_dynamic_True_reverse_True_dim_0_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_pytree_in_out_device_cpu_dynamic_True_reverse_True_dim_2_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_pytree_in_out_device_cpu_dynamic_True_reverse_True_dim_2_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_pytree_in_out_device_cuda_dynamic_False_reverse_False_dim_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_pytree_in_out_device_cuda_dynamic_False_reverse_True_dim_1_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_pytree_in_out_device_cuda_dynamic_False_reverse_True_dim_2_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_pytree_in_out_device_cuda_dynamic_True_reverse_False_dim_0_autograd_False, test/inductor/test_control_flow.py::ScanTests::test_scan_pytree_in_out_device_cuda_dynamic_True_reverse_True_dim_1_autograd_True, test/inductor/test_control_flow.py::ScanTests::test_scan_with_clamp_device_cuda_dynamic_True_autograd_True, test/inductor/test_control_flow.py::MapTests::test_map_nested_with_cond_device_cpu_dynamic_True_autograd_True, test/inductor/test_control_flow.py::MapTests::test_map_pytree_in_out_device_cpu_dynamic_False_autograd_True, test/inductor/test_control_flow.py::MapTests::test_map_pytree_in_out_device_cuda_dynamic_False_autograd_False, test/inductor/test_control_flow.py::MapTests::test_map_simple_device_cpu_dynamic_False_autograd_False, test/inductor/test_control_flow.py::MapTests::test_map_simple_device_cpu_dynamic_True_autograd_True, test/inductor/test_control_flow.py::MapTests::test_map_simple_device_cuda_dynamic_False_autograd_False, test/inductor/test_control_flow.py::MapTests::test_map_simple_linear_with_view_device_cpu_dynamic_True_autograd_False, test/inductor/test_control_flow.py::MapTests::test_map_simple_linear_with_view_device_cpu_dynamic_True_autograd_True 2025-12-04T16:29:39.9455567Z 2025-12-04T16:29:39.9455951Z Finished inductor/test_control_flow 2/4 ... [2025-12-04 16:29:39.944468][26163.634855021], took 13.73min 2025-12-04T16:29:39.9898384Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_control_flow/inductor.test_control_flow-12b6e2c7b80b8695.xml 2025-12-04T16:29:41.7806612Z Uploading artifacts took 1.70 seconds 2025-12-04T16:29:41.7811691Z Running dynamo/test_compile 1/1 ... [2025-12-04 16:29:41.780927][26165.471318609] 2025-12-04T16:29:41.7812536Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:29:41.7818511Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_compile.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:29:41.781468] 2025-12-04T16:29:58.7204863Z 2025-12-04T16:29:58.7206015Z dynamo/test_compile 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_compile_1.1_b494fc7b48d70de4_.log 2025-12-04T16:29:58.7214263Z Running 17 items in this shard: test/dynamo/test_compile.py::InPlaceCompilationTests::test_compilation, test/dynamo/test_compile.py::InPlaceCompilationTests::test_compilation_callback, test/dynamo/test_compile.py::InPlaceCompilationTests::test_compilation_callback_with_graph_break, test/dynamo/test_compile.py::InPlaceCompilationTests::test_compilation_constant_hasattr_fail, test/dynamo/test_compile.py::InPlaceCompilationTests::test_compilation_evnum_hasattr_fail, test/dynamo/test_compile.py::InPlaceCompilationTests::test_compilation_name_error, test/dynamo/test_compile.py::InPlaceCompilationTests::test_compilation_nn_module_invalid_method, test/dynamo/test_compile.py::InPlaceCompilationTests::test_compilation_tensor_invalid_method, test/dynamo/test_compile.py::InPlaceCompilationTests::test_compile_eager_options, test/dynamo/test_compile.py::InPlaceCompilationTests::test_jit_save, test/dynamo/test_compile.py::InPlaceCompilationTests::test_list_bad_access, test/dynamo/test_compile.py::InPlaceCompilationTests::test_overwrite_call_impl, test/dynamo/test_compile.py::InPlaceCompilationTests::test_save, test/dynamo/test_compile.py::InPlaceCompilationTests::test_state_dict_save, test/dynamo/test_compile.py::InPlaceCompilationTests::test_to_sparse_to_dense_with_graph_break, test/dynamo/test_compile.py::InPlaceCompilationTests::test_torch_script_compilation, test/dynamo/test_compile.py::PublicTorchCompilerTests::test_dynamo_signatures 2025-12-04T16:29:58.7222043Z 2025-12-04T16:29:58.7222369Z Finished dynamo/test_compile 1/1 ... [2025-12-04 16:29:58.720277][26182.410671225], took 0.28min 2025-12-04T16:29:58.7649189Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_compile/dynamo.test_compile-1298df40b08e7720.xml 2025-12-04T16:29:58.8519767Z Running dynamo/test_nested_graph_breaks 1/1 ... [2025-12-04 16:29:58.851685][26182.542077201] 2025-12-04T16:29:58.8520379Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:29:58.8523470Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_nested_graph_breaks.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:29:58.852104] 2025-12-04T16:30:10.6332907Z 2025-12-04T16:30:10.6334174Z dynamo/test_nested_graph_breaks 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_nested_graph_breaks_1.1_06faf157c67ea009_.log 2025-12-04T16:30:10.6347959Z Running 23 items in this shard: test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_cells, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_cells_double_graph_break, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_counters, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_dead_nested_cells, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_differing_arg_nums, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_differing_locals_nums, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_disable_nested_graph_breaks, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_doubly_nested_graph_break, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_functorch_with_nested_graph_break, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_generator_nested_graph_break, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_inactive_ctx_manager, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_nested_graph_break_in_loop, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_nested_graph_break_in_try_block, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_nested_step_unsupported, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_no_recompiles, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_return_after_graph_break_deep_nested, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_return_after_graph_break_nested, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_side_effects_cells, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_side_effects_globals, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_side_effects_globals_different_module, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_single_graph_break, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_single_graph_break_repeat, test/dynamo/test_nested_graph_breaks.py::NestedGraphBreakTests::test_supported_ctx_manager 2025-12-04T16:30:10.6360441Z 2025-12-04T16:30:10.6360863Z Finished dynamo/test_nested_graph_breaks 1/1 ... [2025-12-04 16:30:10.633101][26194.32349453], took 0.20min 2025-12-04T16:30:10.6773085Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_nested_graph_breaks/dynamo.test_nested_graph_breaks-3ce3fc0c984a2ae0.xml 2025-12-04T16:30:10.7586007Z Running inductor/test_needs_exact_strides 1/1 ... [2025-12-04 16:30:10.758280][26194.448672297] 2025-12-04T16:30:10.7586708Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:30:10.7589456Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_needs_exact_strides.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:30:10.758692] 2025-12-04T16:30:29.8008347Z 2025-12-04T16:30:29.8009816Z inductor/test_needs_exact_strides 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_needs_exact_strides_1.1_fcd620a2b11e818d_.log 2025-12-04T16:30:29.8012407Z Running 2 items in this shard: test/inductor/test_needs_exact_strides.py::TestNeedsExactStrides::test_custom_op_float32, test/inductor/test_needs_exact_strides.py::TestNeedsExactStrides::test_custom_op_float8_e8m0fnu 2025-12-04T16:30:29.8013614Z 2025-12-04T16:30:29.8014100Z Finished inductor/test_needs_exact_strides 1/1 ... [2025-12-04 16:30:29.800592][26213.490985455], took 0.32min 2025-12-04T16:30:29.8455016Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_needs_exact_strides/inductor.test_needs_exact_strides-afd07b0a4c4dd4e2.xml 2025-12-04T16:30:29.9325778Z Running inductor/test_split_cat_fx_aten_passes 1/1 ... [2025-12-04 16:30:29.932248][26213.622637562] 2025-12-04T16:30:29.9326428Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:30:29.9329911Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_split_cat_fx_aten_passes.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:30:29.932721] 2025-12-04T16:30:49.3260772Z 2025-12-04T16:30:49.3261957Z inductor/test_split_cat_fx_aten_passes 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_split_cat_fx_aten_passes_1.1_c6d6a90bc763f053_.log 2025-12-04T16:30:49.3266067Z Running 5 items in this shard: test/inductor/test_split_cat_fx_aten_passes.py::TestSplitCatAten::test_move_view_after_cat_aten, test/inductor/test_split_cat_fx_aten_passes.py::TestSplitCatAten::test_select_cat_post_grad, test/inductor/test_split_cat_fx_aten_passes.py::TestSplitCatAten::test_split_cat_post_grad, test/inductor/test_split_cat_fx_aten_passes.py::TestSplitCatAten::test_split_cat_post_grad_singular, test/inductor/test_split_cat_fx_aten_passes.py::TestSplitCatAtenNormalizationPasses::test_split_aten_normalization 2025-12-04T16:30:49.3268706Z 2025-12-04T16:30:49.3269127Z Finished inductor/test_split_cat_fx_aten_passes 1/1 ... [2025-12-04 16:30:49.325829][26233.016221085], took 0.32min 2025-12-04T16:30:49.3705582Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_split_cat_fx_aten_passes/inductor.test_split_cat_fx_aten_passes-d90ed49c774dd94e.xml 2025-12-04T16:30:49.4510368Z Running dynamo/test_resume 1/1 ... [2025-12-04 16:30:49.450703][26233.141094167] 2025-12-04T16:30:49.4510986Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:30:49.4513909Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_resume.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:30:49.451126] 2025-12-04T16:30:57.1261160Z 2025-12-04T16:30:57.1262393Z dynamo/test_resume 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_resume_1.1_a3d627742d7240d5_.log 2025-12-04T16:30:57.1263582Z Running 1 items in this shard: test/dynamo/test_resume.py::ResumeFunctionTests::test_freevars 2025-12-04T16:30:57.1264082Z 2025-12-04T16:30:57.1264401Z Finished dynamo/test_resume 1/1 ... [2025-12-04 16:30:57.125892][26240.816285521], took 0.13min 2025-12-04T16:30:57.1709823Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_resume/dynamo.test_resume-1b040f63515910f4.xml 2025-12-04T16:30:57.2452785Z Running dynamo/test_backward_higher_order_ops 1/1 ... [2025-12-04 16:30:57.244921][26240.935312982] 2025-12-04T16:30:57.2453474Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:30:57.2456615Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_backward_higher_order_ops.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:30:57.245361] 2025-12-04T16:31:16.5870463Z 2025-12-04T16:31:16.5871775Z dynamo/test_backward_higher_order_ops 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_backward_higher_order_ops_1.1_c04481e0c8fd9912_.log 2025-12-04T16:31:16.5876629Z Running 7 items in this shard: test/dynamo/test_backward_higher_order_ops.py::BackwardHigherOrderOpTests::test_invoke_in_eager, test/dynamo/test_backward_higher_order_ops.py::BackwardHigherOrderOpTests::test_invoke_in_pt2, test/dynamo/test_backward_higher_order_ops.py::BackwardHigherOrderOpTests::test_invoke_in_pt2_compiled_autograd, test/dynamo/test_backward_higher_order_ops.py::BackwardHigherOrderOpTests::test_invoke_in_pt2_compiled_autograd_graph_breaks, test/dynamo/test_backward_higher_order_ops.py::BackwardHigherOrderOpTests::test_invoke_in_pt2_compiled_autograd_side_effect, test/dynamo/test_backward_higher_order_ops.py::BackwardHigherOrderOpTests::test_invoke_make_bw, test/dynamo/test_backward_higher_order_ops.py::BackwardHigherOrderOpTests::test_invoke_make_fx_forward_contrived 2025-12-04T16:31:16.5880635Z 2025-12-04T16:31:16.5881049Z Finished dynamo/test_backward_higher_order_ops 1/1 ... [2025-12-04 16:31:16.586815][26260.277208192], took 0.32min 2025-12-04T16:31:16.6355373Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_backward_higher_order_ops/dynamo.test_backward_higher_order_ops-8f7cdfdf9264a3b0.xml 2025-12-04T16:31:16.7149353Z Running inductor/test_custom_partitioner_fn 1/1 ... [2025-12-04 16:31:16.714637][26260.405028984] 2025-12-04T16:31:16.7149996Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:31:16.7153359Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_custom_partitioner_fn.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:31:16.715043] 2025-12-04T16:31:35.5066486Z 2025-12-04T16:31:35.5067730Z inductor/test_custom_partitioner_fn 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_custom_partitioner_fn_1.1_2eb6199e2d992792_.log 2025-12-04T16:31:35.5069229Z Running 1 items in this shard: test/inductor/test_custom_partitioner_fn.py::TestCustomPartitionerFn::test_custom_partitioner_fn 2025-12-04T16:31:35.5069895Z 2025-12-04T16:31:35.5070305Z Finished inductor/test_custom_partitioner_fn 1/1 ... [2025-12-04 16:31:35.506440][26279.196833982], took 0.31min 2025-12-04T16:31:35.5514047Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/inductor.test_custom_partitioner_fn/inductor.test_custom_partitioner_fn-44a32a2607e301e4.xml 2025-12-04T16:31:35.6302537Z Running dynamo/test_debug_utils 1/1 ... [2025-12-04 16:31:35.629972][26279.320364195] 2025-12-04T16:31:35.6303126Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:31:35.6306392Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_debug_utils.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:31:35.630386] 2025-12-04T16:31:44.0563575Z 2025-12-04T16:31:44.0564598Z dynamo/test_debug_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_debug_utils_1.1_3242effe45123c25_.log 2025-12-04T16:31:44.0567234Z Running 4 items in this shard: test/dynamo/test_debug_utils.py::TestDebugUtilsCUDA::test_cast_model_to_fp64_dtype_args_cuda, test/dynamo/test_debug_utils.py::TestDebugUtilsCUDA::test_generate_env_vars_string_cuda, test/dynamo/test_debug_utils.py::TestDebugUtilsDeviceCUDA::test_aot_graph_parser_cuda, test/dynamo/test_debug_utils.py::TestDebugUtilsDeviceCUDA::test_sym_aot_graph_parser_cuda 2025-12-04T16:31:44.0569163Z 2025-12-04T16:31:44.0569519Z Finished dynamo/test_debug_utils 1/1 ... [2025-12-04 16:31:44.056136][26287.746530074], took 0.14min 2025-12-04T16:31:44.1013995Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_debug_utils/dynamo.test_debug_utils-d82e5bff075a86e1.xml 2025-12-04T16:31:44.2001999Z Running dynamo/test_base_hop 1/1 ... [2025-12-04 16:31:44.199876][26287.890266583] 2025-12-04T16:31:44.2002551Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:31:44.2006114Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_base_hop.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:31:44.200317] 2025-12-04T16:31:53.5277824Z 2025-12-04T16:31:53.5278781Z dynamo/test_base_hop 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_base_hop_1.1_81fd091fbb10b281_.log 2025-12-04T16:31:53.5283129Z Running 11 items in this shard: test/dynamo/test_base_hop.py::BaseHOPTest::test_aliasing_mutation_error, test/dynamo/test_base_hop.py::BaseHOPTest::test_aot_eager, test/dynamo/test_base_hop.py::BaseHOPTest::test_auto_functionalize, test/dynamo/test_base_hop.py::BaseHOPTest::test_dynamo, test/dynamo/test_base_hop.py::BaseHOPTest::test_eager_call, test/dynamo/test_base_hop.py::BaseHOPTest::test_int_input, test/dynamo/test_base_hop.py::BaseHOPTest::test_none_input, test/dynamo/test_base_hop.py::BaseHOPTest::test_schema_gen_pytree_in_out, test/dynamo/test_base_hop.py::BaseHOPTest::test_schema_gen_pytree_in_out_with_mutation, test/dynamo/test_base_hop.py::BaseHOPTest::test_schema_gen_single_return, test/dynamo/test_base_hop.py::BaseHOPTest::test_schema_gen_single_return_with_mutation 2025-12-04T16:31:53.5286837Z 2025-12-04T16:31:53.5287379Z Finished dynamo/test_base_hop 1/1 ... [2025-12-04 16:31:53.527608][26297.218001517], took 0.16min 2025-12-04T16:31:53.5728172Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_base_hop/dynamo.test_base_hop-6ccb089675742efe.xml 2025-12-04T16:31:53.6548674Z Running dynamo/test_package 1/1 ... [2025-12-04 16:31:53.654563][26297.344953167] 2025-12-04T16:31:53.6549234Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:31:53.6552291Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_package.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:31:53.654991] 2025-12-04T16:32:47.8512104Z 2025-12-04T16:32:47.8513630Z dynamo/test_package 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_package_1.1_bbe95160f77383a1_.log 2025-12-04T16:32:47.8557193Z Running 50 items in this shard: test/dynamo/test_package.py::TestPackage::test_automatic_dynamo_graph_breaks_device_cpu, test/dynamo/test_package.py::TestPackage::test_automatic_dynamo_graph_breaks_device_cuda, test/dynamo/test_package.py::TestPackage::test_automatic_dynamo_graph_breaks_device_xpu, test/dynamo/test_package.py::TestPackage::test_automatic_dynamo_lazy_backward_device_cpu, test/dynamo/test_package.py::TestPackage::test_automatic_dynamo_lazy_backward_device_cuda, test/dynamo/test_package.py::TestPackage::test_automatic_dynamo_lazy_backward_device_xpu, test/dynamo/test_package.py::TestPackage::test_automatic_dynamo_recompiles_device_cpu, test/dynamo/test_package.py::TestPackage::test_automatic_dynamo_recompiles_device_cuda, test/dynamo/test_package.py::TestPackage::test_automatic_dynamo_recompiles_device_xpu, test/dynamo/test_package.py::TestPackage::test_automatic_dynamo_serialize_device_cpu, test/dynamo/test_package.py::TestPackage::test_automatic_dynamo_serialize_device_cuda, test/dynamo/test_package.py::TestPackage::test_automatic_dynamo_serialize_device_xpu, test/dynamo/test_package.py::TestPackage::test_basic_fn_backend_eager_device_cpu, test/dynamo/test_package.py::TestPackage::test_basic_fn_backend_eager_device_cuda, test/dynamo/test_package.py::TestPackage::test_basic_fn_backend_eager_device_xpu, test/dynamo/test_package.py::TestPackage::test_basic_fn_backend_inductor_device_cpu, test/dynamo/test_package.py::TestPackage::test_basic_fn_backend_inductor_device_cuda, test/dynamo/test_package.py::TestPackage::test_basic_fn_backend_inductor_device_xpu, test/dynamo/test_package.py::TestPackage::test_call_function_from_resume_device_cpu, test/dynamo/test_package.py::TestPackage::test_call_function_from_resume_device_cuda, test/dynamo/test_package.py::TestPackage::test_call_function_from_resume_device_xpu, test/dynamo/test_package.py::TestPackage::test_code_with_generator_device_cpu, test/dynamo/test_package.py::TestPackage::test_code_with_generator_device_cuda, test/dynamo/test_package.py::TestPackage::test_code_with_generator_device_xpu, test/dynamo/test_package.py::TestPackage::test_dynamic_shape_backend_eager_device_cpu, test/dynamo/test_package.py::TestPackage::test_dynamic_shape_backend_eager_device_cuda, test/dynamo/test_package.py::TestPackage::test_dynamic_shape_backend_eager_device_xpu, test/dynamo/test_package.py::TestPackage::test_dynamic_shape_backend_inductor_device_cpu, test/dynamo/test_package.py::TestPackage::test_dynamic_shape_backend_inductor_device_cuda, test/dynamo/test_package.py::TestPackage::test_dynamic_shape_backend_inductor_device_xpu, test/dynamo/test_package.py::TestPackage::test_dynamo_cache_manual_load_device_cpu, test/dynamo/test_package.py::TestPackage::test_dynamo_cache_manual_load_device_cuda, test/dynamo/test_package.py::TestPackage::test_dynamo_cache_manual_load_device_xpu, test/dynamo/test_package.py::TestPackage::test_file_change, test/dynamo/test_package.py::TestPackage::test_graph_break_bomb_backend_eager_device_cpu, test/dynamo/test_package.py::TestPackage::test_graph_break_bomb_backend_eager_device_cuda, test/dynamo/test_package.py::TestPackage::test_graph_break_bomb_backend_eager_device_xpu, test/dynamo/test_package.py::TestPackage::test_graph_break_bomb_backend_inductor_device_cpu, test/dynamo/test_package.py::TestPackage::test_graph_break_bomb_backend_inductor_device_cuda, test/dynamo/test_package.py::TestPackage::test_graph_break_bomb_backend_inductor_device_xpu, test/dynamo/test_package.py::TestPackage::test_graph_break_partial_backend_device_cpu, test/dynamo/test_package.py::TestPackage::test_graph_break_partial_backend_device_cuda, test/dynamo/test_package.py::TestPackage::test_graph_break_partial_backend_device_xpu, test/dynamo/test_package.py::TestPackage::test_lazy_backward_backend_eager_device_cpu, test/dynamo/test_package.py::TestPackage::test_lazy_backward_backend_eager_device_cuda, test/dynamo/test_package.py::TestPackage::test_lazy_backward_backend_eager_device_xpu, test/dynamo/test_package.py::TestPackage::test_lazy_backward_backend_inductor_device_cpu, test/dynamo/test_package.py::TestPackage::test_lazy_backward_backend_inductor_device_cuda, test/dynamo/test_package.py::TestPackage::test_lazy_backward_backend_inductor_device_xpu, test/dynamo/test_package.py::TestPackage::test_nn_module 2025-12-04T16:32:47.8598400Z 2025-12-04T16:32:47.8599186Z Finished dynamo/test_package 1/1 ... [2025-12-04 16:32:47.851020][26351.541412791], took 0.90min 2025-12-04T16:32:47.9002680Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_package/dynamo.test_package-9643d20377a48274.xml 2025-12-04T16:32:47.9827169Z Running dynamo/test_aot_autograd_cache 1/1 ... [2025-12-04 16:32:47.982322][26351.672713289] 2025-12-04T16:32:47.9828148Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:32:47.9831922Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_aot_autograd_cache.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:32:47.982805] 2025-12-04T16:35:16.4242958Z 2025-12-04T16:35:16.4244097Z dynamo/test_aot_autograd_cache 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_aot_autograd_cache_1.1_f0d2e6b1214663ab_.log 2025-12-04T16:35:16.4305200Z Running 108 items in this shard: test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_aot_runtime_trace_joint, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_autograd_function, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_autograd_guard_single_entry_device_cuda_bfloat16, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_autograd_guard_single_entry_device_cuda_float16, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_autograd_inductor_guards_device_cuda_bfloat16_requires_grad_False, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_autograd_inductor_guards_device_cuda_bfloat16_requires_grad_True, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_autograd_inductor_guards_device_cuda_float16_requires_grad_False, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_autograd_inductor_guards_device_cuda_float16_requires_grad_True, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_autograd_lazy_backward, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_autograd_no_dynamo_trace_backward, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_basic, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_cache_hot_load_device_cpu_bfloat16_dynamic_False, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_cache_hot_load_device_cpu_bfloat16_dynamic_True, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_cache_hot_load_device_cpu_float32_dynamic_False, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_cache_hot_load_device_cpu_float32_dynamic_True, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_cache_hot_load_device_cuda_bfloat16_dynamic_False, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_cache_hot_load_device_cuda_bfloat16_dynamic_True, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_cache_hot_load_device_cuda_float32_dynamic_False, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_cache_hot_load_device_cuda_float32_dynamic_True, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_cache_lazy_backward_for_compiled_autograd, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_clear_fx_graph_cache, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_compiled_autograd_bypass, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_constant_tensor_device_guards, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_custom_autograd_function, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_custom_autograd_function_miss, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_custom_autograd_function_with_custom_triton_kernel, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_custom_autograd_function_with_custom_triton_kernel_cache_invalidation, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_dynamic_shapes_different_sizes, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_fx_graph_cache_off, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_inference_graph_cache_hit_with_compiled_autograd_enabled, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_invoke_subgraph, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_multi_graph_specialization, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_multiple_compile_triton_kernels, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_nn_module_with_params_global_constant, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_non_bundled_to_bundled_config_change, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_regional_inductor_basic, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_regional_inductor_cache_miss_on_change, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_regional_inductor_with_backward, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_saved_tensors_hooks_autograd_cache, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_saved_tensors_hooks_autograd_cache_symbolic, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_symbol_specialization, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_triton_op_cache_invalidation, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_triton_op_cache_multiple_ops_invalidation, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_unsafe_mark_cacheable_fn_select_allow_in_graph, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_unsafe_mark_cacheable_fn_select_tag_activation_checkpoint, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_view_replay, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheTests::test_vmap, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_aot_runtime_trace_joint, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_autograd_function, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_autograd_guard_single_entry_device_cuda_bfloat16, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_autograd_guard_single_entry_device_cuda_float16, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_autograd_inductor_guards_device_cuda_bfloat16_requires_grad_False, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_autograd_inductor_guards_device_cuda_bfloat16_requires_grad_True, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_autograd_inductor_guards_device_cuda_float16_requires_grad_False, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_autograd_inductor_guards_device_cuda_float16_requires_grad_True, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_autograd_lazy_backward, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_autograd_no_dynamo_trace_backward, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_basic, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_cache_hot_load_device_cpu_bfloat16_dynamic_False, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_cache_hot_load_device_cpu_bfloat16_dynamic_True, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_cache_hot_load_device_cpu_float32_dynamic_False, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_cache_hot_load_device_cpu_float32_dynamic_True, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_cache_hot_load_device_cuda_bfloat16_dynamic_False, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_cache_hot_load_device_cuda_bfloat16_dynamic_True, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_cache_hot_load_device_cuda_float32_dynamic_False, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_cache_hot_load_device_cuda_float32_dynamic_True, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_cache_lazy_backward_for_compiled_autograd, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_clear_fx_graph_cache, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_compiled_autograd_bypass, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_constant_tensor_device_guards, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_custom_autograd_function, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_custom_autograd_function_miss, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_custom_autograd_function_with_custom_triton_kernel, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_custom_autograd_function_with_custom_triton_kernel_cache_invalidation, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_dynamic_shapes_different_sizes, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_fx_graph_cache_off, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_inference_graph_cache_hit_with_compiled_autograd_enabled, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_invoke_subgraph, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_multi_graph_specialization, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_multiple_compile_triton_kernels, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_nn_module_with_params_global_constant, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_non_bundled_to_bundled_config_change, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_regional_inductor_basic, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_regional_inductor_cache_miss_on_change, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_regional_inductor_with_backward, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_saved_tensors_hooks_autograd_cache, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_saved_tensors_hooks_autograd_cache_symbolic, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_symbol_specialization, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_triton_op_cache_invalidation, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_triton_op_cache_multiple_ops_invalidation, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_unsafe_mark_cacheable_fn_select_allow_in_graph, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_unsafe_mark_cacheable_fn_select_tag_activation_checkpoint, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_view_replay, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCacheBundledTests::test_vmap, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCachePicklerTests::test_basic_hash_key, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCachePicklerTests::test_different_configs, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCachePicklerTests::test_different_global_configs, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCachePicklerTests::test_different_graphs, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCachePicklerTests::test_different_inputs, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCachePicklerTests::test_freezing, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCachePicklerTests::test_identical_graphs_and_configs, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCachePicklerTests::test_incompatible_function, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCachePicklerTests::test_nn_module_with_params, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCachePicklerTests::test_normal_torch_function, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCachePicklerTests::test_private_builtin, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCachePicklerTests::test_private_namespace, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCachePicklerTests::test_safe_torchfunction, test/dynamo/test_aot_autograd_cache.py::AOTAutogradCachePicklerTests::test_sanitize_gm_for_cache 2025-12-04T16:35:16.4363938Z 2025-12-04T16:35:16.4364324Z Finished dynamo/test_aot_autograd_cache 1/1 ... [2025-12-04 16:35:16.424308][26500.114697173], took 2.47min 2025-12-04T16:35:16.4704977Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_aot_autograd_cache/dynamo.test_aot_autograd_cache-2e1ac30e1e69e405.xml 2025-12-04T16:35:16.5633131Z Running inductor/test_mps_basic 1/1 ... [2025-12-04 16:35:16.562961][26500.253352125] 2025-12-04T16:35:16.5633732Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:35:16.5636260Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_mps_basic.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:35:16.563378] 2025-12-04T16:35:28.4849924Z 2025-12-04T16:35:28.4851398Z inductor/test_mps_basic 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_mps_basic_1.1_979b26eef19a62b7_.log 2025-12-04T16:35:28.4852220Z 2025-12-04T16:35:28.4852589Z Finished inductor/test_mps_basic 1/1 ... [2025-12-04 16:35:28.484766][26512.175158886], took 0.20min 2025-12-04T16:35:28.5308616Z Running dynamo/test_comptime 1/1 ... [2025-12-04 16:35:28.530587][26512.220981124] 2025-12-04T16:35:28.5309379Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:35:28.5312633Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_comptime.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:35:28.530982] 2025-12-04T16:35:42.1644660Z 2025-12-04T16:35:42.1646027Z dynamo/test_comptime 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_comptime_1.1_a63436f12def3f22_.log 2025-12-04T16:35:42.1650845Z Running 12 items in this shard: test/dynamo/test_comptime.py::ComptimeTests::test_get_local, test/dynamo/test_comptime.py::ComptimeTests::test_get_local_closure_variable, test/dynamo/test_comptime.py::ComptimeTests::test_graph_break, test/dynamo/test_comptime.py::ComptimeTests::test_print_bt, test/dynamo/test_comptime.py::ComptimeTests::test_print_direct, test/dynamo/test_comptime.py::ComptimeTests::test_print_disas, test/dynamo/test_comptime.py::ComptimeTests::test_print_graph, test/dynamo/test_comptime.py::ComptimeTests::test_print_guards, test/dynamo/test_comptime.py::ComptimeTests::test_print_locals, test/dynamo/test_comptime.py::ComptimeTests::test_print_single, test/dynamo/test_comptime.py::ComptimeTests::test_print_value_stack, test/dynamo/test_comptime.py::ComptimeTests::test_sleep 2025-12-04T16:35:42.1654818Z 2025-12-04T16:35:42.1655152Z Finished dynamo/test_comptime 1/1 ... [2025-12-04 16:35:42.164284][26525.854677648], took 0.23min 2025-12-04T16:35:42.2104407Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/dynamo.test_comptime/dynamo.test_comptime-61e4f48b41f3f41b.xml 2025-12-04T16:35:42.2856271Z Running test_sort_and_select 1/1 ... [2025-12-04 16:35:42.285307][26525.975697837] 2025-12-04T16:35:42.2857110Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:35:42.2860632Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_sort_and_select.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:35:42.285736] 2025-12-04T16:35:53.9166611Z 2025-12-04T16:35:53.9168691Z test_sort_and_select 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_sort_and_select_1.1_5068bfed31d9fabf_.log 2025-12-04T16:35:53.9214583Z Running 111 items in this shard: test/test_sort_and_select.py::TestSortAndSelectCUDA::test_complex_unsupported_cpu_cuda, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_isin_cuda_float16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_isin_cuda_float32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_isin_cuda_float64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_isin_cuda_int16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_isin_cuda_int32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_isin_cuda_int64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_isin_cuda_int8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_isin_cuda_uint8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_isin_different_devices_cuda_float32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_isin_different_devices_cuda_float64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_isin_different_devices_cuda_int16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_isin_different_devices_cuda_int32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_isin_different_devices_cuda_int64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_isin_different_devices_cuda_int8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_isin_different_devices_cuda_uint8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_isin_different_dtypes_cuda, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_kthvalue_cuda_float64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_kthvalue_scalar_cuda_float32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_msort_cuda_bfloat16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_msort_cuda_float16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_msort_cuda_float32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_msort_cuda_float64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_msort_cuda_int16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_msort_cuda_int32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_msort_cuda_int64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_msort_cuda_int8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_msort_cuda_uint8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_sort_1d_output_discontiguous_cuda_float32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_sort_1d_parallel_cuda_int16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_sort_1d_parallel_cuda_int32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_sort_1d_parallel_cuda_int64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_sort_1d_parallel_cuda_int8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_sort_1d_parallel_cuda_uint8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_sort_cuda, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_sort_discontiguous_cuda_float32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_sort_discontiguous_slow_cuda_float32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_sort_expanded_tensor_cuda_float32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_sort_large_cuda_float16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_sort_large_slice_cuda, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_sort_overflow_cuda_int16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_sort_overflow_cuda_int32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_sort_overflow_cuda_int64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_sort_overflow_cuda_int8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_sort_overflow_cuda_uint8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_sort_restride_cuda_float32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_sort_stable_none_cuda, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_against_numpy_cuda_bfloat16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_against_numpy_cuda_bool, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_against_numpy_cuda_float16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_against_numpy_cuda_float32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_against_numpy_cuda_float64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_against_numpy_cuda_int16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_against_numpy_cuda_int32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_against_numpy_cuda_int64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_against_numpy_cuda_int8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_against_numpy_cuda_uint8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_cuda_bfloat16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_cuda_bool, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_cuda_float16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_cuda_float32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_cuda_float64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_cuda_int16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_cuda_int32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_cuda_int64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_cuda_int8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_stable_sort_cuda_uint8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_1d_output_discontiguous_cuda_float32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_4d_cuda, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_arguments_cuda, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_cuda, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_integral_cuda_int16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_integral_cuda_int32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_integral_cuda_int64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_integral_cuda_int8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_integral_cuda_uint8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_lower_precision_cuda_bfloat16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_lower_precision_cuda_float16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_noncontiguous_gpu_cuda, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_nonfinite_cuda_bfloat16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_nonfinite_cuda_float16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_nonfinite_cuda_float32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_nonfinite_cuda_float64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_quantized_scalar_input_cuda, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_zero_cuda_bfloat16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_zero_cuda_float32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_zero_cuda_float64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_zero_cuda_int16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_zero_cuda_int32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_zero_cuda_int64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_zero_cuda_int8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_topk_zero_cuda_uint8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_unique_consecutive_cuda_bool, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_unique_consecutive_cuda_float16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_unique_consecutive_cuda_float32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_unique_consecutive_cuda_float64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_unique_consecutive_cuda_int16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_unique_consecutive_cuda_int32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_unique_consecutive_cuda_int64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_unique_consecutive_cuda_int8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_unique_consecutive_cuda_uint8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_unique_cuda_bool, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_unique_cuda_float16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_unique_cuda_float32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_unique_cuda_float64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_unique_cuda_int16, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_unique_cuda_int32, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_unique_cuda_int64, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_unique_cuda_int8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_unique_cuda_uint8, test/test_sort_and_select.py::TestSortAndSelectCUDA::test_unique_dim_cuda 2025-12-04T16:35:53.9259671Z 2025-12-04T16:35:53.9259992Z Finished test_sort_and_select 1/1 ... [2025-12-04 16:35:53.916646][26537.607038253], took 0.19min 2025-12-04T16:35:53.9628575Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_sort_and_select/test_sort_and_select-ab12b6ad99352fe2.xml 2025-12-04T16:35:54.0616977Z Running functorch/test_rearrange 1/1 ... [2025-12-04 16:35:54.061315][26537.751705554] 2025-12-04T16:35:54.0617578Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:35:54.0620643Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/test_rearrange.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:35:54.061763] 2025-12-04T16:35:59.4838745Z 2025-12-04T16:35:59.4840142Z functorch/test_rearrange 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_rearrange_1.1_dc24a0c907a22605_.log 2025-12-04T16:35:59.4845087Z Running 10 items in this shard: test/functorch/test_rearrange.py::TestRearrange::test_0_dim_tensor, test/functorch/test_rearrange.py::TestRearrange::test_collapsed_ellipsis_errors_out, test/functorch/test_rearrange.py::TestRearrange::test_concatenations_and_stacking, test/functorch/test_rearrange.py::TestRearrange::test_dimension_mismatch_no_ellipsis, test/functorch/test_rearrange.py::TestRearrange::test_dimension_mismatch_with_ellipsis, test/functorch/test_rearrange.py::TestRearrange::test_ellipsis_ops, test/functorch/test_rearrange.py::TestRearrange::test_rearrange_consistency, test/functorch/test_rearrange.py::TestRearrange::test_rearrange_permutations, test/functorch/test_rearrange.py::TestRearrange::test_squeeze, test/functorch/test_rearrange.py::TestRearrange::test_unsqueeze 2025-12-04T16:35:59.4849102Z 2025-12-04T16:35:59.4849460Z Finished functorch/test_rearrange 1/1 ... [2025-12-04 16:35:59.483659][26543.174052936], took 0.09min 2025-12-04T16:35:59.5303204Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_rearrange/functorch.test_rearrange-e13c2240713d23f9.xml 2025-12-04T16:35:59.5603818Z Running functorch/test_parsing 1/1 ... [2025-12-04 16:35:59.560119][26543.250510354] 2025-12-04T16:35:59.5604625Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:35:59.5608101Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/test_parsing.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:35:59.560511] 2025-12-04T16:36:04.8826226Z 2025-12-04T16:36:04.8827479Z functorch/test_parsing 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_parsing_1.1_0964d504887b4f27_.log 2025-12-04T16:36:04.8834487Z Running 12 items in this shard: test/functorch/test_parsing.py::TestAnonymousAxis::test_anonymous_axes, test/functorch/test_parsing.py::TestParsedExpression::test_elementary_axis_name, test/functorch/test_parsing.py::TestParsedExpression::test_invalid_expressions, test/functorch/test_parsing.py::TestParsedExpression::test_parse_expression, test/functorch/test_parsing.py::TestParsingUtils::test_ellipsis_invalid_identifier, test/functorch/test_parsing.py::TestParsingUtils::test_ellipsis_matching, test/functorch/test_parsing.py::TestParsingUtils::test_left_parenthesized_ellipsis, test/functorch/test_parsing.py::TestParsingUtils::test_parse_pattern_number_of_arrows, test/functorch/test_parsing.py::TestValidateRearrangeExpressions::test_identifier_mismatch, test/functorch/test_parsing.py::TestValidateRearrangeExpressions::test_non_unitary_anonymous_axes_raises_error, test/functorch/test_parsing.py::TestValidateRearrangeExpressions::test_unexpected_axes_lengths, test/functorch/test_parsing.py::TestValidateRearrangeExpressions::test_validate_axes_lengths_are_integers 2025-12-04T16:36:04.8840581Z 2025-12-04T16:36:04.8840928Z Finished functorch/test_parsing 1/1 ... [2025-12-04 16:36:04.882430][26548.572824005], took 0.09min 2025-12-04T16:36:04.9295277Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_parsing/functorch.test_parsing-a56fd01f0949f839.xml 2025-12-04T16:36:04.9744976Z Running profiler/test_profiler 1/1 ... [2025-12-04 16:36:04.974183][26548.664574457] 2025-12-04T16:36:04.9745577Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:36:04.9748218Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'profiler/test_profiler.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:36:04.974595] 2025-12-04T16:37:35.7795584Z 2025-12-04T16:37:35.7796768Z profiler/test_profiler 1/1 was successful, full logs can be found in artifacts with path test/test-reports/profiler.test_profiler_1.1_f03b04cece9b91aa_.log 2025-12-04T16:37:35.7832162Z Running 83 items in this shard: test/profiler/test_profiler.py::TestProfilerCUDA::test_cudagraph_profiling_workaround, test/profiler/test_profiler.py::TestProfilerCUDA::test_custom_module_input_op_ids, test/profiler/test_profiler.py::TestProfilerCUDA::test_mem_leak, test/profiler/test_profiler.py::TestProfilerITT::test_custom_module_input_op_ids, test/profiler/test_profiler.py::TestProfiler::test_basic_chrome_trace, test/profiler/test_profiler.py::TestProfiler::test_basic_profile, test/profiler/test_profiler.py::TestProfiler::test_concrete_inputs_profiling, test/profiler/test_profiler.py::TestProfiler::test_concrete_inputs_profiling_toggling, test/profiler/test_profiler.py::TestProfiler::test_cpu_annotation_overlap, test/profiler/test_profiler.py::TestProfiler::test_disable_external_correlation, test/profiler/test_profiler.py::TestProfiler::test_dynamic_toggle, test/profiler/test_profiler.py::TestProfiler::test_event_list, test/profiler/test_profiler.py::TestProfiler::test_export_stacks, test/profiler/test_profiler.py::TestProfiler::test_flops, test/profiler/test_profiler.py::TestProfiler::test_forked_process, test/profiler/test_profiler.py::TestProfiler::test_guarded_record_function_fast, test/profiler/test_profiler.py::TestProfiler::test_high_level_trace, test/profiler/test_profiler.py::TestProfiler::test_is_profiler_enabled, test/profiler/test_profiler.py::TestProfiler::test_kineto, test/profiler/test_profiler.py::TestProfiler::test_kineto_multigpu, test/profiler/test_profiler.py::TestProfiler::test_kineto_profiler_api, test/profiler/test_profiler.py::TestProfiler::test_kineto_profiler_multiple_steppers, test/profiler/test_profiler.py::TestProfiler::test_kineto_profiler_with_environment_variable, test/profiler/test_profiler.py::TestProfiler::test_lazy_build_tree, test/profiler/test_profiler.py::TestProfiler::test_memory_profiler, test/profiler/test_profiler.py::TestProfiler::test_module_hierarchy, test/profiler/test_profiler.py::TestProfiler::test_nested_tensor_with_shapes, test/profiler/test_profiler.py::TestProfiler::test_oom_tracing, test/profiler/test_profiler.py::TestProfiler::test_override_time_units, test/profiler/test_profiler.py::TestProfiler::test_profile_all_threads, test/profiler/test_profiler.py::TestProfiler::test_profiler_correlation_id, test/profiler/test_profiler.py::TestProfiler::test_profiler_cuda_sync_events, test/profiler/test_profiler.py::TestProfiler::test_profiler_disable_fwd_bwd_link, test/profiler/test_profiler.py::TestProfiler::test_profiler_fwd_bwd_link, test/profiler/test_profiler.py::TestProfiler::test_profiler_metadata, test/profiler/test_profiler.py::TestProfiler::test_profiler_op_event_args, test/profiler/test_profiler.py::TestProfiler::test_profiler_op_event_kwargs, test/profiler/test_profiler.py::TestProfiler::test_profiler_op_event_kwargs_list_of_strings, test/profiler/test_profiler.py::TestProfiler::test_profiler_strides, test/profiler/test_profiler.py::TestProfiler::test_profiler_time_scale, test/profiler/test_profiler.py::TestProfiler::test_profiler_tracing, test/profiler/test_profiler.py::TestProfiler::test_profiler_type, test/profiler/test_profiler.py::TestProfiler::test_python_gc_event, test/profiler/test_profiler.py::TestProfiler::test_record_function_fast, test/profiler/test_profiler.py::TestProfiler::test_schedule_function_count, test/profiler/test_profiler.py::TestProfiler::test_skip_first_wait, test/profiler/test_profiler.py::TestProfiler::test_source, test/profiler/test_profiler.py::TestProfiler::test_source_multithreaded_basic_work_in_main_thread_False, test/profiler/test_profiler.py::TestProfiler::test_source_multithreaded_basic_work_in_main_thread_True, test/profiler/test_profiler.py::TestProfiler::test_source_multithreaded_close_in_scope_work_in_main_thread_False, test/profiler/test_profiler.py::TestProfiler::test_source_multithreaded_close_in_scope_work_in_main_thread_True, test/profiler/test_profiler.py::TestProfiler::test_source_multithreaded_complex_work_in_main_thread_False, test/profiler/test_profiler.py::TestProfiler::test_source_multithreaded_complex_work_in_main_thread_True, test/profiler/test_profiler.py::TestProfiler::test_source_multithreaded_multiple_preexisting_work_in_main_thread_False, test/profiler/test_profiler.py::TestProfiler::test_source_multithreaded_multiple_preexisting_work_in_main_thread_True, test/profiler/test_profiler.py::TestProfiler::test_source_multithreaded_open_in_scope_work_in_main_thread_False, test/profiler/test_profiler.py::TestProfiler::test_source_multithreaded_open_in_scope_work_in_main_thread_True, test/profiler/test_profiler.py::TestProfiler::test_tensorboard_trace_handler, test/profiler/test_profiler.py::TestProfiler::test_user_annotation, test/profiler/test_profiler.py::TestExperimentalUtils::test_bfs, test/profiler/test_profiler.py::TestExperimentalUtils::test_dfs, test/profiler/test_profiler.py::TestExperimentalUtils::test_expose_kineto_event_metadata, test/profiler/test_profiler.py::TestExperimentalUtils::test_fuzz_symbolize, test/profiler/test_profiler.py::TestExperimentalUtils::test_profiler_conv2d_bias_followed_by_batchnorm2d_pattern, test/profiler/test_profiler.py::TestExperimentalUtils::test_profiler_debug_autotuner, test/profiler/test_profiler.py::TestExperimentalUtils::test_profiler_extra_cuda_copy_pattern, test/profiler/test_profiler.py::TestExperimentalUtils::test_profiler_extra_cuda_copy_pattern_benchmark, test/profiler/test_profiler.py::TestExperimentalUtils::test_profiler_for_loop_indexing_pattern, test/profiler/test_profiler.py::TestExperimentalUtils::test_profiler_fp32_matmul_pattern, test/profiler/test_profiler.py::TestExperimentalUtils::test_profiler_grad_not_set_to_none_pattern, test/profiler/test_profiler.py::TestExperimentalUtils::test_profiler_matmul_dim_fp16_pattern, test/profiler/test_profiler.py::TestExperimentalUtils::test_profiler_name_pattern, test/profiler/test_profiler.py::TestExperimentalUtils::test_profiler_optimizer_single_tensor_pattern, test/profiler/test_profiler.py::TestExperimentalUtils::test_profiler_overload_names, test/profiler/test_profiler.py::TestExperimentalUtils::test_profiler_pattern_match_helper, test/profiler/test_profiler.py::TestExperimentalUtils::test_profiler_pattern_matcher_json_report, test/profiler/test_profiler.py::TestExperimentalUtils::test_profiler_synchronized_dataloader_pattern, test/profiler/test_profiler.py::TestExperimentalUtils::test_utils_compute_idle_time, test/profiler/test_profiler.py::TestExperimentalUtils::test_utils_compute_queue_depth, test/profiler/test_profiler.py::TestExperimentalUtils::test_utils_compute_queue_depth_when_no_cuda_events, test/profiler/test_profiler.py::TestExperimentalUtils::test_utils_compute_self_time, test/profiler/test_profiler.py::TestExperimentalUtils::test_utils_get_optimizable_events, test/profiler/test_profiler.py::TestExperimentalUtils::test_utils_intervals_overlap 2025-12-04T16:37:35.7866425Z 2025-12-04T16:37:35.7866772Z Finished profiler/test_profiler 1/1 ... [2025-12-04 16:37:35.779419][26639.469809841], took 1.51min 2025-12-04T16:37:35.8271514Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/profiler.test_profiler/profiler.test_profiler-ac51a5ffaf89322c.xml 2025-12-04T16:37:35.8962413Z Running torch_np/test_binary_ufuncs 1/1 ... [2025-12-04 16:37:35.895891][26639.58628185] 2025-12-04T16:37:35.8963018Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:37:35.8965941Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/test_binary_ufuncs.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:37:35.896346] 2025-12-04T16:37:41.5192607Z 2025-12-04T16:37:41.5193608Z torch_np/test_binary_ufuncs 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.test_binary_ufuncs_1.1_3f5612601f960528_.log 2025-12-04T16:37:41.5208578Z Running 38 items in this shard: test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_add, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_arctan2, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_bitwise_and, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_bitwise_or, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_bitwise_xor, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_copysign, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_divide, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_equal, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_float_power, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_floor_divide, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_fmax, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_fmin, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_fmod, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_gcd, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_greater, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_greater_equal, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_heaviside, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_hypot, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_lcm, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_ldexp, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_left_shift, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_less, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_less_equal, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_logaddexp, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_logaddexp2, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_logical_and, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_logical_or, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_logical_xor, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_matmul, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_maximum, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_minimum, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_multiply, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_nextafter, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_not_equal, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_power, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_remainder, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_right_shift, test/torch_np/test_binary_ufuncs.py::TestBinaryUfuncBasic::test_subtract 2025-12-04T16:37:41.5222882Z 2025-12-04T16:37:41.5223267Z Finished torch_np/test_binary_ufuncs 1/1 ... [2025-12-04 16:37:41.519110][26645.209501417], took 0.09min 2025-12-04T16:37:41.5660474Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.test_binary_ufuncs/torch_np.test_binary_ufuncs-e1b3ae8f2439c11d.xml 2025-12-04T16:37:41.5951136Z Running torch_np/test_unary_ufuncs 1/1 ... [2025-12-04 16:37:41.594865][26645.285257751] 2025-12-04T16:37:41.5951730Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:37:41.5954932Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/test_unary_ufuncs.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:37:41.595263] 2025-12-04T16:37:47.1178703Z 2025-12-04T16:37:47.1179774Z torch_np/test_unary_ufuncs 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.test_unary_ufuncs_1.1_2c643a870062efc9_.log 2025-12-04T16:37:47.1193927Z Running 42 items in this shard: test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_absolute, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_arccos, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_arccosh, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_arcsin, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_arcsinh, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_arctan, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_arctanh, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_cbrt, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_ceil, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_conjugate, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_cos, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_cosh, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_deg2rad, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_degrees, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_exp, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_exp2, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_expm1, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_fabs, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_floor, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_isfinite, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_isinf, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_isnan, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_log, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_log10, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_log1p, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_log2, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_logical_not, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_negative, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_positive, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_rad2deg, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_radians, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_reciprocal, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_rint, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_sign, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_signbit, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_sin, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_sinh, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_sqrt, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_square, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_tan, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_tanh, test/torch_np/test_unary_ufuncs.py::TestUnaryUfuncs::test_trunc 2025-12-04T16:37:47.1207460Z 2025-12-04T16:37:47.1207825Z Finished torch_np/test_unary_ufuncs 1/1 ... [2025-12-04 16:37:47.117715][26650.808108519], took 0.09min 2025-12-04T16:37:47.1642486Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.test_unary_ufuncs/torch_np.test_unary_ufuncs-0ead42044090ba01.xml 2025-12-04T16:37:47.1950652Z Running test_utils_filelock 1/1 ... [2025-12-04 16:37:47.194844][26650.885236718] 2025-12-04T16:37:47.1951187Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:37:47.1954790Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_utils_filelock.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:37:47.195234] 2025-12-04T16:37:52.6173990Z 2025-12-04T16:37:52.6175022Z test_utils_filelock 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_utils_filelock_1.1_4b9e876ba6fc808a_.log 2025-12-04T16:37:52.6176412Z Running 2 items in this shard: test/test_utils_filelock.py::TestFileLock::test_no_crash, test/test_utils_filelock.py::TestFileLock::test_sequencing 2025-12-04T16:37:52.6177226Z 2025-12-04T16:37:52.6177555Z Finished test_utils_filelock 1/1 ... [2025-12-04 16:37:52.617210][26656.307604487], took 0.09min 2025-12-04T16:37:52.6640777Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_utils_filelock/test_utils_filelock-bc82f2b31ad6d0a9.xml 2025-12-04T16:37:52.6928424Z Running test_extension_utils 1/1 ... [2025-12-04 16:37:52.692571][26656.382963511] 2025-12-04T16:37:52.6928997Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:37:52.6932881Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_extension_utils.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:37:52.692965] 2025-12-04T16:37:58.0651432Z 2025-12-04T16:37:58.0652397Z test_extension_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_extension_utils_1.1_d74913d9452f40a0_.log 2025-12-04T16:37:58.0654163Z Running 2 items in this shard: test/test_extension_utils.py::TestExtensionUtils::test_external_module_register, test/test_extension_utils.py::TestExtensionUtils::test_external_module_register_with_renamed_backend 2025-12-04T16:37:58.0655238Z 2025-12-04T16:37:58.0655565Z Finished test_extension_utils 1/1 ... [2025-12-04 16:37:58.064946][26661.755337573], took 0.09min 2025-12-04T16:37:58.1115882Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_extension_utils/test_extension_utils-1a92a953d2096623.xml 2025-12-04T16:37:58.1448392Z Running test_rename_privateuse1_to_existing_device 1/1 ... [2025-12-04 16:37:58.144556][26661.834946826] 2025-12-04T16:37:58.1449075Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:37:58.1452286Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_rename_privateuse1_to_existing_device.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:37:58.144974] 2025-12-04T16:38:03.5672770Z 2025-12-04T16:38:03.5674124Z test_rename_privateuse1_to_existing_device 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_rename_privateuse1_to_existing_device_1.1_066a744b408e3e3f_.log 2025-12-04T16:38:03.5675950Z Running 1 items in this shard: test/test_rename_privateuse1_to_existing_device.py::TestRenamePrivateuseoneToExistingBackend::test_external_module_register_with_existing_backend 2025-12-04T16:38:03.5676905Z 2025-12-04T16:38:03.5677362Z Finished test_rename_privateuse1_to_existing_device 1/1 ... [2025-12-04 16:38:03.567043][26667.257436714], took 0.09min 2025-12-04T16:38:03.6139476Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_rename_privateuse1_to_existing_device/test_rename_privateuse1_to_existing_device-a35ba0709779b2ea.xml 2025-12-04T16:38:03.6394398Z Running nn/attention/test_fa4 1/1 ... [2025-12-04 16:38:03.639184][26667.329574912] 2025-12-04T16:38:03.6395246Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:38:03.6398796Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'nn/attention/test_fa4.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:38:03.639638] 2025-12-04T16:38:10.8641916Z 2025-12-04T16:38:10.8642976Z nn/attention/test_fa4 1/1 was successful, full logs can be found in artifacts with path test/test-reports/nn.attention.test_fa4_1.1_c563975b7201948d_.log 2025-12-04T16:38:10.8696518Z Running 66 items in this shard: test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_fa4_kernel_called_bfloat16_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_fa4_kernel_called_float16_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_1_seq_len_1024_heads_4_head_dim_128_is_causal_False_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_1_seq_len_1024_heads_4_head_dim_128_is_causal_True_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_1_seq_len_1024_heads_4_head_dim_64_is_causal_False_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_1_seq_len_1024_heads_4_head_dim_64_is_causal_True_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_1_seq_len_1024_heads_8_head_dim_128_is_causal_False_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_1_seq_len_1024_heads_8_head_dim_128_is_causal_True_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_1_seq_len_1024_heads_8_head_dim_64_is_causal_False_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_1_seq_len_1024_heads_8_head_dim_64_is_causal_True_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_1_seq_len_512_heads_4_head_dim_128_is_causal_False_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_1_seq_len_512_heads_4_head_dim_128_is_causal_True_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_1_seq_len_512_heads_4_head_dim_64_is_causal_False_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_1_seq_len_512_heads_4_head_dim_64_is_causal_True_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_1_seq_len_512_heads_8_head_dim_128_is_causal_False_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_1_seq_len_512_heads_8_head_dim_128_is_causal_True_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_1_seq_len_512_heads_8_head_dim_64_is_causal_False_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_1_seq_len_512_heads_8_head_dim_64_is_causal_True_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_2_seq_len_1024_heads_4_head_dim_128_is_causal_False_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_2_seq_len_1024_heads_4_head_dim_128_is_causal_True_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_2_seq_len_1024_heads_4_head_dim_64_is_causal_False_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_2_seq_len_1024_heads_4_head_dim_64_is_causal_True_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_2_seq_len_1024_heads_8_head_dim_128_is_causal_False_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_2_seq_len_1024_heads_8_head_dim_128_is_causal_True_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_2_seq_len_1024_heads_8_head_dim_64_is_causal_False_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_2_seq_len_1024_heads_8_head_dim_64_is_causal_True_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_2_seq_len_512_heads_4_head_dim_128_is_causal_False_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_2_seq_len_512_heads_4_head_dim_128_is_causal_True_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_2_seq_len_512_heads_4_head_dim_64_is_causal_False_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_2_seq_len_512_heads_4_head_dim_64_is_causal_True_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_2_seq_len_512_heads_8_head_dim_128_is_causal_False_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_2_seq_len_512_heads_8_head_dim_128_is_causal_True_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_2_seq_len_512_heads_8_head_dim_64_is_causal_False_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_bfloat16_batch_2_seq_len_512_heads_8_head_dim_64_is_causal_True_cuda_bfloat16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_1_seq_len_1024_heads_4_head_dim_128_is_causal_False_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_1_seq_len_1024_heads_4_head_dim_128_is_causal_True_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_1_seq_len_1024_heads_4_head_dim_64_is_causal_False_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_1_seq_len_1024_heads_4_head_dim_64_is_causal_True_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_1_seq_len_1024_heads_8_head_dim_128_is_causal_False_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_1_seq_len_1024_heads_8_head_dim_128_is_causal_True_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_1_seq_len_1024_heads_8_head_dim_64_is_causal_False_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_1_seq_len_1024_heads_8_head_dim_64_is_causal_True_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_1_seq_len_512_heads_4_head_dim_128_is_causal_False_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_1_seq_len_512_heads_4_head_dim_128_is_causal_True_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_1_seq_len_512_heads_4_head_dim_64_is_causal_False_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_1_seq_len_512_heads_4_head_dim_64_is_causal_True_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_1_seq_len_512_heads_8_head_dim_128_is_causal_False_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_1_seq_len_512_heads_8_head_dim_128_is_causal_True_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_1_seq_len_512_heads_8_head_dim_64_is_causal_False_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_1_seq_len_512_heads_8_head_dim_64_is_causal_True_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_2_seq_len_1024_heads_4_head_dim_128_is_causal_False_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_2_seq_len_1024_heads_4_head_dim_128_is_causal_True_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_2_seq_len_1024_heads_4_head_dim_64_is_causal_False_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_2_seq_len_1024_heads_4_head_dim_64_is_causal_True_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_2_seq_len_1024_heads_8_head_dim_128_is_causal_False_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_2_seq_len_1024_heads_8_head_dim_128_is_causal_True_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_2_seq_len_1024_heads_8_head_dim_64_is_causal_False_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_2_seq_len_1024_heads_8_head_dim_64_is_causal_True_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_2_seq_len_512_heads_4_head_dim_128_is_causal_False_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_2_seq_len_512_heads_4_head_dim_128_is_causal_True_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_2_seq_len_512_heads_4_head_dim_64_is_causal_False_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_2_seq_len_512_heads_4_head_dim_64_is_causal_True_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_2_seq_len_512_heads_8_head_dim_128_is_causal_False_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_2_seq_len_512_heads_8_head_dim_128_is_causal_True_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_2_seq_len_512_heads_8_head_dim_64_is_causal_False_cuda_float16, test/nn/attention/test_fa4.py::TestFlashAttentionFA4CUDA::test_flash_attention_matches_math_float16_batch_2_seq_len_512_heads_8_head_dim_64_is_causal_True_cuda_float16 2025-12-04T16:38:10.8748920Z 2025-12-04T16:38:10.8749253Z Finished nn/attention/test_fa4 1/1 ... [2025-12-04 16:38:10.864190][26674.55458045], took 0.12min 2025-12-04T16:38:10.9113543Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/nn.attention.test_fa4/nn.attention.test_fa4-7e669f8dd97f47a6.xml 2025-12-04T16:38:10.9911318Z Running typing/test_python_operators 1/1 ... [2025-12-04 16:38:10.990887][26674.681280269] 2025-12-04T16:38:10.9912109Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:38:10.9915240Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'typing/test_python_operators.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:38:10.991284] 2025-12-04T16:38:17.1143272Z 2025-12-04T16:38:17.1144743Z typing/test_python_operators 1/1 was successful, full logs can be found in artifacts with path test/test-reports/typing.test_python_operators_1.1_7abeab1ea263e7df_.log 2025-12-04T16:38:17.1277102Z Running 318 items in this shard: test/typing/test_python_operators.py::TestPythonOperators::test_binary_a100_op_%_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a101_op_%_b101, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a102_op_%_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a103_op_%_b103, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a104_op_*_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a105_op_*_b105, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a106_op_*_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a107_op_*_b107, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a108_op_**_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a109_op_**_b109, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a110_op_**_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a111_op_**_b111, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a112_op_+_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a113_op_+_b113, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a114_op_+_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a115_op_+_b115, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a116_op_-_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a117_op_-_b117, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a118_op_-_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a119_op_-_b119, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a120_op_/_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a121_op_/_b121, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a122_op_/_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a123_op_/_b123, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a124_op_//_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a125_op_//_b125, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a126_op_//_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a127_op_//_b127, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a128_op_&_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a129_op_&_b129, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a130_op_&_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a131_op_&_b131, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a132_op_<<_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a133_op_<<_b133, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a134_op_<<_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a135_op_<<_b135, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a136_op_>>_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a137_op_>>_b137, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a138_op_>>_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a139_op_>>_b139, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a140_op_^_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a141_op_^_b141, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a142_op_^_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a143_op_^_b143, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a144_op_|_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a145_op_|_b145, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a146_op_|_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a147_op_|_b147, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a148_op_@_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a149_op_@_b149, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a150_op_@_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a151_op_@_b151, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a228_op_!=_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a229_op_!=_b229, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a230_op_!=_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a231_op_!=_b231, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a232_op_<_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a233_op_<_b233, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a234_op_<_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a235_op_<_b235, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a236_op_<=_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a237_op_<=_b237, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a238_op_<=_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a239_op_<=_b239, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a240_op_==_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a241_op_==_b241, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a242_op_==_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a243_op_==_b243, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a244_op_>_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a245_op_>_b245, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a246_op_>_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a247_op_>_b247, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a248_op_>=_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a249_op_>=_b249, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a250_op_>=_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a251_op_>=_b251, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a252_op_%_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a253_op_%_b253, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a254_op_%_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a255_op_%_b255, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a256_op_*_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a257_op_*_b257, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a258_op_*_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a259_op_*_b259, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a260_op_**_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a261_op_**_b261, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a262_op_**_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a263_op_**_b263, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a264_op_+_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a265_op_+_b265, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a266_op_+_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a267_op_+_b267, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a268_op_-_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a269_op_-_b269, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a270_op_-_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a271_op_-_b271, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a272_op_/_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a273_op_/_b273, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a274_op_/_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a275_op_/_b275, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a276_op_//_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a277_op_//_b277, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a278_op_//_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a279_op_//_b279, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a280_op_&_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a281_op_&_b281, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a282_op_&_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a283_op_&_b283, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a284_op_<<_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a285_op_<<_b285, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a286_op_<<_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a287_op_<<_b287, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a288_op_>>_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a289_op_>>_b289, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a290_op_>>_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a291_op_>>_b291, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a292_op_^_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a293_op_^_b293, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a294_op_^_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a295_op_^_b295, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a296_op_|_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a297_op_|_b297, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a298_op_|_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a299_op_|_b299, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a300_op_@_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a301_op_@_b301, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a302_op_@_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a303_op_@_b303, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a76_op_!=_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a77_op_!=_b77, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a78_op_!=_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a79_op_!=_b79, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a80_op_<_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a81_op_<_b81, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a82_op_<_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a83_op_<_b83, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a84_op_<=_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a85_op_<=_b85, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a86_op_<=_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a87_op_<=_b87, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a88_op_==_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a89_op_==_b89, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a90_op_==_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a91_op_==_b91, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a92_op_>_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a93_op_>_b93, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a94_op_>_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a95_op_>_b95, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a96_op_>=_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a97_op_>=_b97, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a98_op_>=_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a99_op_>=_b99, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_!=_b1, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_!=_b3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_!=_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_!=_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_%_b25, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_%_b27, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_%_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_%_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_&_b53, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_&_b55, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_&_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_&_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_**_b33, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_**_b35, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_**_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_**_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_*_b29, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_*_b31, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_*_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_*_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_+_b37, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_+_b39, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_+_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_+_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_-_b41, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_-_b43, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_-_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_-_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_//_b49, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_//_b51, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_//_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_//_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_/_b45, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_/_b47, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_/_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_/_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_<<_b57, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_<<_b59, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_<<_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_<<_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_<=_b11, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_<=_b9, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_<=_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_<=_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_<_b5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_<_b7, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_<_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_<_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_==_b13, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_==_b15, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_==_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_==_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_>=_b21, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_>=_b23, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_>=_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_>=_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_>>_b61, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_>>_b63, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_>>_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_>>_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_>_b17, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_>_b19, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_>_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_>_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_@_b73, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_@_b75, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_@_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_@_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_^_b65, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_^_b67, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_^_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_^_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_|_b69, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_|_b71, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_|_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_1_5_op_|_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_!=_b153, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_!=_b155, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_!=_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_!=_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_%_b177, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_%_b179, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_%_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_%_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_&_b205, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_&_b207, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_&_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_&_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_**_b185, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_**_b187, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_**_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_**_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_*_b181, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_*_b183, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_*_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_*_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_+_b189, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_+_b191, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_+_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_+_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_-_b193, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_-_b195, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_-_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_-_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_//_b201, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_//_b203, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_//_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_//_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_/_b197, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_/_b199, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_/_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_/_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_<<_b209, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_<<_b211, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_<<_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_<<_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_<=_b161, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_<=_b163, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_<=_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_<=_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_<_b157, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_<_b159, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_<_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_<_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_==_b165, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_==_b167, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_==_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_==_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_>=_b173, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_>=_b175, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_>=_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_>=_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_>>_b213, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_>>_b215, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_>>_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_>>_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_>_b169, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_>_b171, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_>_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_>_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_@_b225, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_@_b227, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_@_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_@_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_^_b217, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_^_b219, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_^_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_^_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_|_b221, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_|_b223, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_|_b_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_binary_a_3_op_|_b_3, test/typing/test_python_operators.py::TestPythonOperators::test_operators_are_correct_and_complete, test/typing/test_python_operators.py::TestPythonOperators::test_type_tests_are_complete, test/typing/test_python_operators.py::TestPythonOperators::test_unary_op_+_a1, test/typing/test_python_operators.py::TestPythonOperators::test_unary_op_+_a3, test/typing/test_python_operators.py::TestPythonOperators::test_unary_op_+_a_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_unary_op_+_a_3, test/typing/test_python_operators.py::TestPythonOperators::test_unary_op_-_a5, test/typing/test_python_operators.py::TestPythonOperators::test_unary_op_-_a7, test/typing/test_python_operators.py::TestPythonOperators::test_unary_op_-_a_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_unary_op_-_a_3, test/typing/test_python_operators.py::TestPythonOperators::test_unary_op_~_a11, test/typing/test_python_operators.py::TestPythonOperators::test_unary_op_~_a9, test/typing/test_python_operators.py::TestPythonOperators::test_unary_op_~_a_1_5, test/typing/test_python_operators.py::TestPythonOperators::test_unary_op_~_a_3 2025-12-04T16:38:17.1406112Z 2025-12-04T16:38:17.1406497Z Finished typing/test_python_operators 1/1 ... [2025-12-04 16:38:17.114676][26680.80506653], took 0.10min 2025-12-04T16:38:17.1620973Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/typing.test_python_operators/typing.test_python_operators-6bf858ca1197a53f.xml 2025-12-04T16:38:17.2589217Z Running test_functionalization 1/1 ... [2025-12-04 16:38:17.258546][26680.948936678] 2025-12-04T16:38:17.2589818Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:38:17.2592587Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_functionalization.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:38:17.258984] 2025-12-04T16:38:27.5890628Z 2025-12-04T16:38:27.5891487Z test_functionalization 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_functionalization_1.1_e1689373b3a112cf_.log 2025-12-04T16:38:27.5942688Z Running 112 items in this shard: test/test_functionalization.py::TestFunctionalization::test_advanced_indexing, test/test_functionalization.py::TestFunctionalization::test_advanced_indexing_correct_strides, test/test_functionalization.py::TestFunctionalization::test_aliases_maintained_after_pass_when_reapplying_views, test/test_functionalization.py::TestFunctionalization::test_as_strided, test/test_functionalization.py::TestFunctionalization::test_batch_norm, test/test_functionalization.py::TestFunctionalization::test_cat, test/test_functionalization.py::TestFunctionalization::test_channels_last_contiguous, test/test_functionalization.py::TestFunctionalization::test_copy_, test/test_functionalization.py::TestFunctionalization::test_copy_stride_mismatch, test/test_functionalization.py::TestFunctionalization::test_diagonal, test/test_functionalization.py::TestFunctionalization::test_diagonal_mutated_input, test/test_functionalization.py::TestFunctionalization::test_everything, test/test_functionalization.py::TestFunctionalization::test_expand_symint, test/test_functionalization.py::TestFunctionalization::test_fill_, test/test_functionalization.py::TestFunctionalization::test_freeze, test/test_functionalization.py::TestFunctionalization::test_index_mutation_on_non_input, test/test_functionalization.py::TestFunctionalization::test_inplace_on_non_view, test/test_functionalization.py::TestFunctionalization::test_instance_norm, test/test_functionalization.py::TestFunctionalization::test_metadata_change, test/test_functionalization.py::TestFunctionalization::test_metadata_change_out_op, test/test_functionalization.py::TestFunctionalization::test_mixed_wrappers_invalid, test/test_functionalization.py::TestFunctionalization::test_mixed_wrappers_valid, test/test_functionalization.py::TestFunctionalization::test_multi_out, test/test_functionalization.py::TestFunctionalization::test_multiple_views_of_same_base, test/test_functionalization.py::TestFunctionalization::test_mutable_op_not_inplace_or_other, test/test_functionalization.py::TestFunctionalization::test_mutation_overlapping_mem, test/test_functionalization.py::TestFunctionalization::test_nested_functions_propagate_updates, test/test_functionalization.py::TestFunctionalization::test_only_one_view, test/test_functionalization.py::TestFunctionalization::test_optional_tensor_list, test/test_functionalization.py::TestFunctionalization::test_python_functionalization, test/test_functionalization.py::TestFunctionalization::test_python_functionalization_conj, test/test_functionalization.py::TestFunctionalization::test_python_functionalization_is_conj, test/test_functionalization.py::TestFunctionalization::test_python_functionalization_is_neg, test/test_functionalization.py::TestFunctionalization::test_python_functionalization_lift_fresh, test/test_functionalization.py::TestFunctionalization::test_python_functionalization_lift_fresh_storage, test/test_functionalization.py::TestFunctionalization::test_python_functionalization_neg, test/test_functionalization.py::TestFunctionalization::test_python_functionalization_zero_tensor, test/test_functionalization.py::TestFunctionalization::test_reapply_views_simple, test/test_functionalization.py::TestFunctionalization::test_resize_larger_invalid, test/test_functionalization.py::TestFunctionalization::test_resize_larger_valid, test/test_functionalization.py::TestFunctionalization::test_resize_same_size_diff_rank, test/test_functionalization.py::TestFunctionalization::test_resize_smaller, test/test_functionalization.py::TestFunctionalization::test_save_for_backwards_segfault, test/test_functionalization.py::TestFunctionalization::test_scalars, test/test_functionalization.py::TestFunctionalization::test_set_, test/test_functionalization.py::TestFunctionalization::test_simple, test/test_functionalization.py::TestFunctionalization::test_simple_out, test/test_functionalization.py::TestFunctionalization::test_slice, test/test_functionalization.py::TestFunctionalization::test_split, test/test_functionalization.py::TestFunctionalization::test_split_with_sizes, test/test_functionalization.py::TestFunctionalization::test_tensor_ctr, test/test_functionalization.py::TestFunctionalization::test_tensor_list_composite, test/test_functionalization.py::TestFunctionalization::test_tensor_list_mixed_functional_nonfunctional, test/test_functionalization.py::TestFunctionalization::test_unbind, test/test_functionalization.py::TestFunctionalization::test_view_clone_view_inplace, test/test_functionalization.py::TestFunctionalization::test_view_inplace, test/test_functionalization.py::TestCrossRefFunctionalization::test_advanced_indexing, test/test_functionalization.py::TestCrossRefFunctionalization::test_advanced_indexing_correct_strides, test/test_functionalization.py::TestCrossRefFunctionalization::test_aliases_maintained_after_pass_when_reapplying_views, test/test_functionalization.py::TestCrossRefFunctionalization::test_as_strided, test/test_functionalization.py::TestCrossRefFunctionalization::test_batch_norm, test/test_functionalization.py::TestCrossRefFunctionalization::test_cat, test/test_functionalization.py::TestCrossRefFunctionalization::test_channels_last_contiguous, test/test_functionalization.py::TestCrossRefFunctionalization::test_copy_, test/test_functionalization.py::TestCrossRefFunctionalization::test_copy_stride_mismatch, test/test_functionalization.py::TestCrossRefFunctionalization::test_diagonal, test/test_functionalization.py::TestCrossRefFunctionalization::test_diagonal_mutated_input, test/test_functionalization.py::TestCrossRefFunctionalization::test_everything, test/test_functionalization.py::TestCrossRefFunctionalization::test_expand_symint, test/test_functionalization.py::TestCrossRefFunctionalization::test_fill_, test/test_functionalization.py::TestCrossRefFunctionalization::test_freeze, test/test_functionalization.py::TestCrossRefFunctionalization::test_index_mutation_on_non_input, test/test_functionalization.py::TestCrossRefFunctionalization::test_inplace_on_non_view, test/test_functionalization.py::TestCrossRefFunctionalization::test_instance_norm, test/test_functionalization.py::TestCrossRefFunctionalization::test_metadata_change, test/test_functionalization.py::TestCrossRefFunctionalization::test_metadata_change_out_op, test/test_functionalization.py::TestCrossRefFunctionalization::test_mixed_wrappers_invalid, test/test_functionalization.py::TestCrossRefFunctionalization::test_mixed_wrappers_valid, test/test_functionalization.py::TestCrossRefFunctionalization::test_multi_out, test/test_functionalization.py::TestCrossRefFunctionalization::test_multiple_views_of_same_base, test/test_functionalization.py::TestCrossRefFunctionalization::test_mutable_op_not_inplace_or_other, test/test_functionalization.py::TestCrossRefFunctionalization::test_mutation_overlapping_mem, test/test_functionalization.py::TestCrossRefFunctionalization::test_nested_functions_propagate_updates, test/test_functionalization.py::TestCrossRefFunctionalization::test_only_one_view, test/test_functionalization.py::TestCrossRefFunctionalization::test_optional_tensor_list, test/test_functionalization.py::TestCrossRefFunctionalization::test_python_functionalization, test/test_functionalization.py::TestCrossRefFunctionalization::test_python_functionalization_conj, test/test_functionalization.py::TestCrossRefFunctionalization::test_python_functionalization_is_conj, test/test_functionalization.py::TestCrossRefFunctionalization::test_python_functionalization_is_neg, test/test_functionalization.py::TestCrossRefFunctionalization::test_python_functionalization_lift_fresh, test/test_functionalization.py::TestCrossRefFunctionalization::test_python_functionalization_lift_fresh_storage, test/test_functionalization.py::TestCrossRefFunctionalization::test_python_functionalization_neg, test/test_functionalization.py::TestCrossRefFunctionalization::test_python_functionalization_zero_tensor, test/test_functionalization.py::TestCrossRefFunctionalization::test_reapply_views_simple, test/test_functionalization.py::TestCrossRefFunctionalization::test_resize_larger_invalid, test/test_functionalization.py::TestCrossRefFunctionalization::test_resize_larger_valid, test/test_functionalization.py::TestCrossRefFunctionalization::test_resize_same_size_diff_rank, test/test_functionalization.py::TestCrossRefFunctionalization::test_resize_smaller, test/test_functionalization.py::TestCrossRefFunctionalization::test_save_for_backwards_segfault, test/test_functionalization.py::TestCrossRefFunctionalization::test_scalars, test/test_functionalization.py::TestCrossRefFunctionalization::test_set_, test/test_functionalization.py::TestCrossRefFunctionalization::test_simple, test/test_functionalization.py::TestCrossRefFunctionalization::test_simple_out, test/test_functionalization.py::TestCrossRefFunctionalization::test_slice, test/test_functionalization.py::TestCrossRefFunctionalization::test_split, test/test_functionalization.py::TestCrossRefFunctionalization::test_split_with_sizes, test/test_functionalization.py::TestCrossRefFunctionalization::test_tensor_ctr, test/test_functionalization.py::TestCrossRefFunctionalization::test_tensor_list_composite, test/test_functionalization.py::TestCrossRefFunctionalization::test_tensor_list_mixed_functional_nonfunctional, test/test_functionalization.py::TestCrossRefFunctionalization::test_unbind, test/test_functionalization.py::TestCrossRefFunctionalization::test_view_clone_view_inplace, test/test_functionalization.py::TestCrossRefFunctionalization::test_view_inplace 2025-12-04T16:38:27.5992916Z 2025-12-04T16:38:27.5993280Z Finished test_functionalization 1/1 ... [2025-12-04 16:38:27.589000][26691.27939157], took 0.17min 2025-12-04T16:38:27.6373214Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_functionalization/test_functionalization-44af17f38403c26c.xml 2025-12-04T16:38:27.7170191Z Running profiler/test_kineto 1/1 ... [2025-12-04 16:38:27.716708][26691.407098488] 2025-12-04T16:38:27.7170763Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:38:27.7173904Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'profiler/test_kineto.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:38:27.717146] 2025-12-04T16:38:45.6570958Z 2025-12-04T16:38:45.6572343Z profiler/test_kineto 1/1 was successful, full logs can be found in artifacts with path test/test-reports/profiler.test_kineto_1.1_5a67f12c59b2d264_.log 2025-12-04T16:38:45.6573813Z Running 1 items in this shard: test/profiler/test_kineto.py::SimpleKinetoInitializationTest::test_kineto_profiler_with_environment_variable 2025-12-04T16:38:45.6574563Z 2025-12-04T16:38:45.6574908Z Finished profiler/test_kineto 1/1 ... [2025-12-04 16:38:45.656860][26709.347253765], took 0.30min 2025-12-04T16:38:45.7048801Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/profiler.test_kineto/profiler.test_kineto-1d87d2a72eae2747.xml 2025-12-04T16:38:45.7780339Z Running test_module_tracker 1/1 ... [2025-12-04 16:38:45.777732][26709.468123979] 2025-12-04T16:38:45.7780895Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:38:45.7783928Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_module_tracker.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:38:45.778157] 2025-12-04T16:38:51.1503419Z 2025-12-04T16:38:51.1504370Z test_module_tracker 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_module_tracker_1.1_ac47ee0370be4ffc_.log 2025-12-04T16:38:51.1506223Z Running 3 items in this shard: test/test_module_tracker.py::TestModuleTracker::test_bw_detection, test/test_module_tracker.py::TestModuleTracker::test_confused_hierarchy, test/test_module_tracker.py::TestModuleTracker::test_module_hierarchy 2025-12-04T16:38:51.1507438Z 2025-12-04T16:38:51.1507791Z Finished test_module_tracker 1/1 ... [2025-12-04 16:38:51.150107][26714.840498469], took 0.09min 2025-12-04T16:38:51.1983130Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_module_tracker/test_module_tracker-fac08a90d6f586d0.xml 2025-12-04T16:38:51.2309964Z Running torch_np/numpy_tests/core/test_scalarinherit 1/1 ... [2025-12-04 16:38:51.230750][26714.921141891] 2025-12-04T16:38:51.2310680Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:38:51.2314222Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/core/test_scalarinherit.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:38:51.231160] 2025-12-04T16:38:56.3029328Z 2025-12-04T16:38:56.3030818Z torch_np/numpy_tests/core/test_scalarinherit 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.core.test_scalarinherit_1.1_98f4b4754a63741f_.log 2025-12-04T16:38:56.3033086Z Running 3 items in this shard: test/torch_np/numpy_tests/core/test_scalarinherit.py::TestInherit::test_gh_15395, test/torch_np/numpy_tests/core/test_scalarinherit.py::TestInherit::test_init, test/torch_np/numpy_tests/core/test_scalarinherit.py::TestInherit::test_init2 2025-12-04T16:38:56.3034448Z 2025-12-04T16:38:56.3034931Z Finished torch_np/numpy_tests/core/test_scalarinherit 1/1 ... [2025-12-04 16:38:56.302728][26719.993121573], took 0.08min 2025-12-04T16:38:56.3507299Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.core.test_scalarinherit/torch_np.numpy_tests.core.test_scalarinherit-9bae58f1bf1858ad.xml 2025-12-04T16:38:56.3788134Z Running test_tensorexpr_pybind 1/1 ... [2025-12-04 16:38:56.378581][26720.068973718] 2025-12-04T16:38:56.3788696Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:38:56.3792155Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_tensorexpr_pybind.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:38:56.378988] 2025-12-04T16:39:01.8512055Z 2025-12-04T16:39:01.8513271Z test_tensorexpr_pybind 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_tensorexpr_pybind_1.1_7dab8d3742d21018_.log 2025-12-04T16:39:01.8520649Z Running 17 items in this shard: test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_alloc_in_loop, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_call_raw, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_dtype_error, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_dynamic_shape, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_dynamic_shape_2d, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_external_calls, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_shape_prop, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_shape_prop_module, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_custom_lowering, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_expand, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_permute, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_scalar_inputs, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_t, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_tensor_inputs, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_kernel_with_transpose, test/test_tensorexpr_pybind.py::TestTensorExprPyBind::test_simple_sum, test/test_tensorexpr_pybind.py::TestExprHandlePyBind::test_unary_ops 2025-12-04T16:39:01.8527710Z 2025-12-04T16:39:01.8528049Z Finished test_tensorexpr_pybind 1/1 ... [2025-12-04 16:39:01.851051][26725.541443739], took 0.09min 2025-12-04T16:39:01.8988650Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_tensorexpr_pybind/test_tensorexpr_pybind-ea4c0e89818d6b23.xml 2025-12-04T16:39:01.9294124Z Running test_fx_experimental 1/1 ... [2025-12-04 16:39:01.929128][26725.619518871] 2025-12-04T16:39:01.9294672Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:39:01.9297950Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_fx_experimental.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:39:01.929545] 2025-12-04T16:39:26.8338728Z 2025-12-04T16:39:26.8340212Z test_fx_experimental 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_fx_experimental_1.1_6141111e1c7aeffb_.log 2025-12-04T16:39:26.8783773Z Running 724 items in this shard: test/test_fx_experimental.py::TestFXExperimental::test_annotate_getitem_node, test/test_fx_experimental.py::TestFXExperimental::test_annotate_returns_with_schema, test/test_fx_experimental.py::TestFXExperimental::test_aot_based_partition, test/test_fx_experimental.py::TestFXExperimental::test_call_to_assert_no_msg, test/test_fx_experimental.py::TestFXExperimental::test_call_to_assert_with_empty_msg, test/test_fx_experimental.py::TestFXExperimental::test_call_to_assert_with_msg, test/test_fx_experimental.py::TestFXExperimental::test_call_to_assert_with_multiline_message, test/test_fx_experimental.py::TestFXExperimental::test_conv_bn_fusion, test/test_fx_experimental.py::TestFXExperimental::test_conv_bn_fusion_mixed_dtype, test/test_fx_experimental.py::TestFXExperimental::test_conv_bn_fusion_not_running_state, test/test_fx_experimental.py::TestFXExperimental::test_cost_aware_partition, test/test_fx_experimental.py::TestFXExperimental::test_fetch, test/test_fx_experimental.py::TestFXExperimental::test_find_single_partition, test/test_fx_experimental.py::TestFXExperimental::test_lack_of_devices, test/test_fx_experimental.py::TestFXExperimental::test_large_node_error, test/test_fx_experimental.py::TestFXExperimental::test_merge_matmuls, test/test_fx_experimental.py::TestFXExperimental::test_meta_tracer, test/test_fx_experimental.py::TestFXExperimental::test_normalize_args, test/test_fx_experimental.py::TestFXExperimental::test_normalize_args_perserve_type, test/test_fx_experimental.py::TestFXExperimental::test_normalize_args_preserve_meta, test/test_fx_experimental.py::TestFXExperimental::test_normalize_binary_operators, test/test_fx_experimental.py::TestFXExperimental::test_normalize_modules_exhaustive, test/test_fx_experimental.py::TestFXExperimental::test_optimize_for_inference_cpu, test/test_fx_experimental.py::TestFXExperimental::test_optimize_for_inference_cpu_torchvision, test/test_fx_experimental.py::TestFXExperimental::test_partition_device_mapping, test/test_fx_experimental.py::TestFXExperimental::test_partition_latency, test/test_fx_experimental.py::TestFXExperimental::test_partition_node_manipulation, test/test_fx_experimental.py::TestFXExperimental::test_replace_target_nodes_with, test/test_fx_experimental.py::TestFXExperimental::test_saturate_host, test/test_fx_experimental.py::TestFXExperimental::test_size_based_partition, test/test_fx_experimental.py::TestFXExperimental::test_sparse_nn_partition, test/test_fx_experimental.py::TestFXExperimental::test_split_module_dead_code, test/test_fx_experimental.py::TestFXExperimental::test_split_module_default_arg, test/test_fx_experimental.py::TestFXExperimental::test_split_module_input_names, test/test_fx_experimental.py::TestFXExperimental::test_split_module_keep_original_order_and_noop_graph, test/test_fx_experimental.py::TestFXExperimental::test_split_module_kwargs_expansion, test/test_fx_experimental.py::TestFXExperimental::test_split_module_return_node, test/test_fx_experimental.py::TestFXExperimental::test_split_module_symint_dependency_handling, test/test_fx_experimental.py::TestFXExperimental::test_split_qualname_mapping, test/test_fx_experimental.py::TestFXExperimental::test_subgraph_creation, test/test_fx_experimental.py::TestFXExperimental::test_subgraph_trivial_resnet, test/test_fx_experimental.py::TestFXExperimental::test_subgraph_uniquename, test/test_fx_experimental.py::TestFXExperimental::test_to_folder, test/test_fx_experimental.py::TestFXExperimental::test_traceable_function_with_nonstandard_name, test/test_fx_experimental.py::TestFXExperimental::test_type_matches, test/test_fx_experimental.py::TestTranslationValidation::test_sat, test/test_fx_experimental.py::TestTranslationValidation::test_sat_bitwise, test/test_fx_experimental.py::TestTranslationValidation::test_sympy_to_z3, test/test_fx_experimental.py::TestTranslationValidation::test_unsat, test/test_fx_experimental.py::TestTranslationValidation::test_z3str, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_args_op_overload_cuda, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_H_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_T_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive___getitem___cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive___radd___cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive___rdiv___cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive___rmatmul___cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive___rmod___cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive___rmul___cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive___rpow___cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive___rsub___cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive__batch_norm_with_update_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive__chunk_cat_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive__native_batch_norm_legit_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive__segment_reduce_lengths_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive__segment_reduce_offsets_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive__softmax_backward_data_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive__unsafe_masked_index_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive__unsafe_masked_index_put_accumulate_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive__upsample_bilinear2d_aa_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_abs_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_acos_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_acosh_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_add_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_addbmm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_addcdiv_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_addcmul_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_addmm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_addmm_decomposed_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_addmv_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_addr_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_alias_copy_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_all_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_allclose_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_amax_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_amin_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_aminmax_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_angle_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_any_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_arange_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_argmax_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_argmin_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_argsort_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_argwhere_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_as_strided_copy_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_as_strided_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_as_strided_partial_views_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_as_strided_scatter_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_asin_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_asinh_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_atan2_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_atan_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_atanh_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_atleast_1d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_atleast_2d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_atleast_3d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_baddbmm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_bernoulli_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_bfloat16_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_block_diag_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_bmm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_bool_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_broadcast_shapes_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_broadcast_tensors_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_broadcast_to_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_bucketize_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_byte_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_cartesian_prod_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_cat_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_cauchy_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_cdist_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_cdouble_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_ceil_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_cfloat_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_chalf_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_char_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_cholesky_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_cholesky_inverse_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_cholesky_solve_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_chunk_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_clamp_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_clamp_max_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_clamp_min_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_clone_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_column_stack_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_combinations_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_complex_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_conj_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_conj_physical_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_constant_pad_nd_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_contiguous_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_copysign_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_corrcoef_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_cos_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_cosh_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_count_nonzero_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_cov_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_cross_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_cummax_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_cummin_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_cumprod_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_cumsum_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_cumulative_trapezoid_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_deg2rad_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_diag_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_diag_embed_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_diagflat_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_diagonal_copy_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_diagonal_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_diagonal_scatter_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_diff_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_digamma_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_dist_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_div_floor_rounding_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_div_no_rounding_mode_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_div_trunc_rounding_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_dot_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_double_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_dsplit_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_dstack_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_einsum_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_empty_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_empty_like_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_empty_permuted_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_empty_strided_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_eq_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_equal_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_erf_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_erfc_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_erfinv_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_exp2_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_exp_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_expand_as_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_expand_copy_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_expand_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_expm1_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_exponential_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_eye_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_fft2_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_fft_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_fftn_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_fftshift_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_hfft2_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_hfft_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_hfftn_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_ifft2_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_ifft_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_ifftn_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_ifftshift_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_ihfft2_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_ihfft_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_ihfftn_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_irfft2_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_irfft_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_irfftn_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_rfft2_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_rfft_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fft_rfftn_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fill_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_flatten_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_flip_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fliplr_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_flipud_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_float_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_float_power_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_floor_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_floor_divide_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fmax_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fmin_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_fmod_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_frac_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_frexp_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_full_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_full_like_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_gather_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_ge_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_geometric_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_geqrf_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_gradient_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_grid_sampler_2d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_grid_sampler_3d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_gt_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_half_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_hash_tensor_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_heaviside_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_histc_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_hsplit_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_hstack_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_hypot_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_i0_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_igamma_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_igammac_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_index_add_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_index_copy_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_index_fill_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_index_put_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_index_reduce_amax_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_index_reduce_amin_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_index_reduce_mean_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_index_reduce_prod_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_index_select_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_inner_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_int_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_isclose_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_isfinite_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_isin_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_isinf_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_isnan_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_isneginf_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_isposinf_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_isreal_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_item_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_jiterator_2inputs_2outputs_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_jiterator_4inputs_with_extra_args_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_jiterator_binary_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_jiterator_binary_return_by_ref_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_jiterator_unary_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_kron_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_kthvalue_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_ldexp_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_le_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_lerp_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_lgamma_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_cholesky_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_cholesky_ex_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_cond_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_cross_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_det_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_diagonal_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_eig_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_eigh_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_eigvals_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_eigvalsh_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_householder_product_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_inv_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_inv_ex_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_ldl_factor_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_ldl_factor_ex_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_ldl_solve_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_lstsq_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_lstsq_grad_oriented_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_lu_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_lu_factor_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_lu_factor_ex_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_lu_solve_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_matrix_norm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_matrix_power_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_matrix_rank_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_matrix_rank_hermitian_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_multi_dot_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_norm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_norm_subgradients_at_zero_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_pinv_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_pinv_hermitian_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_pinv_singular_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_qr_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_slogdet_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_solve_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_solve_ex_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_solve_triangular_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_svd_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_svdvals_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_tensorinv_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_tensorsolve_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_vander_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_vecdot_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linalg_vector_norm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linspace_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_linspace_tensor_overload_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_log10_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_log1p_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_log2_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_log_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_log_normal_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_log_softmax_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_log_softmax_with_dtype_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_logaddexp2_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_logaddexp_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_logcumsumexp_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_logdet_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_logical_and_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_logical_not_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_logical_or_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_logical_xor_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_logit_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_logspace_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_logspace_tensor_overload_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_logsumexp_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_long_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_lt_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_lu_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_lu_solve_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_lu_unpack_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_mH_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_mT_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_amax_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_amin_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_argmax_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_argmin_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_cumprod_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_cumsum_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_fill_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_log_softmax_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_logaddexp_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_logsumexp_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_mean_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_median_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_norm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_normalize_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_prod_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_scatter_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_select_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_softmax_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_softmin_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_std_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_sum_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_masked_var_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_matmul_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_matrix_exp_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_max_binary_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_max_pool2d_with_indices_backward_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_max_reduction_no_dim_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_max_reduction_with_dim_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_maximum_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_mean_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_median_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_meshgrid_list_of_tensors_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_meshgrid_variadic_tensors_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_min_binary_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_min_reduction_no_dim_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_min_reduction_with_dim_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_minimum_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_mm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_mode_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_movedim_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_msort_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_mul_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_multinomial_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_mv_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_mvlgamma_mvlgamma_p_1_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_mvlgamma_mvlgamma_p_3_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_mvlgamma_mvlgamma_p_5_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nan_to_num_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nanmean_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nanmedian_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nanquantile_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nansum_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_narrow_copy_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_narrow_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_native_batch_norm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_native_dropout_backward_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_native_layer_norm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_ne_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_neg_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_new_empty_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_new_empty_strided_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_new_full_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_new_ones_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_new_zeros_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nextafter_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_adaptive_avg_pool1d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_adaptive_avg_pool2d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_adaptive_avg_pool3d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_adaptive_max_pool1d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_adaptive_max_pool2d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_adaptive_max_pool3d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_alpha_dropout_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_avg_pool1d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_avg_pool2d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_avg_pool3d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_batch_norm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_batch_norm_without_cudnn_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_bilinear_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_binary_cross_entropy_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_binary_cross_entropy_with_logits_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_celu_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_channel_shuffle_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_conv1d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_conv2d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_conv3d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_conv_transpose1d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_conv_transpose2d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_conv_transpose3d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_cosine_embedding_loss_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_cosine_similarity_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_cross_entropy_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_ctc_loss_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_dropout2d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_dropout3d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_dropout_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_elu_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_embedding_bag_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_embedding_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_feature_alpha_dropout_with_train_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_fractional_max_pool2d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_fractional_max_pool3d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_gaussian_nll_loss_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_gelu_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_glu_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_grid_sample_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_group_norm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_hardshrink_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_hardsigmoid_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_hardswish_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_hardtanh_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_hinge_embedding_loss_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_huber_loss_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_instance_norm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_interpolate_area_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_interpolate_bicubic_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_interpolate_bilinear_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_interpolate_linear_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_interpolate_nearest-exact_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_interpolate_nearest_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_interpolate_trilinear_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_kl_div_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_l1_loss_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_layer_norm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_leaky_relu_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_linear_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_local_response_norm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_logsigmoid_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_margin_ranking_loss_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_max_pool1d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_max_pool2d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_max_pool3d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_max_unpool1d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_max_unpool1d_grad_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_max_unpool2d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_max_unpool2d_grad_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_max_unpool3d_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_max_unpool3d_grad_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_mish_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_mse_loss_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_multi_head_attention_forward_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_multi_margin_loss_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_multilabel_margin_loss_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_multilabel_soft_margin_loss_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_nll_loss_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_normalize_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_pad_circular_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_pad_constant_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_pad_reflect_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_pad_replicate_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_pad_replicate_negative_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_pairwise_distance_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_pdist_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_pixel_shuffle_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_pixel_unshuffle_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_poisson_nll_loss_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_prelu_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_relu6_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_relu_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_rms_norm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_rrelu_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_scaled_dot_product_attention_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_selu_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_silu_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_smooth_l1_loss_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_soft_margin_loss_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_softmin_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_softmin_with_dtype_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_softplus_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_softshrink_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_softsign_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_tanhshrink_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_threshold_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_triplet_margin_loss_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_triplet_margin_with_distance_loss_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_unfold_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_upsample_bilinear_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nn_functional_upsample_nearest_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nonzero_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_nonzero_static_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_norm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_norm_fro_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_norm_inf_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_norm_nuc_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_normal_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_normal_in_place_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_normal_number_mean_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_ones_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_ones_like_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_ormqr_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_outer_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_pca_lowrank_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_permute_copy_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_permute_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_pinverse_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_polar_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_polygamma_polygamma_n_0_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_polygamma_polygamma_n_1_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_polygamma_polygamma_n_2_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_polygamma_polygamma_n_3_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_polygamma_polygamma_n_4_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_positive_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_pow_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_prod_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_put_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_qr_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_quantile_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_rad2deg_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_rand_like_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_randint_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_randint_like_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_randn_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_randn_like_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_ravel_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_real_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_reciprocal_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_remainder_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_renorm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_repeat_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_repeat_interleave_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_reshape_as_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_reshape_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_resize__cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_resize_as__cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_resolve_conj_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_resolve_neg_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_roll_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_rot90_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_round_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_round_decimals_0_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_round_decimals_3_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_round_decimals_neg_3_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_rsqrt_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_rsub_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_scalar_tensor_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_scatter_add_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_scatter_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_scatter_reduce_amax_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_scatter_reduce_amin_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_scatter_reduce_mean_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_scatter_reduce_prod_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_scatter_reduce_sum_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_searchsorted_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_select_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_select_scatter_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_sgn_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_short_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_sigmoid_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_sign_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_signal_windows_bartlett_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_signal_windows_blackman_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_signal_windows_cosine_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_signal_windows_exponential_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_signal_windows_gaussian_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_signal_windows_general_cosine_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_signal_windows_general_hamming_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_signal_windows_hamming_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_signal_windows_hann_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_signal_windows_kaiser_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_signal_windows_nuttall_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_signbit_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_sin_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_sinc_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_sinh_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_slice_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_slice_scatter_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_softmax_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_softmax_with_dtype_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_sort_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_sparse_mm_reduce_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_sparse_sampled_addmm_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_airy_ai_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_bessel_j0_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_bessel_j1_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_bessel_y0_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_bessel_y1_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_chebyshev_polynomial_t_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_chebyshev_polynomial_u_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_chebyshev_polynomial_v_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_chebyshev_polynomial_w_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_entr_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_erfcx_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_hermite_polynomial_h_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_hermite_polynomial_he_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_i0e_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_i1_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_i1e_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_laguerre_polynomial_l_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_legendre_polynomial_p_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_log_ndtr_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_modified_bessel_i0_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_modified_bessel_i1_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_modified_bessel_k0_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_modified_bessel_k1_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_ndtr_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_ndtri_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_polygamma_special_polygamma_n_0_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_scaled_modified_bessel_k0_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_scaled_modified_bessel_k1_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_shifted_chebyshev_polynomial_t_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_shifted_chebyshev_polynomial_u_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_shifted_chebyshev_polynomial_v_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_shifted_chebyshev_polynomial_w_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_spherical_bessel_j0_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_xlog1py_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_special_zeta_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_split_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_split_list_args_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_split_with_sizes_copy_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_split_with_sizes_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_sqrt_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_square_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_squeeze_copy_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_squeeze_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_squeeze_multiple_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_stack_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_std_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_std_mean_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_std_mean_unbiased_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_std_unbiased_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_stft_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_sub_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_sum_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_sum_to_size_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_svd_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_svd_lowrank_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_t_copy_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_t_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_take_along_dim_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_take_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_tan_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_tanh_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_tensor_split_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_tensordot_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_tile_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_to_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_to_sparse_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_topk_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_torch_ops_aten__efficient_attention_forward_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_torch_ops_aten__safe_softmax_default_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_trace_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_transpose_copy_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_transpose_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_trapezoid_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_trapz_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_triangular_solve_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_tril_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_triu_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_true_divide_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_trunc_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_unbind_copy_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_unbind_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_unflatten_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_unfold_copy_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_unfold_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_uniform_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_unique_consecutive_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_unique_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_unsafe_chunk_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_unsafe_split_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_unsqueeze_copy_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_unsqueeze_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_var_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_var_mean_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_var_mean_unbiased_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_var_unbiased_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_vdot_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_view_as_complex_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_view_as_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_view_copy_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_view_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_vsplit_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_vstack_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_where_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_xlogy_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_zero__cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_zeros_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_operator_exhaustive_zeros_like_cuda_float32, test/test_fx_experimental.py::TestNormalizeOperatorsCUDA::test_normalize_quantized_eb_cuda 2025-12-04T16:39:26.9222658Z 2025-12-04T16:39:26.9223291Z Finished test_fx_experimental 1/1 ... [2025-12-04 16:39:26.834999][26750.525389228], took 0.42min 2025-12-04T16:39:26.9224520Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_fx_experimental/test_fx_experimental-ab6703fb0b80faae.xml 2025-12-04T16:39:26.9834106Z Running test_fx_passes 1/1 ... [2025-12-04 16:39:26.983078][26750.673468667] 2025-12-04T16:39:26.9834827Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:39:26.9837943Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_fx_passes.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:39:26.983518] 2025-12-04T16:39:33.0564927Z 2025-12-04T16:39:33.0566262Z test_fx_passes 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_fx_passes_1.1_01e7f0194a1e970b_.log 2025-12-04T16:39:33.0590535Z Running 53 items in this shard: test/test_fx_passes.py::TestFXGraphPasses::test_fuser_pass_deep_model, test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition0, test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition1, test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition10, test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition11, test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition2, test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition3, test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition4, test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition5, test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition6, test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition7, test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition8, test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_partition9, test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_xfail_partition0, test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_xfail_partition1, test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_xfail_partition2, test/test_fx_passes.py::TestFXGraphPasses::test_fuser_util_xfail_partition3, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn0_expected_partition0_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn10_expected_partition10_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn11_expected_partition11_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn12_expected_partition12_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn13_expected_partition13_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn14_expected_partition14_bookend_non_compute_pass_True, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn15_expected_partition15_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn16_expected_partition16_bookend_non_compute_pass_True, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn17_expected_partition17_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn18_expected_partition18_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn1_expected_partition1_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn2_expected_partition2_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn3_expected_partition3_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn4_expected_partition4_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn5_expected_partition5_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn6_expected_partition6_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn7_expected_partition7_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn8_expected_partition8_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_fn9_expected_partition9_bookend_non_compute_pass_False, test/test_fx_passes.py::TestFXGraphPasses::test_partitioner_independent_output_fn0_expected_partition0, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model0, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model1, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model10, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model11, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model12, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model13, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model14, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model15, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model2, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model3, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model4, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model5, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model6, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model7, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model8, test/test_fx_passes.py::TestFXMatcherUtils::test_subgraph_matcher_test_model9 2025-12-04T16:39:33.0614134Z 2025-12-04T16:39:33.0614421Z Finished test_fx_passes 1/1 ... [2025-12-04 16:39:33.056416][26756.746808503], took 0.10min 2025-12-04T16:39:33.1048375Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_fx_passes/test_fx_passes-268f4eef0e7a6c57.xml 2025-12-04T16:39:33.2071311Z Running functorch/test_logging 1/1 ... [2025-12-04 16:39:33.206822][26756.897212883] 2025-12-04T16:39:33.2072071Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:39:33.2075503Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/test_logging.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:39:33.207269] 2025-12-04T16:39:38.7294754Z 2025-12-04T16:39:38.7296875Z functorch/test_logging 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_logging_1.1_ac84cd84fa23725e_.log 2025-12-04T16:39:38.7299168Z Running 1 items in this shard: test/functorch/test_logging.py::TestAOTLogging::test_logging 2025-12-04T16:39:38.7300119Z 2025-12-04T16:39:38.7300763Z Finished functorch/test_logging 1/1 ... [2025-12-04 16:39:38.729249][26762.419642471], took 0.09min 2025-12-04T16:39:38.7782101Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_logging/functorch.test_logging-94ce26f72ae6dfc2.xml 2025-12-04T16:39:38.8078920Z Running test_namedtensor 1/1 ... [2025-12-04 16:39:38.807532][26762.497924224] 2025-12-04T16:39:38.8079800Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:39:38.8083408Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_namedtensor.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:39:38.807988] 2025-12-04T16:39:46.5836220Z 2025-12-04T16:39:46.5837200Z test_namedtensor 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_namedtensor_1.1_b9881adc6407f0b9_.log 2025-12-04T16:39:46.5866201Z Running 89 items in this shard: test/test_namedtensor.py::TestNamedTensor::test_aaa_must_run_first_check_experimental_warning, test/test_namedtensor.py::TestNamedTensor::test_addcmul_addcdiv, test/test_namedtensor.py::TestNamedTensor::test_addmm, test/test_namedtensor.py::TestNamedTensor::test_addmv, test/test_namedtensor.py::TestNamedTensor::test_align_as, test/test_namedtensor.py::TestNamedTensor::test_align_tensors, test/test_namedtensor.py::TestNamedTensor::test_align_tensors_two_inputs, test/test_namedtensor.py::TestNamedTensor::test_align_to, test/test_namedtensor.py::TestNamedTensor::test_align_to_ellipsis, test/test_namedtensor.py::TestNamedTensor::test_any_all, test/test_namedtensor.py::TestNamedTensor::test_as_strided, test/test_namedtensor.py::TestNamedTensor::test_as_strided_cuda, test/test_namedtensor.py::TestNamedTensor::test_autograd_ignores_names, test/test_namedtensor.py::TestNamedTensor::test_autograd_smoke, test/test_namedtensor.py::TestNamedTensor::test_autograd_warns_named_grad, test/test_namedtensor.py::TestNamedTensor::test_bernoulli, test/test_namedtensor.py::TestNamedTensor::test_big_tensor_repr_has_names, test/test_namedtensor.py::TestNamedTensor::test_binary_ops, test/test_namedtensor.py::TestNamedTensor::test_bitwise_not, test/test_namedtensor.py::TestNamedTensor::test_bmm, test/test_namedtensor.py::TestNamedTensor::test_cat, test/test_namedtensor.py::TestNamedTensor::test_cdist, test/test_namedtensor.py::TestNamedTensor::test_comparison_ops, test/test_namedtensor.py::TestNamedTensor::test_copy_transpose, test/test_namedtensor.py::TestNamedTensor::test_cummax_cummin, test/test_namedtensor.py::TestNamedTensor::test_detach, test/test_namedtensor.py::TestNamedTensor::test_diagonal, test/test_namedtensor.py::TestNamedTensor::test_dot, test/test_namedtensor.py::TestNamedTensor::test_empty_names, test/test_namedtensor.py::TestNamedTensor::test_equal, test/test_namedtensor.py::TestNamedTensor::test_expand, test/test_namedtensor.py::TestNamedTensor::test_factory_coverage, test/test_namedtensor.py::TestNamedTensor::test_factory_edge_cases, test/test_namedtensor.py::TestNamedTensor::test_flatten, test/test_namedtensor.py::TestNamedTensor::test_flatten_index_error, test/test_namedtensor.py::TestNamedTensor::test_flatten_nodims, test/test_namedtensor.py::TestNamedTensor::test_has_names, test/test_namedtensor.py::TestNamedTensor::test_index_fill, test/test_namedtensor.py::TestNamedTensor::test_info_smoke, test/test_namedtensor.py::TestNamedTensor::test_logcumsumexp, test/test_namedtensor.py::TestNamedTensor::test_logical_not, test/test_namedtensor.py::TestNamedTensor::test_logical_ops, test/test_namedtensor.py::TestNamedTensor::test_masked_fill, test/test_namedtensor.py::TestNamedTensor::test_masked_select, test/test_namedtensor.py::TestNamedTensor::test_matmul, test/test_namedtensor.py::TestNamedTensor::test_max_pooling, test/test_namedtensor.py::TestNamedTensor::test_max_pooling_without_names_does_not_warn, test/test_namedtensor.py::TestNamedTensor::test_mm, test/test_namedtensor.py::TestNamedTensor::test_mv, test/test_namedtensor.py::TestNamedTensor::test_no_jit_script_support, test/test_namedtensor.py::TestNamedTensor::test_no_jit_tracer_support, test/test_namedtensor.py::TestNamedTensor::test_no_multiprocessing_support, test/test_namedtensor.py::TestNamedTensor::test_no_pickle_support, test/test_namedtensor.py::TestNamedTensor::test_no_save_support, test/test_namedtensor.py::TestNamedTensor::test_noncontig_contiguous, test/test_namedtensor.py::TestNamedTensor::test_none_names_refcount, test/test_namedtensor.py::TestNamedTensor::test_nyi_dimname_overload_msg, test/test_namedtensor.py::TestNamedTensor::test_out_fn_semantics, test/test_namedtensor.py::TestNamedTensor::test_pow_special, test/test_namedtensor.py::TestNamedTensor::test_py3_ellipsis, test/test_namedtensor.py::TestNamedTensor::test_reduction_fns, test/test_namedtensor.py::TestNamedTensor::test_refine_names, test/test_namedtensor.py::TestNamedTensor::test_rename, test/test_namedtensor.py::TestNamedTensor::test_rename_, test/test_namedtensor.py::TestNamedTensor::test_rename_globber, test/test_namedtensor.py::TestNamedTensor::test_rename_rename_map, test/test_namedtensor.py::TestNamedTensor::test_repr, test/test_namedtensor.py::TestNamedTensor::test_resize, test/test_namedtensor.py::TestNamedTensor::test_select, test/test_namedtensor.py::TestNamedTensor::test_select_cuda, test/test_namedtensor.py::TestNamedTensor::test_set_names_property, test/test_namedtensor.py::TestNamedTensor::test_size, test/test_namedtensor.py::TestNamedTensor::test_split_fns_propagates_names, test/test_namedtensor.py::TestNamedTensor::test_squeeze, test/test_namedtensor.py::TestNamedTensor::test_stride, test/test_namedtensor.py::TestNamedTensor::test_support_device_named_grad, test/test_namedtensor.py::TestNamedTensor::test_tensor_from_lists, test/test_namedtensor.py::TestNamedTensor::test_tensor_from_named_tensor, test/test_namedtensor.py::TestNamedTensor::test_tensor_from_numpy, test/test_namedtensor.py::TestNamedTensor::test_tensor_from_tensor, test/test_namedtensor.py::TestNamedTensor::test_tensor_grad_is_unnamed, test/test_namedtensor.py::TestNamedTensor::test_transpose_variants, test/test_namedtensor.py::TestNamedTensor::test_trivial, test/test_namedtensor.py::TestNamedTensor::test_unary_propagate_names_fns, test/test_namedtensor.py::TestNamedTensor::test_unflatten, test/test_namedtensor.py::TestNamedTensor::test_unsupported_op_error_msg, test/test_namedtensor.py::TestNamedTensor::test_using_seen_interned_string_doesnt_bump_refcount, test/test_namedtensor.py::TestNamedTensor::test_using_unseen_interned_string_bumps_refcount_permanently, test/test_namedtensor.py::TestNamedTensor::test_using_unseen_uninterned_string_refcounts 2025-12-04T16:39:46.5894356Z 2025-12-04T16:39:46.5894667Z Finished test_namedtensor 1/1 ... [2025-12-04 16:39:46.583532][26770.273923618], took 0.13min 2025-12-04T16:39:46.6326827Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_namedtensor/test_namedtensor-9c123d78058198ac.xml 2025-12-04T16:39:46.7226432Z Running test_tensorexpr 1/1 ... [2025-12-04 16:39:46.722326][26770.412716758] 2025-12-04T16:39:46.7226995Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:39:46.7230134Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_tensorexpr.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:39:46.722764] 2025-12-04T16:40:34.2542812Z 2025-12-04T16:40:34.2543748Z test_tensorexpr 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_tensorexpr_1.1_d8551b64b93513ad_.log 2025-12-04T16:40:34.2567965Z Running 74 items in this shard: test/test_tensorexpr.py::TestTensorExprFuser::test_add_const_rhs, test/test_tensorexpr.py::TestTensorExprFuser::test_add_sub, test/test_tensorexpr.py::TestTensorExprFuser::test_alias_analysis_input_and_module, test/test_tensorexpr.py::TestTensorExprFuser::test_alias_analysis_inputs, test/test_tensorexpr.py::TestTensorExprFuser::test_alias_analysis_module, test/test_tensorexpr.py::TestTensorExprFuser::test_all_combos, test/test_tensorexpr.py::TestTensorExprFuser::test_alpha, test/test_tensorexpr.py::TestTensorExprFuser::test_binary_ops, test/test_tensorexpr.py::TestTensorExprFuser::test_bitwise_ops, test/test_tensorexpr.py::TestTensorExprFuser::test_broadcast, test/test_tensorexpr.py::TestTensorExprFuser::test_broadcast3, test/test_tensorexpr.py::TestTensorExprFuser::test_broadcast_2, test/test_tensorexpr.py::TestTensorExprFuser::test_broadcast_big2, test/test_tensorexpr.py::TestTensorExprFuser::test_cat, test/test_tensorexpr.py::TestTensorExprFuser::test_cat_empty_tensors, test/test_tensorexpr.py::TestTensorExprFuser::test_cat_negative_dim, test/test_tensorexpr.py::TestTensorExprFuser::test_cat_only, test/test_tensorexpr.py::TestTensorExprFuser::test_cat_promote_inputs, test/test_tensorexpr.py::TestTensorExprFuser::test_cat_with_constant_dim, test/test_tensorexpr.py::TestTensorExprFuser::test_char, test/test_tensorexpr.py::TestTensorExprFuser::test_chunk, test/test_tensorexpr.py::TestTensorExprFuser::test_clamp, test/test_tensorexpr.py::TestTensorExprFuser::test_constant, test/test_tensorexpr.py::TestTensorExprFuser::test_double, test/test_tensorexpr.py::TestTensorExprFuser::test_double_intrinsics, test/test_tensorexpr.py::TestTensorExprFuser::test_dynamic_shape, test/test_tensorexpr.py::TestTensorExprFuser::test_easy, test/test_tensorexpr.py::TestTensorExprFuser::test_eq, test/test_tensorexpr.py::TestTensorExprFuser::test_exp_pow, test/test_tensorexpr.py::TestTensorExprFuser::test_four_arg, test/test_tensorexpr.py::TestTensorExprFuser::test_ge, test/test_tensorexpr.py::TestTensorExprFuser::test_gt, test/test_tensorexpr.py::TestTensorExprFuser::test_guard_fails, test/test_tensorexpr.py::TestTensorExprFuser::test_half_bn_relu, test/test_tensorexpr.py::TestTensorExprFuser::test_half_gelu, test/test_tensorexpr.py::TestTensorExprFuser::test_int64_promotion, test/test_tensorexpr.py::TestTensorExprFuser::test_int_output, test/test_tensorexpr.py::TestTensorExprFuser::test_le, test/test_tensorexpr.py::TestTensorExprFuser::test_loop, test/test_tensorexpr.py::TestTensorExprFuser::test_lt, test/test_tensorexpr.py::TestTensorExprFuser::test_mask, test/test_tensorexpr.py::TestTensorExprFuser::test_min_max, test/test_tensorexpr.py::TestTensorExprFuser::test_min_max_reduction, test/test_tensorexpr.py::TestTensorExprFuser::test_min_max_reduction2, test/test_tensorexpr.py::TestTensorExprFuser::test_min_max_reduction_dim1, test/test_tensorexpr.py::TestTensorExprFuser::test_min_max_reduction_dim1_2, test/test_tensorexpr.py::TestTensorExprFuser::test_multi_rand, test/test_tensorexpr.py::TestTensorExprFuser::test_multioutput, test/test_tensorexpr.py::TestTensorExprFuser::test_multiple_outputs, test/test_tensorexpr.py::TestTensorExprFuser::test_nans, test/test_tensorexpr.py::TestTensorExprFuser::test_ne, test/test_tensorexpr.py::TestTensorExprFuser::test_promotion, test/test_tensorexpr.py::TestTensorExprFuser::test_propagated_mem_layout, test/test_tensorexpr.py::TestTensorExprFuser::test_rand_like, test/test_tensorexpr.py::TestTensorExprFuser::test_rank_two, test/test_tensorexpr.py::TestTensorExprFuser::test_relu, test/test_tensorexpr.py::TestTensorExprFuser::test_remainder, test/test_tensorexpr.py::TestTensorExprFuser::test_reps, test/test_tensorexpr.py::TestTensorExprFuser::test_round_2, test/test_tensorexpr.py::TestTensorExprFuser::test_scalar, test/test_tensorexpr.py::TestTensorExprFuser::test_short, test/test_tensorexpr.py::TestTensorExprFuser::test_simple_add, test/test_tensorexpr.py::TestTensorExprFuser::test_sin_pow, test/test_tensorexpr.py::TestTensorExprFuser::test_slice, test/test_tensorexpr.py::TestTensorExprFuser::test_sliced_stride, test/test_tensorexpr.py::TestTensorExprFuser::test_softmax_cpu, test/test_tensorexpr.py::TestTensorExprFuser::test_softmax_cuda, test/test_tensorexpr.py::TestTensorExprFuser::test_strided_output_preserved, test/test_tensorexpr.py::TestTensorExprFuser::test_three_arg, test/test_tensorexpr.py::TestTensorExprFuser::test_three_arg2, test/test_tensorexpr.py::TestTensorExprFuser::test_transpose, test/test_tensorexpr.py::TestTensorExprFuser::test_unary_ops, test/test_tensorexpr.py::TestTensorExprFuser::test_unsqueeze, test/test_tensorexpr.py::TestTensorExprFuser::test_where 2025-12-04T16:40:34.2591304Z 2025-12-04T16:40:34.2591621Z Finished test_tensorexpr 1/1 ... [2025-12-04 16:40:34.254201][26817.944593828], took 0.79min 2025-12-04T16:40:34.3038994Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_tensorexpr/test_tensorexpr-0ea46f0a35c352a2.xml 2025-12-04T16:40:34.4013639Z Running functorch/test_minifier 1/1 ... [2025-12-04 16:40:34.401055][26818.091445698] 2025-12-04T16:40:34.4014217Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:40:34.4017110Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/test_minifier.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:40:34.401472] 2025-12-04T16:40:40.2743373Z 2025-12-04T16:40:40.2744610Z functorch/test_minifier 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_minifier_1.1_a1ba5cac68d3f419_.log 2025-12-04T16:40:40.2747147Z Running 5 items in this shard: test/functorch/test_minifier.py::TestMinifier::test_has_add_mul, test/functorch/test_minifier.py::TestMinifier::test_has_mul_minifier, test/functorch/test_minifier.py::TestMinifier::test_input_returned, test/functorch/test_minifier.py::TestMinifier::test_module, test/functorch/test_minifier.py::TestMinifier::test_tup_use 2025-12-04T16:40:40.2748935Z 2025-12-04T16:40:40.2749296Z Finished functorch/test_minifier 1/1 ... [2025-12-04 16:40:40.274117][26823.964511983], took 0.10min 2025-12-04T16:40:40.3231941Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/functorch.test_minifier/functorch.test_minifier-320e05a9203d92d3.xml 2025-12-04T16:40:40.3548952Z Running higher_order_ops/test_invoke_quant 1/1 ... [2025-12-04 16:40:40.354665][26824.04505639] 2025-12-04T16:40:40.3549561Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:40:40.3552787Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'higher_order_ops/test_invoke_quant.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:40:40.355047] 2025-12-04T16:40:58.3951532Z 2025-12-04T16:40:58.3953261Z higher_order_ops/test_invoke_quant 1/1 was successful, full logs can be found in artifacts with path test/test-reports/higher_order_ops.test_invoke_quant_1.1_1d7f8dcadfd047a9_.log 2025-12-04T16:40:58.3960142Z Running 14 items in this shard: test/higher_order_ops/test_invoke_quant.py::TestInvokeQuantEager::test_construct_inline, test/higher_order_ops/test_invoke_quant.py::TestInvokeQuantEager::test_inline, test/higher_order_ops/test_invoke_quant.py::TestInvokeQuantEager::test_multiple, test/higher_order_ops/test_invoke_quant.py::TestInvokeQuantEager::test_simple, test/higher_order_ops/test_invoke_quant.py::TestInvokeQuantAotEager::test_construct_inline, test/higher_order_ops/test_invoke_quant.py::TestInvokeQuantAotEager::test_inline, test/higher_order_ops/test_invoke_quant.py::TestInvokeQuantAotEager::test_multiple, test/higher_order_ops/test_invoke_quant.py::TestInvokeQuantAotEager::test_simple, test/higher_order_ops/test_invoke_quant.py::TestInvokeQuantInductor::test_construct_inline, test/higher_order_ops/test_invoke_quant.py::TestInvokeQuantInductor::test_inline, test/higher_order_ops/test_invoke_quant.py::TestInvokeQuantInductor::test_multiple, test/higher_order_ops/test_invoke_quant.py::TestInvokeQuantInductor::test_pattern_matching, test/higher_order_ops/test_invoke_quant.py::TestInvokeQuantInductor::test_prologue, test/higher_order_ops/test_invoke_quant.py::TestInvokeQuantInductor::test_simple 2025-12-04T16:40:58.3966615Z 2025-12-04T16:40:58.3967006Z Finished higher_order_ops/test_invoke_quant 1/1 ... [2025-12-04 16:40:58.394942][26842.085335075], took 0.30min 2025-12-04T16:40:58.4448527Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/higher_order_ops.test_invoke_quant/higher_order_ops.test_invoke_quant-98f369a4901365e9.xml 2025-12-04T16:40:58.5113470Z Running torch_np/test_basic 1/1 ... [2025-12-04 16:40:58.511063][26842.201456332] 2025-12-04T16:40:58.5114196Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:40:58.5117452Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/test_basic.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:40:58.511454] 2025-12-04T16:41:08.0393047Z 2025-12-04T16:41:08.0394212Z torch_np/test_basic 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.test_basic_1.1_83f086e013fa5452_.log 2025-12-04T16:41:08.0552532Z Running 453 items in this shard: test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func0, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func1, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func10, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func11, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func12, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func13, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func14, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func15, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func16, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func17, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func18, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func19, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func2, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func20, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func21, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func22, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func23, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func24, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func25, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func26, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func27, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func28, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func29, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func3, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func30, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func31, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func32, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func33, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func34, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func35, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func36, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func37, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func38, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func39, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func4, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func40, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func41, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func42, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func43, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func44, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func45, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func46, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func47, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func48, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func49, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func5, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func50, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func51, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func52, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func53, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func54, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func55, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func56, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func57, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func58, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func59, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func6, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func60, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func61, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func62, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func63, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func64, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func65, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func66, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func67, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func68, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func69, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func7, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func70, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func71, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func72, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func73, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func74, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func8, test/torch_np/test_basic.py::TestOneArr::test_asarray_array_func9, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func0, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func1, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func10, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func11, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func12, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func13, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func14, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func15, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func16, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func17, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func18, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func19, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func2, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func20, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func21, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func22, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func23, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func24, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func25, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func26, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func27, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func28, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func29, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func3, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func30, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func31, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func32, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func33, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func34, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func35, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func36, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func37, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func38, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func39, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func4, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func40, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func41, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func42, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func43, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func44, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func45, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func46, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func47, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func48, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func49, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func5, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func50, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func51, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func52, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func53, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func54, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func55, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func56, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func57, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func58, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func59, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func6, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func60, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func61, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func62, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func63, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func64, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func65, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func66, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func67, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func68, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func69, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func7, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func70, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func71, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func72, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func73, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func74, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func8, test/torch_np/test_basic.py::TestOneArr::test_asarray_list_func9, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func0, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func1, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func10, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func11, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func12, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func13, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func14, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func15, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func16, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func17, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func18, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func19, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func2, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func20, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func21, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func22, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func23, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func24, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func25, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func26, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func27, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func28, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func29, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func3, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func30, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func31, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func32, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func33, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func34, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func35, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func36, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func37, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func38, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func39, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func4, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func40, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func41, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func42, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func43, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func44, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func45, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func46, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func47, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func48, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func49, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func5, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func50, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func51, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func52, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func53, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func54, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func55, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func56, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func57, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func58, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func59, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func6, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func60, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func61, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func62, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func63, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func64, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func65, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func66, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func67, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func68, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func69, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func7, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func70, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func71, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func72, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func73, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func74, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func8, test/torch_np/test_basic.py::TestOneArr::test_asarray_tensor_func9, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func0_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func0_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func0_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func0_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func10_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func10_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func10_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func10_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func1_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func1_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func1_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func1_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func2_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func2_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func2_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func2_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func3_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func3_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func3_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func3_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func4_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func4_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func4_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func4_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func5_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func5_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func5_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func5_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func6_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func6_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func6_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func6_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func7_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func7_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func7_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func7_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func8_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func8_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func8_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func8_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func9_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func9_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func9_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_array_func9_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func0_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func0_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func0_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func0_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func10_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func10_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func10_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func10_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func1_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func1_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func1_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func1_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func2_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func2_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func2_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func2_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func3_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func3_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func3_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func3_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func4_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func4_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func4_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func4_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func5_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func5_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func5_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func5_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func6_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func6_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func6_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func6_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func7_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func7_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func7_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func7_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func8_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func8_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func8_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func8_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func9_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func9_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func9_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_list_func9_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func0_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func0_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func0_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func0_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func10_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func10_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func10_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func10_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func1_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func1_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func1_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func1_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func2_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func2_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func2_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func2_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func3_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func3_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func3_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func3_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func4_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func4_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func4_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func4_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func5_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func5_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func5_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func5_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func6_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func6_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func6_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func6_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func7_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func7_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func7_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func7_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func8_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func8_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func8_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func8_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func9_axis3, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func9_axis_-1, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func9_axis_0, test/torch_np/test_basic.py::TestOneArrAndAxis::test_andaxis_tensor_func9_axis_1, test/torch_np/test_basic.py::TestOneArrAndAxesTuple::test_andtuple_array_func0_axes0, test/torch_np/test_basic.py::TestOneArrAndAxesTuple::test_andtuple_array_func0_axes1, test/torch_np/test_basic.py::TestOneArrAndAxesTuple::test_andtuple_array_func0_axes2, test/torch_np/test_basic.py::TestOneArrAndAxesTuple::test_andtuple_list_func0_axes0, test/torch_np/test_basic.py::TestOneArrAndAxesTuple::test_andtuple_list_func0_axes1, test/torch_np/test_basic.py::TestOneArrAndAxesTuple::test_andtuple_list_func0_axes2, test/torch_np/test_basic.py::TestOneArrAndAxesTuple::test_andtuple_tensor_func0_axes0, test/torch_np/test_basic.py::TestOneArrAndAxesTuple::test_andtuple_tensor_func0_axes1, test/torch_np/test_basic.py::TestOneArrAndAxesTuple::test_andtuple_tensor_func0_axes2, test/torch_np/test_basic.py::TestOneArrAndShape::test_andshape_array_func0, test/torch_np/test_basic.py::TestOneArrAndShape::test_andshape_array_func1, test/torch_np/test_basic.py::TestOneArrAndShape::test_andshape_array_func2, test/torch_np/test_basic.py::TestOneArrAndShape::test_andshape_array_func3, test/torch_np/test_basic.py::TestOneArrAndShape::test_andshape_array_func4, test/torch_np/test_basic.py::TestOneArrAndShape::test_andshape_list_func0, test/torch_np/test_basic.py::TestOneArrAndShape::test_andshape_list_func1, test/torch_np/test_basic.py::TestOneArrAndShape::test_andshape_list_func2, test/torch_np/test_basic.py::TestOneArrAndShape::test_andshape_list_func3, test/torch_np/test_basic.py::TestOneArrAndShape::test_andshape_list_func4, test/torch_np/test_basic.py::TestOneArrAndShape::test_andshape_tensor_func0, test/torch_np/test_basic.py::TestOneArrAndShape::test_andshape_tensor_func1, test/torch_np/test_basic.py::TestOneArrAndShape::test_andshape_tensor_func2, test/torch_np/test_basic.py::TestOneArrAndShape::test_andshape_tensor_func3, test/torch_np/test_basic.py::TestOneArrAndShape::test_andshape_tensor_func4, test/torch_np/test_basic.py::TestOneArrToScalar::test_toscalar_array_func0_np_func0, test/torch_np/test_basic.py::TestOneArrToScalar::test_toscalar_array_func1_np_func1, test/torch_np/test_basic.py::TestOneArrToScalar::test_toscalar_array_func2_np_func2, test/torch_np/test_basic.py::TestOneArrToScalar::test_toscalar_list_func0_np_func0, test/torch_np/test_basic.py::TestOneArrToScalar::test_toscalar_list_func1_np_func1, test/torch_np/test_basic.py::TestOneArrToScalar::test_toscalar_list_func2_np_func2, test/torch_np/test_basic.py::TestOneArrToScalar::test_toscalar_tensor_func0_np_func0, test/torch_np/test_basic.py::TestOneArrToScalar::test_toscalar_tensor_func1_np_func1, test/torch_np/test_basic.py::TestOneArrToScalar::test_toscalar_tensor_func2_np_func2, test/torch_np/test_basic.py::TestShapeLikeToArray::test_shape_func0, test/torch_np/test_basic.py::TestShapeLikeToArray::test_shape_func1, test/torch_np/test_basic.py::TestShapeLikeToArray::test_shape_func2, test/torch_np/test_basic.py::TestShapeLikeToArray::test_shape_func3, test/torch_np/test_basic.py::TestSequenceOfArrays::test_several_func0, test/torch_np/test_basic.py::TestSequenceOfArrays::test_several_func1, test/torch_np/test_basic.py::TestSequenceOfArrays::test_several_func2, test/torch_np/test_basic.py::TestSequenceOfArrays::test_several_func3, test/torch_np/test_basic.py::TestSequenceOfArrays::test_single_array_func0, test/torch_np/test_basic.py::TestSequenceOfArrays::test_single_array_func1, test/torch_np/test_basic.py::TestSequenceOfArrays::test_single_array_func2, test/torch_np/test_basic.py::TestSequenceOfArrays::test_single_array_func3, test/torch_np/test_basic.py::TestSequenceOfArrays::test_single_list_func0, test/torch_np/test_basic.py::TestSequenceOfArrays::test_single_list_func1, test/torch_np/test_basic.py::TestSequenceOfArrays::test_single_list_func2, test/torch_np/test_basic.py::TestSequenceOfArrays::test_single_list_func3, test/torch_np/test_basic.py::TestSequenceOfArrays::test_single_tensor_func0, test/torch_np/test_basic.py::TestSequenceOfArrays::test_single_tensor_func1, test/torch_np/test_basic.py::TestSequenceOfArrays::test_single_tensor_func2, test/torch_np/test_basic.py::TestSequenceOfArrays::test_single_tensor_func3, test/torch_np/test_basic.py::TestSequenceOfArraysToSingle::test_several_func0, test/torch_np/test_basic.py::TestSequenceOfArraysToSingle::test_several_func1, test/torch_np/test_basic.py::TestSequenceOfArraysToSingle::test_several_func2, test/torch_np/test_basic.py::TestSequenceOfArraysToSingle::test_several_func3, test/torch_np/test_basic.py::TestSequenceOfArraysToSingle::test_several_func4, test/torch_np/test_basic.py::TestSequenceOfArraysToSingle::test_several_func5, test/torch_np/test_basic.py::TestSequenceOfArraysToSingle::test_several_func6, test/torch_np/test_basic.py::TestArrayToSequence::test_asarray_array_func0, test/torch_np/test_basic.py::TestArrayToSequence::test_asarray_array_func1, test/torch_np/test_basic.py::TestArrayToSequence::test_asarray_list_func0, test/torch_np/test_basic.py::TestArrayToSequence::test_asarray_list_func1, test/torch_np/test_basic.py::TestArrayToSequence::test_asarray_tensor_func0, test/torch_np/test_basic.py::TestArrayToSequence::test_asarray_tensor_func1, test/torch_np/test_basic.py::TestPythonArgsToArray::test_argstoarray_simple_func0_args0, test/torch_np/test_basic.py::TestPythonArgsToArray::test_argstoarray_simple_func1_args1, test/torch_np/test_basic.py::TestPythonArgsToArray::test_argstoarray_simple_func2_args2, test/torch_np/test_basic.py::TestPythonArgsToArray::test_argstoarray_simple_func3_args3, test/torch_np/test_basic.py::TestPythonArgsToArray::test_argstoarray_simple_func4_args4, test/torch_np/test_basic.py::TestPythonArgsToArray::test_argstoarray_simple_func5_args5, test/torch_np/test_basic.py::TestPythonArgsToArray::test_argstoarray_simple_func6_args6, test/torch_np/test_basic.py::TestPythonArgsToArray::test_argstoarray_simple_func7_args7, test/torch_np/test_basic.py::TestPythonArgsToArray::test_argstoarray_simple_func8_args8, test/torch_np/test_basic.py::TestPythonArgsToArray::test_argstoarray_simple_func9_args9, test/torch_np/test_basic.py::TestNormalizations::test_too_few_args_positional, test/torch_np/test_basic.py::TestNormalizations::test_unknown_args, test/torch_np/test_basic.py::TestNormalizations::test_unknown_args_with_defaults, test/torch_np/test_basic.py::TestCopyTo::test_copyto_basic, test/torch_np/test_basic.py::TestCopyTo::test_copyto_typecast, test/torch_np/test_basic.py::TestCopyTo::test_copytobcast, test/torch_np/test_basic.py::TestDivmod::test_divmod_no_out, test/torch_np/test_basic.py::TestDivmod::test_divmod_out, test/torch_np/test_basic.py::TestDivmod::test_divmod_out_both_pos_and_kw, test/torch_np/test_basic.py::TestDivmod::test_divmod_out_list, test/torch_np/test_basic.py::TestDivmod::test_divmod_pos_only, test/torch_np/test_basic.py::TestSmokeNotImpl::test_nimpl_basic, test/torch_np/test_basic.py::TestDefaultDtype::test_defaultdtype_defaults, test/torch_np/test_basic.py::TestDefaultDtype::test_set_default_float_dt_float32, test/torch_np/test_basic.py::TestDefaultDtype::test_set_default_float_dt_pytorch, test/torch_np/test_basic.py::TestDefaultDtype::test_set_default_float_float32, test/torch_np/test_basic.py::TestExport::test_exported_objects, test/torch_np/test_basic.py::TestCtorNested::test_arrays_in_lists, test/torch_np/test_basic.py::TestMisc::test_f16_on_cuda, test/torch_np/test_basic.py::TestMisc::test_ndarrays_to_tensors 2025-12-04T16:41:08.0709049Z 2025-12-04T16:41:08.0709380Z Finished torch_np/test_basic 1/1 ... [2025-12-04 16:41:08.039844][26851.730232961], took 0.16min 2025-12-04T16:41:08.0897782Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.test_basic/torch_np.test_basic-f66c371882977b0d.xml 2025-12-04T16:41:08.1649392Z Running test_jiterator 1/1 ... [2025-12-04 16:41:08.164583][26851.854973541] 2025-12-04T16:41:08.1658997Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:41:08.1660373Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_jiterator.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:41:08.165032] 2025-12-04T16:41:46.3827293Z 2025-12-04T16:41:46.3828583Z test_jiterator 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_jiterator_1.1_8988ac20057f1363_.log 2025-12-04T16:41:46.3989508Z Running 289 items in this shard: test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_bfloat16_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_bfloat16_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_bfloat16_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_bfloat16_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_bfloat16_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_bfloat16_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_bfloat16_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_bfloat16_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_bfloat16_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_bfloat16_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_bfloat16_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex128_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex128_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex128_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex128_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex128_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex128_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex128_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex128_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex128_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex128_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex128_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex64_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex64_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex64_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex64_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex64_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex64_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex64_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex64_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex64_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex64_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_complex64_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float16_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float16_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float16_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float16_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float16_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float16_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float16_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float16_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float16_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float16_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float16_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float32_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float32_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float32_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float32_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float32_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float32_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float32_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float32_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float32_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float32_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float32_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float64_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float64_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float64_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float64_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float64_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float64_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float64_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float64_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float64_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float64_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_float64_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int16_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int16_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int16_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int16_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int16_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int16_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int16_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int16_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int16_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int16_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int16_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int32_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int32_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int32_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int32_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int32_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int32_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int32_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int32_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int32_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int32_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int32_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int64_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int64_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int64_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int64_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int64_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int64_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int64_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int64_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int64_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int64_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int64_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int8_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int8_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int8_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int8_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int8_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int8_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int8_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int8_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int8_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int8_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_int8_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_uint8_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_uint8_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_uint8_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_uint8_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_uint8_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_uint8_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_uint8_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_uint8_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_uint8_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_uint8_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_contiguous_shape_strides0_cuda_uint8_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_bfloat16_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_bfloat16_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_bfloat16_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_bfloat16_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_bfloat16_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_bfloat16_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_bfloat16_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_bfloat16_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_bfloat16_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_bfloat16_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_bfloat16_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex128_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex128_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex128_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex128_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex128_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex128_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex128_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex128_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex128_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex128_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex128_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex64_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex64_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex64_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex64_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex64_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex64_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex64_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex64_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex64_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex64_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_complex64_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float16_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float16_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float16_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float16_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float16_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float16_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float16_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float16_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float16_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float16_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float16_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float32_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float32_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float32_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float32_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float32_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float32_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float32_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float32_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float32_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float32_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float32_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float64_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float64_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float64_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float64_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float64_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float64_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float64_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float64_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float64_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float64_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_float64_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int16_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int16_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int16_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int16_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int16_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int16_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int16_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int16_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int16_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int16_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int16_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int32_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int32_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int32_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int32_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int32_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int32_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int32_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int32_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int32_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int32_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int32_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int64_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int64_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int64_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int64_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int64_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int64_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int64_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int64_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int64_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int64_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int64_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int8_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int8_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int8_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int8_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int8_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int8_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int8_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int8_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int8_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int8_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_int8_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_uint8_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_uint8_complex128, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_uint8_complex64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_uint8_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_uint8_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_uint8_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_uint8_int16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_uint8_int32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_uint8_int64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_uint8_int8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_all_dtype_noncontiguous_shape_strides0_cuda_uint8_uint8, test/test_jiterator.py::TestPythonJiteratorCUDA::test_bool_extra_args_is_train_False_cuda, test/test_jiterator.py::TestPythonJiteratorCUDA::test_bool_extra_args_is_train_True_cuda, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha2_beta2_cuda_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha2_beta2_cuda_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha2_beta2_cuda_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha2_beta2_cuda_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha2_beta_-4_2_cuda_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha2_beta_-4_2_cuda_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha2_beta_-4_2_cuda_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha2_beta_-4_2_cuda_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha2_beta_3_cuda_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha2_beta_3_cuda_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha2_beta_3_cuda_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha2_beta_3_cuda_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_-1_beta2_cuda_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_-1_beta2_cuda_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_-1_beta2_cuda_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_-1_beta2_cuda_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_-1_beta_-4_2_cuda_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_-1_beta_-4_2_cuda_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_-1_beta_-4_2_cuda_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_-1_beta_-4_2_cuda_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_-1_beta_3_cuda_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_-1_beta_3_cuda_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_-1_beta_3_cuda_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_-1_beta_3_cuda_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_2_0_beta2_cuda_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_2_0_beta2_cuda_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_2_0_beta2_cuda_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_2_0_beta2_cuda_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_2_0_beta_-4_2_cuda_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_2_0_beta_-4_2_cuda_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_2_0_beta_-4_2_cuda_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_2_0_beta_-4_2_cuda_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_2_0_beta_3_cuda_bfloat16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_2_0_beta_3_cuda_float16, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_2_0_beta_3_cuda_float32, test/test_jiterator.py::TestPythonJiteratorCUDA::test_extra_args_alpha_2_0_beta_3_cuda_float64, test/test_jiterator.py::TestPythonJiteratorCUDA::test_invalid_function_name_code_string_template T my _kernel(T x) { return x; }_cuda, test/test_jiterator.py::TestPythonJiteratorCUDA::test_invalid_function_name_code_string_template Tmy_kernel(T x) { return x; }_cuda, test/test_jiterator.py::TestPythonJiteratorCUDA::test_multiple_functors_cuda, test/test_jiterator.py::TestPythonJiteratorCUDA::test_various_num_inputs_num_inputs_1_cuda, test/test_jiterator.py::TestPythonJiteratorCUDA::test_various_num_inputs_num_inputs_5_cuda, test/test_jiterator.py::TestPythonJiteratorCUDA::test_various_num_inputs_num_inputs_8_cuda, test/test_jiterator.py::TestPythonJiteratorCUDA::test_various_num_outputs_num_outputs_1_cuda, test/test_jiterator.py::TestPythonJiteratorCUDA::test_various_num_outputs_num_outputs_4_cuda, test/test_jiterator.py::TestPythonJiteratorCUDA::test_various_num_outputs_num_outputs_8_cuda 2025-12-04T16:41:46.4146733Z 2025-12-04T16:41:46.4147052Z Finished test_jiterator 1/1 ... [2025-12-04 16:41:46.383212][26890.073600282], took 0.64min 2025-12-04T16:41:46.4338858Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_jiterator/test_jiterator-faac0f7ad3ce18ff.xml 2025-12-04T16:41:46.5384581Z Running test_native_functions 1/1 ... [2025-12-04 16:41:46.538090][26890.2284796] 2025-12-04T16:41:46.5385167Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:41:46.5387683Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_native_functions.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:41:46.538524] 2025-12-04T16:41:52.2611158Z 2025-12-04T16:41:52.2612352Z test_native_functions 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_native_functions_1.1_be6b38187a1d4fbd_.log 2025-12-04T16:41:52.2617748Z Running 11 items in this shard: test/test_native_functions.py::TestNativeFunctions::test_intlist_error_with_overload, test/test_native_functions.py::TestNativeFunctions::test_optional_filled_intlist, test/test_native_functions.py::TestNativeFunctions::test_optional_floatlist, test/test_native_functions.py::TestNativeFunctions::test_optional_floatlist_invalid, test/test_native_functions.py::TestNativeFunctions::test_optional_intlist, test/test_native_functions.py::TestNativeFunctions::test_optional_intlist_invalid, test/test_native_functions.py::TestNativeFunctions::test_string_defaults, test/test_native_functions.py::TestNativeFunctions::test_symintlist_error, test/test_native_functions.py::TestNativeFunctions::test_symintlist_error_with_overload, test/test_native_functions.py::TestNativeFunctions::test_symintlist_error_with_overload_but_is_unique, test/test_native_functions.py::TestNativeFunctions::test_vararg_symintlist_error 2025-12-04T16:41:52.2622553Z 2025-12-04T16:41:52.2622900Z Finished test_native_functions 1/1 ... [2025-12-04 16:41:52.260930][26895.951321112], took 0.10min 2025-12-04T16:41:52.3113886Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_native_functions/test_native_functions-ed1f170746938cdb.xml 2025-12-04T16:41:52.3467653Z Running test_typing 1/1 ... [2025-12-04 16:41:52.346520][26896.036910975] 2025-12-04T16:41:52.3468166Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:41:52.3472118Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_typing.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:41:52.346960] 2025-12-04T16:43:35.1049078Z 2025-12-04T16:43:35.1049993Z test_typing 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_typing_1.1_d8e70b3c726a1ce6_.log 2025-12-04T16:43:35.1055990Z Running 18 items in this shard: test/test_typing.py::TestTyping::test_fail_arithmetic_ops.py, test/test_typing.py::TestTyping::test_fail_creation_ops.py, test/test_typing.py::TestTyping::test_fail_random.py, test/test_typing.py::TestTyping::test_fail_torch_size.py, test/test_typing.py::TestTyping::test_reveal_module_list.py, test/test_typing.py::TestTyping::test_reveal_namedtuple.py, test/test_typing.py::TestTyping::test_reveal_opt_size.py, test/test_typing.py::TestTyping::test_reveal_size.py, test/test_typing.py::TestTyping::test_reveal_tensor_constructors.py, test/test_typing.py::TestTyping::test_reveal_tensor_copy.py, test/test_typing.py::TestTyping::test_reveal_tensor_sampling.py, test/test_typing.py::TestTyping::test_reveal_torch_optim.py, test/test_typing.py::TestTyping::test_success_arithmetic_ops.py, test/test_typing.py::TestTyping::test_success_creation_ops.py, test/test_typing.py::TestTyping::test_success_cuda_steam.py, test/test_typing.py::TestTyping::test_success_distributions.py, test/test_typing.py::TestTyping::test_success_math_ops.py, test/test_typing.py::TestTyping::test_success_torch_size.py 2025-12-04T16:43:35.1061452Z 2025-12-04T16:43:35.1061729Z Finished test_typing 1/1 ... [2025-12-04 16:43:35.104692][26998.795085962], took 1.71min 2025-12-04T16:43:35.1558492Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_typing/test_typing-029ceb92f90c5fb6.xml 2025-12-04T16:43:35.2143277Z Running lazy/test_functionalization 1/1 ... [2025-12-04 16:43:35.214018][26998.904409243] 2025-12-04T16:43:35.2143895Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:43:35.2147031Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'lazy/test_functionalization.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:43:35.214455] 2025-12-04T16:43:40.6367890Z 2025-12-04T16:43:40.6368952Z lazy/test_functionalization 1/1 was successful, full logs can be found in artifacts with path test/test-reports/lazy.test_functionalization_1.1_2e9665a4a8fbdab9_.log 2025-12-04T16:43:40.6370755Z Running 2 items in this shard: test/lazy/test_functionalization.py::LazyFuncionalizationTest::test_data_assign, test/lazy/test_functionalization.py::LazyFuncionalizationTest::test_lazy_init_with_view 2025-12-04T16:43:40.6371810Z 2025-12-04T16:43:40.6372187Z Finished lazy/test_functionalization 1/1 ... [2025-12-04 16:43:40.636544][27004.326938264], took 0.09min 2025-12-04T16:43:40.6869146Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/lazy.test_functionalization/lazy.test_functionalization-3b38bfe5c8489bba.xml 2025-12-04T16:43:40.7166024Z Running torch_np/test_random 1/1 ... [2025-12-04 16:43:40.716312][27004.406704386] 2025-12-04T16:43:40.7166591Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:43:40.7169500Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/test_random.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:43:40.716718] 2025-12-04T16:43:46.1889636Z 2025-12-04T16:43:46.1890791Z torch_np/test_random 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.test_random_1.1_5e1f3614f612e456_.log 2025-12-04T16:43:46.1908461Z Running 41 items in this shard: test/torch_np/test_random.py::TestScalarReturn::test_rndm_array_use_numpy_False_func0, test/torch_np/test_random.py::TestScalarReturn::test_rndm_array_use_numpy_False_func1, test/torch_np/test_random.py::TestScalarReturn::test_rndm_array_use_numpy_False_func2, test/torch_np/test_random.py::TestScalarReturn::test_rndm_array_use_numpy_False_func3, test/torch_np/test_random.py::TestScalarReturn::test_rndm_array_use_numpy_False_func6, test/torch_np/test_random.py::TestScalarReturn::test_rndm_array_use_numpy_False_func7, test/torch_np/test_random.py::TestScalarReturn::test_rndm_array_use_numpy_False_random_random, test/torch_np/test_random.py::TestScalarReturn::test_rndm_array_use_numpy_False_random_sample, test/torch_np/test_random.py::TestScalarReturn::test_rndm_array_use_numpy_True_func0, test/torch_np/test_random.py::TestScalarReturn::test_rndm_array_use_numpy_True_func1, test/torch_np/test_random.py::TestScalarReturn::test_rndm_array_use_numpy_True_func2, test/torch_np/test_random.py::TestScalarReturn::test_rndm_array_use_numpy_True_func3, test/torch_np/test_random.py::TestScalarReturn::test_rndm_array_use_numpy_True_func6, test/torch_np/test_random.py::TestScalarReturn::test_rndm_array_use_numpy_True_func7, test/torch_np/test_random.py::TestScalarReturn::test_rndm_array_use_numpy_True_random_random, test/torch_np/test_random.py::TestScalarReturn::test_rndm_array_use_numpy_True_random_sample, test/torch_np/test_random.py::TestScalarReturn::test_rndm_scalar_use_numpy_False_func0, test/torch_np/test_random.py::TestScalarReturn::test_rndm_scalar_use_numpy_False_func1, test/torch_np/test_random.py::TestScalarReturn::test_rndm_scalar_use_numpy_False_func2, test/torch_np/test_random.py::TestScalarReturn::test_rndm_scalar_use_numpy_False_func3, test/torch_np/test_random.py::TestScalarReturn::test_rndm_scalar_use_numpy_False_func6, test/torch_np/test_random.py::TestScalarReturn::test_rndm_scalar_use_numpy_False_func7, test/torch_np/test_random.py::TestScalarReturn::test_rndm_scalar_use_numpy_False_random_random, test/torch_np/test_random.py::TestScalarReturn::test_rndm_scalar_use_numpy_False_random_sample, test/torch_np/test_random.py::TestScalarReturn::test_rndm_scalar_use_numpy_True_func0, test/torch_np/test_random.py::TestScalarReturn::test_rndm_scalar_use_numpy_True_func1, test/torch_np/test_random.py::TestScalarReturn::test_rndm_scalar_use_numpy_True_func2, test/torch_np/test_random.py::TestScalarReturn::test_rndm_scalar_use_numpy_True_func3, test/torch_np/test_random.py::TestScalarReturn::test_rndm_scalar_use_numpy_True_func6, test/torch_np/test_random.py::TestScalarReturn::test_rndm_scalar_use_numpy_True_func7, test/torch_np/test_random.py::TestScalarReturn::test_rndm_scalar_use_numpy_True_random_random, test/torch_np/test_random.py::TestScalarReturn::test_rndm_scalar_use_numpy_True_random_sample, test/torch_np/test_random.py::TestShuffle::test_1d_use_numpy_False, test/torch_np/test_random.py::TestShuffle::test_1d_use_numpy_True, test/torch_np/test_random.py::TestShuffle::test_2d_use_numpy_False, test/torch_np/test_random.py::TestShuffle::test_2d_use_numpy_True, test/torch_np/test_random.py::TestShuffle::test_shuffle_list_use_numpy_False, test/torch_np/test_random.py::TestShuffle::test_shuffle_list_use_numpy_True, test/torch_np/test_random.py::TestChoice::test_choice_use_numpy_False, test/torch_np/test_random.py::TestChoice::test_choice_use_numpy_True, test/torch_np/test_random.py::TestNumpyGlobal::test_numpy_global 2025-12-04T16:43:46.1924943Z 2025-12-04T16:43:46.1925302Z Finished torch_np/test_random 1/1 ... [2025-12-04 16:43:46.188776][27009.879169333], took 0.09min 2025-12-04T16:43:46.2395961Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.test_random/torch_np.test_random-92b710d703a1f97f.xml 2025-12-04T16:43:46.3273397Z Running nn/test_multihead_attention 1/1 ... [2025-12-04 16:43:46.327051][27010.017441608] 2025-12-04T16:43:46.3274004Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:43:46.3277371Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'nn/test_multihead_attention.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:43:46.327459] 2025-12-04T16:44:16.3381673Z 2025-12-04T16:44:16.3383205Z nn/test_multihead_attention 1/1 was successful, full logs can be found in artifacts with path test/test-reports/nn.test_multihead_attention_1.1_aa528fb509f77678_.log 2025-12-04T16:44:16.3400115Z Running 20 items in this shard: test/nn/test_multihead_attention.py::TestMultiheadAttentionNN::test_multihead_attention_average_attn_weights_False, test/nn/test_multihead_attention.py::TestMultiheadAttentionNN::test_multihead_attention_average_attn_weights_True, test/nn/test_multihead_attention.py::TestMultiheadAttentionNN::test_multihead_attn_3d_attn_mask, test/nn/test_multihead_attention.py::TestMultiheadAttentionNN::test_multihead_attn_fast_path_invalid_shape, test/nn/test_multihead_attention.py::TestMultiheadAttentionNN::test_multihead_attn_invalid_shape, test/nn/test_multihead_attention.py::TestMultiheadAttentionNN::test_multihead_attn_nested_tensor_outside_fast_path, test/nn/test_multihead_attention.py::TestMultiheadAttentionNN::test_multihead_attn_no_bias, test/nn/test_multihead_attention.py::TestMultiheadAttentionNNDeviceTypeCUDA::test_fast_path_check_with_mask_does_not_break_in_compile_cuda_float64, test/nn/test_multihead_attention.py::TestMultiheadAttentionNNDeviceTypeCUDA::test_multihead_attention_dtype_batch_first_cuda_float16, test/nn/test_multihead_attention.py::TestMultiheadAttentionNNDeviceTypeCUDA::test_multihead_attention_dtype_batch_first_cuda_float32, test/nn/test_multihead_attention.py::TestMultiheadAttentionNNDeviceTypeCUDA::test_multihead_attention_dtype_batch_first_cuda_float64, test/nn/test_multihead_attention.py::TestMultiheadAttentionNNDeviceTypeCUDA::test_multihead_attention_dtype_cuda_float16, test/nn/test_multihead_attention.py::TestMultiheadAttentionNNDeviceTypeCUDA::test_multihead_attention_dtype_cuda_float32, test/nn/test_multihead_attention.py::TestMultiheadAttentionNNDeviceTypeCUDA::test_multihead_attention_dtype_cuda_float64, test/nn/test_multihead_attention.py::TestMultiheadAttentionNNDeviceTypeCUDA::test_multihead_attn_fast_path_query_and_bias_have_different_dtypes_cuda_float64, test/nn/test_multihead_attention.py::TestMultiheadAttentionNNDeviceTypeCUDA::test_multihead_attn_fast_path_small_test_cuda_float64, test/nn/test_multihead_attention.py::TestMultiheadAttentionNNDeviceTypeCUDA::test_multihead_attn_in_proj_bias_none_cuda_float64, test/nn/test_multihead_attention.py::TestMultiheadAttentionNNDeviceTypeCUDA::test_multihead_attn_in_proj_weight_none_cuda_float64, test/nn/test_multihead_attention.py::TestMultiheadAttentionNNDeviceTypeCUDA::test_multihead_self_attn_two_masks_fast_path_cuda, test/nn/test_multihead_attention.py::TestMultiheadAttentionNNDeviceTypeCUDA::test_multihead_self_attn_two_masks_fast_path_mock_cuda 2025-12-04T16:44:16.3413318Z 2025-12-04T16:44:16.3413704Z Finished nn/test_multihead_attention 1/1 ... [2025-12-04 16:44:16.337943][27040.028336352], took 0.50min 2025-12-04T16:44:16.3888486Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/nn.test_multihead_attention/nn.test_multihead_attention-227b95684fd4827a.xml 2025-12-04T16:44:16.5133253Z Running test_legacy_vmap 1/1 ... [2025-12-04 16:44:16.512976][27040.203367543] 2025-12-04T16:44:16.5133816Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:44:16.5136732Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_legacy_vmap.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:44:16.513417] 2025-12-04T16:44:25.6914479Z 2025-12-04T16:44:25.6915873Z test_legacy_vmap 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_legacy_vmap_1.1_88ccad856a60ef1f_.log 2025-12-04T16:44:25.7007073Z Running 124 items in this shard: test/test_legacy_vmap.py::TestVmapAPILegacy::test_accepts_nested_inputs, test/test_legacy_vmap.py::TestVmapAPILegacy::test_backward_unsupported_interaction, test/test_legacy_vmap.py::TestVmapAPILegacy::test_batched_gradient_basic, test/test_legacy_vmap.py::TestVmapAPILegacy::test_constant_function, test/test_legacy_vmap.py::TestVmapAPILegacy::test_different_map_dim_size_raises, test/test_legacy_vmap.py::TestVmapAPILegacy::test_fallback_atan2, test/test_legacy_vmap.py::TestVmapAPILegacy::test_fallback_does_not_warn_by_default, test/test_legacy_vmap.py::TestVmapAPILegacy::test_fallback_masked_fill, test/test_legacy_vmap.py::TestVmapAPILegacy::test_fallback_multiple_returns, test/test_legacy_vmap.py::TestVmapAPILegacy::test_fallback_warns_when_warnings_are_enabled, test/test_legacy_vmap.py::TestVmapAPILegacy::test_fallback_with_undefined_grad, test/test_legacy_vmap.py::TestVmapAPILegacy::test_fallback_zero_dim, test/test_legacy_vmap.py::TestVmapAPILegacy::test_func_with_no_inputs, test/test_legacy_vmap.py::TestVmapAPILegacy::test_functools_partial, test/test_legacy_vmap.py::TestVmapAPILegacy::test_grad_unsupported_interaction, test/test_legacy_vmap.py::TestVmapAPILegacy::test_in_dim_not_in_tensor_err_msg, test/test_legacy_vmap.py::TestVmapAPILegacy::test_in_dims_wrong_type_err_msg, test/test_legacy_vmap.py::TestVmapAPILegacy::test_inplace_fallback_nary_different_levels, test/test_legacy_vmap.py::TestVmapAPILegacy::test_inplace_fallback_nary_same_levels, test/test_legacy_vmap.py::TestVmapAPILegacy::test_inplace_fallback_unary, test/test_legacy_vmap.py::TestVmapAPILegacy::test_integer_in_dim_but_not_tensor_input_err_msg, test/test_legacy_vmap.py::TestVmapAPILegacy::test_multiple_inputs, test/test_legacy_vmap.py::TestVmapAPILegacy::test_multiple_out_dims, test/test_legacy_vmap.py::TestVmapAPILegacy::test_multiple_outputs, test/test_legacy_vmap.py::TestVmapAPILegacy::test_multiple_outputs_error_cases, test/test_legacy_vmap.py::TestVmapAPILegacy::test_nested_non_default_in_dims, test/test_legacy_vmap.py::TestVmapAPILegacy::test_nested_out_dims, test/test_legacy_vmap.py::TestVmapAPILegacy::test_nested_with_different_map_dim, test/test_legacy_vmap.py::TestVmapAPILegacy::test_nested_with_same_map_dim, test/test_legacy_vmap.py::TestVmapAPILegacy::test_nn_module, test/test_legacy_vmap.py::TestVmapAPILegacy::test_non_default_in_dims_out_dims, test/test_legacy_vmap.py::TestVmapAPILegacy::test_non_tensor_output_raises, test/test_legacy_vmap.py::TestVmapAPILegacy::test_non_zero_in_dims, test/test_legacy_vmap.py::TestVmapAPILegacy::test_none_in_dims, test/test_legacy_vmap.py::TestVmapAPILegacy::test_nonzero_out_dims, test/test_legacy_vmap.py::TestVmapAPILegacy::test_noop_in_inner_vmap, test/test_legacy_vmap.py::TestVmapAPILegacy::test_not_enough_in_dims_err_msg, test/test_legacy_vmap.py::TestVmapAPILegacy::test_out_dim_out_of_bounds_err_msg, test/test_legacy_vmap.py::TestVmapAPILegacy::test_out_dims_and_num_outputs_mismatch_err_msg, test/test_legacy_vmap.py::TestVmapAPILegacy::test_out_dims_edge_case, test/test_legacy_vmap.py::TestVmapAPILegacy::test_out_dims_must_be_int_or_tuple_of_int_err_msg, test/test_legacy_vmap.py::TestVmapAPILegacy::test_single_input, test/test_legacy_vmap.py::TestVmapAPILegacy::test_unsupported_op_err_msg, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_T_numpy, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_as_strided, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_binary_pointwise_ops, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_bmm, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_cat, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_chunk, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_clamp, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_clone, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_comparison_ops, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_conj, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_contiguous, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_diagonal, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_dot, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_expand_as, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_fill_and_zero_inplace, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_imag, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_is_complex, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_is_contiguous, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_is_floating_point, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_mm, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_movedim, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_mv, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_narrow, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_new_empty, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_new_empty_strided, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_new_zeros, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_no_random_op_support, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_real, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_reshape, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_reshape_as, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_result_type, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_select, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_slice, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_split, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_squeeze, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_stack, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_stride, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_sum_dim, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_t, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_tensor_split, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_to, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_trace, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_transpose, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_unary_pointwise_ops, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_unbind, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_unfold, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_view, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_view_as, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_view_as_complex, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_view_as_real, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_vmap_fallback_check, test/test_legacy_vmap.py::TestVmapOperatorsLegacy::test_vmap_fallback_check_ok, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_add_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_binary_cross_entropy_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_diagonal_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_div_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_expand_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_index_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_inplace_manyview_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_inplace_on_view_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_lgamma_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_log1p_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_log_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_logsumexp_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_max_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_median_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_min_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_mul_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_permute_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_reshape_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_select_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_sigmoid_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_slice_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_stack_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_sub_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_threshold_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_trace_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_unrelated_output_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_unrelated_output_multiple_grad_cuda, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_vmap_fallback_check, test/test_legacy_vmap.py::TestVmapBatchedGradientLegacyCUDA::test_vmap_fallback_check_ok 2025-12-04T16:44:25.7094530Z 2025-12-04T16:44:25.7095107Z Finished test_legacy_vmap 1/1 ... [2025-12-04 16:44:25.691330][27049.381722341], took 0.15min 2025-12-04T16:44:25.7492208Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_legacy_vmap/test_legacy_vmap-1f56d97bf135e06f.xml 2025-12-04T16:44:25.8334772Z Running lazy/test_bindings 1/1 ... [2025-12-04 16:44:25.833093][27049.523484001] 2025-12-04T16:44:25.8335644Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:44:25.8339424Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'lazy/test_bindings.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:44:25.833571] 2025-12-04T16:44:30.8054398Z 2025-12-04T16:44:30.8055548Z lazy/test_bindings 1/1 was successful, full logs can be found in artifacts with path test/test-reports/lazy.test_bindings_1.1_b260f38330c47014_.log 2025-12-04T16:44:30.8056584Z Running 1 items in this shard: test/lazy/test_bindings.py::test_metrics 2025-12-04T16:44:30.8057039Z 2025-12-04T16:44:30.8057373Z Finished lazy/test_bindings 1/1 ... [2025-12-04 16:44:30.805210][27054.495603952], took 0.08min 2025-12-04T16:44:30.8563617Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/lazy.test_bindings/lazy.test_bindings-48affd3dc097f9a0.xml 2025-12-04T16:44:30.8846894Z Running test_utils 1/1 ... [2025-12-04 16:44:30.884486][27054.574878774] 2025-12-04T16:44:30.8847390Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:44:30.8851191Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_utils.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:44:30.884876] 2025-12-04T16:45:24.3251222Z 2025-12-04T16:45:24.3252131Z test_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_utils_1.1_f544fe8617f67377_.log 2025-12-04T16:45:24.5846543Z Running 6012 items in this shard: test/test_utils.py::TestCheckpoint::test_checkpoint, test/test_utils.py::TestCheckpoint::test_checkpoint_module_list, test/test_utils.py::TestCheckpoint::test_checkpoint_no_tensors, test/test_utils.py::TestCheckpoint::test_checkpoint_non_tensor, test/test_utils.py::TestCheckpoint::test_checkpoint_non_tensor_inputs_outputs, test/test_utils.py::TestCheckpoint::test_checkpoint_not_preserve_rng_state_and_without_reentrant, test/test_utils.py::TestCheckpoint::test_checkpoint_partial_grad, test/test_utils.py::TestCheckpoint::test_checkpoint_rng_cpu, test/test_utils.py::TestCheckpoint::test_checkpoint_rng_gpu, test/test_utils.py::TestCheckpoint::test_checkpoint_sequential_deprecated_multiple_args, test/test_utils.py::TestCheckpoint::test_checkpoint_sequential_deprecated_no_args, test/test_utils.py::TestCheckpoint::test_checkpoint_trigger, test/test_utils.py::TestCheckpoint::test_checkpoint_valid, test/test_utils.py::TestCheckpoint::test_checkpointing_without_reentrant_early_free, test/test_utils.py::TestCheckpoint::test_get_device_states_recursive, test/test_utils.py::TestCheckpoint::test_infer_device_state_recursive_meta, test/test_utils.py::TestCheckpoint::test_infer_device_state_recursive_multi_gpu, test/test_utils.py::TestDataLoaderUtils::test_multi_drop, test/test_utils.py::TestDataLoaderUtils::test_multi_keep, test/test_utils.py::TestDataLoaderUtils::test_random_seed, test/test_utils.py::TestDataLoaderUtils::test_single_drop, test/test_utils.py::TestDataLoaderUtils::test_single_keep, test/test_utils.py::TestCollectEnv::test_smoke, test/test_utils.py::TestHipify::test_import_hipify, test/test_utils.py::TestHipifyTrie::test_add_and_search_trie, test/test_utils.py::TestHipifyTrie::test_add_multiple_and_search_trie, test/test_utils.py::TestHipifyTrie::test_char_export_trie_to_regex, test/test_utils.py::TestHipifyTrie::test_export_trie_to_regex, test/test_utils.py::TestHipifyTrie::test_prefix_words_export_trie_to_regex, test/test_utils.py::TestHipifyTrie::test_quote_escape, test/test_utils.py::TestHipifyTrie::test_single_export_trie_to_regex, test/test_utils.py::TestHipifyTrie::test_special_char_export_trie_to_regex, test/test_utils.py::TestAssert::test_assert_scriptable, test/test_utils.py::TestAssert::test_assert_true, test/test_utils.py::TestStandaloneCPPJIT::test_load_standalone, test/test_utils.py::TestRenderUtils::test_basic, test/test_utils.py::TestDeviceUtilsCUDA::test_basic_cuda, test/test_utils.py::TestDeviceUtilsCUDA::test_decorator_cuda, test/test_utils.py::TestDeviceUtilsCUDA::test_decorator_generator_cuda, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_H_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_H_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_H_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_H_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_H_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_H_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_H_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_H_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_H_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_H_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_H_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_H_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_H_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_T_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_T_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_T_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_T_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_T_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_T_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_T_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_T_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_T_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_T_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_T_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_T_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_T_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___getitem___cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___getitem___cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___getitem___cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___getitem___cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___getitem___cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___getitem___cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___getitem___cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___getitem___cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___getitem___cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___getitem___cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___getitem___cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___getitem___cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___getitem___cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___radd___cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___radd___cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___radd___cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___radd___cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___radd___cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___radd___cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___radd___cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___radd___cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___radd___cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___radd___cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___radd___cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___radd___cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rand___cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rand___cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rand___cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rand___cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rand___cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rand___cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rdiv___cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rdiv___cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rdiv___cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rdiv___cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rdiv___cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rdiv___cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rdiv___cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rdiv___cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rdiv___cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rdiv___cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rdiv___cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rdiv___cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmatmul___cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmatmul___cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmatmul___cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmatmul___cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmatmul___cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmatmul___cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmod___cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmod___cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmod___cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmod___cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmod___cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmod___cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmod___cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmod___cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmod___cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmul___cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmul___cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmul___cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmul___cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmul___cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmul___cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmul___cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmul___cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmul___cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmul___cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmul___cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rmul___cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___ror___cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___ror___cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___ror___cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___ror___cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___ror___cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___ror___cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rpow___cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rpow___cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rpow___cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rpow___cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rpow___cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rpow___cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rpow___cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rpow___cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rpow___cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rpow___cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rpow___cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rsub___cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rsub___cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rsub___cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rsub___cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rsub___cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rsub___cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rsub___cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rsub___cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rsub___cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rsub___cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rsub___cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rxor___cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rxor___cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rxor___cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rxor___cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rxor___cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops___rxor___cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__batch_norm_with_update_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__batch_norm_with_update_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__batch_norm_with_update_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__batch_norm_with_update_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__chunk_cat_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__chunk_cat_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__chunk_cat_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__chunk_cat_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__chunk_cat_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__chunk_cat_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__chunk_cat_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__chunk_cat_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__chunk_cat_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__chunk_cat_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__chunk_cat_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__chunk_cat_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__chunk_cat_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__native_batch_norm_legit_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__native_batch_norm_legit_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__native_batch_norm_legit_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__native_batch_norm_legit_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__segment_reduce_lengths_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__segment_reduce_lengths_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__segment_reduce_lengths_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__segment_reduce_lengths_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__segment_reduce_offsets_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__segment_reduce_offsets_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__segment_reduce_offsets_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__segment_reduce_offsets_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__softmax_backward_data_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__softmax_backward_data_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__softmax_backward_data_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__softmax_backward_data_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_put_accumulate_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_put_accumulate_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_put_accumulate_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_put_accumulate_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_put_accumulate_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_put_accumulate_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_put_accumulate_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_put_accumulate_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_put_accumulate_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_put_accumulate_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_put_accumulate_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__unsafe_masked_index_put_accumulate_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__upsample_bilinear2d_aa_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__upsample_bilinear2d_aa_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__upsample_bilinear2d_aa_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops__upsample_bilinear2d_aa_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_abs_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_abs_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_abs_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_abs_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_abs_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_abs_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_abs_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_abs_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_abs_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_abs_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_abs_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_abs_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_abs_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acos_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acos_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acos_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acos_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acos_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acos_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acos_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acos_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acos_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acos_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acos_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acos_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acos_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acosh_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acosh_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acosh_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acosh_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acosh_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acosh_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acosh_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acosh_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acosh_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acosh_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acosh_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acosh_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_acosh_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_add_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_add_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_add_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_add_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_add_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_add_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_add_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_add_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_add_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_add_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_add_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_add_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_add_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addbmm_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addbmm_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addbmm_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addbmm_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addbmm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addbmm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addcdiv_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addcdiv_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addcdiv_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addcdiv_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addcdiv_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addcdiv_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addcmul_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addcmul_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addcmul_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addcmul_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addcmul_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addcmul_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addcmul_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addcmul_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addcmul_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addcmul_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addcmul_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addmm_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addmm_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addmm_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addmm_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addmm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addmm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addmm_decomposed_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addmm_decomposed_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addmm_decomposed_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addmm_decomposed_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addmm_decomposed_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addmm_decomposed_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addmv_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addmv_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addmv_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addmv_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addmv_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addmv_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addr_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addr_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addr_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addr_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addr_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addr_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addr_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addr_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addr_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addr_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addr_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_addr_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_alias_copy_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_alias_copy_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_alias_copy_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_alias_copy_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_alias_copy_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_alias_copy_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_alias_copy_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_alias_copy_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_alias_copy_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_alias_copy_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_alias_copy_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_alias_copy_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_alias_copy_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_all_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_all_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_all_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_all_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_all_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_all_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_all_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_all_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_all_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_all_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_all_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_all_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_allclose_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_allclose_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_allclose_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_allclose_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_allclose_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_allclose_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amax_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amax_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amax_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amax_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amax_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amax_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amax_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amax_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amax_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amax_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amin_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amin_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amin_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amin_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amin_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amin_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amin_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amin_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amin_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_amin_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_aminmax_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_aminmax_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_aminmax_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_aminmax_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_aminmax_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_aminmax_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_aminmax_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_aminmax_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_aminmax_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_aminmax_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_angle_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_angle_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_angle_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_angle_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_angle_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_angle_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_angle_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_angle_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_angle_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_angle_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_angle_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_any_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_any_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_any_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_any_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_any_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_any_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_any_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_any_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_any_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_any_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_any_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_any_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_arange_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_arange_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_arange_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_arange_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_arange_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_arange_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_arange_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_arange_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_arange_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argmax_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argmax_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argmax_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argmax_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argmax_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argmax_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argmax_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argmax_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argmax_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argmin_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argmin_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argmin_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argmin_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argmin_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argmin_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argmin_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argmin_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argmin_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argsort_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argsort_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argsort_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argsort_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argsort_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argsort_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argsort_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argsort_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argsort_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argsort_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argwhere_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argwhere_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argwhere_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argwhere_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argwhere_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argwhere_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argwhere_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argwhere_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argwhere_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argwhere_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argwhere_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_argwhere_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_copy_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_copy_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_copy_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_copy_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_copy_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_copy_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_copy_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_copy_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_copy_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_copy_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_copy_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_copy_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_copy_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_partial_views_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_partial_views_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_partial_views_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_partial_views_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_partial_views_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_partial_views_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_partial_views_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_partial_views_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_partial_views_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_partial_views_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_partial_views_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_partial_views_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_partial_views_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_scatter_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_scatter_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_scatter_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_scatter_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_scatter_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_scatter_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_scatter_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_scatter_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_scatter_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_scatter_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_scatter_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_scatter_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_as_strided_scatter_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asin_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asin_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asin_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asin_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asin_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asin_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asin_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asin_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asin_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asin_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asin_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asin_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asin_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asinh_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asinh_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asinh_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asinh_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asinh_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asinh_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asinh_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asinh_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asinh_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asinh_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asinh_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asinh_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_asinh_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan2_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan2_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan2_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan2_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan2_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan2_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan2_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan2_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan2_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan2_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atan_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atanh_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atanh_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atanh_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atanh_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atanh_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atanh_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atanh_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atanh_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atanh_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atanh_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atanh_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atanh_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atanh_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_1d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_1d_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_1d_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_1d_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_1d_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_1d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_1d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_1d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_1d_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_1d_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_1d_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_1d_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_1d_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_2d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_2d_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_2d_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_2d_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_2d_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_2d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_2d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_2d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_2d_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_2d_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_2d_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_2d_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_2d_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_3d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_3d_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_3d_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_3d_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_3d_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_3d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_3d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_3d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_3d_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_3d_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_3d_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_3d_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_atleast_3d_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_baddbmm_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_baddbmm_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_baddbmm_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_baddbmm_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_baddbmm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_baddbmm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bernoulli_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bernoulli_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bernoulli_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bernoulli_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bfloat16_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bfloat16_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bfloat16_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bfloat16_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bfloat16_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bfloat16_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bfloat16_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bfloat16_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bfloat16_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bfloat16_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bfloat16_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bfloat16_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bfloat16_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bincount_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bincount_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bincount_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bincount_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bincount_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_and_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_and_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_and_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_and_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_and_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_and_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_left_shift_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_left_shift_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_left_shift_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_left_shift_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_left_shift_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_not_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_not_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_not_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_not_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_not_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_not_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_or_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_or_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_or_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_or_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_or_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_or_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_right_shift_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_right_shift_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_right_shift_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_right_shift_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_right_shift_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_xor_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_xor_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_xor_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_xor_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_xor_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bitwise_xor_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_block_diag_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_block_diag_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_block_diag_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_block_diag_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_block_diag_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_block_diag_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_block_diag_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_block_diag_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_block_diag_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_block_diag_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_block_diag_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_block_diag_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_block_diag_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bmm_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bmm_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bmm_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bmm_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bmm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bmm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bool_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bool_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bool_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bool_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bool_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bool_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bool_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bool_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bool_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bool_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bool_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bool_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bool_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_shapes_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_tensors_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_tensors_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_tensors_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_tensors_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_tensors_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_tensors_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_tensors_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_tensors_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_tensors_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_tensors_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_tensors_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_tensors_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_to_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_to_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_to_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_to_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_to_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_to_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_to_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_to_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_to_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_to_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_to_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_broadcast_to_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bucketize_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bucketize_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bucketize_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bucketize_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bucketize_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bucketize_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bucketize_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bucketize_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_bucketize_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_byte_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_byte_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_byte_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_byte_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_byte_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_byte_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_byte_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_byte_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_byte_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_byte_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_byte_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_byte_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cartesian_prod_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cartesian_prod_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cartesian_prod_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cartesian_prod_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cartesian_prod_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cartesian_prod_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cartesian_prod_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cartesian_prod_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cartesian_prod_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cartesian_prod_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cartesian_prod_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cartesian_prod_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cat_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cat_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cat_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cat_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cat_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cat_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cat_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cat_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cat_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cat_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cat_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cat_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cat_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cauchy_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cauchy_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cauchy_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cauchy_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cdist_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cdist_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cdouble_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cdouble_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cdouble_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cdouble_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cdouble_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cdouble_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cdouble_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cdouble_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cdouble_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cdouble_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cdouble_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cdouble_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cdouble_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ceil_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ceil_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ceil_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ceil_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ceil_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ceil_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ceil_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ceil_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ceil_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cfloat_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cfloat_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cfloat_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cfloat_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cfloat_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cfloat_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cfloat_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cfloat_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cfloat_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cfloat_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cfloat_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cfloat_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cfloat_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chalf_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chalf_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chalf_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chalf_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chalf_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chalf_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chalf_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chalf_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chalf_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chalf_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chalf_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chalf_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chalf_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_char_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_char_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_char_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_char_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_char_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_char_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_char_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_char_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_char_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_char_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_char_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_char_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_char_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cholesky_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cholesky_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cholesky_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cholesky_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cholesky_inverse_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cholesky_inverse_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cholesky_inverse_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cholesky_inverse_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cholesky_solve_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cholesky_solve_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cholesky_solve_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cholesky_solve_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chunk_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chunk_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chunk_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chunk_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chunk_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chunk_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chunk_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chunk_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chunk_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chunk_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chunk_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chunk_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_chunk_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_max_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_max_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_max_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_max_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_max_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_max_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_max_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_max_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_max_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_max_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_min_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_min_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_min_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_min_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_min_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_min_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_min_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_min_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_min_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clamp_min_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clone_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clone_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clone_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clone_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clone_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clone_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clone_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clone_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clone_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clone_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clone_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clone_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_clone_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_column_stack_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_column_stack_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_column_stack_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_column_stack_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_column_stack_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_column_stack_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_column_stack_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_column_stack_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_column_stack_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_column_stack_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_column_stack_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_column_stack_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_column_stack_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_combinations_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_combinations_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_combinations_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_combinations_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_combinations_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_combinations_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_combinations_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_combinations_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_combinations_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_combinations_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_combinations_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_combinations_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_complex_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_complex_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_complex_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_physical_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_physical_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_physical_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_physical_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_physical_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_physical_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_physical_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_physical_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_physical_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_physical_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_physical_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_physical_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_conj_physical_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_constant_pad_nd_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_constant_pad_nd_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_constant_pad_nd_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_constant_pad_nd_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_constant_pad_nd_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_constant_pad_nd_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_constant_pad_nd_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_constant_pad_nd_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_constant_pad_nd_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_constant_pad_nd_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_constant_pad_nd_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_constant_pad_nd_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_contiguous_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_contiguous_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_contiguous_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_contiguous_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_contiguous_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_contiguous_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_contiguous_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_contiguous_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_contiguous_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_contiguous_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_contiguous_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_contiguous_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_contiguous_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_copysign_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_copysign_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_copysign_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_copysign_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_copysign_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_copysign_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_copysign_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_copysign_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_copysign_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_copysign_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_corrcoef_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_corrcoef_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_corrcoef_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_corrcoef_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_corrcoef_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_corrcoef_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_corrcoef_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_corrcoef_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_corrcoef_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_corrcoef_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_corrcoef_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cos_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cos_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cos_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cos_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cos_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cos_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cos_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cos_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cos_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cos_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cos_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cos_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cos_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cosh_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cosh_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cosh_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cosh_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cosh_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cosh_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cosh_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cosh_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cosh_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cosh_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cosh_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cosh_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cosh_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_count_nonzero_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_count_nonzero_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_count_nonzero_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_count_nonzero_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_count_nonzero_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_count_nonzero_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_count_nonzero_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_count_nonzero_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_count_nonzero_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_count_nonzero_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_count_nonzero_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_count_nonzero_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cov_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cov_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cov_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cov_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cov_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cov_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cov_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cov_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cov_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cov_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cov_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cross_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cross_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cross_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cross_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cross_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cross_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cross_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cross_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cross_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cross_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cross_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummax_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummax_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummax_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummax_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummax_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummax_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummax_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummax_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummax_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummax_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummin_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummin_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummin_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummin_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummin_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummin_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummin_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummin_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummin_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cummin_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumprod_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumprod_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumprod_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumprod_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumprod_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumprod_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumprod_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumprod_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumprod_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumprod_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumprod_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumsum_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumsum_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumsum_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumsum_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumsum_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumsum_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumsum_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumsum_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumsum_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumsum_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumsum_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumulative_trapezoid_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumulative_trapezoid_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumulative_trapezoid_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumulative_trapezoid_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumulative_trapezoid_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumulative_trapezoid_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumulative_trapezoid_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumulative_trapezoid_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumulative_trapezoid_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumulative_trapezoid_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_cumulative_trapezoid_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_deg2rad_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_deg2rad_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_deg2rad_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_deg2rad_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_deg2rad_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_deg2rad_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_deg2rad_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_deg2rad_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_deg2rad_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_deg2rad_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_embed_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_embed_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_embed_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_embed_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_embed_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_embed_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_embed_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_embed_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_embed_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_embed_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_embed_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_embed_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diag_embed_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagflat_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagflat_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagflat_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagflat_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagflat_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagflat_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagflat_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagflat_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagflat_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagflat_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagflat_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagflat_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_copy_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_copy_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_copy_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_copy_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_copy_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_copy_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_copy_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_copy_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_copy_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_copy_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_copy_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_copy_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_copy_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_scatter_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_scatter_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_scatter_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_scatter_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_scatter_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_scatter_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_scatter_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_scatter_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_scatter_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_scatter_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_scatter_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diagonal_scatter_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diff_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diff_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diff_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diff_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diff_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diff_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diff_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diff_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diff_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diff_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diff_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_diff_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_digamma_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_digamma_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_digamma_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_digamma_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_digamma_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_digamma_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_digamma_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_digamma_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_digamma_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_digamma_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dist_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dist_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dist_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dist_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dist_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dist_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_floor_rounding_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_floor_rounding_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_floor_rounding_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_floor_rounding_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_floor_rounding_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_floor_rounding_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_floor_rounding_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_floor_rounding_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_floor_rounding_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_no_rounding_mode_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_no_rounding_mode_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_no_rounding_mode_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_no_rounding_mode_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_no_rounding_mode_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_no_rounding_mode_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_no_rounding_mode_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_no_rounding_mode_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_no_rounding_mode_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_no_rounding_mode_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_no_rounding_mode_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_no_rounding_mode_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_no_rounding_mode_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_trunc_rounding_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_trunc_rounding_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_trunc_rounding_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_trunc_rounding_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_trunc_rounding_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_trunc_rounding_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_trunc_rounding_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_trunc_rounding_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_div_trunc_rounding_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dot_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dot_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dot_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dot_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dot_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dot_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_double_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_double_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_double_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_double_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_double_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_double_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_double_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_double_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_double_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_double_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_double_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_double_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_double_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dsplit_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dsplit_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dsplit_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dsplit_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dsplit_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dsplit_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dsplit_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dsplit_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dsplit_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dsplit_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dsplit_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dsplit_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dsplit_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dstack_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dstack_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dstack_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dstack_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dstack_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dstack_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dstack_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dstack_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dstack_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dstack_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dstack_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dstack_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_dstack_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_einsum_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_einsum_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_einsum_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_einsum_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_einsum_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_einsum_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_like_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_like_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_like_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_like_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_like_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_like_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_like_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_like_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_like_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_like_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_like_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_like_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_like_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_permuted_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_permuted_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_permuted_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_permuted_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_permuted_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_permuted_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_permuted_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_permuted_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_permuted_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_permuted_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_permuted_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_permuted_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_permuted_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_strided_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_strided_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_strided_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_strided_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_strided_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_strided_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_strided_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_strided_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_strided_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_strided_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_strided_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_empty_strided_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eq_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eq_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eq_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eq_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eq_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eq_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eq_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eq_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eq_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eq_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eq_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eq_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eq_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_equal_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_equal_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_equal_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_equal_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_equal_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_equal_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_equal_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_equal_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_equal_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_equal_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_equal_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_equal_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erf_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erf_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erf_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erf_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erf_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erf_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erf_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erf_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erf_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erf_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfc_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfc_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfc_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfc_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfc_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfc_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfc_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfc_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfc_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfc_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfinv_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfinv_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfinv_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfinv_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfinv_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfinv_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfinv_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfinv_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfinv_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_erfinv_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp2_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp2_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp2_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp2_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp2_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp2_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp2_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp2_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp2_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp2_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp2_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp2_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exp_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_as_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_as_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_as_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_as_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_as_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_as_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_as_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_as_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_as_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_as_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_as_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_as_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_copy_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_copy_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_copy_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_copy_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_copy_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_copy_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_copy_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_copy_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_copy_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_copy_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_copy_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_copy_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expand_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expm1_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expm1_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expm1_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expm1_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expm1_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expm1_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expm1_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expm1_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expm1_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expm1_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expm1_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_expm1_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exponential_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exponential_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exponential_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_exponential_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eye_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eye_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eye_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eye_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eye_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eye_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eye_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eye_cuda_float8_e4m3fn, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eye_cuda_float8_e4m3fnuz, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eye_cuda_float8_e5m2, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eye_cuda_float8_e5m2fnuz, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eye_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eye_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eye_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eye_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_eye_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft2_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft2_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft2_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft2_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft2_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft2_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft2_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft2_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft2_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft2_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft2_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft2_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fft_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftn_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftn_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftn_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftn_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftn_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftn_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftn_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftn_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftn_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftn_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftn_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftn_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftshift_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftshift_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftshift_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftshift_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftshift_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftshift_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftshift_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftshift_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftshift_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftshift_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftshift_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftshift_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_fftshift_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft2_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft2_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft2_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft2_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft2_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft2_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft2_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft2_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft2_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft2_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft2_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft2_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfft_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfftn_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfftn_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfftn_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfftn_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfftn_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfftn_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfftn_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfftn_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfftn_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfftn_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfftn_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_hfftn_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft2_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft2_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft2_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft2_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft2_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft2_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft2_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft2_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft2_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft2_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft2_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft2_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifft_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftn_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftn_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftn_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftn_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftn_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftn_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftn_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftn_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftn_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftn_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftn_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftn_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftshift_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftshift_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftshift_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftshift_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftshift_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftshift_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftshift_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftshift_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftshift_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftshift_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftshift_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftshift_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ifftshift_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfft2_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfft2_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfft2_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfft2_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfft2_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfft2_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfft2_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfft2_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfft2_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfft_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfft_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfft_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfft_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfft_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfft_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfft_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfft_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfft_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfftn_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfftn_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfftn_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfftn_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfftn_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfftn_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfftn_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfftn_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_ihfftn_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft2_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft2_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft2_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft2_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft2_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft2_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft2_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft2_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft2_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft2_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft2_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft2_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfft_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfftn_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfftn_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfftn_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfftn_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfftn_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfftn_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfftn_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfftn_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfftn_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfftn_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfftn_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_irfftn_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfft2_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfft2_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfft2_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfft2_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfft2_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfft2_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfft2_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfft2_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfft2_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfft_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfft_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfft_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfft_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfft_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfft_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfft_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfft_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfft_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfftn_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfftn_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfftn_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfftn_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfftn_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfftn_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfftn_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfftn_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fft_rfftn_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fill_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fill_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fill_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fill_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fill_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fill_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fill_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fill_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fill_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fill_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fill_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fill_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fill_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flatten_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flatten_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flatten_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flatten_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flatten_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flatten_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flatten_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flatten_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flatten_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flatten_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flatten_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flatten_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flatten_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flip_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flip_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flip_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flip_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flip_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flip_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flip_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flip_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flip_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flip_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flip_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flip_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fliplr_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fliplr_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fliplr_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fliplr_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fliplr_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fliplr_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fliplr_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fliplr_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fliplr_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fliplr_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fliplr_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fliplr_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flipud_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flipud_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flipud_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flipud_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flipud_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flipud_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flipud_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flipud_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flipud_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flipud_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flipud_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_flipud_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_power_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_power_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_power_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_power_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_power_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_power_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_power_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_power_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_power_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_power_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_power_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_float_power_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_floor_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_floor_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_floor_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_floor_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_floor_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_floor_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_floor_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_floor_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_floor_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_floor_divide_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_floor_divide_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_floor_divide_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_floor_divide_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_floor_divide_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_floor_divide_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_floor_divide_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_floor_divide_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_floor_divide_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmax_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmax_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmax_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmax_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmax_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmax_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmax_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmax_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmax_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmax_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmin_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmin_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmin_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmin_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmin_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmin_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmin_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmin_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmin_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmin_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmod_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmod_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmod_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmod_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmod_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmod_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmod_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmod_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_fmod_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_frac_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_frac_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_frac_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_frac_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_frexp_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_frexp_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_frexp_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_frexp_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_like_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_like_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_like_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_like_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_like_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_like_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_like_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_like_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_like_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_like_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_like_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_like_cuda_uint16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_like_cuda_uint32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_full_like_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gather_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gather_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gather_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gather_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gather_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gather_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gather_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gather_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gather_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gather_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gather_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gather_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gcd_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gcd_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gcd_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gcd_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gcd_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ge_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ge_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ge_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ge_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ge_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ge_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ge_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ge_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ge_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ge_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_geometric_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_geometric_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_geometric_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_geometric_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_geometric_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_geometric_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_geometric_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_geometric_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_geometric_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_geqrf_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_geqrf_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_geqrf_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_geqrf_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gradient_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gradient_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gradient_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gradient_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gradient_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gradient_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gradient_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gradient_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gradient_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gradient_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_grid_sampler_2d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_grid_sampler_2d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_grid_sampler_2d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_grid_sampler_2d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_grid_sampler_3d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_grid_sampler_3d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_grid_sampler_3d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_grid_sampler_3d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gt_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gt_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gt_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gt_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gt_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gt_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gt_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gt_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gt_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_gt_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_half_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_half_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_half_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_half_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_half_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_half_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_half_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_half_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_half_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_half_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_half_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_half_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hash_tensor_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hash_tensor_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hash_tensor_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hash_tensor_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hash_tensor_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hash_tensor_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hash_tensor_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hash_tensor_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hash_tensor_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hash_tensor_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_heaviside_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_heaviside_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_heaviside_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_heaviside_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_heaviside_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_heaviside_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_heaviside_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_heaviside_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_heaviside_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_heaviside_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_histc_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_histc_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_histc_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_histc_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_histc_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_histc_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_histc_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hsplit_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hsplit_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hsplit_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hsplit_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hsplit_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hsplit_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hsplit_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hsplit_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hsplit_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hsplit_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hsplit_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hsplit_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hsplit_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hstack_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hstack_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hstack_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hstack_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hstack_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hstack_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hstack_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hstack_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hstack_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hstack_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hstack_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hstack_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hstack_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hypot_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hypot_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hypot_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_hypot_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_i0_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_i0_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_i0_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_i0_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_i0_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_i0_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_i0_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_i0_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_i0_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_i0_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_igamma_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_igamma_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_igammac_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_igammac_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_imag_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_imag_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_imag_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_add_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_add_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_add_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_add_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_add_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_add_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_add_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_add_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_add_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_add_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_add_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_add_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_add_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_copy_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_copy_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_copy_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_copy_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_copy_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_copy_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_copy_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_copy_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_copy_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_copy_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_copy_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_copy_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_copy_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_fill_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_fill_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_fill_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_fill_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_fill_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_fill_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_fill_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_fill_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_fill_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_fill_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_fill_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_fill_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_fill_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_put_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_put_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_put_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_put_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_put_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_put_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_put_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_put_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_put_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_put_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_put_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_put_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_put_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_amax_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_amax_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_amax_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_amax_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_amax_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_amax_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_amax_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_amax_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_amax_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_amin_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_amin_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_amin_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_amin_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_amin_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_amin_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_amin_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_amin_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_amin_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_mean_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_mean_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_mean_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_mean_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_mean_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_mean_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_mean_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_mean_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_mean_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_prod_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_prod_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_prod_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_prod_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_prod_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_prod_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_prod_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_prod_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_reduce_prod_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_select_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_select_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_select_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_select_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_select_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_select_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_select_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_select_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_select_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_select_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_select_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_select_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_index_select_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_inner_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_inner_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_inner_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_inner_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_inner_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_inner_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_int_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_int_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_int_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_int_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_int_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_int_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_int_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_int_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_int_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_int_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_int_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_int_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isclose_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isclose_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isclose_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isclose_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isclose_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isclose_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isclose_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isclose_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isclose_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isclose_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isclose_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isclose_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isfinite_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isfinite_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isfinite_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isfinite_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isfinite_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isfinite_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isfinite_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isfinite_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isfinite_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isfinite_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isfinite_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isfinite_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isfinite_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isin_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isin_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isin_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isin_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isin_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isin_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isin_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isin_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isin_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isinf_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isinf_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isinf_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isinf_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isinf_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isinf_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isinf_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isinf_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isinf_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isinf_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isinf_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isinf_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isinf_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isnan_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isnan_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isnan_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isnan_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isnan_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isnan_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isnan_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isnan_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isnan_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isnan_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isnan_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isnan_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isneginf_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isneginf_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isneginf_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isneginf_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isneginf_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isneginf_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isneginf_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isneginf_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isneginf_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isneginf_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isposinf_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isposinf_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isposinf_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isposinf_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isposinf_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isposinf_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isposinf_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isposinf_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isposinf_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isposinf_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isreal_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isreal_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isreal_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isreal_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isreal_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isreal_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isreal_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isreal_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isreal_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isreal_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isreal_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isreal_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_isreal_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_istft_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_istft_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_item_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_item_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_item_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_item_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_item_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_item_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_item_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_item_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_item_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_item_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_item_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_item_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_item_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_2inputs_2outputs_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_2inputs_2outputs_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_2inputs_2outputs_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_2inputs_2outputs_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_2inputs_2outputs_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_2inputs_2outputs_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_2inputs_2outputs_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_2inputs_2outputs_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_2inputs_2outputs_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_2inputs_2outputs_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_2inputs_2outputs_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_2inputs_2outputs_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_4inputs_with_extra_args_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_4inputs_with_extra_args_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_4inputs_with_extra_args_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_4inputs_with_extra_args_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_4inputs_with_extra_args_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_4inputs_with_extra_args_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_4inputs_with_extra_args_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_4inputs_with_extra_args_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_4inputs_with_extra_args_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_4inputs_with_extra_args_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_4inputs_with_extra_args_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_4inputs_with_extra_args_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_return_by_ref_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_return_by_ref_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_return_by_ref_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_return_by_ref_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_return_by_ref_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_return_by_ref_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_return_by_ref_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_return_by_ref_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_return_by_ref_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_return_by_ref_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_return_by_ref_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_binary_return_by_ref_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_unary_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_unary_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_unary_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_unary_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_unary_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_unary_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_unary_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_unary_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_unary_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_unary_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_unary_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_jiterator_unary_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kron_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kron_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kron_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kron_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kron_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kron_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kron_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kron_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kron_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kron_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kron_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kron_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kthvalue_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kthvalue_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kthvalue_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kthvalue_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kthvalue_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kthvalue_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kthvalue_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kthvalue_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_kthvalue_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lcm_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lcm_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lcm_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lcm_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lcm_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ldexp_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ldexp_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ldexp_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ldexp_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ldexp_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ldexp_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ldexp_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ldexp_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ldexp_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ldexp_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ldexp_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ldexp_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_le_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_le_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_le_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_le_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_le_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_le_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_le_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_le_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_le_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_le_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lerp_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lerp_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lerp_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lerp_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lerp_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lerp_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lerp_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lgamma_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lgamma_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lgamma_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lgamma_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lgamma_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lgamma_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lgamma_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lgamma_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lgamma_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lgamma_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cholesky_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cholesky_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cholesky_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cholesky_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cholesky_ex_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cholesky_ex_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cholesky_ex_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cholesky_ex_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cond_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cond_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cond_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cond_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cross_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cross_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cross_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cross_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cross_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cross_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cross_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cross_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cross_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cross_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_cross_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_det_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_det_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_det_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_det_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_diagonal_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_diagonal_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_diagonal_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_diagonal_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_diagonal_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_diagonal_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_diagonal_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_diagonal_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_diagonal_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_diagonal_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_diagonal_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_diagonal_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_diagonal_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_eig_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_eig_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_eig_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_eig_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_eigh_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_eigh_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_eigh_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_eigh_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_eigvals_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_eigvals_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_eigvals_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_eigvals_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_eigvalsh_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_eigvalsh_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_eigvalsh_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_eigvalsh_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_householder_product_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_householder_product_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_householder_product_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_householder_product_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_inv_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_inv_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_inv_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_inv_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_inv_ex_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_inv_ex_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_inv_ex_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_inv_ex_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_ldl_factor_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_ldl_factor_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_ldl_factor_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_ldl_factor_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_ldl_factor_ex_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_ldl_factor_ex_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_ldl_factor_ex_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_ldl_factor_ex_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_ldl_solve_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_ldl_solve_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_ldl_solve_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_ldl_solve_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lstsq_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lstsq_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lstsq_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lstsq_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lstsq_grad_oriented_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lstsq_grad_oriented_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lstsq_grad_oriented_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lstsq_grad_oriented_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lu_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lu_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lu_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lu_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lu_factor_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lu_factor_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lu_factor_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lu_factor_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lu_factor_ex_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lu_factor_ex_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lu_factor_ex_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lu_factor_ex_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lu_solve_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lu_solve_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lu_solve_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_lu_solve_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_matrix_norm_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_matrix_norm_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_matrix_norm_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_matrix_norm_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_matrix_norm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_matrix_norm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_matrix_power_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_matrix_power_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_matrix_power_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_matrix_power_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_matrix_rank_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_matrix_rank_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_matrix_rank_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_matrix_rank_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_matrix_rank_hermitian_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_matrix_rank_hermitian_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_matrix_rank_hermitian_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_matrix_rank_hermitian_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_multi_dot_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_multi_dot_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_multi_dot_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_multi_dot_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_multi_dot_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_multi_dot_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_norm_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_norm_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_norm_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_norm_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_norm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_norm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_norm_subgradients_at_zero_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_norm_subgradients_at_zero_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_norm_subgradients_at_zero_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_norm_subgradients_at_zero_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_norm_subgradients_at_zero_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_norm_subgradients_at_zero_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_pinv_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_pinv_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_pinv_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_pinv_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_pinv_hermitian_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_pinv_hermitian_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_pinv_hermitian_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_pinv_hermitian_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_pinv_singular_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_pinv_singular_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_pinv_singular_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_pinv_singular_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_qr_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_qr_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_qr_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_qr_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_slogdet_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_slogdet_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_slogdet_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_slogdet_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_solve_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_solve_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_solve_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_solve_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_solve_ex_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_solve_ex_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_solve_ex_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_solve_ex_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_solve_triangular_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_solve_triangular_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_solve_triangular_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_solve_triangular_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_svd_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_svd_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_svd_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_svd_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_svdvals_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_svdvals_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_svdvals_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_svdvals_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_tensorinv_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_tensorinv_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_tensorinv_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_tensorinv_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_tensorsolve_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_tensorsolve_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_tensorsolve_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_tensorsolve_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vander_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vander_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vander_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vander_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vander_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vander_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vander_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vander_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vander_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vecdot_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vecdot_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vecdot_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vecdot_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vecdot_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vecdot_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vector_norm_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vector_norm_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vector_norm_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vector_norm_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vector_norm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linalg_vector_norm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_tensor_overload_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_tensor_overload_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_tensor_overload_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_tensor_overload_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_tensor_overload_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_tensor_overload_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_tensor_overload_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_tensor_overload_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_tensor_overload_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_tensor_overload_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_linspace_tensor_overload_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log10_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log10_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log10_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log10_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log10_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log10_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log10_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log10_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log10_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log10_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log10_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log10_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log1p_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log1p_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log1p_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log1p_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log1p_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log1p_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log1p_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log1p_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log1p_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log1p_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log1p_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log1p_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log2_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log2_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log2_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log2_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log2_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log2_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log2_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log2_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log2_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log2_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log2_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log2_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_normal_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_normal_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_normal_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_normal_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_softmax_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_softmax_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_softmax_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_softmax_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_softmax_with_dtype_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_softmax_with_dtype_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_softmax_with_dtype_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_softmax_with_dtype_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_softmax_with_dtype_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_softmax_with_dtype_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_softmax_with_dtype_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_softmax_with_dtype_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_softmax_with_dtype_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_softmax_with_dtype_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_softmax_with_dtype_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_softmax_with_dtype_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_log_softmax_with_dtype_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logaddexp2_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logaddexp2_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logaddexp2_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logaddexp2_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logaddexp_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logaddexp_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logaddexp_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logaddexp_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logaddexp_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logaddexp_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logaddexp_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logcumsumexp_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logcumsumexp_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logcumsumexp_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logcumsumexp_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logcumsumexp_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logcumsumexp_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logdet_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logdet_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logdet_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logdet_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_and_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_and_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_and_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_and_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_and_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_and_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_and_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_and_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_and_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_and_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_and_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_and_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_not_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_not_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_not_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_not_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_not_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_not_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_not_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_not_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_not_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_not_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_not_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_not_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_or_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_or_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_or_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_or_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_or_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_or_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_or_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_or_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_or_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_or_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_or_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_or_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_xor_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_xor_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_xor_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_xor_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_xor_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_xor_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_xor_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_xor_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_xor_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_xor_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_xor_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logical_xor_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logit_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logit_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logit_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logit_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logit_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logit_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logit_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logit_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logit_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logit_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_tensor_overload_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_tensor_overload_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_tensor_overload_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_tensor_overload_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_tensor_overload_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_tensor_overload_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_tensor_overload_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_tensor_overload_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_tensor_overload_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_tensor_overload_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logspace_tensor_overload_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logsumexp_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logsumexp_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logsumexp_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logsumexp_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logsumexp_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logsumexp_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logsumexp_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logsumexp_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logsumexp_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logsumexp_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logsumexp_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_logsumexp_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_long_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_long_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_long_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_long_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_long_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_long_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_long_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_long_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_long_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_long_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_long_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_long_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_long_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lt_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lt_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lt_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lt_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lt_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lt_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lt_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lt_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lt_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lt_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lu_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lu_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lu_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lu_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lu_solve_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lu_solve_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lu_solve_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lu_solve_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lu_unpack_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lu_unpack_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lu_unpack_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_lu_unpack_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mH_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mH_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mH_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mH_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mH_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mH_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mH_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mH_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mH_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mH_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mH_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mH_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mH_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mT_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mT_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mT_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mT_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mT_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mT_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mT_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mT_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mT_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mT_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mT_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mT_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mT_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_amax_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_amax_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_amax_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_amax_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_amax_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_amax_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_amax_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_amax_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_amax_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_amin_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_amin_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_amin_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_amin_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_amin_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_amin_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_amin_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_amin_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_amin_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_argmax_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_argmax_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_argmax_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_argmax_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_argmax_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_argmax_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_argmax_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_argmax_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_argmax_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_argmin_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_argmin_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_argmin_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_argmin_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_argmin_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_argmin_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_argmin_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_argmin_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_argmin_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumprod_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumprod_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumprod_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumprod_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumprod_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumprod_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumprod_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumprod_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumprod_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumprod_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumprod_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumsum_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumsum_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumsum_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumsum_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumsum_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumsum_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumsum_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumsum_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumsum_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumsum_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_cumsum_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_fill_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_fill_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_fill_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_fill_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_fill_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_fill_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_fill_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_fill_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_fill_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_fill_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_fill_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_fill_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_fill_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_log_softmax_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_log_softmax_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_log_softmax_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_log_softmax_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_logaddexp_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_logaddexp_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_logaddexp_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_logaddexp_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_logsumexp_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_logsumexp_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_logsumexp_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_logsumexp_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_logsumexp_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_logsumexp_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_logsumexp_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_logsumexp_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_logsumexp_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_logsumexp_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_logsumexp_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_mean_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_mean_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_mean_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_mean_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_mean_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_mean_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_median_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_median_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_median_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_median_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_norm_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_norm_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_norm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_norm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_normalize_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_normalize_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_normalize_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_normalize_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_normalize_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_normalize_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_prod_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_prod_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_prod_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_prod_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_prod_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_prod_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_prod_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_prod_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_prod_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_prod_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_prod_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_prod_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_scatter_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_scatter_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_scatter_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_scatter_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_scatter_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_scatter_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_scatter_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_scatter_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_scatter_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_scatter_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_scatter_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_scatter_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_select_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_select_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_select_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_select_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_select_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_select_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_select_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_select_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_select_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_select_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_select_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_select_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_softmax_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_softmax_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_softmax_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_softmax_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_softmin_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_softmin_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_softmin_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_softmin_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_std_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_std_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_std_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_std_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_std_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_std_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_std_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_std_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_std_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_std_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_std_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_sum_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_sum_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_sum_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_sum_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_sum_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_sum_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_sum_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_sum_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_sum_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_sum_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_sum_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_sum_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_var_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_var_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_var_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_var_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_var_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_var_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_var_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_var_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_var_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_var_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_masked_var_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_matmul_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_matmul_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_matmul_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_matmul_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_matmul_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_matmul_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_matrix_exp_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_matrix_exp_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_matrix_exp_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_matrix_exp_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_matrix_exp_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_matrix_exp_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_binary_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_binary_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_binary_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_binary_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_binary_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_binary_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_binary_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_binary_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_binary_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_binary_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_pool2d_with_indices_backward_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_pool2d_with_indices_backward_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_pool2d_with_indices_backward_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_pool2d_with_indices_backward_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_no_dim_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_no_dim_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_no_dim_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_no_dim_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_no_dim_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_no_dim_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_no_dim_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_no_dim_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_no_dim_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_no_dim_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_with_dim_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_with_dim_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_with_dim_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_with_dim_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_with_dim_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_with_dim_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_with_dim_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_with_dim_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_with_dim_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_max_reduction_with_dim_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_maximum_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_maximum_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_maximum_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_maximum_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_maximum_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_maximum_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_maximum_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_maximum_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_maximum_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_maximum_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mean_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mean_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mean_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mean_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mean_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mean_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_median_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_median_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_median_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_median_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_median_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_median_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_median_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_median_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_median_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_list_of_tensors_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_list_of_tensors_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_list_of_tensors_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_list_of_tensors_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_list_of_tensors_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_list_of_tensors_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_list_of_tensors_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_list_of_tensors_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_list_of_tensors_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_list_of_tensors_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_list_of_tensors_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_list_of_tensors_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_variadic_tensors_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_variadic_tensors_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_variadic_tensors_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_variadic_tensors_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_variadic_tensors_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_variadic_tensors_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_variadic_tensors_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_variadic_tensors_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_variadic_tensors_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_variadic_tensors_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_variadic_tensors_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_meshgrid_variadic_tensors_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_binary_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_binary_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_binary_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_binary_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_binary_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_binary_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_binary_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_binary_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_binary_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_binary_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_no_dim_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_no_dim_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_no_dim_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_no_dim_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_no_dim_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_no_dim_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_no_dim_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_no_dim_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_no_dim_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_no_dim_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_with_dim_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_with_dim_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_with_dim_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_with_dim_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_with_dim_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_with_dim_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_with_dim_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_with_dim_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_with_dim_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_min_reduction_with_dim_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_minimum_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_minimum_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_minimum_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_minimum_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_minimum_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_minimum_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_minimum_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_minimum_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_minimum_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_minimum_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mm_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mm_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mm_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mm_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mode_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mode_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mode_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mode_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mode_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mode_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mode_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mode_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mode_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mode_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_movedim_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_movedim_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_movedim_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_movedim_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_movedim_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_movedim_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_movedim_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_movedim_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_movedim_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_movedim_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_movedim_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_movedim_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_movedim_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_msort_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_msort_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_msort_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_msort_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_msort_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_msort_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_msort_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_msort_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_msort_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_msort_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mul_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mul_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mul_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mul_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mul_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mul_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mul_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mul_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mul_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mul_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mul_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mul_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mul_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_multinomial_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_multinomial_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_multinomial_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_multinomial_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mv_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mv_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mv_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mv_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mv_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mv_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_1_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_1_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_1_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_1_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_1_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_1_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_1_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_1_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_1_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_3_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_3_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_3_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_3_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_3_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_3_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_3_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_3_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_3_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_5_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_5_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_5_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_5_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_5_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_5_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_5_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_5_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_mvlgamma_mvlgamma_p_5_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nan_to_num_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nan_to_num_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nan_to_num_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nan_to_num_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nan_to_num_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nan_to_num_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nan_to_num_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nan_to_num_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nan_to_num_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nan_to_num_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nanmean_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nanmean_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nanmean_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nanmean_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nanmean_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nanmean_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nanmean_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nanmedian_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nanmedian_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nanmedian_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nanmedian_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nanmedian_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nanmedian_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nanmedian_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nanmedian_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nanmedian_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nanquantile_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nanquantile_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nansum_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nansum_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nansum_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nansum_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nansum_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nansum_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nansum_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nansum_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nansum_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nansum_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nansum_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nansum_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nansum_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_copy_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_copy_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_copy_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_copy_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_copy_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_copy_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_copy_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_copy_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_copy_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_copy_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_copy_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_copy_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_copy_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_narrow_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_native_batch_norm_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_native_batch_norm_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_native_batch_norm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_native_batch_norm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_native_dropout_backward_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_native_dropout_backward_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_native_dropout_backward_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_native_dropout_backward_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_native_layer_norm_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_native_layer_norm_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_native_layer_norm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_native_layer_norm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ne_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ne_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ne_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ne_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ne_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ne_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ne_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ne_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ne_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ne_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ne_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ne_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_neg_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_neg_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_neg_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_neg_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_neg_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_neg_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_neg_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_neg_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_neg_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_neg_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_neg_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_neg_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_strided_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_strided_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_strided_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_strided_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_strided_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_strided_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_strided_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_strided_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_strided_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_strided_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_strided_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_strided_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_empty_strided_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_full_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_full_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_full_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_full_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_full_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_full_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_full_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_full_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_full_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_full_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_full_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_full_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_full_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_ones_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_ones_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_ones_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_ones_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_ones_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_ones_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_ones_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_ones_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_ones_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_ones_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_ones_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_ones_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_ones_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_zeros_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_zeros_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_zeros_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_zeros_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_zeros_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_zeros_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_zeros_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_zeros_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_zeros_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_zeros_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_zeros_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_zeros_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_new_zeros_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nextafter_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nextafter_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nextafter_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nextafter_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_avg_pool1d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_avg_pool1d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_avg_pool1d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_avg_pool1d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_avg_pool2d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_avg_pool2d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_avg_pool2d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_avg_pool2d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_avg_pool3d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_avg_pool3d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_avg_pool3d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_avg_pool3d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_max_pool1d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_max_pool1d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_max_pool1d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_max_pool1d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_max_pool2d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_max_pool2d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_max_pool2d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_max_pool2d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_max_pool3d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_max_pool3d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_max_pool3d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_adaptive_max_pool3d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_alpha_dropout_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_alpha_dropout_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_alpha_dropout_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_alpha_dropout_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_avg_pool1d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_avg_pool1d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_avg_pool1d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_avg_pool1d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_avg_pool2d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_avg_pool2d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_avg_pool2d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_avg_pool2d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_avg_pool3d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_avg_pool3d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_avg_pool3d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_avg_pool3d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_batch_norm_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_batch_norm_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_batch_norm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_batch_norm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_batch_norm_without_cudnn_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_batch_norm_without_cudnn_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_batch_norm_without_cudnn_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_batch_norm_without_cudnn_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_bilinear_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_bilinear_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_bilinear_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_bilinear_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_binary_cross_entropy_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_binary_cross_entropy_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_binary_cross_entropy_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_binary_cross_entropy_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_binary_cross_entropy_with_logits_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_binary_cross_entropy_with_logits_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_binary_cross_entropy_with_logits_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_binary_cross_entropy_with_logits_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_celu_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_celu_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_celu_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_celu_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_channel_shuffle_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_channel_shuffle_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_channel_shuffle_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_channel_shuffle_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_channel_shuffle_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_channel_shuffle_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_channel_shuffle_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_channel_shuffle_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_channel_shuffle_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_channel_shuffle_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_channel_shuffle_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_channel_shuffle_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv1d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv1d_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv1d_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv1d_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv1d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv1d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv1d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv2d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv2d_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv2d_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv2d_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv2d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv2d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv2d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv3d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv3d_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv3d_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv3d_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv3d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv3d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv3d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose1d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose1d_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose1d_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose1d_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose1d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose1d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose1d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose2d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose2d_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose2d_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose2d_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose2d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose2d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose2d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose3d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose3d_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose3d_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose3d_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose3d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose3d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_conv_transpose3d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_cosine_embedding_loss_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_cosine_embedding_loss_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_cosine_embedding_loss_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_cosine_embedding_loss_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_cosine_embedding_loss_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_cosine_embedding_loss_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_cosine_embedding_loss_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_cosine_embedding_loss_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_cosine_embedding_loss_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_cosine_embedding_loss_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_cosine_similarity_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_cosine_similarity_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_cosine_similarity_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_cosine_similarity_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_cross_entropy_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_cross_entropy_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_cross_entropy_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_cross_entropy_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_ctc_loss_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_ctc_loss_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_dropout2d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_dropout2d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_dropout2d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_dropout2d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_dropout3d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_dropout3d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_dropout3d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_dropout3d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_dropout_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_dropout_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_dropout_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_dropout_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_elu_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_elu_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_elu_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_elu_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_embedding_bag_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_embedding_bag_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_embedding_bag_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_embedding_bag_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_embedding_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_embedding_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_embedding_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_embedding_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_feature_alpha_dropout_with_train_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_feature_alpha_dropout_with_train_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_feature_alpha_dropout_with_train_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_feature_alpha_dropout_with_train_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_feature_alpha_dropout_without_train_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_feature_alpha_dropout_without_train_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_feature_alpha_dropout_without_train_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_feature_alpha_dropout_without_train_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_feature_alpha_dropout_without_train_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_feature_alpha_dropout_without_train_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_feature_alpha_dropout_without_train_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_feature_alpha_dropout_without_train_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_feature_alpha_dropout_without_train_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_feature_alpha_dropout_without_train_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_feature_alpha_dropout_without_train_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_fractional_max_pool2d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_fractional_max_pool2d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_fractional_max_pool2d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_fractional_max_pool2d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_fractional_max_pool3d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_fractional_max_pool3d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_fractional_max_pool3d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_fractional_max_pool3d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_gaussian_nll_loss_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_gaussian_nll_loss_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_gaussian_nll_loss_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_gaussian_nll_loss_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_gelu_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_gelu_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_gelu_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_gelu_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_glu_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_glu_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_glu_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_glu_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_grid_sample_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_grid_sample_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_grid_sample_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_grid_sample_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_group_norm_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_group_norm_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_group_norm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_group_norm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardshrink_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardshrink_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardshrink_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardshrink_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardsigmoid_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardsigmoid_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardsigmoid_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardsigmoid_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardswish_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardswish_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardswish_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardswish_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardtanh_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardtanh_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardtanh_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardtanh_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardtanh_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardtanh_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardtanh_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hardtanh_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hinge_embedding_loss_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hinge_embedding_loss_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hinge_embedding_loss_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_hinge_embedding_loss_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_huber_loss_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_huber_loss_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_huber_loss_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_huber_loss_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_instance_norm_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_instance_norm_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_instance_norm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_instance_norm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_area_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_area_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_area_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_area_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_bicubic_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_bicubic_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_bicubic_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_bicubic_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_bilinear_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_bilinear_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_bilinear_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_bilinear_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_linear_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_linear_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_linear_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_linear_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_nearest-exact_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_nearest-exact_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_nearest-exact_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_nearest-exact_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_nearest-exact_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_nearest_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_nearest_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_nearest_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_nearest_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_nearest_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_trilinear_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_trilinear_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_trilinear_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_interpolate_trilinear_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_kl_div_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_kl_div_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_kl_div_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_kl_div_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_l1_loss_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_l1_loss_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_l1_loss_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_l1_loss_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_l1_loss_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_l1_loss_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_layer_norm_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_layer_norm_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_layer_norm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_layer_norm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_leaky_relu_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_leaky_relu_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_leaky_relu_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_leaky_relu_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_linear_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_linear_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_linear_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_linear_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_linear_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_linear_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_local_response_norm_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_local_response_norm_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_local_response_norm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_local_response_norm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_logsigmoid_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_logsigmoid_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_logsigmoid_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_logsigmoid_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_margin_ranking_loss_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_margin_ranking_loss_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_margin_ranking_loss_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_margin_ranking_loss_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_margin_ranking_loss_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_margin_ranking_loss_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_margin_ranking_loss_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_margin_ranking_loss_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_margin_ranking_loss_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_pool1d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_pool1d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_pool1d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_pool1d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_pool2d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_pool2d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_pool2d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_pool2d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_pool3d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_pool3d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_pool3d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_pool3d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool1d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool1d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool1d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool1d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool1d_grad_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool1d_grad_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool1d_grad_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool1d_grad_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool2d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool2d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool2d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool2d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool2d_grad_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool2d_grad_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool2d_grad_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool2d_grad_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool3d_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool3d_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool3d_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool3d_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool3d_grad_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool3d_grad_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool3d_grad_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_max_unpool3d_grad_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_mish_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_mish_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_mish_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_mish_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_mse_loss_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_mse_loss_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_mse_loss_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_mse_loss_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_multi_head_attention_forward_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_multi_head_attention_forward_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_multi_head_attention_forward_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_multi_head_attention_forward_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_multi_margin_loss_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_multi_margin_loss_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_multi_margin_loss_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_multi_margin_loss_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_multilabel_margin_loss_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_multilabel_margin_loss_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_multilabel_margin_loss_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_multilabel_margin_loss_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_multilabel_soft_margin_loss_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_multilabel_soft_margin_loss_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_multilabel_soft_margin_loss_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_multilabel_soft_margin_loss_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_nll_loss_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_nll_loss_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_nll_loss_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_nll_loss_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_normalize_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_normalize_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_normalize_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_normalize_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_normalize_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_normalize_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_one_hot_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_circular_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_circular_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_circular_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_circular_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_circular_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_circular_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_circular_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_circular_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_circular_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_circular_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_circular_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_circular_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_constant_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_constant_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_constant_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_constant_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_constant_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_constant_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_constant_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_constant_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_constant_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_constant_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_constant_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_constant_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_reflect_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_reflect_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_reflect_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_reflect_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_reflect_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_reflect_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_reflect_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_reflect_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_reflect_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_reflect_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_reflect_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_negative_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_negative_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_negative_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_negative_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_negative_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_negative_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_negative_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_negative_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_negative_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_negative_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pad_replicate_negative_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pairwise_distance_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pairwise_distance_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pairwise_distance_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pairwise_distance_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pairwise_distance_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pairwise_distance_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pairwise_distance_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pairwise_distance_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pairwise_distance_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pairwise_distance_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pairwise_distance_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pdist_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pdist_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_shuffle_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_shuffle_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_shuffle_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_shuffle_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_shuffle_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_shuffle_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_shuffle_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_shuffle_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_shuffle_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_shuffle_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_shuffle_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_shuffle_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_unshuffle_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_unshuffle_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_unshuffle_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_unshuffle_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_unshuffle_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_unshuffle_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_unshuffle_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_unshuffle_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_unshuffle_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_unshuffle_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_unshuffle_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_pixel_unshuffle_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_poisson_nll_loss_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_poisson_nll_loss_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_poisson_nll_loss_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_poisson_nll_loss_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_poisson_nll_loss_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_poisson_nll_loss_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_poisson_nll_loss_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_poisson_nll_loss_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_poisson_nll_loss_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_prelu_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_prelu_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_prelu_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_prelu_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_relu6_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_relu6_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_relu6_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_relu6_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_relu6_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_relu6_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_relu6_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_relu6_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_relu6_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_relu_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_relu_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_relu_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_relu_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_relu_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_relu_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_relu_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_relu_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_relu_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_rms_norm_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_rms_norm_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_rms_norm_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_rms_norm_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_rms_norm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_rms_norm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_rrelu_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_rrelu_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_rrelu_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_rrelu_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_scaled_dot_product_attention_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_scaled_dot_product_attention_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_scaled_dot_product_attention_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_scaled_dot_product_attention_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_selu_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_selu_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_selu_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_selu_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_silu_complex_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_silu_complex_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_silu_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_silu_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_silu_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_silu_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_smooth_l1_loss_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_smooth_l1_loss_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_smooth_l1_loss_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_smooth_l1_loss_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_soft_margin_loss_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_soft_margin_loss_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_soft_margin_loss_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_soft_margin_loss_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softmin_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softmin_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softmin_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softmin_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softmin_with_dtype_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softmin_with_dtype_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softmin_with_dtype_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softmin_with_dtype_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softmin_with_dtype_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softmin_with_dtype_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softmin_with_dtype_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softmin_with_dtype_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softmin_with_dtype_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softmin_with_dtype_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softmin_with_dtype_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softplus_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softplus_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softplus_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softplus_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softshrink_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softshrink_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softshrink_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softshrink_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softsign_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softsign_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softsign_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softsign_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softsign_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softsign_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softsign_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softsign_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softsign_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softsign_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softsign_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_softsign_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_tanhshrink_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_tanhshrink_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_tanhshrink_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_tanhshrink_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_tanhshrink_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_tanhshrink_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_tanhshrink_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_tanhshrink_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_tanhshrink_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_tanhshrink_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_tanhshrink_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_threshold_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_threshold_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_threshold_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_threshold_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_threshold_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_threshold_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_threshold_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_threshold_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_threshold_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_loss_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_loss_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_loss_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_loss_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_loss_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_loss_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_loss_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_loss_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_loss_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_loss_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_loss_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_with_distance_loss_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_with_distance_loss_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_with_distance_loss_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_with_distance_loss_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_with_distance_loss_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_with_distance_loss_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_with_distance_loss_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_with_distance_loss_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_with_distance_loss_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_with_distance_loss_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_triplet_margin_with_distance_loss_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_unfold_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_unfold_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_unfold_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_unfold_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_unfold_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_unfold_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_unfold_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_upsample_bilinear_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_upsample_bilinear_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_upsample_bilinear_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_upsample_bilinear_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_upsample_nearest_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_upsample_nearest_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_upsample_nearest_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_upsample_nearest_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nn_functional_upsample_nearest_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_static_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_static_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_static_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_static_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_static_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_static_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_static_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_static_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_static_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_static_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_static_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_static_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_nonzero_static_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_fro_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_fro_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_fro_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_fro_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_fro_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_fro_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_inf_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_inf_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_inf_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_inf_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_inf_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_inf_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_nuc_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_nuc_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_nuc_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_norm_nuc_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_normal_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_normal_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_normal_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_normal_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_normal_in_place_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_normal_in_place_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_normal_in_place_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_normal_in_place_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_normal_in_place_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_normal_in_place_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_normal_number_mean_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_normal_number_mean_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_normal_number_mean_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_normal_number_mean_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_like_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_like_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_like_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_like_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_like_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_like_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_like_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_like_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_like_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_like_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_like_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_like_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ones_like_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ormqr_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ormqr_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ormqr_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ormqr_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_outer_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_outer_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_outer_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_outer_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_outer_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_outer_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_outer_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_outer_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_outer_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_outer_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_outer_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_outer_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pca_lowrank_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pca_lowrank_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pca_lowrank_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pca_lowrank_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_copy_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_copy_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_copy_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_copy_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_copy_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_copy_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_copy_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_copy_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_copy_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_copy_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_copy_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_copy_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_copy_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_permute_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pinverse_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pinverse_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pinverse_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pinverse_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polar_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polar_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_0_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_0_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_0_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_0_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_0_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_0_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_0_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_0_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_0_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_0_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_1_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_1_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_1_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_1_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_1_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_1_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_1_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_1_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_1_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_1_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_2_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_2_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_2_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_2_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_2_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_2_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_2_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_2_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_2_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_2_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_3_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_3_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_3_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_3_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_3_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_3_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_3_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_3_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_3_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_3_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_4_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_4_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_4_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_4_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_4_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_4_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_4_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_4_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_4_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_polygamma_polygamma_n_4_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_positive_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_positive_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_positive_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_positive_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_positive_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_positive_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_positive_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_positive_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_positive_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_positive_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_positive_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_positive_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pow_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pow_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pow_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pow_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pow_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pow_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pow_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pow_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pow_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pow_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pow_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_pow_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_prod_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_prod_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_prod_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_prod_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_prod_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_prod_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_prod_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_prod_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_prod_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_prod_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_prod_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_prod_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_prod_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_put_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_put_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_put_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_put_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_put_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_put_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_put_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_put_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_put_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_put_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_put_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_put_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_qr_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_qr_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_qr_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_qr_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_quantile_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_quantile_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rad2deg_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rad2deg_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rad2deg_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rad2deg_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rad2deg_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rad2deg_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rad2deg_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rad2deg_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rad2deg_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rad2deg_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rand_like_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rand_like_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rand_like_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rand_like_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rand_like_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rand_like_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rand_like_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randint_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randint_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randint_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randint_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randint_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randint_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randint_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randint_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randint_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randint_like_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randint_like_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randint_like_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randint_like_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randint_like_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randint_like_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randint_like_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randint_like_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randint_like_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randn_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randn_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randn_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randn_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randn_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randn_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randn_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randn_like_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randn_like_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randn_like_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randn_like_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randn_like_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randn_like_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_randn_like_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ravel_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ravel_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ravel_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ravel_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ravel_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ravel_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ravel_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ravel_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ravel_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ravel_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ravel_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ravel_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_ravel_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_real_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_real_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_real_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_real_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_real_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_real_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_real_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_real_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_real_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_real_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_real_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_real_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_real_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reciprocal_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reciprocal_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reciprocal_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reciprocal_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reciprocal_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reciprocal_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reciprocal_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reciprocal_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reciprocal_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reciprocal_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reciprocal_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reciprocal_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_remainder_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_remainder_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_remainder_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_remainder_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_remainder_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_remainder_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_remainder_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_remainder_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_remainder_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_renorm_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_renorm_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_renorm_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_renorm_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_renorm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_renorm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_interleave_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_interleave_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_interleave_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_interleave_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_interleave_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_interleave_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_interleave_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_interleave_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_interleave_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_interleave_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_interleave_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_interleave_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_repeat_interleave_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_as_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_as_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_as_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_as_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_as_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_as_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_as_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_as_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_as_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_as_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_as_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_as_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_as_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_reshape_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize__cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize__cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize__cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize__cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize__cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize__cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize__cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize__cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize__cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize__cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize__cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize__cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize_as__cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize_as__cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize_as__cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize_as__cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize_as__cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize_as__cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize_as__cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize_as__cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize_as__cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize_as__cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize_as__cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resize_as__cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_conj_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_conj_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_conj_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_conj_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_conj_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_conj_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_conj_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_conj_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_conj_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_conj_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_conj_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_conj_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_neg_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_neg_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_neg_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_neg_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_neg_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_neg_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_neg_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_neg_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_neg_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_neg_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_neg_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_neg_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_resolve_neg_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_roll_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_roll_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_roll_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_roll_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_roll_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_roll_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_roll_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_roll_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_roll_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_roll_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_roll_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_roll_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_roll_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rot90_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rot90_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rot90_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rot90_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rot90_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rot90_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rot90_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rot90_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rot90_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rot90_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rot90_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rot90_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_decimals_0_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_decimals_0_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_decimals_0_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_decimals_0_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_decimals_3_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_decimals_3_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_decimals_3_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_decimals_3_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_decimals_neg_3_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_decimals_neg_3_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_decimals_neg_3_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_round_decimals_neg_3_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsqrt_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsqrt_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsqrt_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsqrt_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsqrt_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsqrt_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsqrt_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsqrt_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsqrt_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsqrt_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsqrt_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsqrt_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsqrt_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsub_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsub_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsub_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsub_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsub_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsub_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsub_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsub_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsub_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsub_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_rsub_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scalar_tensor_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scalar_tensor_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scalar_tensor_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scalar_tensor_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scalar_tensor_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scalar_tensor_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scalar_tensor_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scalar_tensor_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scalar_tensor_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scalar_tensor_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scalar_tensor_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scalar_tensor_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scalar_tensor_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_add_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_add_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_add_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_add_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_add_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_add_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_add_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_add_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_add_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_add_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_add_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_add_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_amax_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_amax_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_amax_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_amax_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_amax_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_amax_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_amax_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_amax_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_amax_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_amin_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_amin_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_amin_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_amin_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_amin_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_amin_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_amin_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_amin_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_amin_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_mean_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_mean_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_mean_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_mean_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_mean_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_mean_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_mean_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_mean_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_mean_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_prod_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_prod_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_prod_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_prod_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_prod_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_prod_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_prod_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_prod_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_prod_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_sum_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_sum_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_sum_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_sum_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_sum_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_sum_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_sum_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_sum_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_sum_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_scatter_reduce_sum_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_searchsorted_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_searchsorted_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_searchsorted_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_searchsorted_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_searchsorted_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_searchsorted_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_searchsorted_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_searchsorted_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_searchsorted_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_scatter_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_scatter_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_scatter_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_scatter_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_scatter_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_scatter_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_scatter_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_scatter_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_scatter_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_select_scatter_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sgn_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sgn_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sgn_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sgn_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sgn_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sgn_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sgn_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sgn_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sgn_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sgn_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sgn_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sgn_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sgn_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_short_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_short_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_short_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_short_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_short_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_short_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_short_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_short_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_short_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_short_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_short_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_short_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sigmoid_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sigmoid_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sigmoid_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sigmoid_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sigmoid_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sigmoid_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sigmoid_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sigmoid_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sigmoid_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sigmoid_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sigmoid_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sigmoid_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sigmoid_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sign_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sign_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sign_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sign_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sign_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sign_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sign_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sign_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sign_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sign_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_bartlett_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_bartlett_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_blackman_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_blackman_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_cosine_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_cosine_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_exponential_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_exponential_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_gaussian_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_gaussian_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_general_cosine_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_general_cosine_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_general_hamming_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_general_hamming_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_hamming_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_hamming_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_hann_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_hann_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_kaiser_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_kaiser_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_nuttall_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signal_windows_nuttall_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signbit_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signbit_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signbit_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signbit_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signbit_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signbit_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signbit_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signbit_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signbit_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_signbit_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sin_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sin_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sin_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sin_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sin_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sin_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sin_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sin_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sin_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sin_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sin_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sin_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sin_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinc_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinc_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinc_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinc_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinc_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinc_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinc_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinc_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinc_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinc_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinc_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinc_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinh_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinh_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinh_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinh_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinh_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinh_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinh_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinh_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinh_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinh_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinh_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinh_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sinh_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_scatter_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_scatter_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_scatter_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_scatter_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_scatter_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_scatter_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_scatter_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_scatter_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_scatter_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_slice_scatter_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_softmax_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_softmax_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_softmax_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_softmax_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_softmax_with_dtype_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_softmax_with_dtype_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_softmax_with_dtype_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_softmax_with_dtype_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_softmax_with_dtype_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_softmax_with_dtype_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_softmax_with_dtype_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_softmax_with_dtype_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_softmax_with_dtype_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_softmax_with_dtype_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_softmax_with_dtype_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_softmax_with_dtype_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sort_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sort_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sort_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sort_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sort_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sort_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sort_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sort_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sort_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sort_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sparse_mm_reduce_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sparse_mm_reduce_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sparse_mm_reduce_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sparse_mm_reduce_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sparse_sampled_addmm_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sparse_sampled_addmm_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sparse_sampled_addmm_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sparse_sampled_addmm_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_airy_ai_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_airy_ai_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_airy_ai_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_airy_ai_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_airy_ai_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_airy_ai_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_airy_ai_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_airy_ai_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_j0_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_j0_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_j0_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_j0_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_j0_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_j0_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_j0_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_j0_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_j1_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_j1_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_j1_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_j1_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_j1_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_j1_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_j1_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_j1_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_y0_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_y0_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_y0_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_y0_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_y0_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_y0_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_y0_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_y0_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_y1_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_y1_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_y1_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_y1_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_y1_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_y1_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_y1_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_bessel_y1_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_t_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_t_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_t_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_t_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_t_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_t_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_t_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_t_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_u_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_u_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_u_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_u_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_u_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_u_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_u_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_u_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_v_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_v_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_v_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_v_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_v_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_v_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_v_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_v_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_w_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_w_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_w_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_w_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_w_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_w_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_w_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_chebyshev_polynomial_w_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_entr_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_entr_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_entr_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_entr_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_entr_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_entr_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_entr_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_entr_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_entr_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_entr_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_erfcx_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_erfcx_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_erfcx_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_erfcx_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_erfcx_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_erfcx_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_erfcx_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_erfcx_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_hermite_polynomial_h_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_hermite_polynomial_h_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_hermite_polynomial_h_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_hermite_polynomial_h_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_hermite_polynomial_h_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_hermite_polynomial_h_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_hermite_polynomial_h_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_hermite_polynomial_h_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_hermite_polynomial_he_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_hermite_polynomial_he_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_hermite_polynomial_he_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_hermite_polynomial_he_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_hermite_polynomial_he_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_hermite_polynomial_he_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_hermite_polynomial_he_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_hermite_polynomial_he_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i0e_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i0e_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i0e_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i0e_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i0e_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i0e_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i0e_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i0e_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i0e_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i0e_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1e_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1e_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1e_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1e_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1e_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1e_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1e_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1e_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1e_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_i1e_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_laguerre_polynomial_l_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_laguerre_polynomial_l_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_laguerre_polynomial_l_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_laguerre_polynomial_l_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_laguerre_polynomial_l_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_laguerre_polynomial_l_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_laguerre_polynomial_l_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_laguerre_polynomial_l_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_legendre_polynomial_p_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_legendre_polynomial_p_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_legendre_polynomial_p_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_legendre_polynomial_p_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_legendre_polynomial_p_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_legendre_polynomial_p_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_legendre_polynomial_p_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_legendre_polynomial_p_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_log_ndtr_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_log_ndtr_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_log_ndtr_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_log_ndtr_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_log_ndtr_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_log_ndtr_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_log_ndtr_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_log_ndtr_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_i0_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_i0_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_i0_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_i0_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_i0_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_i0_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_i0_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_i0_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_i1_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_i1_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_i1_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_i1_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_i1_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_i1_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_i1_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_i1_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_k0_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_k0_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_k0_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_k0_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_k0_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_k0_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_k0_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_k0_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_k1_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_k1_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_k1_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_k1_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_k1_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_k1_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_k1_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_modified_bessel_k1_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_ndtr_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_ndtr_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_ndtr_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_ndtr_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_ndtr_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_ndtr_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_ndtr_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_ndtr_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_ndtr_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_ndtr_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_ndtri_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_ndtri_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_ndtri_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_ndtri_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_ndtri_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_ndtri_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_ndtri_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_ndtri_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_polygamma_special_polygamma_n_0_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_polygamma_special_polygamma_n_0_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_polygamma_special_polygamma_n_0_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_polygamma_special_polygamma_n_0_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_polygamma_special_polygamma_n_0_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_polygamma_special_polygamma_n_0_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_polygamma_special_polygamma_n_0_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_polygamma_special_polygamma_n_0_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_polygamma_special_polygamma_n_0_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_polygamma_special_polygamma_n_0_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_scaled_modified_bessel_k0_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_scaled_modified_bessel_k0_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_scaled_modified_bessel_k0_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_scaled_modified_bessel_k0_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_scaled_modified_bessel_k0_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_scaled_modified_bessel_k0_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_scaled_modified_bessel_k0_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_scaled_modified_bessel_k0_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_scaled_modified_bessel_k1_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_scaled_modified_bessel_k1_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_scaled_modified_bessel_k1_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_scaled_modified_bessel_k1_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_scaled_modified_bessel_k1_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_scaled_modified_bessel_k1_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_scaled_modified_bessel_k1_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_scaled_modified_bessel_k1_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_t_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_t_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_t_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_t_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_t_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_t_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_t_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_t_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_u_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_u_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_u_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_u_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_u_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_u_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_u_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_u_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_v_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_v_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_v_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_v_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_v_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_v_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_v_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_v_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_w_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_w_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_w_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_w_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_w_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_w_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_w_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_shifted_chebyshev_polynomial_w_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_spherical_bessel_j0_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_spherical_bessel_j0_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_spherical_bessel_j0_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_spherical_bessel_j0_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_spherical_bessel_j0_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_spherical_bessel_j0_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_spherical_bessel_j0_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_spherical_bessel_j0_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_xlog1py_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_xlog1py_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_xlog1py_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_xlog1py_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_xlog1py_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_xlog1py_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_xlog1py_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_xlog1py_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_xlog1py_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_xlog1py_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_zeta_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_zeta_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_zeta_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_zeta_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_zeta_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_zeta_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_zeta_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_special_zeta_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_list_args_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_list_args_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_list_args_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_list_args_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_list_args_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_list_args_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_list_args_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_list_args_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_list_args_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_list_args_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_list_args_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_list_args_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_copy_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_copy_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_copy_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_copy_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_copy_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_copy_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_copy_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_copy_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_copy_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_copy_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_copy_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_copy_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_copy_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_split_with_sizes_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sqrt_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sqrt_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sqrt_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sqrt_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sqrt_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sqrt_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sqrt_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sqrt_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sqrt_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sqrt_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sqrt_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sqrt_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sqrt_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_square_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_square_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_square_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_square_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_square_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_square_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_square_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_square_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_square_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_square_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_square_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_square_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_copy_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_copy_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_copy_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_copy_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_copy_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_copy_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_copy_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_copy_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_copy_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_copy_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_copy_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_copy_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_copy_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_multiple_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_multiple_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_multiple_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_multiple_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_multiple_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_multiple_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_multiple_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_multiple_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_multiple_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_multiple_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_multiple_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_multiple_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_squeeze_multiple_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_stack_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_stack_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_stack_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_stack_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_stack_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_stack_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_stack_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_stack_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_stack_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_stack_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_stack_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_stack_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_stack_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_mean_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_mean_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_mean_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_mean_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_mean_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_mean_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_mean_unbiased_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_mean_unbiased_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_mean_unbiased_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_mean_unbiased_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_mean_unbiased_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_mean_unbiased_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_unbiased_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_unbiased_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_unbiased_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_unbiased_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_unbiased_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_std_unbiased_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_stft_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_stft_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_stft_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_stft_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sub_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sub_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sub_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sub_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sub_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sub_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sub_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sub_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sub_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sub_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sub_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sub_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_to_size_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_to_size_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_to_size_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_to_size_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_to_size_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_to_size_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_to_size_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_to_size_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_to_size_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_to_size_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_to_size_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_sum_to_size_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_svd_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_svd_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_svd_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_svd_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_svd_lowrank_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_svd_lowrank_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_svd_lowrank_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_svd_lowrank_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_copy_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_copy_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_copy_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_copy_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_copy_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_copy_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_copy_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_copy_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_copy_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_copy_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_copy_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_copy_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_t_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_along_dim_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_along_dim_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_along_dim_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_along_dim_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_along_dim_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_along_dim_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_along_dim_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_along_dim_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_along_dim_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_along_dim_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_along_dim_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_along_dim_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_take_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tan_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tan_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tan_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tan_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tan_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tan_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tan_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tan_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tan_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tan_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tan_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tan_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tan_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tanh_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tanh_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tanh_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tanh_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tanh_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tanh_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tanh_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tanh_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tanh_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tanh_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tanh_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tanh_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tanh_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tensor_split_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tensor_split_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tensor_split_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tensor_split_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tensor_split_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tensor_split_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tensor_split_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tensor_split_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tensor_split_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tensor_split_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tensor_split_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tensor_split_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tensordot_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tensordot_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tensordot_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tensordot_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tensordot_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tensordot_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tile_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tile_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tile_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tile_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tile_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tile_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tile_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tile_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tile_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tile_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tile_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tile_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_sparse_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_sparse_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_sparse_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_sparse_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_sparse_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_sparse_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_sparse_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_sparse_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_sparse_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_sparse_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_sparse_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_to_sparse_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_topk_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_topk_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_topk_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_topk_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_topk_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_topk_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_topk_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_topk_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_topk_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_torch__scaled_mm_cuda_float8_e4m3fn, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_torch__scaled_mm_v2_cuda_float8_e4m3fn, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_torch_ops_aten__efficient_attention_forward_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_torch_ops_aten__efficient_attention_forward_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_torch_ops_aten__flash_attention_forward_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_torch_ops_aten__safe_softmax_default_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_torch_ops_aten__safe_softmax_default_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_torch_ops_aten__safe_softmax_default_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_torch_ops_aten__safe_softmax_default_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_torch_ops_aten__safe_softmax_default_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_torch_ops_aten__safe_softmax_default_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_torch_ops_aten__safe_softmax_default_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_torch_ops_aten__safe_softmax_default_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_torch_ops_aten__safe_softmax_default_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_torch_ops_aten__safe_softmax_default_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trace_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trace_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trace_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trace_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trace_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trace_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trace_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trace_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trace_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trace_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trace_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trace_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trace_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_copy_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_copy_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_copy_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_copy_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_copy_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_copy_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_copy_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_copy_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_copy_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_copy_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_copy_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_copy_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_copy_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_transpose_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapezoid_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapezoid_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapezoid_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapezoid_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapezoid_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapezoid_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapezoid_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapezoid_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapezoid_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapezoid_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapezoid_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapz_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapz_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapz_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapz_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapz_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapz_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapz_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapz_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapz_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapz_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trapz_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_triangular_solve_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_triangular_solve_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_triangular_solve_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_triangular_solve_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tril_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tril_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tril_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tril_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tril_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tril_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tril_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tril_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tril_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tril_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tril_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tril_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tril_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tril_indices_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_tril_indices_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_triu_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_triu_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_triu_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_triu_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_triu_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_triu_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_triu_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_triu_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_triu_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_triu_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_triu_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_triu_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_triu_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_triu_indices_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_triu_indices_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_true_divide_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_true_divide_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_true_divide_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_true_divide_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_true_divide_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_true_divide_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_true_divide_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_true_divide_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_true_divide_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_true_divide_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_true_divide_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_true_divide_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_true_divide_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trunc_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trunc_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trunc_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trunc_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trunc_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trunc_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trunc_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trunc_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_trunc_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_copy_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_copy_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_copy_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_copy_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_copy_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_copy_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_copy_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_copy_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_copy_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_copy_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_copy_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_copy_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_copy_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unbind_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unflatten_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unflatten_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unflatten_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unflatten_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unflatten_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unflatten_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unflatten_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unflatten_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unflatten_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unflatten_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unflatten_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unflatten_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unflatten_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_copy_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_copy_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_copy_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_copy_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_copy_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_copy_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_copy_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_copy_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_copy_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_copy_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_copy_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_copy_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_copy_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unfold_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_uniform_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_uniform_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_uniform_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_uniform_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_uniform_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_uniform_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_consecutive_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_consecutive_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_consecutive_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_consecutive_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_consecutive_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_consecutive_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_consecutive_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_consecutive_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_consecutive_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_consecutive_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_cuda_uint16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_cuda_uint32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_cuda_uint64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unique_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unravel_index_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unravel_index_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unravel_index_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unravel_index_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unravel_index_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_chunk_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_chunk_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_chunk_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_chunk_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_chunk_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_chunk_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_chunk_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_chunk_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_chunk_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_chunk_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_chunk_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_chunk_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_chunk_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_split_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_split_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_split_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_split_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_split_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_split_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_split_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_split_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_split_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_split_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_split_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_split_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsafe_split_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_copy_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_copy_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_copy_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_copy_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_copy_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_copy_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_copy_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_copy_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_copy_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_copy_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_copy_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_copy_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_copy_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_unsqueeze_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_mean_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_mean_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_mean_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_mean_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_mean_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_mean_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_mean_unbiased_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_mean_unbiased_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_mean_unbiased_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_mean_unbiased_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_mean_unbiased_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_mean_unbiased_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_unbiased_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_unbiased_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_unbiased_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_unbiased_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_unbiased_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_var_unbiased_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vdot_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vdot_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vdot_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vdot_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vdot_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vdot_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_as_complex_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_as_complex_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_as_complex_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_as_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_as_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_as_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_as_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_as_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_as_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_as_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_as_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_as_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_as_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_as_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_as_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_as_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_as_real_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_as_real_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_copy_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_copy_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_copy_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_copy_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_copy_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_copy_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_copy_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_copy_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_copy_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_copy_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_copy_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_copy_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_view_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vsplit_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vsplit_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vsplit_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vsplit_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vsplit_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vsplit_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vsplit_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vsplit_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vsplit_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vsplit_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vsplit_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vsplit_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vsplit_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vstack_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vstack_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vstack_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vstack_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vstack_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vstack_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vstack_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vstack_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vstack_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vstack_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vstack_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vstack_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_vstack_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_where_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_where_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_where_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_where_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_where_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_where_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_where_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_where_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_where_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_where_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_where_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_where_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_where_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_xlogy_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_xlogy_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_xlogy_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_xlogy_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_xlogy_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_xlogy_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_xlogy_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_xlogy_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_xlogy_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_xlogy_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zero__cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zero__cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zero__cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zero__cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zero__cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zero__cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zero__cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zero__cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zero__cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zero__cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zero__cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zero__cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_like_cuda_bfloat16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_like_cuda_bool, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_like_cuda_complex128, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_like_cuda_complex32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_like_cuda_complex64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_like_cuda_float16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_like_cuda_float32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_like_cuda_float64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_like_cuda_int16, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_like_cuda_int32, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_like_cuda_int64, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_like_cuda_int8, test/test_utils.py::TestDeviceUtilsCUDA::test_device_mode_ops_zeros_like_cuda_uint8, test/test_utils.py::TestDeviceUtilsCUDA::test_get_default_device_cuda, test/test_utils.py::TestDeviceUtilsCUDA::test_get_default_device_more_cuda, test/test_utils.py::TestDeviceUtilsCUDA::test_nn_module_cuda, test/test_utils.py::TestDeviceUtilsCUDA::test_set_default_device_cuda, test/test_utils.py::TestCppExtensionUtils::test_cc_compiler_is_ok, test/test_utils.py::TestCppExtensionUtils::test_cpp_compiler_is_ok, test/test_utils.py::TestTraceback::test_basic, test/test_utils.py::TestTraceback::test_captured_traceback, test/test_utils.py::TestTraceback::test_captured_traceback_format_all, test/test_utils.py::TestTraceback::test_captured_traceback_format_all_cached, test/test_utils.py::TestTraceback::test_format_traceback_short, test/test_utils.py::TestTryImport::test_import_existing, test/test_utils.py::TestTryImport::test_import_imported, test/test_utils.py::TestTryImport::test_import_missing, test/test_utils.py::TestDeprecate::test_deprecated 2025-12-04T16:45:24.8362131Z 2025-12-04T16:45:24.8362471Z Finished test_utils 1/1 ... [2025-12-04 16:45:24.333624][27108.024010596], took 0.89min 2025-12-04T16:45:24.8363558Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_utils/test_utils-cb6667809c56b359.xml 2025-12-04T16:45:24.8364559Z Running test_pytree 1/1 ... [2025-12-04 16:45:24.622698][27108.313087324] 2025-12-04T16:45:24.8365056Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:45:24.8366610Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_pytree.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:45:24.623169] 2025-12-04T16:45:32.3488537Z 2025-12-04T16:45:32.3489458Z test_pytree 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_pytree_1.1_52fd485987cd9dff_.log 2025-12-04T16:45:32.3524660Z Running 100 items in this shard: test/test_pytree.py::TestGenericPytree::test_aligned_public_apis, test/test_pytree.py::TestGenericPytree::test_broadcast_to_and_flatten_cxx, test/test_pytree.py::TestGenericPytree::test_broadcast_to_and_flatten_python, test/test_pytree.py::TestGenericPytree::test_enum_treespec_roundtrip_cxx, test/test_pytree.py::TestGenericPytree::test_enum_treespec_roundtrip_python, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_defaultdict_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_defaultdict_python, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_deque_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_deque_python, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_dict_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_dict_python, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_leaf_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_leaf_python, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_list_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_list_python, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_namedtuple_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_namedtuple_python, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_nested_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_nested_python, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_ordereddict_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_ordereddict_python, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_return_types_max_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_return_types_max_python, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_return_types_min_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_return_types_min_python, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_tuple_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_tuple_python, test/test_pytree.py::TestGenericPytree::test_flatten_with_is_leaf_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_with_is_leaf_python, test/test_pytree.py::TestGenericPytree::test_is_namedtuple_cxx, test/test_pytree.py::TestGenericPytree::test_is_namedtuple_python, test/test_pytree.py::TestGenericPytree::test_is_structseq_cxx, test/test_pytree.py::TestGenericPytree::test_is_structseq_python, test/test_pytree.py::TestGenericPytree::test_pytree_serialize_bad_input_cxx, test/test_pytree.py::TestGenericPytree::test_pytree_serialize_bad_input_python, test/test_pytree.py::TestGenericPytree::test_register_pytree_node_cxx, test/test_pytree.py::TestGenericPytree::test_register_pytree_node_python, test/test_pytree.py::TestGenericPytree::test_tree_all_any_cxx, test/test_pytree.py::TestGenericPytree::test_tree_all_any_python, test/test_pytree.py::TestGenericPytree::test_tree_map_cxx, test/test_pytree.py::TestGenericPytree::test_tree_map_dict_order_cxx, test/test_pytree.py::TestGenericPytree::test_tree_map_dict_order_python, test/test_pytree.py::TestGenericPytree::test_tree_map_multi_inputs_cxx, test/test_pytree.py::TestGenericPytree::test_tree_map_multi_inputs_python, test/test_pytree.py::TestGenericPytree::test_tree_map_only_cxx, test/test_pytree.py::TestGenericPytree::test_tree_map_only_predicate_fn_cxx, test/test_pytree.py::TestGenericPytree::test_tree_map_only_predicate_fn_python, test/test_pytree.py::TestGenericPytree::test_tree_map_only_python, test/test_pytree.py::TestGenericPytree::test_tree_map_python, test/test_pytree.py::TestPythonPytree::test_constant, test/test_pytree.py::TestPythonPytree::test_constant_default_eq_error, test/test_pytree.py::TestPythonPytree::test_constant_default_hash_error, test/test_pytree.py::TestPythonPytree::test_dataclass, test/test_pytree.py::TestPythonPytree::test_deprecated_register_pytree_node, test/test_pytree.py::TestPythonPytree::test_flatten_flatten_with_key_consistency, test/test_pytree.py::TestPythonPytree::test_import_pytree_doesnt_import_optree, test/test_pytree.py::TestPythonPytree::test_key_access, test/test_pytree.py::TestPythonPytree::test_key_str, test/test_pytree.py::TestPythonPytree::test_pytree_context_serialize_bad, test/test_pytree.py::TestPythonPytree::test_pytree_custom_type_serialize, test/test_pytree.py::TestPythonPytree::test_pytree_custom_type_serialize_bad, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_bad_protocol, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_defaultdict_enum, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_enum, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_namedtuple, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_namedtuple_bad, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_register_bad, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec0, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec1, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec2, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec3, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec4, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec5, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec6, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec7, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec8, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec9, test/test_pytree.py::TestPythonPytree::test_register_dataclass_class, test/test_pytree.py::TestPythonPytree::test_saved_serialized, test/test_pytree.py::TestPythonPytree::test_tree_flatten_with_path_is_leaf, test/test_pytree.py::TestPythonPytree::test_tree_flatten_with_path_roundtrip, test/test_pytree.py::TestPythonPytree::test_tree_leaves_with_path, test/test_pytree.py::TestPythonPytree::test_tree_map_with_path, test/test_pytree.py::TestPythonPytree::test_tree_map_with_path_multiple_trees, test/test_pytree.py::TestPythonPytree::test_treespec_equality, test/test_pytree.py::TestPythonPytree::test_treespec_repr, test/test_pytree.py::TestCxxPytree::test_pytree_custom_type_serialize, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_namedtuple, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec0, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec1, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec2, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec3, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec4, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec5, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec6, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec7, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec8, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec9, test/test_pytree.py::TestCxxPytree::test_treespec_equality, test/test_pytree.py::TestCxxPytree::test_treespec_repr 2025-12-04T16:45:32.3559157Z 2025-12-04T16:45:32.3559453Z Finished test_pytree 1/1 ... [2025-12-04 16:45:32.348806][27116.039198727], took 0.13min 2025-12-04T16:45:32.3997735Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_pytree/test_pytree-b5c87da58bd7db26.xml 2025-12-04T16:45:32.4770792Z Running test_namedtuple_return_api 1/1 ... [2025-12-04 16:45:32.476730][27116.167121337] 2025-12-04T16:45:32.4771414Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:45:32.4774711Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_namedtuple_return_api.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:45:32.477182] 2025-12-04T16:45:39.6014578Z 2025-12-04T16:45:39.6015630Z test_namedtuple_return_api 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_namedtuple_return_api_1.1_ed47513e0b44d1fb_.log 2025-12-04T16:45:39.6017787Z Running 3 items in this shard: test/test_namedtuple_return_api.py::TestNamedTupleAPI::test_import_return_types, test/test_namedtuple_return_api.py::TestNamedTupleAPI::test_namedtuple_return, test/test_namedtuple_return_api.py::TestNamedTupleAPI::test_native_functions_yaml 2025-12-04T16:45:39.6019193Z 2025-12-04T16:45:39.6019551Z Finished test_namedtuple_return_api 1/1 ... [2025-12-04 16:45:39.601253][27123.291647512], took 0.12min 2025-12-04T16:45:39.6520268Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_namedtuple_return_api/test_namedtuple_return_api-9fcfafd59257c871.xml 2025-12-04T16:45:39.7307032Z Running profiler/test_record_function 1/1 ... [2025-12-04 16:45:39.730404][27123.420794734] 2025-12-04T16:45:39.7307641Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:45:39.7311009Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'profiler/test_record_function.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:45:39.730842] 2025-12-04T16:45:45.1529326Z 2025-12-04T16:45:45.1530437Z profiler/test_record_function 1/1 was successful, full logs can be found in artifacts with path test/test-reports/profiler.test_record_function_1.1_ca949195f6b456e2_.log 2025-12-04T16:45:45.1534279Z Running 6 items in this shard: test/profiler/test_record_function.py::TestRecordFunction::test_datapipe_delegation_with_profiler, test/profiler/test_record_function.py::TestRecordFunction::test_datapipe_with_record_function, test/profiler/test_record_function.py::TestRecordFunction::test_datapipe_with_record_function_fork, test/profiler/test_record_function.py::TestRecordFunction::test_python_dispatch_mode_record_function, test/profiler/test_record_function.py::TestRecordFunction::test_python_subclass_record_function, test/profiler/test_record_function.py::TestRecordFunction::test_record_function 2025-12-04T16:45:45.1537427Z 2025-12-04T16:45:45.1537815Z Finished profiler/test_record_function 1/1 ... [2025-12-04 16:45:45.152703][27128.843097206], took 0.09min 2025-12-04T16:45:45.2035168Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/profiler.test_record_function/profiler.test_record_function-5b633e88461adafa.xml 2025-12-04T16:45:45.2347342Z Running test_compile_benchmark_util 1/1 ... [2025-12-04 16:45:45.234484][27128.924875754] 2025-12-04T16:45:45.2347937Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:45:45.2351365Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_compile_benchmark_util.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:45:45.234890] 2025-12-04T16:46:00.8220821Z 2025-12-04T16:46:00.8221904Z test_compile_benchmark_util 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_compile_benchmark_util_1.1_fb69b29a06036afc_.log 2025-12-04T16:46:00.8223628Z Running 1 items in this shard: test/test_compile_benchmark_util.py::TestCompileBenchmarkUtil::test_training_and_inference 2025-12-04T16:46:00.8224252Z 2025-12-04T16:46:00.8224636Z Finished test_compile_benchmark_util 1/1 ... [2025-12-04 16:46:00.821893][27144.512284829], took 0.26min 2025-12-04T16:46:00.8733867Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_compile_benchmark_util/test_compile_benchmark_util-df68b65dcac6dfa0.xml 2025-12-04T16:46:00.9546802Z Running test_set_default_mobile_cpu_allocator 1/1 ... [2025-12-04 16:46:00.954341][27144.644731582] 2025-12-04T16:46:00.9547653Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:46:00.9550551Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_set_default_mobile_cpu_allocator.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:46:00.954789] 2025-12-04T16:46:06.3268572Z 2025-12-04T16:46:06.3269748Z test_set_default_mobile_cpu_allocator 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_set_default_mobile_cpu_allocator_1.1_a92884857a29fe5f_.log 2025-12-04T16:46:06.3271752Z Running 2 items in this shard: test/test_set_default_mobile_cpu_allocator.py::TestSetDefaultMobileCPUAllocator::test_exception, test/test_set_default_mobile_cpu_allocator.py::TestSetDefaultMobileCPUAllocator::test_no_exception 2025-12-04T16:46:06.3272950Z 2025-12-04T16:46:06.3273362Z Finished test_set_default_mobile_cpu_allocator 1/1 ... [2025-12-04 16:46:06.326672][27150.017064362], took 0.09min 2025-12-04T16:46:06.3778335Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_set_default_mobile_cpu_allocator/test_set_default_mobile_cpu_allocator-e4da27e112b70d28.xml 2025-12-04T16:46:06.4146259Z Running test_fake_tensor 1/1 ... [2025-12-04 16:46:06.414369][27150.104760668] 2025-12-04T16:46:06.4146787Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:46:06.4150416Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_fake_tensor.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:46:06.414794] 2025-12-04T16:46:46.3861797Z 2025-12-04T16:46:46.3863300Z test_fake_tensor 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_fake_tensor_1.1_51e2ac03aee772c5_.log 2025-12-04T16:46:46.3998845Z Running 288 items in this shard: test/test_fake_tensor.py::FakeTensorTest::test__adaptive_avg_pool2d_backward, test/test_fake_tensor.py::FakeTensorTest::test_alias_call, test/test_fake_tensor.py::FakeTensorTest::test_allow_meta, test/test_fake_tensor.py::FakeTensorTest::test_aten_copy_multi_device, test/test_fake_tensor.py::FakeTensorTest::test_aten_index_multi_device, test/test_fake_tensor.py::FakeTensorTest::test_aten_slice_scatter_multi_device, test/test_fake_tensor.py::FakeTensorTest::test_basic, test/test_fake_tensor.py::FakeTensorTest::test_batch_tensor, test/test_fake_tensor.py::FakeTensorTest::test_binary_op_type_promotion, test/test_fake_tensor.py::FakeTensorTest::test_constructor, test/test_fake_tensor.py::FakeTensorTest::test_conv_nhwc, test/test_fake_tensor.py::FakeTensorTest::test_convert_fake_to_real, test/test_fake_tensor.py::FakeTensorTest::test_cpu_fallback, test/test_fake_tensor.py::FakeTensorTest::test_cuda_initialized, test/test_fake_tensor.py::FakeTensorTest::test_cuda_lstm, test/test_fake_tensor.py::FakeTensorTest::test_cudnn_rnn_with_fallback, test/test_fake_tensor.py::FakeTensorTest::test_cudnn_rnn_without_fallback, test/test_fake_tensor.py::FakeTensorTest::test_custom_op_fallback, test/test_fake_tensor.py::FakeTensorTest::test_data_dependent_operator, test/test_fake_tensor.py::FakeTensorTest::test_deepcopy, test/test_fake_tensor.py::FakeTensorTest::test_device_inplace_copy, test/test_fake_tensor.py::FakeTensorTest::test_embedding_bag_meta, test/test_fake_tensor.py::FakeTensorTest::test_export_numpy, test/test_fake_tensor.py::FakeTensorTest::test_fake_device, test/test_fake_tensor.py::FakeTensorTest::test_fake_dispatch_keys, test/test_fake_tensor.py::FakeTensorTest::test_fake_grad_copy, test/test_fake_tensor.py::FakeTensorTest::test_fake_mode_error, test/test_fake_tensor.py::FakeTensorTest::test_fast_div, test/test_fake_tensor.py::FakeTensorTest::test_fast_div_int_to_float, test/test_fake_tensor.py::FakeTensorTest::test_from_numpy, test/test_fake_tensor.py::FakeTensorTest::test_fsdp_flat_param, test/test_fake_tensor.py::FakeTensorTest::test_full, test/test_fake_tensor.py::FakeTensorTest::test_index_cuda_with_cpu_complex128, test/test_fake_tensor.py::FakeTensorTest::test_index_cuda_with_cpu_complex64, test/test_fake_tensor.py::FakeTensorTest::test_index_cuda_with_cpu_float32, test/test_fake_tensor.py::FakeTensorTest::test_index_cuda_with_cpu_float64, test/test_fake_tensor.py::FakeTensorTest::test_index_cuda_with_cpu_float8_e4m3fn, test/test_fake_tensor.py::FakeTensorTest::test_index_cuda_with_cpu_float8_e4m3fnuz, test/test_fake_tensor.py::FakeTensorTest::test_index_cuda_with_cpu_float8_e5m2, test/test_fake_tensor.py::FakeTensorTest::test_index_cuda_with_cpu_float8_e5m2fnuz, test/test_fake_tensor.py::FakeTensorTest::test_index_cuda_with_cpu_int16, test/test_fake_tensor.py::FakeTensorTest::test_index_cuda_with_cpu_int32, test/test_fake_tensor.py::FakeTensorTest::test_index_cuda_with_cpu_int64, test/test_fake_tensor.py::FakeTensorTest::test_index_cuda_with_cpu_int8, test/test_fake_tensor.py::FakeTensorTest::test_index_cuda_with_cpu_uint8, test/test_fake_tensor.py::FakeTensorTest::test_index_put_error, test/test_fake_tensor.py::FakeTensorTest::test_jagged_fake_to_fake_preserved, test/test_fake_tensor.py::FakeTensorTest::test_like_constructor, test/test_fake_tensor.py::FakeTensorTest::test_mixed_real_and_fake_inputs, test/test_fake_tensor.py::FakeTensorTest::test_mode, test/test_fake_tensor.py::FakeTensorTest::test_nan_to_num, test/test_fake_tensor.py::FakeTensorTest::test_nanmean_out, test/test_fake_tensor.py::FakeTensorTest::test_new, test/test_fake_tensor.py::FakeTensorTest::test_no_tag_func, test/test_fake_tensor.py::FakeTensorTest::test_non_kwarg_device, test/test_fake_tensor.py::FakeTensorTest::test_non_overlapping_stride_zero, test/test_fake_tensor.py::FakeTensorTest::test_non_parameter_grad, test/test_fake_tensor.py::FakeTensorTest::test_normalize_device, test/test_fake_tensor.py::FakeTensorTest::test_op_with_zero_dim_bypassed, test/test_fake_tensor.py::FakeTensorTest::test_out_multi_device, test/test_fake_tensor.py::FakeTensorTest::test_parameter_instantiation, test/test_fake_tensor.py::FakeTensorTest::test_parameter_view, test/test_fake_tensor.py::FakeTensorTest::test_print_in_fake_mode, test/test_fake_tensor.py::FakeTensorTest::test_randperm, test/test_fake_tensor.py::FakeTensorTest::test_recursive_invocation, test/test_fake_tensor.py::FakeTensorTest::test_repr, test/test_fake_tensor.py::FakeTensorTest::test_same_shape_env_preserved, test/test_fake_tensor.py::FakeTensorTest::test_scalar_inputs, test/test_fake_tensor.py::FakeTensorTest::test_scan_reverse_False, test/test_fake_tensor.py::FakeTensorTest::test_scan_reverse_True, test/test_fake_tensor.py::FakeTensorTest::test_setitem, test/test_fake_tensor.py::FakeTensorTest::test_shape_take_not_device, test/test_fake_tensor.py::FakeTensorTest::test_split_return_self, test/test_fake_tensor.py::FakeTensorTest::test_throw, test/test_fake_tensor.py::FakeTensorTest::test_tolist, test/test_fake_tensor.py::FakeTensorTest::test_type_as, test/test_fake_tensor.py::FakeTensorTest::test_unbind_copy_out, test/test_fake_tensor.py::FakeTensorTest::test_unsqueeze_copy, test/test_fake_tensor.py::FakeTensorTest::test_upsample_bilinear_small_channels, test/test_fake_tensor.py::FakeTensorTest::test_zero_dim, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test__adaptive_avg_pool2d_backward_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_alias_call_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_allow_meta_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_aten_copy_multi_device_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_aten_index_multi_device_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_aten_slice_scatter_multi_device_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_basic_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_batch_tensor_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_binary_op_type_promotion_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_constructor_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_conv_nhwc_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_convert_fake_to_real_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_cpu_fallback_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_cuda_initialized_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_cuda_lstm_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_cudnn_rnn_with_fallback_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_cudnn_rnn_without_fallback_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_custom_op_fallback_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_data_dependent_operator_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_deepcopy_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_device_inplace_copy_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_embedding_bag_meta_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_export_numpy_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_fake_device_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_fake_dispatch_keys_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_fake_grad_copy_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_fake_mode_error_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_fast_div_int_to_float_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_fast_div_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_from_numpy_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_fsdp_flat_param_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_full_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_index_cuda_with_cpu_complex128_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_index_cuda_with_cpu_complex64_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_index_cuda_with_cpu_float32_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_index_cuda_with_cpu_float64_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_index_cuda_with_cpu_float8_e4m3fn_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_index_cuda_with_cpu_float8_e4m3fnuz_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_index_cuda_with_cpu_float8_e5m2_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_index_cuda_with_cpu_float8_e5m2fnuz_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_index_cuda_with_cpu_int16_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_index_cuda_with_cpu_int32_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_index_cuda_with_cpu_int64_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_index_cuda_with_cpu_int8_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_index_cuda_with_cpu_uint8_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_index_put_error_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_jagged_fake_to_fake_preserved_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_like_constructor_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_mixed_real_and_fake_inputs_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_mode_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_nan_to_num_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_nanmean_out_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_new_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_no_tag_func_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_non_kwarg_device_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_non_overlapping_stride_zero_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_non_parameter_grad_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_normalize_device_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_op_with_zero_dim_bypassed_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_out_multi_device_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_parameter_instantiation_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_parameter_view_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_print_in_fake_mode_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_randperm_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_recursive_invocation_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_repr_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_same_shape_env_preserved_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_scalar_inputs_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_scan_reverse_False_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_scan_reverse_True_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_setitem_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_shape_take_not_device_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_split_return_self_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_throw_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_tolist_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_type_as_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_unbind_copy_out_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_unsqueeze_copy_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_upsample_bilinear_small_channels_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorTest::test_zero_dim_propagate_real_tensors, test/test_fake_tensor.py::FakeTensorConstHandling::test_aliased_const_write, test/test_fake_tensor.py::FakeTensorConstHandling::test_constant_invalidation, test/test_fake_tensor.py::FakeTensorConstHandling::test_constant_propagate_through_functions, test/test_fake_tensor.py::FakeTensorConstHandling::test_fake_tensor_batch_norm_cpu, test/test_fake_tensor.py::FakeTensorConstHandling::test_fake_tensor_in_intlist_repro, test/test_fake_tensor.py::FakeTensorConstHandling::test_inplace_add, test/test_fake_tensor.py::FakeTensorConstHandling::test_inplace_view_invalidation, test/test_fake_tensor.py::FakeTensorConstHandling::test_shared_storage_invalidation, test/test_fake_tensor.py::FakeTensorConstHandling::test_shared_storages, test/test_fake_tensor.py::FakeTensorConstHandling::test_simple, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConstHandling::test_aliased_const_write_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConstHandling::test_constant_invalidation_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConstHandling::test_constant_propagate_through_functions_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConstHandling::test_fake_tensor_batch_norm_cpu_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConstHandling::test_fake_tensor_in_intlist_repro_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConstHandling::test_inplace_add_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConstHandling::test_inplace_view_invalidation_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConstHandling::test_shared_storage_invalidation_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConstHandling::test_shared_storages_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConstHandling::test_simple_propagate_real_tensors, test/test_fake_tensor.py::FakeTensorOpInfoTestCUDA::test_fake_NumpyCatCustomOp_cuda_float32, test/test_fake_tensor.py::FakeTensorOpInfoTestCUDA::test_fake_NumpyCubeCustomOp_cuda_float32, test/test_fake_tensor.py::FakeTensorOpInfoTestCUDA::test_fake_NumpyMulCustomOp_cuda_float32, test/test_fake_tensor.py::FakeTensorOpInfoTestCUDA::test_fake_NumpyMulScalarCustomOp_cuda_float32, test/test_fake_tensor.py::FakeTensorOpInfoTestCUDA::test_fake_NumpyNMSCustomOp_cuda_float32, test/test_fake_tensor.py::FakeTensorOpInfoTestCUDA::test_fake_NumpyNonzeroCustomOp_cuda_float32, test/test_fake_tensor.py::FakeTensorOpInfoTestCUDA::test_fake_NumpySortCustomOp_cuda_float32, test/test_fake_tensor.py::FakeTensorOpInfoTestCUDA::test_fake_NumpySplitCopyCustomOp_cuda_float32, test/test_fake_tensor.py::FakeTensorOpInfoTestCUDA::test_fake_NumpySplitCopyWithIntCustomOp_cuda_float32, test/test_fake_tensor.py::FakeTensorOpInfoTestCUDA::test_fake_NumpyTakeCustomOp_cuda_float32, test/test_fake_tensor.py::FakeTensorOpInfoTestCUDA::test_fake_NumpyViewCopyCustomOp_cuda_float32, test/test_fake_tensor.py::FakeTensorConverterTest::test_dead_key, test/test_fake_tensor.py::FakeTensorConverterTest::test_dead_weak_ref, test/test_fake_tensor.py::FakeTensorConverterTest::test_memoized_conversion_from_meta, test/test_fake_tensor.py::FakeTensorConverterTest::test_memoized_conversion_to_meta, test/test_fake_tensor.py::FakeTensorConverterTest::test_multiple_modes, test/test_fake_tensor.py::FakeTensorConverterTest::test_no_active_mode, test/test_fake_tensor.py::FakeTensorConverterTest::test_no_ref_cycle, test/test_fake_tensor.py::FakeTensorConverterTest::test_separate_mode_error, test/test_fake_tensor.py::FakeTensorConverterTest::test_separate_tensor_storages_non_view, test/test_fake_tensor.py::FakeTensorConverterTest::test_separate_tensor_storages_view, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConverterTest::test_dead_key_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConverterTest::test_dead_weak_ref_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConverterTest::test_memoized_conversion_from_meta_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConverterTest::test_memoized_conversion_to_meta_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConverterTest::test_multiple_modes_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConverterTest::test_no_active_mode_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConverterTest::test_no_ref_cycle_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConverterTest::test_separate_mode_error_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConverterTest::test_separate_tensor_storages_non_view_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorConverterTest::test_separate_tensor_storages_view_propagate_real_tensors, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_conv_c1_backward, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_cross_entropy_loss, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_embedding_bag_private, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_fake_gpu_no_init, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_flash_attention, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_like_ops, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_module_to, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_move_meta_tensor, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_move_module_under_fake, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_no_dispatch_with_like_function, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_non_kwarg_only_device, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_sparse_new, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_str_storage, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_tensor_constructors_all_have_kwarg_device, test/test_fake_tensor.py::FakeTensorOperatorInvariants::test_tensor_new, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorOperatorInvariants::test_conv_c1_backward_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorOperatorInvariants::test_cross_entropy_loss_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorOperatorInvariants::test_embedding_bag_private_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorOperatorInvariants::test_fake_gpu_no_init_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorOperatorInvariants::test_flash_attention_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorOperatorInvariants::test_like_ops_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorOperatorInvariants::test_module_to_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorOperatorInvariants::test_move_meta_tensor_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorOperatorInvariants::test_move_module_under_fake_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorOperatorInvariants::test_no_dispatch_with_like_function_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorOperatorInvariants::test_non_kwarg_only_device_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorOperatorInvariants::test_sparse_new_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorOperatorInvariants::test_str_storage_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorOperatorInvariants::test_tensor_constructors_all_have_kwarg_device_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorOperatorInvariants::test_tensor_new_propagate_real_tensors, test/test_fake_tensor.py::FakeTensorPropTest::test_fake_tensor_prop_on_nn_module, test/test_fake_tensor.py::FakeTensorPropTest::test_fake_tensor_prop_on_nn_module_with_optional_args, test/test_fake_tensor.py::FakeTensorPropTest::test_nan_to_num, test/test_fake_tensor.py::FakeTensorPropTest::test_nonzero_stride, test/test_fake_tensor.py::FakeTensorPropTest::test_torch_load_with_fake_mode, test/test_fake_tensor.py::FakeTensorPropTest::test_unbacked_shape_realloc, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorPropTest::test_fake_tensor_prop_on_nn_module_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorPropTest::test_fake_tensor_prop_on_nn_module_with_optional_args_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorPropTest::test_nan_to_num_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorPropTest::test_nonzero_stride_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorPropTest::test_torch_load_with_fake_mode_propagate_real_tensors, test/test_fake_tensor.py::PropagateRealTensorsFakeTensorPropTest::test_unbacked_shape_realloc_propagate_real_tensors, test/test_fake_tensor.py::FakeTensorSerialization::test_serialization, test/test_fake_tensor.py::FakeTensorSerialization::test_serialization_with_tracing, test/test_fake_tensor.py::FakeTensorDispatchCache::test__upsample_bilinear2d_aa_backward_dynamic_shapes, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_aten_index, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_bypass, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_default_device, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_default_dtype, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_dispatch_key_set, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_hit, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_inplace_op, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_key_constants, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_key_device, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_key_dtype, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_key_is_conj, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_key_is_inference, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_key_is_neg, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_key_memory_format, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_key_requires_grad, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_key_shape, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_key_storage_offset, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_key_stride, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_tuple_outputs, test/test_fake_tensor.py::FakeTensorDispatchCache::test_cache_view_op, test/test_fake_tensor.py::FakeTensorDispatchCache::test_empty_list, test/test_fake_tensor.py::FakeTensorDispatchCache::test_fft_hfft2_issue145522, test/test_fake_tensor.py::FakeTensorDispatchCache::test_from_buffer, test/test_fake_tensor.py::FakeTensorDispatchCache::test_inference_mode, test/test_fake_tensor.py::FakeTensorDispatchCache::test_invoke_subgraph, test/test_fake_tensor.py::FakeTensorDispatchCache::test_invoke_subgraph_cacheable_inplace, test/test_fake_tensor.py::FakeTensorDispatchCache::test_meta_tensor_to_fake_cpu, test/test_fake_tensor.py::FakeTensorDispatchCache::test_shape_env_settings, test/test_fake_tensor.py::FakeTensorDispatchCache::test_unbacked_output, test/test_fake_tensor.py::FakeTensorDispatchCache::test_wrapper_tensor_subclass_different_device, test/test_fake_tensor.py::FakeTensorPreferDeviceType::test_fake_tensor_prefer_device_type, test/test_fake_tensor.py::FakeTensorPreferDeviceType::test_fake_tensor_prefer_device_type_cpu_only 2025-12-04T16:46:46.4130015Z 2025-12-04T16:46:46.4130338Z Finished test_fake_tensor 1/1 ... [2025-12-04 16:46:46.386457][27190.076844908], took 0.67min 2025-12-04T16:46:46.4382111Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_fake_tensor/test_fake_tensor-03d6e27aa07cc9f8.xml 2025-12-04T16:46:46.5339285Z Running higher_order_ops/test_print 1/1 ... [2025-12-04 16:46:46.533568][27190.223959678] 2025-12-04T16:46:46.5340119Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:46:46.5343037Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'higher_order_ops/test_print.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:46:46.534031] 2025-12-04T16:46:54.8104307Z 2025-12-04T16:46:54.8105498Z higher_order_ops/test_print 1/1 was successful, full logs can be found in artifacts with path test/test-reports/higher_order_ops.test_print_1.1_da02828dbc3b3692_.log 2025-12-04T16:46:54.8110423Z Running 10 items in this shard: test/higher_order_ops/test_print.py::TestHopPrint::test_base_print, test/higher_order_ops/test_print.py::TestHopPrint::test_constant_mutation_backend_aot_eager, test/higher_order_ops/test_print.py::TestHopPrint::test_constant_mutation_backend_eager, test/higher_order_ops/test_print.py::TestHopPrint::test_para_print, test/higher_order_ops/test_print.py::TestHopPrint::test_print_gen_schema, test/higher_order_ops/test_print.py::TestHopPrint::test_print_with_input_mutations, test/higher_order_ops/test_print.py::TestHopPrint::test_print_with_proxy_graph, test/higher_order_ops/test_print.py::TestHopPrint::test_print_with_side_effect, test/higher_order_ops/test_print.py::TestHopPrint::test_reorder_print_no_graph_break_backend_aot_eager, test/higher_order_ops/test_print.py::TestHopPrint::test_reorder_print_no_graph_break_backend_eager 2025-12-04T16:46:54.8114808Z 2025-12-04T16:46:54.8115176Z Finished higher_order_ops/test_print 1/1 ... [2025-12-04 16:46:54.810202][27198.500595341], took 0.14min 2025-12-04T16:46:54.8620326Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/higher_order_ops.test_print/higher_order_ops.test_print-b7576a32603564cc.xml 2025-12-04T16:46:54.9610334Z Running test_per_overload_api 1/1 ... [2025-12-04 16:46:54.960701][27198.651091962] 2025-12-04T16:46:54.9610931Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:46:54.9613710Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_per_overload_api.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:46:54.961114] 2025-12-04T16:47:00.3330314Z 2025-12-04T16:47:00.3331298Z test_per_overload_api 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_per_overload_api_1.1_c9215af73fca3909_.log 2025-12-04T16:47:00.3333450Z Running 3 items in this shard: test/test_per_overload_api.py::TestPerOverloadAPI::test_basics_opoverload, test/test_per_overload_api.py::TestPerOverloadAPI::test_basics_opoverloadpacket, test/test_per_overload_api.py::TestPerOverloadAPI::test_decompose 2025-12-04T16:47:00.3334749Z 2025-12-04T16:47:00.3335076Z Finished test_per_overload_api 1/1 ... [2025-12-04 16:47:00.332835][27204.023229766], took 0.09min 2025-12-04T16:47:00.3848134Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_per_overload_api/test_per_overload_api-e690ac2b4ba1f976.xml 2025-12-04T16:47:00.4268618Z Running torch_np/numpy_tests/core/test_einsum 1/1 ... [2025-12-04 16:47:00.426586][27204.116977277] 2025-12-04T16:47:00.4269272Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:47:00.4272173Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/core/test_einsum.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:47:00.426975] 2025-12-04T16:47:20.3200274Z 2025-12-04T16:47:20.3201667Z torch_np/numpy_tests/core/test_einsum 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.core.test_einsum_1.1_62576103f884ccec_.log 2025-12-04T16:47:20.3222955Z Running 50 items in this shard: test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_broadcasting_dot_cases, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_collapse, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_combined_views_mapping, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_complex, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_different_paths_dtype_B, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_different_paths_dtype_D, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_different_paths_dtype_F, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_different_paths_dtype_b, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_different_paths_dtype_d, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_different_paths_dtype_e, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_different_paths_dtype_f, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_different_paths_dtype_h, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_different_paths_dtype_i, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_different_paths_dtype_l, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_edge_cases, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_einsum_all_contig_non_contig_output, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_einsum_broadcast, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_einsum_errors, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_einsum_failed_on_p9_and_s390x, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_einsum_fixed_collapsingbug, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_einsum_fixedstridebug, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_einsum_misc, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_einsum_sums_cfloat128, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_einsum_sums_cfloat64, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_einsum_sums_float16, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_einsum_sums_float32, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_einsum_sums_float64, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_einsum_sums_int16, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_einsum_sums_int32, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_einsum_sums_int64, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_einsum_sums_int8, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_einsum_sums_uint8, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_einsum_views, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_expand, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_hadamard_like_products, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_index_transformations, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_inner_product, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_out_is_res, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_output_order, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_random_cases, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_small_boolean_arrays, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsum::test_subscript_range, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsumPath::test_edge_paths, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsumPath::test_long_paths, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsumPath::test_memory_contraints, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsumPath::test_path_type_input, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsumPath::test_path_type_input_internal_trace, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsumPath::test_path_type_input_invalid, test/torch_np/numpy_tests/core/test_einsum.py::TestEinsumPath::test_spaces, test/torch_np/numpy_tests/core/test_einsum.py::TestMisc::test_overlap 2025-12-04T16:47:20.3243435Z 2025-12-04T16:47:20.3243854Z Finished torch_np/numpy_tests/core/test_einsum 1/1 ... [2025-12-04 16:47:20.319887][27224.010279424], took 0.33min 2025-12-04T16:47:20.3722144Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.core.test_einsum/torch_np.numpy_tests.core.test_einsum-21e2123caec3dd16.xml 2025-12-04T16:47:20.4481733Z Running test_multiprocessing 1/1 ... [2025-12-04 16:47:20.447866][27224.138256406] 2025-12-04T16:47:20.4482304Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:47:20.4485281Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_multiprocessing.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:47:20.448285] 2025-12-04T16:50:33.3451187Z 2025-12-04T16:50:33.3452209Z test_multiprocessing 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_multiprocessing_1.1_9528eb91822da171_.log 2025-12-04T16:50:33.3469491Z Running 42 items in this shard: test/test_multiprocessing.py::TestMultiprocessing::test_autograd_errors, test/test_multiprocessing.py::TestMultiprocessing::test_autograd_fine_with_spawn, test/test_multiprocessing.py::TestMultiprocessing::test_cuda_bad_call, test/test_multiprocessing.py::TestMultiprocessing::test_cuda_ipc_deadlock, test/test_multiprocessing.py::TestMultiprocessing::test_cuda_memory_allocation, test/test_multiprocessing.py::TestMultiprocessing::test_cuda_parameter_sharing, test/test_multiprocessing.py::TestMultiprocessing::test_cuda_send_many, test/test_multiprocessing.py::TestMultiprocessing::test_cuda_simple, test/test_multiprocessing.py::TestMultiprocessing::test_cuda_small_tensors, test/test_multiprocessing.py::TestMultiprocessing::test_cuda_variable_sharing, test/test_multiprocessing.py::TestMultiprocessing::test_empty_shared, test/test_multiprocessing.py::TestMultiprocessing::test_empty_tensor_sharing, test/test_multiprocessing.py::TestMultiprocessing::test_empty_tensor_sharing_cuda, test/test_multiprocessing.py::TestMultiprocessing::test_empty_tensor_sharing_meta, test/test_multiprocessing.py::TestMultiprocessing::test_event, test/test_multiprocessing.py::TestMultiprocessing::test_event_handle_exporter, test/test_multiprocessing.py::TestMultiprocessing::test_event_handle_importer, test/test_multiprocessing.py::TestMultiprocessing::test_event_handle_multi_gpu, test/test_multiprocessing.py::TestMultiprocessing::test_event_multiprocess, test/test_multiprocessing.py::TestMultiprocessing::test_fd_pool, test/test_multiprocessing.py::TestMultiprocessing::test_fd_preserve_sharing, test/test_multiprocessing.py::TestMultiprocessing::test_fd_sharing, test/test_multiprocessing.py::TestMultiprocessing::test_fs, test/test_multiprocessing.py::TestMultiprocessing::test_fs_is_shared, test/test_multiprocessing.py::TestMultiprocessing::test_fs_pool, test/test_multiprocessing.py::TestMultiprocessing::test_fs_preserve_sharing, test/test_multiprocessing.py::TestMultiprocessing::test_fs_sharing, test/test_multiprocessing.py::TestMultiprocessing::test_inherit_tensor, test/test_multiprocessing.py::TestMultiprocessing::test_integer_parameter_serialization_cpu, test/test_multiprocessing.py::TestMultiprocessing::test_integer_parameter_serialization_cuda, test/test_multiprocessing.py::TestMultiprocessing::test_is_shared, test/test_multiprocessing.py::TestMultiprocessing::test_is_shared_cuda, test/test_multiprocessing.py::TestMultiprocessing::test_leaf_variable_sharing, test/test_multiprocessing.py::TestMultiprocessing::test_meta_simple, test/test_multiprocessing.py::TestMultiprocessing::test_mixed_types_cuda_sharing, test/test_multiprocessing.py::TestMultiprocessing::test_non_leaf_variable_sharing, test/test_multiprocessing.py::TestMultiprocessing::test_parameter_sharing, test/test_multiprocessing.py::TestMultiprocessing::test_rebuild_cuda_tensor, test/test_multiprocessing.py::TestMultiprocessing::test_set_thread_name, test/test_multiprocessing.py::TestMultiprocessing::test_tensor_sharing_meta, test/test_multiprocessing.py::TestMultiprocessing::test_variable_sharing, test/test_multiprocessing.py::TestMultiprocessing::test_wrong_cuda_fork 2025-12-04T16:50:33.3485729Z 2025-12-04T16:50:33.3486089Z Finished test_multiprocessing 1/1 ... [2025-12-04 16:50:33.344627][27417.035020388], took 3.21min 2025-12-04T16:50:33.3976289Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_multiprocessing/test_multiprocessing-66d2144d604c4acf.xml 2025-12-04T16:50:35.0339838Z Uploading artifacts took 1.56 seconds 2025-12-04T16:50:35.0343555Z Running test_modules 1/1 ... [2025-12-04 16:50:35.034184][27418.724574927] 2025-12-04T16:50:35.0344172Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:50:35.0349079Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_modules.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:50:35.034640] 2025-12-04T16:58:04.1839002Z 2025-12-04T16:58:04.1840428Z test_modules 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_modules_1.1_8b9aed1ec5d9ad87_.log 2025-12-04T16:58:04.3738384Z Running 3621 items in this shard: test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_CELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_CELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_ELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_ELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_Hardswish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_Hardswish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_Hardtanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_Hardtanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_LeakyReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_LeakyReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_Mish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_Mish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_ReLU6_cuda_float32, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_ReLU6_cuda_float64, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_ReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_ReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_SELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_SELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_SiLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_SiLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_Threshold_cuda_float32, test/test_modules.py::TestModuleCUDA::test_check_inplace_nn_Threshold_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_AdaptiveAvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_AdaptiveAvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_AdaptiveAvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_AdaptiveAvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_AdaptiveAvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_AdaptiveAvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_AdaptiveMaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_AdaptiveMaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_AdaptiveMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_AdaptiveMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_AdaptiveMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_AdaptiveMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_AvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_AvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_AvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_AvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_AvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_AvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_BCELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_BCELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_BCEWithLogitsLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_BCEWithLogitsLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_BatchNorm1d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_BatchNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_BatchNorm1d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_BatchNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_BatchNorm2d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_BatchNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_BatchNorm2d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_BatchNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_BatchNorm3d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_BatchNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_BatchNorm3d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_BatchNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Bilinear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Bilinear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_CELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_CELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_CTCLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_CTCLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_CircularPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_CircularPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_CircularPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_CircularPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_CircularPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_CircularPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConstantPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConstantPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConstantPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConstantPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConstantPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConstantPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Conv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Conv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Conv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Conv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Conv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Conv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConvTranspose1d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConvTranspose1d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConvTranspose1d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConvTranspose2d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConvTranspose2d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConvTranspose3d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConvTranspose3d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConvTranspose3d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_CosineEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_CosineEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_CrossEntropyLoss_cuda_float16, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_CrossEntropyLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_CrossEntropyLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Embedding_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Embedding_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_FractionalMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_FractionalMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_FractionalMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_FractionalMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_GELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_GELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_GLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_GLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_GRUCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_GRUCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_GRU_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_GRU_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_GRU_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_GRU_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_GaussianNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_GaussianNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_GroupNorm_cuda_bfloat16, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_GroupNorm_cuda_float16, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_GroupNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_GroupNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Hardshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Hardshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Hardswish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Hardswish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Hardtanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Hardtanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_HingeEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_HingeEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_HuberLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_HuberLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_InstanceNorm1d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_InstanceNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_InstanceNorm1d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_InstanceNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_InstanceNorm2d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_InstanceNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_InstanceNorm2d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_InstanceNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_InstanceNorm3d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_InstanceNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_InstanceNorm3d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_InstanceNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_KLDivLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_KLDivLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_L1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_L1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LPPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LPPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LPPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LPPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LPPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LPPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LSTMCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LSTMCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LSTM_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LSTM_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LSTM_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LSTM_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LayerNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LayerNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LazyConv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LazyConv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LazyConv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LazyConv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LazyConv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LazyConv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LazyConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LazyConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LazyConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LazyConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LazyConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LazyConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LeakyReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LeakyReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Linear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Linear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LocalResponseNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LocalResponseNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LogSigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LogSigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LogSoftmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_LogSoftmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MSELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MSELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MarginRankingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MarginRankingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Mish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Mish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MultiLabelMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MultiLabelMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MultiLabelSoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MultiLabelSoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MultiMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MultiMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MultiheadAttention_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MultiheadAttention_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MultiheadAttention_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_MultiheadAttention_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_NLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_NLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_PReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_PReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_PoissonNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_PoissonNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_RMSNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_RMSNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_RNNCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_RNNCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_RNN_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_RNN_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_RNN_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_RNN_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ReLU6_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ReLU6_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ReflectionPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ReflectionPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ReflectionPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ReflectionPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ReflectionPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ReflectionPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ReplicationPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ReplicationPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ReplicationPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ReplicationPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ReplicationPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ReplicationPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_SELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_SELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_SiLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_SiLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Sigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Sigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_SmoothL1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_SmoothL1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_SoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_SoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Softmax2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Softmax2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Softmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Softmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Softmin_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Softmin_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Softplus_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Softplus_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Softshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Softshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Softsign_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Softsign_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Tanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Tanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Tanhshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Tanhshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Threshold_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Threshold_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_TransformerDecoderLayer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_TransformerDecoderLayer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_TransformerEncoderLayer_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_TransformerEncoderLayer_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_TransformerEncoderLayer_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_TransformerEncoderLayer_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_TransformerEncoder_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_TransformerEncoder_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_TransformerEncoder_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_TransformerEncoder_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Transformer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_Transformer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ZeroPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ZeroPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ZeroPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ZeroPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ZeroPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_cpu_gpu_parity_nn_ZeroPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_AdaptiveAvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_AdaptiveAvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_AdaptiveAvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_AdaptiveAvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_AdaptiveAvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_AdaptiveAvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_AdaptiveMaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_AdaptiveMaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_AdaptiveMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_AdaptiveMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_AdaptiveMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_AdaptiveMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_AvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_AvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_AvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_AvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_AvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_AvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_BCELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_BCELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_BCEWithLogitsLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_BCEWithLogitsLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_BatchNorm1d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_BatchNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_BatchNorm1d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_BatchNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_BatchNorm2d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_BatchNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_BatchNorm2d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_BatchNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_BatchNorm3d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_BatchNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_BatchNorm3d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_BatchNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Bilinear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Bilinear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_CELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_CELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_CTCLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_CTCLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_CircularPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_CircularPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_CircularPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_CircularPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_CircularPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_CircularPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConstantPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConstantPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConstantPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConstantPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConstantPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConstantPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Conv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Conv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Conv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Conv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Conv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Conv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConvTranspose1d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConvTranspose1d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConvTranspose1d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConvTranspose2d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConvTranspose2d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConvTranspose3d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConvTranspose3d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConvTranspose3d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_CosineEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_CosineEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_CrossEntropyLoss_cuda_float16, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_CrossEntropyLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_CrossEntropyLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Embedding_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Embedding_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_FractionalMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_FractionalMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_FractionalMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_FractionalMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_GELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_GELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_GLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_GLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_GRUCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_GRUCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_GRU_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_GRU_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_GRU_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_GRU_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_GaussianNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_GaussianNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_GroupNorm_cuda_bfloat16, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_GroupNorm_cuda_float16, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_GroupNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_GroupNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Hardshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Hardshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Hardswish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Hardswish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Hardtanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Hardtanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_HingeEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_HingeEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_HuberLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_HuberLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_InstanceNorm1d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_InstanceNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_InstanceNorm1d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_InstanceNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_InstanceNorm2d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_InstanceNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_InstanceNorm2d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_InstanceNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_InstanceNorm3d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_InstanceNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_InstanceNorm3d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_InstanceNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_KLDivLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_KLDivLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_L1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_L1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LPPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LPPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LPPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LPPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LPPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LPPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LSTMCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LSTMCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LSTM_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LSTM_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LSTM_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LSTM_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LayerNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LayerNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LazyConv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LazyConv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LazyConv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LazyConv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LazyConv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LazyConv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LazyConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LazyConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LazyConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LazyConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LazyConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LazyConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LeakyReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LeakyReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Linear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Linear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LocalResponseNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LocalResponseNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LogSigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LogSigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LogSoftmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_LogSoftmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MSELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MSELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MarginRankingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MarginRankingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Mish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Mish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MultiLabelMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MultiLabelMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MultiLabelSoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MultiLabelSoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MultiMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MultiMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MultiheadAttention_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MultiheadAttention_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MultiheadAttention_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_MultiheadAttention_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_NLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_NLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_PReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_PReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_PoissonNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_PoissonNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_RMSNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_RMSNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_RNNCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_RNNCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_RNN_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_RNN_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_RNN_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_RNN_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ReLU6_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ReLU6_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ReflectionPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ReflectionPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ReflectionPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ReflectionPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ReflectionPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ReflectionPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ReplicationPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ReplicationPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ReplicationPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ReplicationPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ReplicationPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ReplicationPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_SELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_SELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_SiLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_SiLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Sigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Sigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_SmoothL1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_SmoothL1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_SoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_SoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Softmax2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Softmax2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Softmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Softmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Softmin_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Softmin_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Softplus_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Softplus_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Softshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Softshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Softsign_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Softsign_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Tanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Tanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Tanhshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Tanhshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Threshold_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Threshold_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_TransformerDecoderLayer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_TransformerDecoderLayer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_TransformerEncoderLayer_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_TransformerEncoderLayer_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_TransformerEncoderLayer_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_TransformerEncoderLayer_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_TransformerEncoder_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_TransformerEncoder_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_TransformerEncoder_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_TransformerEncoder_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Transformer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_Transformer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ZeroPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ZeroPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ZeroPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ZeroPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ZeroPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_device_ctx_init_nn_ZeroPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_errors_nn_BatchNorm1d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_errors_nn_BatchNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_errors_nn_BatchNorm1d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_errors_nn_BatchNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_errors_nn_BatchNorm2d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_errors_nn_BatchNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_errors_nn_BatchNorm2d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_errors_nn_BatchNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_errors_nn_BatchNorm3d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_errors_nn_BatchNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_errors_nn_BatchNorm3d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_errors_nn_BatchNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_errors_nn_CircularPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_errors_nn_CircularPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_errors_nn_CircularPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_errors_nn_CircularPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_errors_nn_CircularPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_errors_nn_CircularPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_errors_nn_GRUCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_errors_nn_GRUCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_errors_nn_GRU_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_errors_nn_GRU_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_errors_nn_GRU_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_errors_nn_GRU_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_errors_nn_GroupNorm_cuda_bfloat16, test/test_modules.py::TestModuleCUDA::test_errors_nn_GroupNorm_cuda_float16, test/test_modules.py::TestModuleCUDA::test_errors_nn_GroupNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_errors_nn_GroupNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_errors_nn_LSTMCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_errors_nn_LSTMCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_errors_nn_LSTM_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_errors_nn_LSTM_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_errors_nn_LSTM_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_errors_nn_LSTM_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_errors_nn_RNNCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_errors_nn_RNNCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_errors_nn_RNN_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_errors_nn_RNN_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_errors_nn_RNN_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_errors_nn_RNN_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_AdaptiveAvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_AdaptiveAvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_AdaptiveAvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_AdaptiveAvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_AdaptiveAvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_AdaptiveAvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_AdaptiveMaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_AdaptiveMaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_AdaptiveMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_AdaptiveMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_AdaptiveMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_AdaptiveMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_AvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_AvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_AvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_AvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_AvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_AvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_BCELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_BCELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_BCEWithLogitsLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_BCEWithLogitsLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_BatchNorm1d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_BatchNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_BatchNorm1d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_BatchNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_BatchNorm2d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_BatchNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_BatchNorm2d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_BatchNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_BatchNorm3d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_BatchNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_BatchNorm3d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_BatchNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Bilinear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Bilinear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_CELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_CELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_CTCLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_CTCLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_CircularPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_CircularPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_CircularPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_CircularPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_CircularPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_CircularPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConstantPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConstantPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConstantPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConstantPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConstantPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConstantPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Conv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Conv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Conv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Conv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Conv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Conv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConvTranspose1d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConvTranspose1d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConvTranspose1d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConvTranspose2d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConvTranspose2d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConvTranspose3d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConvTranspose3d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConvTranspose3d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_CosineEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_CosineEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_CrossEntropyLoss_cuda_float16, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_CrossEntropyLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_CrossEntropyLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Embedding_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Embedding_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_FractionalMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_FractionalMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_FractionalMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_FractionalMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_GELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_GELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_GLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_GLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_GRUCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_GRUCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_GRU_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_GRU_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_GRU_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_GRU_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_GaussianNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_GaussianNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_GroupNorm_cuda_bfloat16, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_GroupNorm_cuda_float16, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_GroupNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_GroupNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Hardshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Hardshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Hardswish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Hardswish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Hardtanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Hardtanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_HingeEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_HingeEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_HuberLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_HuberLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_InstanceNorm1d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_InstanceNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_InstanceNorm1d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_InstanceNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_InstanceNorm2d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_InstanceNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_InstanceNorm2d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_InstanceNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_InstanceNorm3d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_InstanceNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_InstanceNorm3d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_InstanceNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_KLDivLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_KLDivLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_L1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_L1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LPPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LPPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LPPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LPPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LPPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LPPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LSTMCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LSTMCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LSTM_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LSTM_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LSTM_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LSTM_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LayerNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LayerNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LazyConv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LazyConv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LazyConv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LazyConv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LazyConv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LazyConv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LazyConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LazyConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LazyConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LazyConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LazyConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LazyConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LeakyReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LeakyReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Linear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Linear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LocalResponseNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LocalResponseNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LogSigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LogSigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LogSoftmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_LogSoftmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MSELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MSELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MarginRankingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MarginRankingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Mish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Mish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MultiLabelMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MultiLabelMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MultiLabelSoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MultiLabelSoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MultiMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MultiMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MultiheadAttention_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MultiheadAttention_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MultiheadAttention_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_MultiheadAttention_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_NLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_NLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_PReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_PReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_PoissonNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_PoissonNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_RMSNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_RMSNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_RNNCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_RNNCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_RNN_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_RNN_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_RNN_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_RNN_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ReLU6_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ReLU6_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ReflectionPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ReflectionPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ReflectionPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ReflectionPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ReflectionPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ReflectionPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ReplicationPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ReplicationPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ReplicationPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ReplicationPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ReplicationPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ReplicationPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_SELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_SELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_SiLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_SiLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Sigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Sigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_SmoothL1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_SmoothL1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_SoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_SoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Softmax2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Softmax2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Softmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Softmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Softmin_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Softmin_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Softplus_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Softplus_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Softshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Softshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Softsign_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Softsign_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Tanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Tanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Tanhshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Tanhshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Threshold_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Threshold_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_TransformerDecoderLayer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_TransformerDecoderLayer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_TransformerEncoderLayer_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_TransformerEncoderLayer_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_TransformerEncoderLayer_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_TransformerEncoderLayer_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_TransformerEncoder_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_TransformerEncoder_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_TransformerEncoder_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_TransformerEncoder_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Transformer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_Transformer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ZeroPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ZeroPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ZeroPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ZeroPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ZeroPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_factory_kwargs_nn_ZeroPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_AdaptiveAvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_AdaptiveAvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_AdaptiveAvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_AdaptiveAvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_AdaptiveAvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_AdaptiveAvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_AdaptiveMaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_AdaptiveMaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_AdaptiveMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_AdaptiveMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_AdaptiveMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_AdaptiveMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_AvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_AvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_AvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_AvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_AvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_AvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_BCELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_BCELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_BCEWithLogitsLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_BCEWithLogitsLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_BatchNorm1d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_BatchNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_BatchNorm1d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_BatchNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_BatchNorm2d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_BatchNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_BatchNorm2d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_BatchNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_BatchNorm3d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_BatchNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_BatchNorm3d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_BatchNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Bilinear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Bilinear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_CELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_CELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_CTCLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_CTCLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_CircularPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_CircularPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_CircularPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_CircularPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_CircularPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_CircularPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConstantPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConstantPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConstantPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConstantPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConstantPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConstantPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Conv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Conv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Conv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Conv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Conv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Conv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConvTranspose1d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConvTranspose1d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConvTranspose1d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConvTranspose2d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConvTranspose2d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConvTranspose3d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConvTranspose3d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConvTranspose3d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_CosineEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_CosineEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_CrossEntropyLoss_cuda_float16, test/test_modules.py::TestModuleCUDA::test_forward_nn_CrossEntropyLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_CrossEntropyLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Embedding_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Embedding_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_FractionalMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_FractionalMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_FractionalMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_FractionalMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_GELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_GELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_GLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_GLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_GRUCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_GRUCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_GRU_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_GRU_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_GRU_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_GRU_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_GaussianNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_GaussianNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_GroupNorm_cuda_bfloat16, test/test_modules.py::TestModuleCUDA::test_forward_nn_GroupNorm_cuda_float16, test/test_modules.py::TestModuleCUDA::test_forward_nn_GroupNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_GroupNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Hardshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Hardshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Hardswish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Hardswish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Hardtanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Hardtanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_HingeEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_HingeEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_HuberLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_HuberLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_InstanceNorm1d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_InstanceNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_InstanceNorm1d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_InstanceNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_InstanceNorm2d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_InstanceNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_InstanceNorm2d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_InstanceNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_InstanceNorm3d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_InstanceNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_InstanceNorm3d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_InstanceNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_KLDivLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_KLDivLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_L1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_L1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_LPPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_LPPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_LPPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_LPPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_LPPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_LPPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_LSTMCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_LSTMCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_LSTM_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_LSTM_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_LSTM_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_LSTM_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_LayerNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_LayerNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_LazyConv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_LazyConv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_LazyConv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_LazyConv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_LazyConv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_LazyConv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_LazyConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_LazyConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_LazyConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_LazyConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_LazyConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_LazyConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_LeakyReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_LeakyReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Linear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Linear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_LocalResponseNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_LocalResponseNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_LogSigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_LogSigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_LogSoftmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_LogSoftmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_MSELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_MSELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_MarginRankingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_MarginRankingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_MaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_MaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_MaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_MaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_MaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_MaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Mish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Mish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_MultiLabelMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_MultiLabelMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_MultiLabelSoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_MultiLabelSoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_MultiMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_MultiMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_MultiheadAttention_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_MultiheadAttention_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_MultiheadAttention_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_MultiheadAttention_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_NLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_NLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_PReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_PReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_PoissonNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_PoissonNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_RMSNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_RMSNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_RNNCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_RNNCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_RNN_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_RNN_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_RNN_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_RNN_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ReLU6_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ReLU6_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ReflectionPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ReflectionPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ReflectionPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ReflectionPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ReflectionPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ReflectionPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ReplicationPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ReplicationPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ReplicationPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ReplicationPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ReplicationPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ReplicationPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_SELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_SELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_SiLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_SiLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Sigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Sigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_SmoothL1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_SmoothL1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_SoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_SoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Softmax2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Softmax2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Softmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Softmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Softmin_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Softmin_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Softplus_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Softplus_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Softshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Softshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Softsign_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Softsign_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Tanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Tanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Tanhshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Tanhshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Threshold_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Threshold_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_TransformerDecoderLayer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_TransformerDecoderLayer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_TransformerEncoderLayer_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_TransformerEncoderLayer_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_TransformerEncoderLayer_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_TransformerEncoderLayer_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_TransformerEncoder_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_TransformerEncoder_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_TransformerEncoder_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_TransformerEncoder_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_Transformer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_Transformer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ZeroPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ZeroPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ZeroPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ZeroPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_forward_nn_ZeroPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_forward_nn_ZeroPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_AdaptiveAvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_AdaptiveAvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_AdaptiveAvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_AdaptiveMaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_AdaptiveMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_AdaptiveMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_AvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_AvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_AvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_BCELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_BCEWithLogitsLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_BatchNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_BatchNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_BatchNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_BatchNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_BatchNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_BatchNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Bilinear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_CELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_CTCLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_CircularPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_CircularPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_CircularPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_ConstantPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_ConstantPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_ConstantPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Conv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Conv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Conv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_ConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_ConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_ConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_CosineEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_CrossEntropyLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_ELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Embedding_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_FractionalMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_FractionalMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_GELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_GLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_GRUCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_GRU_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_GRU_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_GaussianNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_GroupNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Hardshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Hardswish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Hardtanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_HingeEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_HuberLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_InstanceNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_InstanceNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_InstanceNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_InstanceNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_InstanceNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_InstanceNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_KLDivLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_L1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_LPPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_LPPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_LPPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_LSTMCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_LSTM_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_LSTM_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_LayerNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_LazyConv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_LazyConv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_LazyConv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_LazyConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_LazyConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_LazyConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_LeakyReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Linear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_LocalResponseNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_LogSigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_LogSoftmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_MSELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_MarginRankingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_MaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_MaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_MaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Mish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_MultiLabelMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_MultiLabelSoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_MultiMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_MultiheadAttention_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_MultiheadAttention_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_NLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_PReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_PoissonNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_RMSNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_RNNCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_RNN_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_RNN_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_ReLU6_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_ReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_ReflectionPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_ReflectionPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_ReflectionPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_ReplicationPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_ReplicationPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_ReplicationPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_SELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_SiLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Sigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_SmoothL1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_SoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Softmax2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Softmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Softmin_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Softplus_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Softshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Softsign_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Tanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Tanhshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Threshold_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_TransformerDecoderLayer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_TransformerEncoderLayer_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_TransformerEncoderLayer_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_TransformerEncoder_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_TransformerEncoder_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_Transformer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_ZeroPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_ZeroPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_grad_nn_ZeroPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_AdaptiveAvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_AdaptiveAvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_AdaptiveAvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_AdaptiveMaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_AdaptiveMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_AdaptiveMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_AvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_AvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_AvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_BCELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_BCEWithLogitsLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_BatchNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_BatchNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_BatchNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_BatchNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_BatchNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_BatchNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Bilinear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_CELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_CTCLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_CircularPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_CircularPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_CircularPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_ConstantPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_ConstantPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_ConstantPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Conv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Conv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Conv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_ConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_ConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_ConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_CosineEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_CrossEntropyLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_ELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Embedding_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_FractionalMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_FractionalMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_GELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_GLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_GRUCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_GRU_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_GRU_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_GaussianNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_GroupNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Hardshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Hardtanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_HingeEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_HuberLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_InstanceNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_InstanceNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_InstanceNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_InstanceNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_InstanceNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_InstanceNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_KLDivLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_L1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_LPPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_LPPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_LPPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_LSTMCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_LSTM_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_LSTM_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_LayerNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_LazyConv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_LazyConv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_LazyConv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_LazyConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_LazyConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_LazyConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_LeakyReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Linear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_LocalResponseNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_LogSigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_LogSoftmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_MSELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_MarginRankingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_MaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_MaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_MaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Mish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_MultiLabelMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_MultiLabelSoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_MultiMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_MultiheadAttention_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_MultiheadAttention_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_NLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_PReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_PoissonNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_RMSNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_RNNCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_RNN_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_RNN_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_ReLU6_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_ReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_ReflectionPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_ReflectionPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_ReflectionPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_ReplicationPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_ReplicationPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_ReplicationPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_SELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_SiLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Sigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_SmoothL1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_SoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Softmax2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Softmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Softmin_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Softplus_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Softshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Softsign_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Tanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Tanhshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Threshold_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_TransformerDecoderLayer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_TransformerEncoderLayer_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_TransformerEncoderLayer_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_TransformerEncoder_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_TransformerEncoder_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_Transformer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_ZeroPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_ZeroPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_gradgrad_nn_ZeroPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_AdaptiveAvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_AdaptiveAvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_AdaptiveAvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_AdaptiveAvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_AdaptiveAvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_AdaptiveAvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_AdaptiveMaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_AdaptiveMaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_AdaptiveMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_AdaptiveMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_AdaptiveMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_AdaptiveMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_AvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_AvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_AvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_AvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_AvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_AvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_BCELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_BCELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_BCEWithLogitsLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_BCEWithLogitsLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_BatchNorm1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_BatchNorm1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_BatchNorm2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_BatchNorm2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_BatchNorm3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_BatchNorm3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Bilinear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Bilinear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_CELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_CELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_CTCLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_CTCLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_CircularPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_CircularPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_CircularPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_CircularPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_CircularPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_CircularPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConstantPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConstantPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConstantPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConstantPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConstantPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConstantPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Conv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Conv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Conv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Conv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Conv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Conv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConvTranspose1d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConvTranspose1d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConvTranspose1d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConvTranspose2d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConvTranspose2d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConvTranspose3d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConvTranspose3d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConvTranspose3d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_CosineEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_CosineEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_CrossEntropyLoss_cuda_float16, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_CrossEntropyLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_CrossEntropyLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Embedding_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Embedding_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_FractionalMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_FractionalMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_FractionalMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_FractionalMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_GELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_GELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_GLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_GLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_GRUCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_GRUCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_GRU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_GRU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_GaussianNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_GaussianNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_GroupNorm_cuda_bfloat16, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_GroupNorm_cuda_float16, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_GroupNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_GroupNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Hardshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Hardshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Hardswish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Hardswish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Hardtanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Hardtanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_HingeEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_HingeEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_HuberLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_HuberLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_InstanceNorm1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_InstanceNorm1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_InstanceNorm2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_InstanceNorm2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_InstanceNorm3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_InstanceNorm3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_KLDivLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_KLDivLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_L1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_L1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LPPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LPPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LPPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LPPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LPPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LPPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LSTMCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LSTMCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LSTM_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LSTM_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LayerNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LayerNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LazyConv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LazyConv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LazyConv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LazyConv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LazyConv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LazyConv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LazyConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LazyConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LazyConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LazyConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LazyConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LazyConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LeakyReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LeakyReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Linear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Linear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LocalResponseNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LocalResponseNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LogSigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LogSigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LogSoftmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_LogSoftmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_MSELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_MSELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_MarginRankingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_MarginRankingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_MaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_MaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_MaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_MaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_MaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_MaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Mish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Mish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_MultiLabelMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_MultiLabelMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_MultiLabelSoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_MultiLabelSoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_MultiMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_MultiMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_MultiheadAttention_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_MultiheadAttention_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_NLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_NLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_PReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_PReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_PoissonNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_PoissonNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_RMSNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_RMSNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_RNNCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_RNNCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_RNN_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_RNN_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ReLU6_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ReLU6_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ReflectionPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ReflectionPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ReflectionPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ReflectionPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ReflectionPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ReflectionPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ReplicationPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ReplicationPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ReplicationPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ReplicationPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ReplicationPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ReplicationPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_SELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_SELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_SiLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_SiLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Sigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Sigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_SmoothL1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_SmoothL1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_SoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_SoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Softmax2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Softmax2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Softmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Softmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Softmin_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Softmin_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Softplus_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Softplus_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Softshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Softshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Softsign_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Softsign_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Tanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Tanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Tanhshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Tanhshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Threshold_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Threshold_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_TransformerDecoderLayer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_TransformerDecoderLayer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_TransformerEncoderLayer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_TransformerEncoderLayer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_TransformerEncoder_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_TransformerEncoder_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Transformer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_Transformer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ZeroPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ZeroPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ZeroPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ZeroPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ZeroPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_if_train_and_eval_modes_differ_nn_ZeroPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_AdaptiveAvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_AdaptiveAvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_AdaptiveAvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_AdaptiveAvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_AdaptiveAvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_AdaptiveAvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_AdaptiveMaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_AdaptiveMaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_AdaptiveMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_AdaptiveMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_AdaptiveMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_AdaptiveMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_AvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_AvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_AvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_AvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_AvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_AvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_BCELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_BCELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_BCEWithLogitsLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_BCEWithLogitsLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_BatchNorm1d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_BatchNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_BatchNorm1d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_BatchNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_BatchNorm2d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_BatchNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_BatchNorm2d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_BatchNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_BatchNorm3d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_BatchNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_BatchNorm3d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_BatchNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Bilinear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Bilinear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_CELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_CELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_CTCLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_CTCLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_CircularPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_CircularPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_CircularPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_CircularPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_CircularPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_CircularPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConstantPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConstantPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConstantPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConstantPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConstantPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConstantPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Conv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Conv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Conv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Conv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Conv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Conv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose1d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose1d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose1d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose3d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose3d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose3d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_CosineEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_CosineEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_CrossEntropyLoss_cuda_float16, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_CrossEntropyLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_CrossEntropyLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Embedding_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Embedding_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_FractionalMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_FractionalMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_FractionalMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_FractionalMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_GELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_GELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_GLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_GLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_GRUCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_GRUCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_GRU_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_GRU_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_GRU_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_GRU_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_GaussianNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_GaussianNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_GroupNorm_cuda_bfloat16, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_GroupNorm_cuda_float16, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_GroupNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_GroupNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Hardshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Hardshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Hardswish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Hardswish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Hardtanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Hardtanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_HingeEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_HingeEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_HuberLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_HuberLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_InstanceNorm1d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_InstanceNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_InstanceNorm1d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_InstanceNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_InstanceNorm2d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_InstanceNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_InstanceNorm2d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_InstanceNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_InstanceNorm3d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_InstanceNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_InstanceNorm3d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_InstanceNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_KLDivLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_KLDivLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_L1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_L1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LPPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LPPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LPPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LPPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LPPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LPPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LSTMCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LSTMCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LSTM_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LSTM_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LSTM_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LSTM_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LayerNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LayerNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LazyConv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LazyConv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LazyConv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LazyConv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LazyConv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LazyConv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LazyConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LazyConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LazyConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LazyConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LazyConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LazyConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LeakyReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LeakyReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Linear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Linear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LocalResponseNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LocalResponseNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LogSigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LogSigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LogSoftmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_LogSoftmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MSELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MSELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MarginRankingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MarginRankingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Mish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Mish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MultiLabelMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MultiLabelMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MultiLabelSoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MultiLabelSoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MultiMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MultiMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MultiheadAttention_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MultiheadAttention_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MultiheadAttention_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_MultiheadAttention_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_NLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_NLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_PReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_PReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_PoissonNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_PoissonNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_RMSNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_RMSNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_RNNCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_RNNCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_RNN_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_RNN_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_RNN_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_RNN_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ReLU6_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ReLU6_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ReflectionPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ReflectionPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ReflectionPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ReflectionPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ReflectionPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ReflectionPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ReplicationPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ReplicationPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ReplicationPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ReplicationPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ReplicationPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ReplicationPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_SELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_SELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_SiLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_SiLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Sigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Sigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_SmoothL1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_SmoothL1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_SoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_SoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Softmax2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Softmax2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Softmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Softmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Softmin_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Softmin_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Softplus_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Softplus_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Softshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Softshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Softsign_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Softsign_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Tanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Tanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Tanhshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Tanhshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Threshold_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Threshold_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_TransformerDecoderLayer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_TransformerDecoderLayer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_TransformerEncoderLayer_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_TransformerEncoderLayer_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_TransformerEncoderLayer_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_TransformerEncoderLayer_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_TransformerEncoder_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_TransformerEncoder_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_TransformerEncoder_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_TransformerEncoder_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Transformer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_Transformer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ZeroPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ZeroPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ZeroPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ZeroPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ZeroPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_memory_format_nn_ZeroPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_AdaptiveAvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_AdaptiveAvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_AdaptiveAvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_AdaptiveAvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_AdaptiveAvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_AdaptiveAvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_AdaptiveMaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_AdaptiveMaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_AdaptiveMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_AdaptiveMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_AdaptiveMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_AdaptiveMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_AvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_AvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_AvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_AvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_AvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_AvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_BCELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_BCELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_BCEWithLogitsLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_BCEWithLogitsLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_BatchNorm1d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_BatchNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_BatchNorm1d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_BatchNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_BatchNorm2d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_BatchNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_BatchNorm2d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_BatchNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_BatchNorm3d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_BatchNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_BatchNorm3d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_BatchNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Bilinear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Bilinear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_CELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_CELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_CTCLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_CTCLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_CircularPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_CircularPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_CircularPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_CircularPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_CircularPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_CircularPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConstantPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConstantPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConstantPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConstantPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConstantPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConstantPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Conv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Conv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Conv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Conv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Conv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Conv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConvTranspose1d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConvTranspose1d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConvTranspose1d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConvTranspose2d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConvTranspose2d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConvTranspose3d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConvTranspose3d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConvTranspose3d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_CosineEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_CosineEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_CrossEntropyLoss_cuda_float16, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_CrossEntropyLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_CrossEntropyLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Embedding_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Embedding_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_FractionalMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_FractionalMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_FractionalMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_FractionalMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_GELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_GELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_GLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_GLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_GRUCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_GRUCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_GRU_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_GRU_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_GRU_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_GRU_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_GaussianNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_GaussianNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_GroupNorm_cuda_bfloat16, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_GroupNorm_cuda_float16, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_GroupNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_GroupNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Hardshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Hardshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Hardswish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Hardswish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Hardtanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Hardtanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_HingeEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_HingeEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_HuberLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_HuberLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_InstanceNorm1d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_InstanceNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_InstanceNorm1d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_InstanceNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_InstanceNorm2d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_InstanceNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_InstanceNorm2d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_InstanceNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_InstanceNorm3d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_InstanceNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_InstanceNorm3d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_InstanceNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_KLDivLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_KLDivLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_L1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_L1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LPPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LPPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LPPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LPPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LPPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LPPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LSTMCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LSTMCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LSTM_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LSTM_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LSTM_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LSTM_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LayerNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LayerNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LazyConv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LazyConv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LazyConv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LazyConv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LazyConv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LazyConv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LazyConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LazyConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LazyConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LazyConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LazyConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LazyConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LeakyReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LeakyReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Linear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Linear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LocalResponseNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LocalResponseNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LogSigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LogSigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LogSoftmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_LogSoftmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MSELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MSELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MarginRankingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MarginRankingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Mish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Mish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MultiLabelMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MultiLabelMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MultiLabelSoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MultiLabelSoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MultiMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MultiMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MultiheadAttention_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MultiheadAttention_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MultiheadAttention_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_MultiheadAttention_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_NLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_NLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_PReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_PReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_PoissonNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_PoissonNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_RMSNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_RMSNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_RNNCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_RNNCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_RNN_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_RNN_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_RNN_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_RNN_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ReLU6_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ReLU6_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ReflectionPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ReflectionPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ReflectionPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ReflectionPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ReflectionPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ReflectionPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ReplicationPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ReplicationPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ReplicationPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ReplicationPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ReplicationPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ReplicationPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_SELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_SELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_SiLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_SiLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Sigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Sigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_SmoothL1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_SmoothL1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_SoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_SoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Softmax2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Softmax2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Softmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Softmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Softmin_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Softmin_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Softplus_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Softplus_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Softshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Softshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Softsign_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Softsign_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Tanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Tanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Tanhshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Tanhshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Threshold_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Threshold_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_TransformerDecoderLayer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_TransformerDecoderLayer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_TransformerEncoderLayer_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_TransformerEncoderLayer_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_TransformerEncoderLayer_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_TransformerEncoderLayer_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_TransformerEncoder_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_TransformerEncoder_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_TransformerEncoder_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_TransformerEncoder_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Transformer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_Transformer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ZeroPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ZeroPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ZeroPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ZeroPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ZeroPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_multiple_device_transfer_nn_ZeroPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_AdaptiveAvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_AdaptiveAvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_AdaptiveAvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_AdaptiveAvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_AdaptiveAvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_AdaptiveAvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_AdaptiveMaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_AdaptiveMaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_AdaptiveMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_AdaptiveMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_AdaptiveMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_AdaptiveMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_AvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_AvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_AvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_AvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_AvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_AvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_BCELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_BCELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_BCEWithLogitsLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_BCEWithLogitsLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_BatchNorm1d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_BatchNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_BatchNorm1d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_BatchNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_BatchNorm2d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_BatchNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_BatchNorm2d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_BatchNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_BatchNorm3d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_BatchNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_BatchNorm3d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_BatchNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Bilinear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Bilinear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_CELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_CELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_CTCLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_CTCLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_CircularPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_CircularPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_CircularPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_CircularPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_CircularPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_CircularPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConstantPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConstantPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConstantPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConstantPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConstantPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConstantPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Conv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Conv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Conv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Conv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Conv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Conv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConvTranspose1d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConvTranspose1d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConvTranspose1d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConvTranspose2d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConvTranspose2d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConvTranspose3d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConvTranspose3d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConvTranspose3d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_CosineEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_CosineEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_CrossEntropyLoss_cuda_float16, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_CrossEntropyLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_CrossEntropyLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Embedding_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Embedding_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_FractionalMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_FractionalMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_FractionalMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_FractionalMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_GELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_GELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_GLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_GLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_GRUCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_GRUCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_GRU_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_GRU_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_GRU_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_GRU_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_GaussianNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_GaussianNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_GroupNorm_cuda_bfloat16, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_GroupNorm_cuda_float16, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_GroupNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_GroupNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Hardshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Hardshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Hardswish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Hardswish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Hardtanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Hardtanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_HingeEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_HingeEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_HuberLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_HuberLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_InstanceNorm1d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_InstanceNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_InstanceNorm1d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_InstanceNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_InstanceNorm2d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_InstanceNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_InstanceNorm2d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_InstanceNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_InstanceNorm3d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_InstanceNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_InstanceNorm3d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_InstanceNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_KLDivLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_KLDivLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_L1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_L1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LPPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LPPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LPPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LPPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LPPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LPPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LSTMCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LSTMCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LSTM_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LSTM_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LSTM_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LSTM_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LayerNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LayerNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LazyConv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LazyConv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LazyConv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LazyConv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LazyConv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LazyConv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LazyConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LazyConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LazyConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LazyConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LazyConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LazyConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LeakyReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LeakyReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Linear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Linear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LocalResponseNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LocalResponseNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LogSigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LogSigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LogSoftmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_LogSoftmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MSELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MSELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MarginRankingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MarginRankingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Mish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Mish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MultiLabelMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MultiLabelMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MultiLabelSoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MultiLabelSoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MultiMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MultiMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MultiheadAttention_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MultiheadAttention_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MultiheadAttention_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_MultiheadAttention_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_NLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_NLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_PReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_PReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_PoissonNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_PoissonNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_RMSNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_RMSNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_RNNCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_RNNCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_RNN_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_RNN_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_RNN_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_RNN_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ReLU6_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ReLU6_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ReflectionPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ReflectionPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ReflectionPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ReflectionPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ReflectionPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ReflectionPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ReplicationPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ReplicationPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ReplicationPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ReplicationPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ReplicationPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ReplicationPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_SELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_SELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_SiLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_SiLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Sigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Sigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_SmoothL1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_SmoothL1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_SoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_SoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Softmax2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Softmax2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Softmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Softmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Softmin_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Softmin_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Softplus_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Softplus_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Softshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Softshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Softsign_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Softsign_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Tanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Tanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Tanhshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Tanhshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Threshold_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Threshold_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_TransformerDecoderLayer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_TransformerDecoderLayer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_TransformerEncoderLayer_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_TransformerEncoderLayer_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_TransformerEncoderLayer_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_TransformerEncoderLayer_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_TransformerEncoder_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_TransformerEncoder_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_TransformerEncoder_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_TransformerEncoder_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Transformer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_Transformer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ZeroPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ZeroPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ZeroPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ZeroPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ZeroPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_non_contiguous_tensors_nn_ZeroPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_AdaptiveAvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_AdaptiveAvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_AdaptiveAvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_AdaptiveAvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_AdaptiveAvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_AdaptiveAvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_AdaptiveMaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_AdaptiveMaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_AdaptiveMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_AdaptiveMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_AdaptiveMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_AdaptiveMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_AvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_AvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_AvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_AvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_AvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_AvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_BCELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_BCELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_BCEWithLogitsLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_BCEWithLogitsLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_BatchNorm1d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_BatchNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_BatchNorm1d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_BatchNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_BatchNorm2d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_BatchNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_BatchNorm2d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_BatchNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_BatchNorm3d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_BatchNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_BatchNorm3d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_BatchNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Bilinear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Bilinear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_CELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_CELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_CTCLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_CTCLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_CircularPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_CircularPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_CircularPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_CircularPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_CircularPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_CircularPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConstantPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConstantPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConstantPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConstantPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConstantPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConstantPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Conv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Conv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Conv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Conv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Conv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Conv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConvTranspose1d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConvTranspose1d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConvTranspose1d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConvTranspose2d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConvTranspose2d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConvTranspose3d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConvTranspose3d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConvTranspose3d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_CosineEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_CosineEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_CrossEntropyLoss_cuda_float16, test/test_modules.py::TestModuleCUDA::test_repr_nn_CrossEntropyLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_CrossEntropyLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Embedding_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Embedding_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_FractionalMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_FractionalMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_FractionalMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_FractionalMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_GELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_GELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_GLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_GLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_GRUCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_GRUCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_GRU_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_GRU_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_GRU_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_GRU_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_GaussianNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_GaussianNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_GroupNorm_cuda_bfloat16, test/test_modules.py::TestModuleCUDA::test_repr_nn_GroupNorm_cuda_float16, test/test_modules.py::TestModuleCUDA::test_repr_nn_GroupNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_GroupNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Hardshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Hardshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Hardswish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Hardswish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Hardtanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Hardtanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_HingeEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_HingeEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_HuberLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_HuberLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_InstanceNorm1d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_InstanceNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_InstanceNorm1d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_InstanceNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_InstanceNorm2d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_InstanceNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_InstanceNorm2d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_InstanceNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_InstanceNorm3d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_InstanceNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_InstanceNorm3d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_InstanceNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_KLDivLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_KLDivLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_L1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_L1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_LPPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_LPPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_LPPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_LPPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_LPPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_LPPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_LSTMCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_LSTMCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_LSTM_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_LSTM_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_LSTM_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_LSTM_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_LayerNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_LayerNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_LazyConv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_LazyConv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_LazyConv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_LazyConv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_LazyConv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_LazyConv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_LazyConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_LazyConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_LazyConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_LazyConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_LazyConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_LazyConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_LeakyReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_LeakyReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Linear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Linear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_LocalResponseNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_LocalResponseNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_LogSigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_LogSigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_LogSoftmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_LogSoftmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_MSELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_MSELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_MarginRankingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_MarginRankingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_MaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_MaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_MaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_MaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_MaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_MaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Mish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Mish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_MultiLabelMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_MultiLabelMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_MultiLabelSoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_MultiLabelSoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_MultiMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_MultiMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_MultiheadAttention_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_MultiheadAttention_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_MultiheadAttention_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_MultiheadAttention_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_NLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_NLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_PReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_PReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_PoissonNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_PoissonNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_RMSNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_RMSNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_RNNCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_RNNCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_RNN_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_RNN_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_RNN_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_RNN_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ReLU6_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ReLU6_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ReflectionPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ReflectionPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ReflectionPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ReflectionPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ReflectionPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ReflectionPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ReplicationPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ReplicationPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ReplicationPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ReplicationPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ReplicationPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ReplicationPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_SELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_SELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_SiLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_SiLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Sigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Sigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_SmoothL1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_SmoothL1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_SoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_SoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Softmax2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Softmax2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Softmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Softmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Softmin_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Softmin_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Softplus_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Softplus_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Softshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Softshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Softsign_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Softsign_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Tanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Tanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Tanhshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Tanhshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Threshold_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Threshold_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_TransformerDecoderLayer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_TransformerDecoderLayer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_TransformerEncoderLayer_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_TransformerEncoderLayer_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_TransformerEncoderLayer_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_TransformerEncoderLayer_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_TransformerEncoder_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_TransformerEncoder_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_TransformerEncoder_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_TransformerEncoder_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_Transformer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_Transformer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ZeroPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ZeroPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ZeroPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ZeroPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_repr_nn_ZeroPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_repr_nn_ZeroPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_AdaptiveAvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_AdaptiveAvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_AdaptiveAvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_AdaptiveAvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_AdaptiveAvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_AdaptiveAvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_AdaptiveMaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_AdaptiveMaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_AdaptiveMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_AdaptiveMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_AdaptiveMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_AdaptiveMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_AvgPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_AvgPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_AvgPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_AvgPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_AvgPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_AvgPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_BCELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_BCELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_BCEWithLogitsLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_BCEWithLogitsLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_BatchNorm1d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_BatchNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_BatchNorm1d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_BatchNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_BatchNorm2d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_BatchNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_BatchNorm2d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_BatchNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_BatchNorm3d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_BatchNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_BatchNorm3d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_BatchNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Bilinear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Bilinear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_CELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_CELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_CTCLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_CTCLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_CircularPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_CircularPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_CircularPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_CircularPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_CircularPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_CircularPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConstantPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConstantPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConstantPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConstantPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConstantPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConstantPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Conv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Conv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Conv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Conv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Conv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Conv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConvTranspose1d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConvTranspose1d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConvTranspose1d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConvTranspose2d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConvTranspose2d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConvTranspose2d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConvTranspose3d_cuda_complex128, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConvTranspose3d_cuda_complex32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConvTranspose3d_cuda_complex64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_CosineEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_CosineEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_CrossEntropyLoss_cuda_float16, test/test_modules.py::TestModuleCUDA::test_save_load_nn_CrossEntropyLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_CrossEntropyLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Embedding_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Embedding_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_FractionalMaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_FractionalMaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_FractionalMaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_FractionalMaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_GELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_GELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_GLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_GLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_GRUCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_GRUCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_GRU_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_GRU_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_GRU_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_GRU_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_GaussianNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_GaussianNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_GroupNorm_cuda_bfloat16, test/test_modules.py::TestModuleCUDA::test_save_load_nn_GroupNorm_cuda_float16, test/test_modules.py::TestModuleCUDA::test_save_load_nn_GroupNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_GroupNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Hardshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Hardshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Hardswish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Hardswish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Hardtanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Hardtanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_HingeEmbeddingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_HingeEmbeddingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_HuberLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_HuberLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_InstanceNorm1d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_InstanceNorm1d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_InstanceNorm1d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_InstanceNorm1d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_InstanceNorm2d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_InstanceNorm2d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_InstanceNorm2d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_InstanceNorm2d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_InstanceNorm3d_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_InstanceNorm3d_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_InstanceNorm3d_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_InstanceNorm3d_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_KLDivLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_KLDivLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_L1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_L1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LPPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LPPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LPPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LPPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LPPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LPPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LSTMCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LSTMCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LSTM_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LSTM_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LSTM_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LSTM_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LayerNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LayerNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LazyConv1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LazyConv1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LazyConv2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LazyConv2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LazyConv3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LazyConv3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LazyConvTranspose1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LazyConvTranspose1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LazyConvTranspose2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LazyConvTranspose2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LazyConvTranspose3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LazyConvTranspose3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LeakyReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LeakyReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Linear_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Linear_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LocalResponseNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LocalResponseNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LogSigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LogSigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LogSoftmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_LogSoftmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MSELoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MSELoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MarginRankingLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MarginRankingLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MaxPool1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MaxPool1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MaxPool2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MaxPool2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MaxPool3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MaxPool3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Mish_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Mish_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MultiLabelMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MultiLabelMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MultiLabelSoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MultiLabelSoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MultiMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MultiMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MultiheadAttention_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MultiheadAttention_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MultiheadAttention_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_MultiheadAttention_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_NLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_NLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_PReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_PReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_PoissonNLLLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_PoissonNLLLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_RMSNorm_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_RMSNorm_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_RNNCell_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_RNNCell_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_RNN_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_RNN_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_RNN_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_RNN_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ReLU6_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ReLU6_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ReLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ReLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ReflectionPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ReflectionPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ReflectionPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ReflectionPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ReflectionPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ReflectionPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ReplicationPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ReplicationPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ReplicationPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ReplicationPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ReplicationPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ReplicationPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_SELU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_SELU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_SiLU_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_SiLU_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Sigmoid_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Sigmoid_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_SmoothL1Loss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_SmoothL1Loss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_SoftMarginLoss_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_SoftMarginLoss_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Softmax2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Softmax2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Softmax_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Softmax_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Softmin_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Softmin_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Softplus_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Softplus_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Softshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Softshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Softsign_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Softsign_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Tanh_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Tanh_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Tanhshrink_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Tanhshrink_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Threshold_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Threshold_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_TransformerDecoderLayer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_TransformerDecoderLayer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_TransformerEncoderLayer_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_TransformerEncoderLayer_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_TransformerEncoderLayer_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_TransformerEncoderLayer_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_TransformerEncoder_eval_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_TransformerEncoder_eval_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_TransformerEncoder_train_mode_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_TransformerEncoder_train_mode_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Transformer_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_Transformer_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ZeroPad1d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ZeroPad1d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ZeroPad2d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ZeroPad2d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ZeroPad3d_cuda_float32, test/test_modules.py::TestModuleCUDA::test_save_load_nn_ZeroPad3d_cuda_float64, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_AdaptiveAvgPool1d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_AdaptiveAvgPool1d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_AdaptiveAvgPool2d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_AdaptiveAvgPool2d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_AdaptiveAvgPool3d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_AdaptiveAvgPool3d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_AdaptiveMaxPool1d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_AdaptiveMaxPool1d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_AdaptiveMaxPool2d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_AdaptiveMaxPool2d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_AdaptiveMaxPool3d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_AdaptiveMaxPool3d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_AvgPool1d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_AvgPool1d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_AvgPool2d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_AvgPool2d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_AvgPool3d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_AvgPool3d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_BCELoss_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_BCELoss_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_BCEWithLogitsLoss_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_BCEWithLogitsLoss_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_BatchNorm1d_eval_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_BatchNorm1d_eval_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_BatchNorm1d_train_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_BatchNorm1d_train_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_BatchNorm2d_eval_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_BatchNorm2d_eval_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_BatchNorm2d_train_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_BatchNorm2d_train_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_BatchNorm3d_eval_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_BatchNorm3d_eval_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_BatchNorm3d_train_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_BatchNorm3d_train_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Bilinear_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Bilinear_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_CELU_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_CELU_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_CTCLoss_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_CTCLoss_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_CircularPad1d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_CircularPad1d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_CircularPad2d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_CircularPad2d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_CircularPad3d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_CircularPad3d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ConstantPad1d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ConstantPad1d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ConstantPad2d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ConstantPad2d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ConstantPad3d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ConstantPad3d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Conv1d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Conv1d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Conv2d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Conv2d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Conv3d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Conv3d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ConvTranspose1d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ConvTranspose1d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ConvTranspose2d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ConvTranspose2d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ConvTranspose3d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ConvTranspose3d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_CosineEmbeddingLoss_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_CosineEmbeddingLoss_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_CrossEntropyLoss_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_CrossEntropyLoss_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ELU_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ELU_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Embedding_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Embedding_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_FractionalMaxPool2d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_FractionalMaxPool2d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_FractionalMaxPool3d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_FractionalMaxPool3d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_GELU_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_GELU_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_GLU_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_GLU_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_GRUCell_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_GRUCell_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_GRU_eval_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_GRU_eval_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_GRU_train_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_GRU_train_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_GaussianNLLLoss_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_GaussianNLLLoss_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_GroupNorm_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_GroupNorm_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Hardshrink_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Hardshrink_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Hardswish_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Hardswish_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Hardtanh_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Hardtanh_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_HingeEmbeddingLoss_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_HingeEmbeddingLoss_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_HuberLoss_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_HuberLoss_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_InstanceNorm1d_eval_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_InstanceNorm1d_eval_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_InstanceNorm1d_train_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_InstanceNorm1d_train_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_InstanceNorm2d_eval_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_InstanceNorm2d_eval_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_InstanceNorm2d_train_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_InstanceNorm2d_train_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_InstanceNorm3d_eval_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_InstanceNorm3d_eval_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_InstanceNorm3d_train_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_InstanceNorm3d_train_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_KLDivLoss_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_KLDivLoss_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_L1Loss_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_L1Loss_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LPPool1d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LPPool1d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LPPool2d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LPPool2d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LPPool3d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LPPool3d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LSTMCell_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LSTMCell_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LSTM_eval_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LSTM_eval_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LSTM_train_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LSTM_train_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LayerNorm_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LayerNorm_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LeakyReLU_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LeakyReLU_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Linear_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Linear_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LocalResponseNorm_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LocalResponseNorm_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LogSigmoid_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LogSigmoid_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LogSoftmax_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_LogSoftmax_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MSELoss_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MSELoss_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MarginRankingLoss_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MarginRankingLoss_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MaxPool1d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MaxPool1d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MaxPool2d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MaxPool2d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MaxPool3d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MaxPool3d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Mish_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Mish_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MultiLabelMarginLoss_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MultiLabelMarginLoss_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MultiLabelSoftMarginLoss_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MultiLabelSoftMarginLoss_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MultiMarginLoss_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MultiMarginLoss_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MultiheadAttention_eval_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MultiheadAttention_eval_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MultiheadAttention_train_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_MultiheadAttention_train_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_NLLLoss_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_NLLLoss_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_PReLU_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_PReLU_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_PoissonNLLLoss_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_PoissonNLLLoss_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_RMSNorm_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_RMSNorm_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_RNNCell_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_RNNCell_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_RNN_eval_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_RNN_eval_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_RNN_train_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_RNN_train_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ReLU6_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ReLU6_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ReLU_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ReLU_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ReflectionPad1d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ReflectionPad1d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ReflectionPad2d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ReflectionPad2d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ReflectionPad3d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ReflectionPad3d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ReplicationPad1d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ReplicationPad1d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ReplicationPad2d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ReplicationPad2d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ReplicationPad3d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ReplicationPad3d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_SELU_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_SELU_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_SiLU_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_SiLU_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Sigmoid_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Sigmoid_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_SmoothL1Loss_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_SmoothL1Loss_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_SoftMarginLoss_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_SoftMarginLoss_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Softmax2d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Softmax2d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Softmax_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Softmax_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Softmin_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Softmin_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Softplus_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Softplus_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Softshrink_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Softshrink_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Softsign_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Softsign_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Tanh_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Tanh_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Tanhshrink_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Tanhshrink_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Threshold_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Threshold_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_TransformerDecoderLayer_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_TransformerDecoderLayer_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_TransformerEncoderLayer_eval_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_TransformerEncoderLayer_eval_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_TransformerEncoderLayer_train_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_TransformerEncoderLayer_train_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_TransformerEncoder_eval_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_TransformerEncoder_eval_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_TransformerEncoder_train_mode_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_TransformerEncoder_train_mode_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Transformer_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_Transformer_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ZeroPad1d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ZeroPad1d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ZeroPad2d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ZeroPad2d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ZeroPad3d_swap_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_empty_nn_ZeroPad3d_swap_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveAvgPool1d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveAvgPool1d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveAvgPool1d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveAvgPool1d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveAvgPool2d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveAvgPool2d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveAvgPool2d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveAvgPool2d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveAvgPool3d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveAvgPool3d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveAvgPool3d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveAvgPool3d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveMaxPool1d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveMaxPool1d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveMaxPool1d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveMaxPool1d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveMaxPool2d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveMaxPool2d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveMaxPool2d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveMaxPool2d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveMaxPool3d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveMaxPool3d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveMaxPool3d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AdaptiveMaxPool3d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AvgPool1d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AvgPool1d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AvgPool1d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AvgPool1d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AvgPool2d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AvgPool2d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AvgPool2d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AvgPool2d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AvgPool3d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AvgPool3d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AvgPool3d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_AvgPool3d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BCELoss_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BCELoss_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BCELoss_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BCELoss_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BCEWithLogitsLoss_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BCEWithLogitsLoss_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BCEWithLogitsLoss_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BCEWithLogitsLoss_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm1d_eval_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm1d_eval_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm1d_eval_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm1d_eval_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm1d_train_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm1d_train_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm1d_train_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm1d_train_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm2d_eval_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm2d_eval_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm2d_eval_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm2d_eval_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm2d_train_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm2d_train_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm2d_train_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm2d_train_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm3d_eval_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm3d_eval_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm3d_eval_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm3d_eval_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm3d_train_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm3d_train_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm3d_train_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_BatchNorm3d_train_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Bilinear_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Bilinear_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Bilinear_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Bilinear_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CELU_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CELU_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CELU_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CELU_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CTCLoss_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CTCLoss_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CTCLoss_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CTCLoss_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CircularPad1d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CircularPad1d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CircularPad1d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CircularPad1d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CircularPad2d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CircularPad2d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CircularPad2d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CircularPad2d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CircularPad3d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CircularPad3d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CircularPad3d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CircularPad3d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConstantPad1d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConstantPad1d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConstantPad1d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConstantPad1d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConstantPad2d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConstantPad2d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConstantPad2d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConstantPad2d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConstantPad3d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConstantPad3d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConstantPad3d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConstantPad3d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Conv1d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Conv1d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Conv1d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Conv1d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Conv2d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Conv2d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Conv2d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Conv2d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Conv3d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Conv3d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Conv3d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Conv3d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConvTranspose1d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConvTranspose1d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConvTranspose1d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConvTranspose1d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConvTranspose2d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConvTranspose2d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConvTranspose2d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConvTranspose2d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConvTranspose3d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConvTranspose3d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConvTranspose3d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ConvTranspose3d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CosineEmbeddingLoss_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CosineEmbeddingLoss_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CosineEmbeddingLoss_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CosineEmbeddingLoss_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CrossEntropyLoss_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CrossEntropyLoss_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CrossEntropyLoss_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_CrossEntropyLoss_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ELU_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ELU_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ELU_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ELU_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Embedding_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Embedding_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Embedding_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Embedding_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_FractionalMaxPool2d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_FractionalMaxPool2d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_FractionalMaxPool2d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_FractionalMaxPool2d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_FractionalMaxPool3d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_FractionalMaxPool3d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_FractionalMaxPool3d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_FractionalMaxPool3d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GELU_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GELU_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GELU_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GELU_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GLU_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GLU_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GLU_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GLU_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GRUCell_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GRUCell_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GRUCell_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GRUCell_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GRU_eval_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GRU_eval_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GRU_eval_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GRU_eval_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GRU_train_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GRU_train_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GRU_train_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GRU_train_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GaussianNLLLoss_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GaussianNLLLoss_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GaussianNLLLoss_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GaussianNLLLoss_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GroupNorm_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GroupNorm_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GroupNorm_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_GroupNorm_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Hardshrink_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Hardshrink_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Hardshrink_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Hardshrink_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Hardswish_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Hardswish_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Hardswish_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Hardswish_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Hardtanh_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Hardtanh_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Hardtanh_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Hardtanh_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_HingeEmbeddingLoss_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_HingeEmbeddingLoss_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_HingeEmbeddingLoss_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_HingeEmbeddingLoss_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_HuberLoss_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_HuberLoss_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_HuberLoss_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_HuberLoss_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm1d_eval_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm1d_eval_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm1d_eval_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm1d_eval_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm1d_train_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm1d_train_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm1d_train_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm1d_train_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm2d_eval_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm2d_eval_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm2d_eval_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm2d_eval_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm2d_train_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm2d_train_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm2d_train_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm2d_train_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm3d_eval_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm3d_eval_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm3d_eval_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm3d_eval_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm3d_train_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm3d_train_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm3d_train_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_InstanceNorm3d_train_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_KLDivLoss_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_KLDivLoss_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_KLDivLoss_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_KLDivLoss_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_L1Loss_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_L1Loss_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_L1Loss_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_L1Loss_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LPPool1d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LPPool1d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LPPool1d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LPPool1d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LPPool2d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LPPool2d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LPPool2d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LPPool2d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LPPool3d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LPPool3d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LPPool3d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LPPool3d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LSTMCell_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LSTMCell_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LSTMCell_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LSTMCell_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LSTM_eval_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LSTM_eval_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LSTM_eval_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LSTM_eval_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LSTM_train_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LSTM_train_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LSTM_train_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LSTM_train_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LayerNorm_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LayerNorm_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LayerNorm_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LayerNorm_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LeakyReLU_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LeakyReLU_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LeakyReLU_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LeakyReLU_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Linear_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Linear_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Linear_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Linear_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LocalResponseNorm_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LocalResponseNorm_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LocalResponseNorm_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LocalResponseNorm_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LogSigmoid_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LogSigmoid_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LogSigmoid_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LogSigmoid_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LogSoftmax_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LogSoftmax_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LogSoftmax_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_LogSoftmax_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MSELoss_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MSELoss_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MSELoss_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MSELoss_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MarginRankingLoss_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MarginRankingLoss_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MarginRankingLoss_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MarginRankingLoss_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MaxPool1d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MaxPool1d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MaxPool1d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MaxPool1d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MaxPool2d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MaxPool2d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MaxPool2d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MaxPool2d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MaxPool3d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MaxPool3d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MaxPool3d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MaxPool3d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Mish_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Mish_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Mish_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Mish_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiLabelMarginLoss_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiLabelMarginLoss_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiLabelMarginLoss_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiLabelMarginLoss_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiLabelSoftMarginLoss_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiLabelSoftMarginLoss_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiLabelSoftMarginLoss_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiLabelSoftMarginLoss_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiMarginLoss_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiMarginLoss_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiMarginLoss_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiMarginLoss_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiheadAttention_eval_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiheadAttention_eval_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiheadAttention_eval_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiheadAttention_eval_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiheadAttention_train_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiheadAttention_train_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiheadAttention_train_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_MultiheadAttention_train_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_NLLLoss_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_NLLLoss_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_NLLLoss_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_NLLLoss_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_PReLU_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_PReLU_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_PReLU_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_PReLU_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_PoissonNLLLoss_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_PoissonNLLLoss_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_PoissonNLLLoss_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_PoissonNLLLoss_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_RMSNorm_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_RMSNorm_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_RMSNorm_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_RMSNorm_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_RNNCell_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_RNNCell_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_RNNCell_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_RNNCell_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_RNN_eval_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_RNN_eval_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_RNN_eval_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_RNN_eval_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_RNN_train_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_RNN_train_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_RNN_train_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_RNN_train_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReLU6_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReLU6_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReLU6_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReLU6_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReLU_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReLU_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReLU_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReLU_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReflectionPad1d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReflectionPad1d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReflectionPad1d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReflectionPad1d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReflectionPad2d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReflectionPad2d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReflectionPad2d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReflectionPad2d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReflectionPad3d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReflectionPad3d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReflectionPad3d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReflectionPad3d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReplicationPad1d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReplicationPad1d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReplicationPad1d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReplicationPad1d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReplicationPad2d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReplicationPad2d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReplicationPad2d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReplicationPad2d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReplicationPad3d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReplicationPad3d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReplicationPad3d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ReplicationPad3d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_SELU_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_SELU_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_SELU_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_SELU_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_SiLU_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_SiLU_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_SiLU_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_SiLU_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Sigmoid_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Sigmoid_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Sigmoid_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Sigmoid_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_SmoothL1Loss_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_SmoothL1Loss_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_SmoothL1Loss_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_SmoothL1Loss_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_SoftMarginLoss_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_SoftMarginLoss_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_SoftMarginLoss_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_SoftMarginLoss_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softmax2d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softmax2d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softmax2d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softmax2d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softmax_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softmax_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softmax_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softmax_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softmin_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softmin_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softmin_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softmin_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softplus_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softplus_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softplus_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softplus_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softshrink_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softshrink_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softshrink_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softshrink_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softsign_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softsign_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softsign_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Softsign_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Tanh_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Tanh_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Tanh_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Tanh_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Tanhshrink_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Tanhshrink_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Tanhshrink_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Tanhshrink_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Threshold_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Threshold_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Threshold_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Threshold_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerDecoderLayer_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerDecoderLayer_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerDecoderLayer_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerDecoderLayer_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerEncoderLayer_eval_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerEncoderLayer_eval_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerEncoderLayer_eval_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerEncoderLayer_eval_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerEncoderLayer_train_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerEncoderLayer_train_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerEncoderLayer_train_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerEncoderLayer_train_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerEncoder_eval_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerEncoder_eval_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerEncoder_eval_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerEncoder_eval_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerEncoder_train_mode_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerEncoder_train_mode_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerEncoder_train_mode_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_TransformerEncoder_train_mode_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Transformer_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Transformer_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Transformer_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_Transformer_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ZeroPad1d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ZeroPad1d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ZeroPad1d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ZeroPad1d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ZeroPad2d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ZeroPad2d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ZeroPad2d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ZeroPad2d_swap_True_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ZeroPad3d_swap_False_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ZeroPad3d_swap_False_set_grad_True_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ZeroPad3d_swap_True_set_grad_False_cuda_float32, test/test_modules.py::TestModuleCUDA::test_to_nn_ZeroPad3d_swap_True_set_grad_True_cuda_float32 2025-12-04T16:58:04.5297850Z 2025-12-04T16:58:04.5298192Z Finished test_modules 1/1 ... [2025-12-04 16:58:04.190813][27867.881199429], took 7.49min 2025-12-04T16:58:04.5299305Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_modules/test_modules-5eb6bc8f38dae4a5.xml 2025-12-04T16:58:04.5300479Z Running complex_tensor/test_complex_tensor 1/1 ... [2025-12-04 16:58:04.417601][27868.107989864] 2025-12-04T16:58:04.5301092Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T16:58:04.5302378Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'complex_tensor/test_complex_tensor.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 16:58:04.418068] 2025-12-04T17:09:05.1501456Z 2025-12-04T17:09:05.1503363Z complex_tensor/test_complex_tensor 1/1 was successful, full logs can be found in artifacts with path test/test-reports/complex_tensor.test_complex_tensor_1.1_4feb438836f89d3d_.log 2025-12-04T17:09:05.2085449Z Running 578 items in this shard: test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_abs_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_abs_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_abs_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_acos_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_acos_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_acos_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_acosh_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_acosh_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_acosh_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_add_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_add_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_add_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_addmm_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_addmm_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_addmm_decomposed_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_addmm_decomposed_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_all_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_all_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_allclose_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_allclose_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_angle_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_angle_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_angle_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_any_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_any_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_asin_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_asin_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_asin_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_asinh_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_asinh_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_asinh_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_atan_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_atan_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_atan_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_atanh_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_atanh_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_atanh_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_bmm_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_bmm_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_cat_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_cat_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_cat_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_clone_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_clone_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_clone_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_conj_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_conj_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_conj_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_conj_physical_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_conj_physical_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_conj_physical_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_constant_pad_nd_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_constant_pad_nd_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_cos_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_cos_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_cos_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_cosh_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_cosh_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_cosh_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_cumprod_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_cumprod_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_cumsum_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_cumsum_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_diagonal_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_diagonal_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_diagonal_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_diagonal_scatter_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_diagonal_scatter_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_div_no_rounding_mode_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_div_no_rounding_mode_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_div_no_rounding_mode_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_dot_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_dot_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_empty_like_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_empty_like_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_empty_like_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_eq_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_eq_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_eq_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_exp_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_exp_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_exp_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_expand_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_expand_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_expm1_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_expm1_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_flatten_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_flatten_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_flatten_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_flip_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_flip_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_full_like_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_full_like_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_gather_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_gather_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_imag_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_imag_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_imag_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_index_add_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_index_add_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_index_add_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_index_select_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_index_select_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_index_select_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_isclose_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_isclose_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_isfinite_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_isfinite_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_isfinite_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_isinf_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_isinf_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_isinf_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_isnan_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_isnan_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_log1p_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_log1p_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_log_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_log_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_log_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_logical_and_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_logical_and_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_logical_not_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_logical_not_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_logical_or_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_logical_or_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_logical_xor_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_logical_xor_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_masked_fill_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_masked_fill_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_masked_fill_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_masked_scatter_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_masked_scatter_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_mean_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_mean_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_mm_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_mm_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_mul_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_mul_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_mul_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_ne_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_ne_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_neg_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_neg_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_neg_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_new_zeros_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_new_zeros_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_new_zeros_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_permute_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_permute_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_permute_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_pow_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_pow_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_pow_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_prod_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_prod_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_prod_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_randn_like_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_randn_like_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_randn_like_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_real_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_real_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_real_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_reciprocal_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_reciprocal_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_repeat_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_repeat_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_rsqrt_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_rsqrt_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_rsqrt_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_rsub_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_rsub_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_scatter_add_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_scatter_add_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_select_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_select_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_select_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_sgn_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_sgn_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_sgn_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_sin_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_sin_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_sin_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_sinh_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_sinh_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_sinh_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_slice_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_slice_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_slice_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_split_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_split_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_split_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_split_list_args_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_split_list_args_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_split_with_sizes_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_split_with_sizes_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_split_with_sizes_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_sqrt_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_sqrt_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_sqrt_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_squeeze_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_squeeze_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_squeeze_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_squeeze_multiple_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_squeeze_multiple_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_squeeze_multiple_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_stack_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_stack_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_stack_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_sub_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_sub_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_sub_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_sum_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_sum_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_sum_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_t_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_t_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_tan_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_tan_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_tan_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_tanh_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_tanh_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_tanh_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_to_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_to_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_transpose_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_transpose_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_transpose_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_true_divide_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_true_divide_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_true_divide_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_unsqueeze_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_unsqueeze_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_unsqueeze_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_var_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_var_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_var_unbiased_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_var_unbiased_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_view_as_real_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_view_as_real_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_view_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_view_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_view_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_where_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_where_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_where_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_zero__cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_zero__cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_zeros_like_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_zeros_like_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexTensorCUDA::test_consistency_zeros_like_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_abs_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_acos_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_acosh_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_add_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_addmm_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_addmm_decomposed_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_angle_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_asin_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_asinh_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_atan_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_atanh_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_bmm_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_cat_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_clone_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_conj_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_conj_physical_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_constant_pad_nd_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_cos_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_cosh_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_cumprod_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_cumsum_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_diagonal_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_diagonal_scatter_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_div_no_rounding_mode_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_dot_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_exp_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_expand_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_expm1_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_flatten_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_flip_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_gather_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_imag_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_index_add_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_index_select_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_log1p_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_log_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_masked_fill_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_masked_scatter_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_mean_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_mm_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_mul_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_neg_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_permute_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_pow_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_prod_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_real_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_reciprocal_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_repeat_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_rsqrt_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_rsub_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_scatter_add_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_select_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_sgn_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_sin_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_sinh_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_slice_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_split_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_split_list_args_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_split_with_sizes_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_sqrt_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_squeeze_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_squeeze_multiple_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_stack_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_sub_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_sum_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_t_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_tan_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_tanh_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_to_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_transpose_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_true_divide_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_unsqueeze_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_var_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_var_unbiased_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_view_as_real_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_view_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_where_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexBwdGradientsCUDA::test_fn_grad_zero__cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_abs_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_abs_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_abs_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_acos_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_acos_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_acos_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_acosh_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_acosh_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_acosh_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_add_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_add_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_add_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_addmm_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_addmm_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_addmm_decomposed_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_addmm_decomposed_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_all_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_all_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_allclose_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_allclose_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_angle_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_angle_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_angle_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_any_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_any_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_asin_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_asin_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_asin_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_asinh_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_asinh_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_asinh_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_atan_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_atan_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_atan_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_atanh_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_atanh_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_atanh_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_bmm_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_bmm_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_cat_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_cat_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_cat_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_clone_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_clone_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_clone_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_conj_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_conj_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_conj_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_conj_physical_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_conj_physical_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_conj_physical_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_constant_pad_nd_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_constant_pad_nd_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_cos_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_cos_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_cos_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_cosh_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_cosh_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_cosh_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_cumprod_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_cumprod_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_cumsum_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_cumsum_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_diagonal_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_diagonal_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_diagonal_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_diagonal_scatter_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_diagonal_scatter_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_div_no_rounding_mode_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_div_no_rounding_mode_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_div_no_rounding_mode_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_dot_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_dot_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_empty_like_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_empty_like_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_empty_like_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_eq_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_eq_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_eq_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_exp_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_exp_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_exp_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_expand_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_expand_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_expm1_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_expm1_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_flatten_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_flatten_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_flatten_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_flip_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_flip_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_full_like_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_full_like_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_gather_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_gather_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_imag_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_imag_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_imag_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_index_add_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_index_add_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_index_add_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_index_select_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_index_select_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_index_select_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_isclose_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_isclose_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_isfinite_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_isfinite_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_isfinite_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_isinf_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_isinf_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_isinf_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_isnan_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_isnan_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_log1p_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_log1p_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_log_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_log_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_log_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_logical_and_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_logical_and_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_logical_not_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_logical_not_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_logical_or_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_logical_or_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_logical_xor_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_logical_xor_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_masked_fill_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_masked_fill_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_masked_fill_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_masked_scatter_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_masked_scatter_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_mean_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_mean_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_mm_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_mm_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_mul_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_mul_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_mul_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_ne_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_ne_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_neg_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_neg_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_neg_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_new_zeros_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_new_zeros_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_new_zeros_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_permute_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_permute_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_permute_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_pow_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_pow_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_pow_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_prod_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_prod_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_prod_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_randn_like_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_randn_like_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_randn_like_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_real_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_real_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_real_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_reciprocal_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_reciprocal_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_repeat_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_repeat_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_rsqrt_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_rsqrt_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_rsqrt_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_rsub_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_rsub_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_scatter_add_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_scatter_add_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_select_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_select_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_select_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_sgn_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_sgn_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_sgn_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_sin_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_sin_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_sin_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_sinh_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_sinh_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_sinh_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_slice_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_slice_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_slice_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_split_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_split_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_split_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_split_list_args_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_split_list_args_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_split_with_sizes_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_split_with_sizes_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_split_with_sizes_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_sqrt_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_sqrt_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_sqrt_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_squeeze_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_squeeze_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_squeeze_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_squeeze_multiple_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_squeeze_multiple_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_squeeze_multiple_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_stack_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_stack_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_stack_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_sub_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_sub_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_sub_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_sum_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_sum_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_sum_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_t_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_t_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_tan_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_tan_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_tan_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_tanh_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_tanh_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_tanh_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_to_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_to_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_transpose_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_transpose_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_transpose_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_true_divide_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_true_divide_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_true_divide_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_unsqueeze_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_unsqueeze_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_unsqueeze_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_var_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_var_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_var_unbiased_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_var_unbiased_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_view_as_real_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_view_as_real_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_view_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_view_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_view_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_where_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_where_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_where_cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_zero__cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_zero__cuda_complex64, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_zeros_like_cuda_complex128, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_zeros_like_cuda_complex32, test/complex_tensor/test_complex_tensor.py::TestComplexDistributedCUDA::test_distributed_zeros_like_cuda_complex64 2025-12-04T17:09:05.2408425Z 2025-12-04T17:09:05.2408889Z Finished complex_tensor/test_complex_tensor 1/1 ... [2025-12-04 17:09:05.151336][28528.841725813], took 11.01min 2025-12-04T17:09:05.2410346Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/complex_tensor.test_complex_tensor/complex_tensor.test_complex_tensor-00d0e28f118bdf92.xml 2025-12-04T17:09:05.3115402Z Running torch_np/numpy_tests/core/test_indexing 1/1 ... [2025-12-04 17:09:05.311245][28529.001635117] 2025-12-04T17:09:05.3116195Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T17:09:05.3119208Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/core/test_indexing.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 17:09:05.311654] 2025-12-04T17:09:10.9841467Z 2025-12-04T17:09:10.9842660Z torch_np/numpy_tests/core/test_indexing 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.core.test_indexing_1.1_33940d3a3f5ad817_.log 2025-12-04T17:09:10.9877436Z Running 67 items in this shard: test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_boolean_assignment_value_mismatch, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_boolean_indexing_list, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_boolean_indexing_onedim, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_boolean_indexing_twodim, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_boolean_shape_mismatch, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_broaderrors_indexing, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_broken_sequence_not_nd_index, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_ellipsis_index, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_ellipsis_index_2, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_empty_fancy_index, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_empty_tuple_index, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_everything_returns_views, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_index_no_array_to_index, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_index_no_floats, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_indexing_array_negative_strides, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_indexing_array_weird_strides, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_memory_order, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_none_index, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_nontuple_ndindex, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_reverse_strides_and_subspace_bufferinit, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_reversed_strides_result_allocation, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_same_kind_index_casting, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_scalar_array_bool, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_single_bool_index, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_single_int_index, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_slicing_no_floats, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_small_regressions, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_too_many_advanced_indices_index2_num_32_original_ndim_1, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_too_many_advanced_indices_index2_num_32_original_ndim_32, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_too_many_advanced_indices_index2_num_40_original_ndim_1, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_too_many_advanced_indices_index2_num_40_original_ndim_32, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_too_many_advanced_indices_index_False_num_32_original_ndim_1, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_too_many_advanced_indices_index_False_num_32_original_ndim_32, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_too_many_advanced_indices_index_False_num_40_original_ndim_1, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_too_many_advanced_indices_index_False_num_40_original_ndim_32, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_too_many_advanced_indices_index_True_num_32_original_ndim_1, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_too_many_advanced_indices_index_True_num_32_original_ndim_32, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_too_many_advanced_indices_index_True_num_40_original_ndim_1, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_too_many_advanced_indices_index_True_num_40_original_ndim_32, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_too_many_fancy_indices_special_case, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_trivial_fancy_not_possible, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_trivial_fancy_out_of_bounds, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_tuple_subclass, test/torch_np/numpy_tests/core/test_indexing.py::TestIndexing::test_uncontiguous_subspace_assignment, test/torch_np/numpy_tests/core/test_indexing.py::TestBroadcastedAssignments::test_broadcast_error_reports_correct_shape_index0, test/torch_np/numpy_tests/core/test_indexing.py::TestBroadcastedAssignments::test_broadcast_error_reports_correct_shape_index1, test/torch_np/numpy_tests/core/test_indexing.py::TestBroadcastedAssignments::test_broadcast_error_reports_correct_shape_index2, test/torch_np/numpy_tests/core/test_indexing.py::TestBroadcastedAssignments::test_broadcast_subspace, test/torch_np/numpy_tests/core/test_indexing.py::TestBroadcastedAssignments::test_index_is_larger, test/torch_np/numpy_tests/core/test_indexing.py::TestBroadcastedAssignments::test_prepend_not_one, test/torch_np/numpy_tests/core/test_indexing.py::TestBroadcastedAssignments::test_prepending_ones, test/torch_np/numpy_tests/core/test_indexing.py::TestBroadcastedAssignments::test_simple_broadcasting_errors, test/torch_np/numpy_tests/core/test_indexing.py::TestFancyIndexingCast::test_boolean_index_cast_assign, test/torch_np/numpy_tests/core/test_indexing.py::TestMultiIndexingAutomated::test_1d, test/torch_np/numpy_tests/core/test_indexing.py::TestMultiIndexingAutomated::test_boolean, test/torch_np/numpy_tests/core/test_indexing.py::TestMultiIndexingAutomated::test_multidim, test/torch_np/numpy_tests/core/test_indexing.py::TestFloatNonIntegerArgument::test_non_integer_argument_errors, test/torch_np/numpy_tests/core/test_indexing.py::TestFloatNonIntegerArgument::test_non_integer_sequence_multiplication, test/torch_np/numpy_tests/core/test_indexing.py::TestFloatNonIntegerArgument::test_reduce_axis_float_index, test/torch_np/numpy_tests/core/test_indexing.py::TestFloatNonIntegerArgument::test_valid_indexing, test/torch_np/numpy_tests/core/test_indexing.py::TestFloatNonIntegerArgument::test_valid_slicing, test/torch_np/numpy_tests/core/test_indexing.py::TestBooleanIndexing::test_bool_as_int_argument_errors, test/torch_np/numpy_tests/core/test_indexing.py::TestBooleanIndexing::test_boolean_indexing_fast_path, test/torch_np/numpy_tests/core/test_indexing.py::TestBooleanIndexing::test_boolean_indexing_weirdness, test/torch_np/numpy_tests/core/test_indexing.py::TestArrayToIndexDeprecation::test_array_to_index_error, test/torch_np/numpy_tests/core/test_indexing.py::TestNonIntegerArrayLike::test_basic, test/torch_np/numpy_tests/core/test_indexing.py::TestMultipleEllipsisError::test_basic 2025-12-04T17:09:10.9911558Z 2025-12-04T17:09:10.9911986Z Finished torch_np/numpy_tests/core/test_indexing 1/1 ... [2025-12-04 17:09:10.984034][28534.674428001], took 0.09min 2025-12-04T17:09:11.0371406Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.core.test_indexing/torch_np.numpy_tests.core.test_indexing-3b590877a50ad13d.xml 2025-12-04T17:09:11.0719275Z Running test_futures 1/1 ... [2025-12-04 17:09:11.071680][28534.762073104] 2025-12-04T17:09:11.0719888Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T17:09:11.0722760Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_futures.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 17:09:11.072046] 2025-12-04T17:09:17.0450404Z 2025-12-04T17:09:17.0451336Z test_futures 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_futures_1.1_2500c18a32eb03f9_.log 2025-12-04T17:09:17.0458362Z Running 22 items in this shard: test/test_futures.py::TestFuture::test_add_done_callback_error_is_ignored, test/test_futures.py::TestFuture::test_add_done_callback_maintains_callback_order, test/test_futures.py::TestFuture::test_add_done_callback_no_arg_error_is_ignored, test/test_futures.py::TestFuture::test_add_done_callback_simple, test/test_futures.py::TestFuture::test_chained_then, test/test_futures.py::TestFuture::test_collect_all, test/test_futures.py::TestFuture::test_done, test/test_futures.py::TestFuture::test_done_exception, test/test_futures.py::TestFuture::test_interleaving_then_and_add_done_callback_maintains_callback_order, test/test_futures.py::TestFuture::test_interleaving_then_and_add_done_callback_propagates_error, test/test_futures.py::TestFuture::test_mark_future_twice, test/test_futures.py::TestFuture::test_pickle_future, test/test_futures.py::TestFuture::test_set_exception, test/test_futures.py::TestFuture::test_set_exception_multithreading, test/test_futures.py::TestFuture::test_then, test/test_futures.py::TestFuture::test_then_no_arg, test/test_futures.py::TestFuture::test_then_raise, test/test_futures.py::TestFuture::test_then_wrong_arg, test/test_futures.py::TestFuture::test_wait, test/test_futures.py::TestFuture::test_wait_all, test/test_futures.py::TestFuture::test_wait_multi_thread, test/test_futures.py::TestFuture::test_wait_none 2025-12-04T17:09:17.0465272Z 2025-12-04T17:09:17.0465551Z Finished test_futures 1/1 ... [2025-12-04 17:09:17.044917][28540.73531009], took 0.10min 2025-12-04T17:09:17.0981598Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_futures/test_futures-8abfe477c1d087d4.xml 2025-12-04T17:09:17.1688240Z Running test_tensor_creation_ops 1/1 ... [2025-12-04 17:09:17.168543][28540.858934727] 2025-12-04T17:09:17.1688803Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T17:09:17.1691912Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_tensor_creation_ops.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 17:09:17.168948] 2025-12-04T17:10:38.2956072Z 2025-12-04T17:10:38.2957437Z test_tensor_creation_ops 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_tensor_creation_ops_1.1_8ce9396a5d7487a9_.log 2025-12-04T17:10:38.3199696Z Running 536 items in this shard: test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_arange_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_arange_device_vs_cpu_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_arange_device_vs_cpu_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_arange_device_vs_cpu_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_arange_device_vs_cpu_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_arange_inference_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_arange_lowp_cuda_bfloat16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_arange_lowp_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_as_strided_neg_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_as_tensor_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_block_diag_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_block_diag_scipy_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cartesian_prod_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat2_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat2_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat2_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_all_dtypes_and_devices_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_big_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_channels_last_large_inputs_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_empty_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_empty_legacy_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_in_channels_last_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_large_tensor_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_mem_overlap_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_misaligned_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_multi_batch_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_out_channels_last_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_out_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_out_fast_path_dim0_dim1_cuda_complex128, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_out_fast_path_dim0_dim1_cuda_complex64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_out_fast_path_dim0_dim1_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_out_fast_path_dim0_dim1_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_out_fast_path_dim0_dim1_cuda_int16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_out_fast_path_dim0_dim1_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_out_fast_path_dim0_dim1_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_out_fast_path_dim0_dim1_cuda_int8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_out_fast_path_dim0_dim1_cuda_uint16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_out_fast_path_dim0_dim1_cuda_uint32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_out_fast_path_dim0_dim1_cuda_uint64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_out_fast_path_dim0_dim1_cuda_uint8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_out_memory_format_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_preserve_channels_last_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_size1_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_stack_cross_devices_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_cat_trailing_dim_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_combinations_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_complex_type_conversions_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_concat_empty_list_error_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_constructor_device_legacy_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_constructor_dtypes_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_ctor_with_numpy_array_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_device_rounding_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_device_rounding_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_device_rounding_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_diag_embed_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_diagflat_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_dsplit_cuda_complex64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_dsplit_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_dsplit_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_dstack_cuda_complex128, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_dstack_cuda_complex64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_dstack_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_dstack_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_dstack_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_dstack_cuda_int16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_dstack_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_dstack_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_dstack_cuda_int8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_dstack_cuda_uint8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_empty_full_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_empty_overflow_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_empty_strided_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_empty_tensor_props_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_eye_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_fill_all_dtypes_and_devices_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_float_to_int_conversion_finite_cuda_bool, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_float_to_int_conversion_finite_cuda_int16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_float_to_int_conversion_finite_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_float_to_int_conversion_finite_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_float_to_int_conversion_finite_cuda_int8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_float_to_int_conversion_finite_cuda_uint8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_float_to_int_conversion_nonfinite_cuda_bool, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_float_to_int_conversion_nonfinite_cuda_int16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_float_to_int_conversion_nonfinite_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_float_to_int_conversion_nonfinite_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_float_to_int_conversion_nonfinite_cuda_int8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_float_to_int_conversion_nonfinite_cuda_uint8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_from_file_shared_False_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_from_file_shared_True_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_full_inference_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_full_inference_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_full_inference_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_full_out_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_hsplit_cuda_complex64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_hsplit_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_hsplit_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_hstack_column_stack_cuda_complex128, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_hstack_column_stack_cuda_complex64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_hstack_column_stack_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_hstack_column_stack_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_hstack_column_stack_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_hstack_column_stack_cuda_int16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_hstack_column_stack_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_hstack_column_stack_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_hstack_column_stack_cuda_int8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_hstack_column_stack_cuda_uint8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_kaiser_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_kaiser_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_kaiser_window_cuda_bfloat16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_kaiser_window_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_kaiser_window_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_kaiser_window_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_kaiser_window_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_large_linspace_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_large_linspace_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_like_fn_stride_proparation_vs_tensoriterator_unary_op_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linlogspace_mem_overlap_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_cuda_bfloat16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_cuda_complex128, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_cuda_complex64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_cuda_int16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_cuda_int8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_cuda_uint8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_deduction_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_device_vs_cpu_cuda_bfloat16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_device_vs_cpu_cuda_complex128, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_device_vs_cpu_cuda_complex64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_device_vs_cpu_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_device_vs_cpu_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_device_vs_cpu_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_special_steps_cuda_bfloat16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_special_steps_cuda_complex128, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_special_steps_cuda_complex64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_special_steps_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_special_steps_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_special_steps_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_vs_numpy_complex_cuda_complex64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_vs_numpy_cuda_complex128, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_vs_numpy_cuda_complex64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_vs_numpy_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_vs_numpy_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_vs_numpy_integral_cuda_int16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_vs_numpy_integral_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_vs_numpy_integral_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_vs_numpy_integral_cuda_int8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_linspace_vs_numpy_integral_cuda_uint8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_base2_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_base2_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_base2_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_cuda_bfloat16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_cuda_int16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_cuda_int8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_cuda_uint8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_deduction_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_device_vs_cpu_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_device_vs_cpu_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_device_vs_cpu_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_special_steps_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_special_steps_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_special_steps_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_vs_numpy_complex_cuda_complex64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_vs_numpy_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_logspace_vs_numpy_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_meshgrid_default_indexing_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_meshgrid_empty_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_meshgrid_ij_indexing_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_meshgrid_ij_indexing_is_default_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_meshgrid_inconsistent_device_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_meshgrid_inconsistent_dtype_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_meshgrid_non_1d_tensor_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_meshgrid_unsupported_indexing_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_meshgrid_vs_numpy_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_meshgrid_warns_if_no_indexing_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_meshgrid_xy_indexing_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_new_empty_strided_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_new_methods_requires_grad_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_new_tensor_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_new_tensor_device_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_offset_scalar_cast_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_ones_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_bool_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_cuda_int16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_cuda_int8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_default_cuda_bfloat16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_default_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_default_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_default_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_default_cuda_int16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_default_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_default_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_default_cuda_int8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_default_cuda_uint8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_from_to_bool_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_from_to_cuda_bfloat16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_from_to_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_from_to_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_from_to_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_from_to_cuda_int16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_from_to_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_from_to_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_from_to_cuda_int8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_from_to_cuda_uint16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_from_to_cuda_uint32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_from_to_cuda_uint8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_full_range_cuda_bfloat16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_full_range_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_full_range_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_full_range_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_full_range_cuda_int16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_full_range_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_full_range_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_full_range_cuda_int8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_full_range_cuda_uint16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_full_range_cuda_uint32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_full_range_cuda_uint8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_to_cuda_bfloat16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_to_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_to_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_to_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_to_cuda_int16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_to_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_to_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_to_cuda_int8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_to_cuda_uint16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_to_cuda_uint32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_random_to_cuda_uint8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_range_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_range_factories_64bit_indexing_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_range_warning_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_refs_tensor_cuda_bfloat16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_refs_tensor_cuda_bool, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_refs_tensor_cuda_complex128, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_refs_tensor_cuda_complex64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_refs_tensor_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_refs_tensor_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_refs_tensor_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_refs_tensor_cuda_int16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_refs_tensor_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_refs_tensor_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_refs_tensor_cuda_int8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_refs_tensor_cuda_uint8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_repeat_interleave_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_roll_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_bartlett_cuda_bfloat16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_bartlett_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_bartlett_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_bartlett_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_bartlett_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_blackman_cuda_bfloat16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_blackman_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_blackman_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_blackman_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_blackman_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_hamming_cuda_bfloat16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_hamming_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_hamming_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_hamming_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_hamming_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_hann_cuda_bfloat16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_hann_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_hann_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_hann_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_window_functions_window_hann_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_windows_functions_window_bartlett_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_windows_functions_window_bartlett_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_windows_functions_window_blackman_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_windows_functions_window_blackman_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_windows_functions_window_cosine_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_windows_functions_window_cosine_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_windows_functions_window_hamming_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_windows_functions_window_hamming_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_windows_functions_window_hann_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_windows_functions_window_hann_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_windows_functions_window_nuttall_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_signal_windows_functions_window_nuttall_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_simple_scalar_cast_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_stack_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_stack_out_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_storage_filename_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_strided_mismatched_stride_shape_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_tensor_ctor_device_inference_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_tensor_device_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_tensor_factories_empty_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_tensor_factory_copy_var_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_tensor_factory_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_tensor_factory_gpu_type_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_tensor_factory_gpu_type_inference_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_tensor_factory_type_inference_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_tensor_from_non_writable_numpy_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_tensor_from_sequence_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_torch_complex_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_torch_complex_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_torch_complex_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_torch_complex_floating_dtype_error_cuda_bool, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_torch_complex_floating_dtype_error_cuda_complex128, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_torch_complex_floating_dtype_error_cuda_complex64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_torch_complex_floating_dtype_error_cuda_int16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_torch_complex_floating_dtype_error_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_torch_complex_floating_dtype_error_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_torch_complex_floating_dtype_error_cuda_int8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_torch_complex_floating_dtype_error_cuda_uint8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_torch_complex_out_dtype_error_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_torch_complex_out_dtype_error_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_torch_complex_same_dtype_error_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_torch_complex_same_dtype_error_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_torch_polar_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_torch_polar_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_unpack_double_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_unpack_double_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vander_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vander_types_cuda_bool, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vander_types_cuda_complex128, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vander_types_cuda_complex64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vander_types_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vander_types_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vander_types_cuda_int16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vander_types_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vander_types_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vander_types_cuda_int8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vander_types_cuda_uint8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vsplit_cuda_complex64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vsplit_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vsplit_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vstack_row_stack_cuda_complex128, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vstack_row_stack_cuda_complex64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vstack_row_stack_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vstack_row_stack_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vstack_row_stack_cuda_float64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vstack_row_stack_cuda_int16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vstack_row_stack_cuda_int32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vstack_row_stack_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vstack_row_stack_cuda_int8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_vstack_row_stack_cuda_uint8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_zeros_bounds_checking_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_zeros_cuda, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_zeros_dtype_layout_device_match_cuda_bool, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_zeros_dtype_layout_device_match_cuda_complex64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_zeros_dtype_layout_device_match_cuda_float16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_zeros_dtype_layout_device_match_cuda_float32, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_zeros_dtype_layout_device_match_cuda_int16, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_zeros_dtype_layout_device_match_cuda_int64, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_zeros_dtype_layout_device_match_cuda_uint8, test/test_tensor_creation_ops.py::TestTensorCreationCUDA::test_zeros_out_cuda, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_normal_cuda_float32, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_normal_cuda_float64, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_normal_std_error_cuda, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_rand_cuda_complex128, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_rand_cuda_complex32, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_rand_cuda_complex64, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_rand_cuda_float32, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_rand_cuda_float64, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_randint_cuda, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_randint_distribution_cuda, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_randint_inference_cuda, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_randn_cuda_bfloat16, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_randn_cuda_complex128, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_randn_cuda_complex32, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_randn_cuda_complex64, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_randn_cuda_float16, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_randn_cuda_float32, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_randn_cuda_float64, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_random_neg_values_cuda, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_randperm_cuda, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_randperm_device_compatibility_cuda, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_randperm_large_cuda, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_uniform_from_to_cuda_bfloat16, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_uniform_from_to_cuda_float16, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_uniform_from_to_cuda_float32, test/test_tensor_creation_ops.py::TestRandomTensorCreationCUDA::test_uniform_from_to_cuda_float64, test/test_tensor_creation_ops.py::TestLikeTensorCreationCUDA::test_empty_like_cuda, test/test_tensor_creation_ops.py::TestLikeTensorCreationCUDA::test_full_like_inference_cuda, test/test_tensor_creation_ops.py::TestLikeTensorCreationCUDA::test_ones_like_cuda, test/test_tensor_creation_ops.py::TestLikeTensorCreationCUDA::test_ones_like_multiple_device_cuda, test/test_tensor_creation_ops.py::TestLikeTensorCreationCUDA::test_rand_like_cuda, test/test_tensor_creation_ops.py::TestLikeTensorCreationCUDA::test_randint_like_cuda, test/test_tensor_creation_ops.py::TestLikeTensorCreationCUDA::test_randn_like_cuda, test/test_tensor_creation_ops.py::TestLikeTensorCreationCUDA::test_zeros_like_cuda, test/test_tensor_creation_ops.py::TestLikeTensorCreationCUDA::test_zeros_like_multiple_device_cuda, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_buffer_cuda_bool, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_buffer_cuda_complex128, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_buffer_cuda_complex64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_buffer_cuda_float16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_buffer_cuda_float32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_buffer_cuda_float64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_buffer_cuda_int16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_buffer_cuda_int32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_buffer_cuda_int64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_buffer_cuda_int8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_buffer_cuda_uint16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_buffer_cuda_uint32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_buffer_cuda_uint64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_buffer_cuda_uint8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_dlpack_cuda_bfloat16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_dlpack_cuda_complex128, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_dlpack_cuda_complex64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_dlpack_cuda_float16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_dlpack_cuda_float32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_dlpack_cuda_float64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_dlpack_cuda_int16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_dlpack_cuda_int32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_dlpack_cuda_int64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_dlpack_cuda_int8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_dlpack_cuda_uint8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_numpy_cuda_bool, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_numpy_cuda_complex128, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_numpy_cuda_complex64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_numpy_cuda_float16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_numpy_cuda_float32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_numpy_cuda_float64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_numpy_cuda_int16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_numpy_cuda_int32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_numpy_cuda_int64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_numpy_cuda_int8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_numpy_cuda_uint16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_numpy_cuda_uint32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_numpy_cuda_uint64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_numpy_cuda_uint8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_tensor_cuda_bfloat16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_tensor_cuda_bool, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_tensor_cuda_complex128, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_tensor_cuda_complex64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_tensor_cuda_float16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_tensor_cuda_float32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_tensor_cuda_float64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_tensor_cuda_int16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_tensor_cuda_int32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_tensor_cuda_int64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_tensor_cuda_int8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_alias_from_tensor_cuda_uint8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_astensor_consistency_cuda, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_buffer_cuda_bool, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_buffer_cuda_complex128, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_buffer_cuda_complex64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_buffer_cuda_float16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_buffer_cuda_float32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_buffer_cuda_float64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_buffer_cuda_int16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_buffer_cuda_int32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_buffer_cuda_int64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_buffer_cuda_int8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_buffer_cuda_uint16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_buffer_cuda_uint32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_buffer_cuda_uint64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_buffer_cuda_uint8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_cuda_bfloat16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_cuda_complex128, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_cuda_complex64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_cuda_float16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_cuda_float32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_cuda_float64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_cuda_int16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_cuda_int32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_cuda_int64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_cuda_int8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_cuda_uint8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_mult_devices_cuda_bfloat16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_mult_devices_cuda_complex128, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_mult_devices_cuda_complex64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_mult_devices_cuda_float16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_mult_devices_cuda_float32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_mult_devices_cuda_float64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_mult_devices_cuda_int16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_mult_devices_cuda_int32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_mult_devices_cuda_int64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_mult_devices_cuda_int8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_dlpack_mult_devices_cuda_uint8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_numpy_cuda_bool, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_numpy_cuda_complex128, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_numpy_cuda_complex64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_numpy_cuda_float16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_numpy_cuda_float32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_numpy_cuda_float64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_numpy_cuda_int16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_numpy_cuda_int32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_numpy_cuda_int64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_numpy_cuda_int8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_numpy_cuda_uint16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_numpy_cuda_uint32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_numpy_cuda_uint64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_numpy_cuda_uint8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_tensor_mult_devices_cuda_bfloat16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_tensor_mult_devices_cuda_complex128, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_tensor_mult_devices_cuda_complex64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_tensor_mult_devices_cuda_float16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_tensor_mult_devices_cuda_float32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_tensor_mult_devices_cuda_float64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_tensor_mult_devices_cuda_int16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_tensor_mult_devices_cuda_int32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_tensor_mult_devices_cuda_int64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_tensor_mult_devices_cuda_int8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_from_tensor_mult_devices_cuda_uint8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_list_cuda_bfloat16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_list_cuda_bool, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_list_cuda_complex128, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_list_cuda_complex64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_list_cuda_float16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_list_cuda_float32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_list_cuda_float64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_list_cuda_int16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_list_cuda_int32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_list_cuda_int64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_list_cuda_int8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_list_cuda_uint8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_tensor_cuda_bfloat16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_tensor_cuda_bool, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_tensor_cuda_complex128, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_tensor_cuda_complex64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_tensor_cuda_float16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_tensor_cuda_float32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_tensor_cuda_float64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_tensor_cuda_int16, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_tensor_cuda_int32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_tensor_cuda_int64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_tensor_cuda_int8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_copy_tensor_cuda_uint8, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_default_device_cuda, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_device_without_index_cuda, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_numpy_scalars_cuda, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_retain_autograd_history_cuda_complex64, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_retain_autograd_history_cuda_float32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_unsupported_alias_cuda_float32, test/test_tensor_creation_ops.py::TestAsArrayCUDA::test_unsupported_alias_mult_devices_cuda_float32 2025-12-04T17:10:38.3437384Z 2025-12-04T17:10:38.3437742Z Finished test_tensor_creation_ops 1/1 ... [2025-12-04 17:10:38.296288][28621.986677702], took 1.35min 2025-12-04T17:10:38.3502671Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_tensor_creation_ops/test_tensor_creation_ops-c328fea3f872cecd.xml 2025-12-04T17:10:44.8663608Z Uploading artifacts took 6.44 seconds 2025-12-04T17:10:44.8668215Z Running torch_np/numpy_tests/core/test_scalarmath 1/1 ... [2025-12-04 17:10:44.866617][28628.557009386] 2025-12-04T17:10:44.8668953Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T17:10:44.8673475Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/core/test_scalarmath.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 17:10:44.867078] 2025-12-04T17:11:14.1228240Z 2025-12-04T17:11:14.1229569Z torch_np/numpy_tests/core/test_scalarmath 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.core.test_scalarmath_1.1_7a9b3c7e6c8b9fbf_.log 2025-12-04T17:11:14.1344177Z Running 186 items in this shard: test/torch_np/numpy_tests/core/test_scalarmath.py::TestTypes::test_leak, test/torch_np/numpy_tests/core/test_scalarmath.py::TestTypes::test_type_add, test/torch_np/numpy_tests/core/test_scalarmath.py::TestTypes::test_type_create, test/torch_np/numpy_tests/core/test_scalarmath.py::TestTypes::test_types, test/torch_np/numpy_tests/core/test_scalarmath.py::TestBaseMath::test_blocked, test/torch_np/numpy_tests/core/test_scalarmath.py::TestBaseMath::test_lower_align, test/torch_np/numpy_tests/core/test_scalarmath.py::TestPower::test_integers_to_negative_integer_power, test/torch_np/numpy_tests/core/test_scalarmath.py::TestPower::test_large_types, test/torch_np/numpy_tests/core/test_scalarmath.py::TestPower::test_mixed_types, test/torch_np/numpy_tests/core/test_scalarmath.py::TestPower::test_modular_power, test/torch_np/numpy_tests/core/test_scalarmath.py::TestPower::test_small_types, test/torch_np/numpy_tests/core/test_scalarmath.py::TestModulus::test_float_modulus_corner_cases_dt_d, test/torch_np/numpy_tests/core/test_scalarmath.py::TestModulus::test_float_modulus_corner_cases_dt_e, test/torch_np/numpy_tests/core/test_scalarmath.py::TestModulus::test_float_modulus_corner_cases_dt_f, test/torch_np/numpy_tests/core/test_scalarmath.py::TestModulus::test_float_modulus_exact, test/torch_np/numpy_tests/core/test_scalarmath.py::TestModulus::test_float_modulus_roundoff, test/torch_np/numpy_tests/core/test_scalarmath.py::TestModulus::test_modulus_basic, test/torch_np/numpy_tests/core/test_scalarmath.py::TestComplexDivision::test_branches, test/torch_np/numpy_tests/core/test_scalarmath.py::TestComplexDivision::test_signed_zeros, test/torch_np/numpy_tests/core/test_scalarmath.py::TestComplexDivision::test_zero_division, test/torch_np/numpy_tests/core/test_scalarmath.py::TestConversion::test_iinfo_long_values_1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestConversion::test_iinfo_long_values_2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestConversion::test_int_from_long, test/torch_np/numpy_tests/core/test_scalarmath.py::TestConversion::test_int_raise_behaviour, test/torch_np/numpy_tests/core/test_scalarmath.py::TestConversion::test_numpy_scalar_relational_operators, test/torch_np/numpy_tests/core/test_scalarmath.py::TestConversion::test_numpy_scalar_relational_operators_2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestConversion::test_scalar_comparison_to_none, test/torch_np/numpy_tests/core/test_scalarmath.py::TestRepr::test_float_repr, test/torch_np/numpy_tests/core/test_scalarmath.py::TestMultiply::test_no_seq_repeat_basic_array_like, test/torch_np/numpy_tests/core/test_scalarmath.py::TestMultiply::test_seq_repeat, test/torch_np/numpy_tests/core/test_scalarmath.py::TestNegative::test_exceptions, test/torch_np/numpy_tests/core/test_scalarmath.py::TestNegative::test_result, test/torch_np/numpy_tests/core/test_scalarmath.py::TestSubtract::test_exceptions, test/torch_np/numpy_tests/core/test_scalarmath.py::TestSubtract::test_result, test/torch_np/numpy_tests/core/test_scalarmath.py::TestAbs::test_builtin_abs_dtype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestAbs::test_builtin_abs_dtype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestAbs::test_builtin_abs_dtype2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestAbs::test_builtin_abs_dtype3, test/torch_np/numpy_tests/core/test_scalarmath.py::TestAbs::test_builtin_abs_dtype4, test/torch_np/numpy_tests/core/test_scalarmath.py::TestAbs::test_numpy_abs_dtype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestAbs::test_numpy_abs_dtype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestAbs::test_numpy_abs_dtype2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestAbs::test_numpy_abs_dtype3, test/torch_np/numpy_tests/core/test_scalarmath.py::TestAbs::test_numpy_abs_dtype4, test/torch_np/numpy_tests/core/test_scalarmath.py::TestBitShifts::test_shift_all_bits_type_code_B_op0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestBitShifts::test_shift_all_bits_type_code_B_op1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestBitShifts::test_shift_all_bits_type_code_b_op0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestBitShifts::test_shift_all_bits_type_code_b_op1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestBitShifts::test_shift_all_bits_type_code_h_op0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestBitShifts::test_shift_all_bits_type_code_h_op1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestBitShifts::test_shift_all_bits_type_code_i_op0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestBitShifts::test_shift_all_bits_type_code_i_op1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestBitShifts::test_shift_all_bits_type_code_l_op0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestBitShifts::test_shift_all_bits_type_code_l_op1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestHash::test_complex_hashes_type_code_D, test/torch_np/numpy_tests/core/test_scalarmath.py::TestHash::test_complex_hashes_type_code_F, test/torch_np/numpy_tests/core/test_scalarmath.py::TestHash::test_float_and_complex_hashes_type_code_D, test/torch_np/numpy_tests/core/test_scalarmath.py::TestHash::test_float_and_complex_hashes_type_code_F, test/torch_np/numpy_tests/core/test_scalarmath.py::TestHash::test_float_and_complex_hashes_type_code_d, test/torch_np/numpy_tests/core/test_scalarmath.py::TestHash::test_float_and_complex_hashes_type_code_e, test/torch_np/numpy_tests/core/test_scalarmath.py::TestHash::test_float_and_complex_hashes_type_code_f, test/torch_np/numpy_tests/core/test_scalarmath.py::TestHash::test_integer_hashes_type_code_B, test/torch_np/numpy_tests/core/test_scalarmath.py::TestHash::test_integer_hashes_type_code_b, test/torch_np/numpy_tests/core/test_scalarmath.py::TestHash::test_integer_hashes_type_code_h, test/torch_np/numpy_tests/core/test_scalarmath.py::TestHash::test_integer_hashes_type_code_i, test/torch_np/numpy_tests/core/test_scalarmath.py::TestHash::test_integer_hashes_type_code_l, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_divbyzero_dtype_B_operation0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_divbyzero_dtype_B_operation1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_divbyzero_dtype_b_operation0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_divbyzero_dtype_b_operation1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_divbyzero_dtype_h_operation0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_divbyzero_dtype_h_operation1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_divbyzero_dtype_i_operation0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_divbyzero_dtype_i_operation1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_divbyzero_dtype_l_operation0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_divbyzero_dtype_l_operation1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_overflow_dtype_B_operation0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_overflow_dtype_B_operation1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_overflow_dtype_B_operation2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_overflow_dtype_b_operation0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_overflow_dtype_b_operation1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_overflow_dtype_b_operation2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_overflow_dtype_h_operation0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_overflow_dtype_h_operation1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_overflow_dtype_h_operation2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_overflow_dtype_i_operation0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_overflow_dtype_i_operation1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_overflow_dtype_i_operation2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_overflow_dtype_l_operation0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_overflow_dtype_l_operation1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_integer_operation_overflow_dtype_l_operation2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_signed_integer_overflow_dtype_b_operation0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_signed_integer_overflow_dtype_b_operation1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_signed_integer_overflow_dtype_b_operation2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_signed_integer_overflow_dtype_b_operation3, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_signed_integer_overflow_dtype_h_operation0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_signed_integer_overflow_dtype_h_operation1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_signed_integer_overflow_dtype_h_operation2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_signed_integer_overflow_dtype_h_operation3, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_signed_integer_overflow_dtype_i_operation0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_signed_integer_overflow_dtype_i_operation1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_signed_integer_overflow_dtype_i_operation2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_signed_integer_overflow_dtype_i_operation3, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_signed_integer_overflow_dtype_l_operation0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_signed_integer_overflow_dtype_l_operation1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_signed_integer_overflow_dtype_l_operation2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_signed_integer_overflow_dtype_l_operation3, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarOpsMisc::test_scalar_unsigned_integer_overflow_dtype_B, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____add_____rop_____radd___op8_cmp_False_subtype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____add_____rop_____radd___op8_cmp_False_subtype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____add_____rop_____radd___op8_cmp_False_subtype2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____add_____rop_____radd___op8_cmp_False_subtype3, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____eq_____rop_____eq___op2_cmp_True_subtype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____eq_____rop_____eq___op2_cmp_True_subtype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____eq_____rop_____eq___op2_cmp_True_subtype2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____eq_____rop_____eq___op2_cmp_True_subtype3, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____floordiv_____rop_____rfloordiv___op6_cmp_False_subtype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____floordiv_____rop_____rfloordiv___op6_cmp_False_subtype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____floordiv_____rop_____rfloordiv___op6_cmp_False_subtype2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____floordiv_____rop_____rfloordiv___op6_cmp_False_subtype3, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____ge_____rop_____le___op5_cmp_True_subtype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____ge_____rop_____le___op5_cmp_True_subtype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____ge_____rop_____le___op5_cmp_True_subtype2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____ge_____rop_____le___op5_cmp_True_subtype3, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____gt_____rop_____lt___op4_cmp_True_subtype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____gt_____rop_____lt___op4_cmp_True_subtype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____gt_____rop_____lt___op4_cmp_True_subtype2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____gt_____rop_____lt___op4_cmp_True_subtype3, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____le_____rop_____ge___op1_cmp_True_subtype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____le_____rop_____ge___op1_cmp_True_subtype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____le_____rop_____ge___op1_cmp_True_subtype2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____le_____rop_____ge___op1_cmp_True_subtype3, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____lt_____rop_____gt___op0_cmp_True_subtype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____lt_____rop_____gt___op0_cmp_True_subtype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____lt_____rop_____gt___op0_cmp_True_subtype2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____lt_____rop_____gt___op0_cmp_True_subtype3, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____mod_____rop_____rmod___op9_cmp_False_subtype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____mod_____rop_____rmod___op9_cmp_False_subtype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____mod_____rop_____rmod___op9_cmp_False_subtype2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____mod_____rop_____rmod___op9_cmp_False_subtype3, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____mul_____rop_____rmul___op10_cmp_False_subtype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____mul_____rop_____rmul___op10_cmp_False_subtype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____mul_____rop_____rmul___op10_cmp_False_subtype2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____mul_____rop_____rmul___op10_cmp_False_subtype3, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____ne_____rop_____ne___op3_cmp_True_subtype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____ne_____rop_____ne___op3_cmp_True_subtype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____ne_____rop_____ne___op3_cmp_True_subtype2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____ne_____rop_____ne___op3_cmp_True_subtype3, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____pow_____rop_____rpow___op11_cmp_False_subtype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____pow_____rop_____rpow___op11_cmp_False_subtype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____pow_____rop_____rpow___op11_cmp_False_subtype2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____pow_____rop_____rpow___op11_cmp_False_subtype3, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____sub_____rop_____rsub___op12_cmp_False_subtype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____sub_____rop_____rsub___op12_cmp_False_subtype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____sub_____rop_____rsub___op12_cmp_False_subtype2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____sub_____rop_____rsub___op12_cmp_False_subtype3, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____truediv_____rop_____rtruediv___op7_cmp_False_subtype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____truediv_____rop_____rtruediv___op7_cmp_False_subtype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____truediv_____rop_____rtruediv___op7_cmp_False_subtype2, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_pyscalar_subclasses___op_____truediv_____rop_____rtruediv___op7_cmp_False_subtype3, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____add_____rop_____radd___op8_cmp_False_sctype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____add_____rop_____radd___op8_cmp_False_sctype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____eq_____rop_____eq___op2_cmp_True_sctype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____eq_____rop_____eq___op2_cmp_True_sctype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____floordiv_____rop_____rfloordiv___op6_cmp_False_sctype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____floordiv_____rop_____rfloordiv___op6_cmp_False_sctype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____ge_____rop_____le___op5_cmp_True_sctype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____ge_____rop_____le___op5_cmp_True_sctype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____gt_____rop_____lt___op4_cmp_True_sctype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____gt_____rop_____lt___op4_cmp_True_sctype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____le_____rop_____ge___op1_cmp_True_sctype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____le_____rop_____ge___op1_cmp_True_sctype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____lt_____rop_____gt___op0_cmp_True_sctype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____lt_____rop_____gt___op0_cmp_True_sctype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____mod_____rop_____rmod___op9_cmp_False_sctype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____mod_____rop_____rmod___op9_cmp_False_sctype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____mul_____rop_____rmul___op10_cmp_False_sctype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____mul_____rop_____rmul___op10_cmp_False_sctype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____ne_____rop_____ne___op3_cmp_True_sctype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____ne_____rop_____ne___op3_cmp_True_sctype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____pow_____rop_____rpow___op11_cmp_False_sctype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____pow_____rop_____rpow___op11_cmp_False_sctype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____sub_____rop_____rsub___op12_cmp_False_sctype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____sub_____rop_____rsub___op12_cmp_False_sctype1, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____truediv_____rop_____rtruediv___op7_cmp_False_sctype0, test/torch_np/numpy_tests/core/test_scalarmath.py::TestScalarSubclassingMisc::test_subclass_deferral___op_____truediv_____rop_____rtruediv___op7_cmp_False_sctype1 2025-12-04T17:11:14.1456721Z 2025-12-04T17:11:14.1457223Z Finished torch_np/numpy_tests/core/test_scalarmath 1/1 ... [2025-12-04 17:11:14.122974][28657.813364578], took 0.49min 2025-12-04T17:11:14.1771952Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.core.test_scalarmath/torch_np.numpy_tests.core.test_scalarmath-a9b7ea367b3cdb41.xml 2025-12-04T17:11:14.2501610Z Running test_serialization 1/1 ... [2025-12-04 17:11:14.249814][28657.940204595] 2025-12-04T17:11:14.2502171Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T17:11:14.2505179Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_serialization.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 17:11:14.250261] 2025-12-04T17:12:14.3983879Z 2025-12-04T17:12:14.3987410Z test_serialization 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_serialization_1.1_39980f3b0bf1286b_.log 2025-12-04T17:12:14.4174225Z Running 206 items in this shard: test/test_serialization.py::TestOldSerialization::test_debug_set_in_ci, test/test_serialization.py::TestOldSerialization::test_load_error_msg, test/test_serialization.py::TestOldSerialization::test_load_nonexistent_device, test/test_serialization.py::TestOldSerialization::test_load_python2_unicode_module, test/test_serialization.py::TestOldSerialization::test_load_unicode_error_msg, test/test_serialization.py::TestOldSerialization::test_pickle_module, test/test_serialization.py::TestOldSerialization::test_safe_load_basic_types, test/test_serialization.py::TestOldSerialization::test_save_different_dtype_error, test/test_serialization.py::TestOldSerialization::test_save_different_dtype_unallocated, test/test_serialization.py::TestOldSerialization::test_serialization, test/test_serialization.py::TestOldSerialization::test_serialization_backwards_compat, test/test_serialization.py::TestOldSerialization::test_serialization_backwards_compat_safe, test/test_serialization.py::TestOldSerialization::test_serialization_container, test/test_serialization.py::TestOldSerialization::test_serialization_container_filelike, test/test_serialization.py::TestOldSerialization::test_serialization_dill, test/test_serialization.py::TestOldSerialization::test_serialization_dill_version_not_supported, test/test_serialization.py::TestOldSerialization::test_serialization_fake_zip, test/test_serialization.py::TestOldSerialization::test_serialization_filelike, test/test_serialization.py::TestOldSerialization::test_serialization_filelike_api_requirements, test/test_serialization.py::TestOldSerialization::test_serialization_filelike_exceptions, test/test_serialization.py::TestOldSerialization::test_serialization_filelike_missing_attrs, test/test_serialization.py::TestOldSerialization::test_serialization_filelike_stress, test/test_serialization.py::TestOldSerialization::test_serialization_filelike_uses_readinto, test/test_serialization.py::TestOldSerialization::test_serialization_gzip, test/test_serialization.py::TestOldSerialization::test_serialization_map_location, test/test_serialization.py::TestOldSerialization::test_serialization_offset, test/test_serialization.py::TestOldSerialization::test_serialization_offset_filelike_weights_only_False, test/test_serialization.py::TestOldSerialization::test_serialization_offset_filelike_weights_only_True, test/test_serialization.py::TestOldSerialization::test_serialization_offset_gzip, test/test_serialization.py::TestOldSerialization::test_serialization_safe, test/test_serialization.py::TestOldSerialization::test_serialization_save_warnings, test/test_serialization.py::TestOldSerialization::test_serialization_sparse, test/test_serialization.py::TestOldSerialization::test_serialization_sparse_bsc_invalid, test/test_serialization.py::TestOldSerialization::test_serialization_sparse_bsr_invalid, test/test_serialization.py::TestOldSerialization::test_serialization_sparse_csc_invalid, test/test_serialization.py::TestOldSerialization::test_serialization_sparse_csr_invalid, test/test_serialization.py::TestOldSerialization::test_serialization_sparse_invalid, test/test_serialization.py::TestOldSerialization::test_serialization_sparse_invalid_legacy_ctor, test/test_serialization.py::TestOldSerialization::test_serialization_sparse_safe, test/test_serialization.py::TestOldSerialization::test_serialization_storage_slice, test/test_serialization.py::TestOldSerialization::test_serialization_zipfile_utils, test/test_serialization.py::TestOldSerialization::test_serialize_device, test/test_serialization.py::TestOldSerialization::test_skip_data_load, test/test_serialization.py::TestSerialization::test_crc32_options_compute_crc32_False_filename_False, test/test_serialization.py::TestSerialization::test_crc32_options_compute_crc32_False_filename_True, test/test_serialization.py::TestSerialization::test_crc32_options_compute_crc32_True_filename_False, test/test_serialization.py::TestSerialization::test_crc32_options_compute_crc32_True_filename_True, test/test_serialization.py::TestSerialization::test_debug_set_in_ci, test/test_serialization.py::TestSerialization::test_filewriter_metadata_writing_filename_False, test/test_serialization.py::TestSerialization::test_filewriter_metadata_writing_filename_True, test/test_serialization.py::TestSerialization::test_get_unsafe_globals_in_checkpoint, test/test_serialization.py::TestSerialization::test_has_format_version, test/test_serialization.py::TestSerialization::test_load_error_msg, test/test_serialization.py::TestSerialization::test_load_njt_weights_only_should_import_False, test/test_serialization.py::TestSerialization::test_load_njt_weights_only_should_import_True, test/test_serialization.py::TestSerialization::test_load_nonexistent_device, test/test_serialization.py::TestSerialization::test_load_python2_unicode_module, test/test_serialization.py::TestSerialization::test_load_unicode_error_msg, test/test_serialization.py::TestSerialization::test_lr_scheduler_serialization, test/test_serialization.py::TestSerialization::test_meta_serialization_weights_only_False, test/test_serialization.py::TestSerialization::test_meta_serialization_weights_only_True, test/test_serialization.py::TestSerialization::test_mmap_load_offset_calculation_path_type0, test/test_serialization.py::TestSerialization::test_mmap_load_offset_calculation_path_type1, test/test_serialization.py::TestSerialization::test_pathlike_serialization_weights_only_False, test/test_serialization.py::TestSerialization::test_pathlike_serialization_weights_only_True, test/test_serialization.py::TestSerialization::test_pickle_module, test/test_serialization.py::TestSerialization::test_safe_load_basic_types, test/test_serialization.py::TestSerialization::test_save_different_dtype_error, test/test_serialization.py::TestSerialization::test_save_different_dtype_unallocated, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_bfloat16_weights_only_False, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_bfloat16_weights_only_True, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_bool_weights_only_False, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_bool_weights_only_True, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_complex128_weights_only_False, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_complex128_weights_only_True, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_complex64_weights_only_False, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_complex64_weights_only_True, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_float16_weights_only_False, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_float16_weights_only_True, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_float32_weights_only_False, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_float32_weights_only_True, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_float64_weights_only_False, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_float64_weights_only_True, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_int16_weights_only_False, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_int16_weights_only_True, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_int32_weights_only_False, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_int32_weights_only_True, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_int64_weights_only_False, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_int64_weights_only_True, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_int8_weights_only_False, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_int8_weights_only_True, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_uint8_weights_only_False, test/test_serialization.py::TestSerialization::test_save_load_preserves_dtype_uint8_weights_only_True, test/test_serialization.py::TestSerialization::test_serializaion_no_storage_leak, test/test_serialization.py::TestSerialization::test_serialization, test/test_serialization.py::TestSerialization::test_serialization_2gb_file, test/test_serialization.py::TestSerialization::test_serialization_4gb_file, test/test_serialization.py::TestSerialization::test_serialization_backwards_compat, test/test_serialization.py::TestSerialization::test_serialization_backwards_compat_safe, test/test_serialization.py::TestSerialization::test_serialization_byte_literal_byte_literals0_weights_only_False, test/test_serialization.py::TestSerialization::test_serialization_byte_literal_byte_literals0_weights_only_True, test/test_serialization.py::TestSerialization::test_serialization_byte_literal_byte_literals1_weights_only_False, test/test_serialization.py::TestSerialization::test_serialization_byte_literal_byte_literals1_weights_only_True, test/test_serialization.py::TestSerialization::test_serialization_byteorder_mark, test/test_serialization.py::TestSerialization::test_serialization_dill, test/test_serialization.py::TestSerialization::test_serialization_dill_version_not_supported, test/test_serialization.py::TestSerialization::test_serialization_dtype_complex32_weights_only_False, test/test_serialization.py::TestSerialization::test_serialization_dtype_complex32_weights_only_True, test/test_serialization.py::TestSerialization::test_serialization_dtype_float8_e4m3fn_weights_only_False, test/test_serialization.py::TestSerialization::test_serialization_dtype_float8_e4m3fn_weights_only_True, test/test_serialization.py::TestSerialization::test_serialization_dtype_float8_e5m2_weights_only_False, test/test_serialization.py::TestSerialization::test_serialization_dtype_float8_e5m2_weights_only_True, test/test_serialization.py::TestSerialization::test_serialization_dtype_uint16_weights_only_False, test/test_serialization.py::TestSerialization::test_serialization_dtype_uint16_weights_only_True, test/test_serialization.py::TestSerialization::test_serialization_dtype_uint32_weights_only_False, test/test_serialization.py::TestSerialization::test_serialization_dtype_uint32_weights_only_True, test/test_serialization.py::TestSerialization::test_serialization_dtype_uint64_weights_only_False, test/test_serialization.py::TestSerialization::test_serialization_dtype_uint64_weights_only_True, test/test_serialization.py::TestSerialization::test_serialization_efficient_zerotensor_weights_only_False, test/test_serialization.py::TestSerialization::test_serialization_efficient_zerotensor_weights_only_True, test/test_serialization.py::TestSerialization::test_serialization_fake_zip, test/test_serialization.py::TestSerialization::test_serialization_filelike, test/test_serialization.py::TestSerialization::test_serialization_filelike_api_requirements, test/test_serialization.py::TestSerialization::test_serialization_filelike_exceptions, test/test_serialization.py::TestSerialization::test_serialization_filelike_missing_attrs, test/test_serialization.py::TestSerialization::test_serialization_filelike_stress, test/test_serialization.py::TestSerialization::test_serialization_filelike_uses_readinto, test/test_serialization.py::TestSerialization::test_serialization_gzip, test/test_serialization.py::TestSerialization::test_serialization_load_bom_data, test/test_serialization.py::TestSerialization::test_serialization_load_bom_data_bfloat16, test/test_serialization.py::TestSerialization::test_serialization_load_bom_data_bool, test/test_serialization.py::TestSerialization::test_serialization_load_bom_data_cdouble, test/test_serialization.py::TestSerialization::test_serialization_load_bom_data_cfloat, test/test_serialization.py::TestSerialization::test_serialization_load_bom_data_double, test/test_serialization.py::TestSerialization::test_serialization_load_bom_data_float, test/test_serialization.py::TestSerialization::test_serialization_load_bom_data_half, test/test_serialization.py::TestSerialization::test_serialization_load_bom_data_int, test/test_serialization.py::TestSerialization::test_serialization_load_bom_data_int16, test/test_serialization.py::TestSerialization::test_serialization_load_bom_data_int8, test/test_serialization.py::TestSerialization::test_serialization_load_bom_data_long, test/test_serialization.py::TestSerialization::test_serialization_load_bom_data_uint8, test/test_serialization.py::TestSerialization::test_serialization_map_location, test/test_serialization.py::TestSerialization::test_serialization_math_bits_weights_only_False, test/test_serialization.py::TestSerialization::test_serialization_math_bits_weights_only_True, test/test_serialization.py::TestSerialization::test_serialization_mmap_loading, test/test_serialization.py::TestSerialization::test_serialization_mmap_loading_ctx, test/test_serialization.py::TestSerialization::test_serialization_mmap_loading_options_path_type0_weights_only_False, test/test_serialization.py::TestSerialization::test_serialization_mmap_loading_options_path_type0_weights_only_True, test/test_serialization.py::TestSerialization::test_serialization_mmap_loading_options_path_type1_weights_only_False, test/test_serialization.py::TestSerialization::test_serialization_mmap_loading_options_path_type1_weights_only_True, test/test_serialization.py::TestSerialization::test_serialization_mmap_loading_with_map_location, test/test_serialization.py::TestSerialization::test_serialization_nested_class, test/test_serialization.py::TestSerialization::test_serialization_offset_gzip, test/test_serialization.py::TestSerialization::test_serialization_python_attr_weights_only_False, test/test_serialization.py::TestSerialization::test_serialization_python_attr_weights_only_True, test/test_serialization.py::TestSerialization::test_serialization_safe, test/test_serialization.py::TestSerialization::test_serialization_save_warnings, test/test_serialization.py::TestSerialization::test_serialization_sparse, test/test_serialization.py::TestSerialization::test_serialization_sparse_bsc_invalid, test/test_serialization.py::TestSerialization::test_serialization_sparse_bsr_invalid, test/test_serialization.py::TestSerialization::test_serialization_sparse_csc_invalid, test/test_serialization.py::TestSerialization::test_serialization_sparse_csr_invalid, test/test_serialization.py::TestSerialization::test_serialization_sparse_invalid, test/test_serialization.py::TestSerialization::test_serialization_sparse_invalid_legacy_ctor, test/test_serialization.py::TestSerialization::test_serialization_sparse_safe, test/test_serialization.py::TestSerialization::test_serialization_storage_slice, test/test_serialization.py::TestSerialization::test_serialization_uintx_intx, test/test_serialization.py::TestSerialization::test_serialization_warning_s390x, test/test_serialization.py::TestSerialization::test_serialization_with_header, test/test_serialization.py::TestSerialization::test_serialization_zipfile_actually_jit, test/test_serialization.py::TestSerialization::test_serialization_zipfile_utils, test/test_serialization.py::TestSerialization::test_serialization_zipfile_weights_only_False, test/test_serialization.py::TestSerialization::test_serialization_zipfile_weights_only_True, test/test_serialization.py::TestSerialization::test_serialize_device, test/test_serialization.py::TestSerialization::test_skip_data_load, test/test_serialization.py::TestSerialization::test_skip_data_serialization_error_cases, test/test_serialization.py::TestSerialization::test_skip_data_serialization_materialize_fake_False, test/test_serialization.py::TestSerialization::test_skip_data_serialization_materialize_fake_True, test/test_serialization.py::TestSerialization::test_skip_data_serialization_preserves_views_materialize_fake_False, test/test_serialization.py::TestSerialization::test_skip_data_serialization_preserves_views_materialize_fake_True, test/test_serialization.py::TestSerialization::test_storage_alignment, test/test_serialization.py::TestSerialization::test_use_pinned_memory_for_d2h, test/test_serialization.py::TestSerialization::test_weights_only_assert, test/test_serialization.py::TestSerialization::test_weights_only_blocked_func_error_msg, test/test_serialization.py::TestSerialization::test_weights_only_env_variables_force_weights_only_False, test/test_serialization.py::TestSerialization::test_weights_only_env_variables_force_weights_only_True, test/test_serialization.py::TestSerialization::test_weights_only_error_unsafe_global_False, test/test_serialization.py::TestSerialization::test_weights_only_error_unsafe_global_True, test/test_serialization.py::TestSerialization::test_weights_only_safe_globals_blocklist, test/test_serialization.py::TestSerialization::test_weights_only_safe_globals_build, test/test_serialization.py::TestSerialization::test_weights_only_safe_globals_build_with_slots_slots_all, test/test_serialization.py::TestSerialization::test_weights_only_safe_globals_build_with_slots_slots_some, test/test_serialization.py::TestSerialization::test_weights_only_safe_globals_newobj, test/test_serialization.py::TestSerialization::test_weights_only_with_zoneinfo_unpickle_registration_success, test/test_serialization.py::TestSubclassSerialization::test_cloned_deepcopy_requires_grad_False, test/test_serialization.py::TestSubclassSerialization::test_cloned_deepcopy_requires_grad_True, test/test_serialization.py::TestSubclassSerialization::test_empty_class_serialization, test/test_serialization.py::TestSubclassSerialization::test_safe_globals_context_manager_weights_only, test/test_serialization.py::TestSubclassSerialization::test_safe_globals_for_weights_only, test/test_serialization.py::TestSubclassSerialization::test_sets_are_loadable_with_weights_only, test/test_serialization.py::TestSubclassSerialization::test_tensor_subclass_deepcopy, test/test_serialization.py::TestSubclassSerialization::test_tensor_subclass_getstate_overwrite, test/test_serialization.py::TestSubclassSerialization::test_tensor_subclass_map_location, test/test_serialization.py::TestSubclassSerialization::test_tensor_subclass_wrapper_serialization, test/test_serialization.py::TestBothSerializationCUDA::test_serialization_new_format_old_format_compat_weights_only_False_cuda, test/test_serialization.py::TestBothSerializationCUDA::test_serialization_new_format_old_format_compat_weights_only_True_cuda 2025-12-04T17:12:14.4357432Z 2025-12-04T17:12:14.4358040Z Finished test_serialization 1/1 ... [2025-12-04 17:12:14.398576][28718.08896104], took 1.00min 2025-12-04T17:12:14.4663236Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_serialization/test_serialization-43b1fedba2a2b650.xml 2025-12-04T17:12:14.5356729Z Running torch_np/numpy_tests/core/test_dlpack 1/1 ... [2025-12-04 17:12:14.535342][28718.225733177] 2025-12-04T17:12:14.5357401Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T17:12:14.5360949Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/core/test_dlpack.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 17:12:14.535835] 2025-12-04T17:12:19.6578014Z 2025-12-04T17:12:19.6579396Z torch_np/numpy_tests/core/test_dlpack 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.core.test_dlpack_1.1_be350d36e188bf08_.log 2025-12-04T17:12:19.6601899Z Running 53 items in this shard: test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_dlpack_destructor_exception, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_dlpack_device, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_dtype_passthrough_dtype0, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_dtype_passthrough_dtype1, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_dtype_passthrough_dtype2, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_dtype_passthrough_dtype3, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_dtype_passthrough_dtype4, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_dtype_passthrough_dtype5, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_dtype_passthrough_dtype6, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_dtype_passthrough_dtype7, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_dtype_passthrough_dtype8, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_dtype_passthrough_dtype9, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_dunder_dlpack_refcount, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_dunder_dlpack_stream, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_from_dlpack_refcount, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_from_torch, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_0, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_1, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_10, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_11, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_12, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_13, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_14, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_15, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_16, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_17, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_18, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_19, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_2, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_20, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_21, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_22, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_23, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_24, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_25, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_26, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_27, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_28, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_29, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_3, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_30, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_31, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_32, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_4, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_5, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_6, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_7, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_8, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_higher_dims_ndim_9, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_ndim0, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_non_contiguous, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_readonly, test/torch_np/numpy_tests/core/test_dlpack.py::TestDLPack::test_to_torch 2025-12-04T17:12:19.6623447Z 2025-12-04T17:12:19.6623863Z Finished torch_np/numpy_tests/core/test_dlpack 1/1 ... [2025-12-04 17:12:19.657654][28723.348046357], took 0.09min 2025-12-04T17:12:19.7112138Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.numpy_tests.core.test_dlpack/torch_np.numpy_tests.core.test_dlpack-350d972e2ff20e05.xml 2025-12-04T17:12:19.7409148Z Running test_multiprocessing_spawn 1/1 ... [2025-12-04 17:12:19.740644][28723.431036144] 2025-12-04T17:12:19.7409742Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T17:12:19.7413558Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_multiprocessing_spawn.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 17:12:19.741076] 2025-12-04T17:15:07.1883501Z 2025-12-04T17:15:07.1884602Z test_multiprocessing_spawn 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_multiprocessing_spawn_1.1_a0b8d9d26431815a_.log 2025-12-04T17:15:07.1898524Z Running 31 items in this shard: test/test_multiprocessing_spawn.py::SpawnTest::test_exception_all, test/test_multiprocessing_spawn.py::SpawnTest::test_exception_raises, test/test_multiprocessing_spawn.py::SpawnTest::test_exception_single, test/test_multiprocessing_spawn.py::SpawnTest::test_first_argument_index, test/test_multiprocessing_spawn.py::SpawnTest::test_signal_raises, test/test_multiprocessing_spawn.py::SpawnTest::test_success, test/test_multiprocessing_spawn.py::SpawnTest::test_success_first_then_exception, test/test_multiprocessing_spawn.py::SpawnTest::test_success_non_blocking, test/test_multiprocessing_spawn.py::SpawnTest::test_terminate_exit_grace_period0, test/test_multiprocessing_spawn.py::SpawnTest::test_terminate_exit_grace_period_20, test/test_multiprocessing_spawn.py::SpawnTest::test_terminate_signal, test/test_multiprocessing_spawn.py::ForkTest::test_exception_all, test/test_multiprocessing_spawn.py::ForkTest::test_exception_single, test/test_multiprocessing_spawn.py::ForkTest::test_first_argument_index, test/test_multiprocessing_spawn.py::ForkTest::test_success, test/test_multiprocessing_spawn.py::ForkTest::test_success_first_then_exception, test/test_multiprocessing_spawn.py::ForkTest::test_success_non_blocking, test/test_multiprocessing_spawn.py::ForkTest::test_terminate_exit_grace_period0, test/test_multiprocessing_spawn.py::ForkTest::test_terminate_exit_grace_period_20, test/test_multiprocessing_spawn.py::ForkTest::test_terminate_signal, test/test_multiprocessing_spawn.py::ParallelForkServerShouldWorkTest::test_exception_all, test/test_multiprocessing_spawn.py::ParallelForkServerShouldWorkTest::test_exception_single, test/test_multiprocessing_spawn.py::ParallelForkServerShouldWorkTest::test_first_argument_index, test/test_multiprocessing_spawn.py::ParallelForkServerShouldWorkTest::test_success, test/test_multiprocessing_spawn.py::ParallelForkServerShouldWorkTest::test_success_first_then_exception, test/test_multiprocessing_spawn.py::ParallelForkServerShouldWorkTest::test_success_non_blocking, test/test_multiprocessing_spawn.py::ParallelForkServerShouldWorkTest::test_terminate_exit_grace_period0, test/test_multiprocessing_spawn.py::ParallelForkServerShouldWorkTest::test_terminate_exit_grace_period_20, test/test_multiprocessing_spawn.py::ParallelForkServerShouldWorkTest::test_terminate_signal, test/test_multiprocessing_spawn.py::ParallelForkServerPerfTest::test_forkserver_perf, test/test_multiprocessing_spawn.py::ErrorTest::test_errors_pickleable 2025-12-04T17:15:07.1911390Z 2025-12-04T17:15:07.1911767Z Finished test_multiprocessing_spawn 1/1 ... [2025-12-04 17:15:07.188115][28890.878508497], took 2.79min 2025-12-04T17:15:07.2428331Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_multiprocessing_spawn/test_multiprocessing_spawn-b04898e1c1418acf.xml 2025-12-04T17:15:07.3251517Z Running test_cuda_nvml_based_avail 1/1 ... [2025-12-04 17:15:07.324844][28891.015235283] 2025-12-04T17:15:07.3252099Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T17:15:07.3255241Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_cuda_nvml_based_avail.py', '--shard-id=1', '--num-shards=1', '-v', '--subprocess', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 17:15:07.325276] 2025-12-04T17:15:51.1749140Z 2025-12-04T17:15:51.1750150Z test_cuda_nvml_based_avail 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_cuda_nvml_based_avail_1.1_c66f584abc6fe991_.log 2025-12-04T17:15:51.1755683Z Running 9 items in this shard: test/test_cuda_nvml_based_avail.py::TestExtendedCUDAIsAvail::test_cuda_is_available_nvml_avail_False_avoid_init2, test/test_cuda_nvml_based_avail.py::TestExtendedCUDAIsAvail::test_cuda_is_available_nvml_avail_False_avoid_init_0, test/test_cuda_nvml_based_avail.py::TestExtendedCUDAIsAvail::test_cuda_is_available_nvml_avail_False_avoid_init_1, test/test_cuda_nvml_based_avail.py::TestExtendedCUDAIsAvail::test_cuda_is_available_nvml_avail_True_avoid_init2, test/test_cuda_nvml_based_avail.py::TestExtendedCUDAIsAvail::test_cuda_is_available_nvml_avail_True_avoid_init_0, test/test_cuda_nvml_based_avail.py::TestExtendedCUDAIsAvail::test_cuda_is_available_nvml_avail_True_avoid_init_1, test/test_cuda_nvml_based_avail.py::TestVisibleDeviceParses::test_env_var_parsing, test/test_cuda_nvml_based_avail.py::TestVisibleDeviceParses::test_ordinal_parse_visible_devices, test/test_cuda_nvml_based_avail.py::TestVisibleDeviceParses::test_partial_uuid_resolver 2025-12-04T17:15:51.1760965Z Running 1 items in this shard: test/test_cuda_nvml_based_avail.py::TestExtendedCUDAIsAvail::test_cuda_is_available_nvml_avail_False_avoid_init2 2025-12-04T17:15:51.1762288Z Running 1 items in this shard: test/test_cuda_nvml_based_avail.py::TestExtendedCUDAIsAvail::test_cuda_is_available_nvml_avail_False_avoid_init_0 2025-12-04T17:15:51.1763877Z Running 1 items in this shard: test/test_cuda_nvml_based_avail.py::TestExtendedCUDAIsAvail::test_cuda_is_available_nvml_avail_False_avoid_init_1 2025-12-04T17:15:51.1765190Z Running 1 items in this shard: test/test_cuda_nvml_based_avail.py::TestExtendedCUDAIsAvail::test_cuda_is_available_nvml_avail_True_avoid_init2 2025-12-04T17:15:51.1766505Z Running 1 items in this shard: test/test_cuda_nvml_based_avail.py::TestExtendedCUDAIsAvail::test_cuda_is_available_nvml_avail_True_avoid_init_0 2025-12-04T17:15:51.1767807Z Running 1 items in this shard: test/test_cuda_nvml_based_avail.py::TestExtendedCUDAIsAvail::test_cuda_is_available_nvml_avail_True_avoid_init_1 2025-12-04T17:15:51.1768967Z Running 1 items in this shard: test/test_cuda_nvml_based_avail.py::TestVisibleDeviceParses::test_env_var_parsing 2025-12-04T17:15:51.1770054Z Running 1 items in this shard: test/test_cuda_nvml_based_avail.py::TestVisibleDeviceParses::test_ordinal_parse_visible_devices 2025-12-04T17:15:51.1771234Z Running 1 items in this shard: test/test_cuda_nvml_based_avail.py::TestVisibleDeviceParses::test_partial_uuid_resolver 2025-12-04T17:15:51.1771829Z 2025-12-04T17:15:51.1772173Z Finished test_cuda_nvml_based_avail 1/1 ... [2025-12-04 17:15:51.174921][28934.865309168], took 0.73min 2025-12-04T17:15:51.2292334Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-ee0ad83af64551d9.xml 2025-12-04T17:15:51.3048340Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-445675b8d8f15e5e.xml 2025-12-04T17:15:51.3348697Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-d82ae01551c43307.xml 2025-12-04T17:15:51.3636733Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-4e71d02e98da078f.xml 2025-12-04T17:15:51.3978462Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-233c3fc007b40619.xml 2025-12-04T17:15:51.4281930Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-a0f26f1a84f3180c.xml 2025-12-04T17:15:51.4611639Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-4a2a7cd181a00ba9.xml 2025-12-04T17:15:51.5072070Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-24fabc8827e58618.xml 2025-12-04T17:15:51.5352326Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-f87021e84032e697.xml 2025-12-04T17:15:51.5631497Z Running test_mobile_optimizer 1/1 ... [2025-12-04 17:15:51.562901][28935.253292708] 2025-12-04T17:15:51.5632061Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T17:15:51.5635606Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_mobile_optimizer.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 17:15:51.563316] 2025-12-04T17:15:59.2389785Z 2025-12-04T17:15:59.2390893Z test_mobile_optimizer 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_mobile_optimizer_1.1_f99b91a9a53c0fc9_.log 2025-12-04T17:15:59.2394610Z Running 7 items in this shard: test/test_mobile_optimizer.py::TestOptimizer::test_clone_module_with_class, test/test_mobile_optimizer.py::TestOptimizer::test_generate_mobile_module_lints, test/test_mobile_optimizer.py::TestOptimizer::test_hoist_conv_packed_params, test/test_mobile_optimizer.py::TestOptimizer::test_mobilenet_optimize_for_mobile, test/test_mobile_optimizer.py::TestOptimizer::test_optimize_for_mobile, test/test_mobile_optimizer.py::TestOptimizer::test_preserve_bundled_inputs_methods, test/test_mobile_optimizer.py::TestOptimizer::test_quantized_conv_no_asan_failures 2025-12-04T17:15:59.2397667Z 2025-12-04T17:15:59.2398007Z Finished test_mobile_optimizer 1/1 ... [2025-12-04 17:15:59.238772][28942.929163667], took 0.13min 2025-12-04T17:15:59.2943795Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_mobile_optimizer/test_mobile_optimizer-8da56fb35697f793.xml 2025-12-04T17:15:59.3753915Z Running torch_np/test_function_base 1/1 ... [2025-12-04 17:15:59.375076][28943.065466799] 2025-12-04T17:15:59.3754509Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T17:15:59.3757831Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/test_function_base.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 17:15:59.375512] 2025-12-04T17:16:04.7977799Z 2025-12-04T17:16:04.7979197Z torch_np/test_function_base 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.test_function_base_1.1_2a583a88d944c488_.log 2025-12-04T17:16:04.7980946Z Running 2 items in this shard: test/torch_np/test_function_base.py::TestAppend::test_basic, test/torch_np/test_function_base.py::TestMisc::test_broadcast_shapes 2025-12-04T17:16:04.7981765Z 2025-12-04T17:16:04.7982126Z Finished torch_np/test_function_base 1/1 ... [2025-12-04 17:16:04.797560][28948.487953319], took 0.09min 2025-12-04T17:16:04.8530783Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/torch_np.test_function_base/torch_np.test_function_base-fc4ff98d96269d3c.xml 2025-12-04T17:16:04.8820065Z Running test_type_promotion 1/1 ... [2025-12-04 17:16:04.881710][28948.572101944] 2025-12-04T17:16:04.8820890Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T17:16:04.8824073Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_type_promotion.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 17:16:04.882130] 2025-12-04T17:16:22.3206717Z 2025-12-04T17:16:22.3208325Z test_type_promotion 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_type_promotion_1.1_fda622c3de5d55b3_.log 2025-12-04T17:16:22.3414182Z Running 423 items in this shard: test/test_type_promotion.py::TestTypePromotionCUDA::test_add_wrapped_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_alpha_mismatch_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_alternate_result_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_bfloat16_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_booleans_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_can_cast_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_cat_different_dtypes_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_cat_out_different_dtypes_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_bool_bool_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_bool_bool_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_bool_bool_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_bool_bool_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_bool_float32_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_bool_float32_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_bool_float32_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_bool_float32_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_bool_float64_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_bool_float64_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_bool_float64_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_bool_float64_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_bool_int32_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_bool_int32_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_bool_int32_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_bool_int32_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float32_bool_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float32_bool_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float32_bool_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float32_bool_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float32_float32_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float32_float32_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float32_float32_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float32_float32_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float32_float64_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float32_float64_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float32_float64_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float32_float64_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float32_int32_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float32_int32_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float32_int32_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float32_int32_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float64_bool_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float64_bool_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float64_bool_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float64_bool_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float64_float32_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float64_float32_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float64_float32_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float64_float32_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float64_float64_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float64_float64_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float64_float64_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float64_float64_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float64_int32_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float64_int32_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float64_int32_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_float64_int32_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_int32_bool_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_int32_bool_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_int32_bool_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_int32_bool_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_int32_float32_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_int32_float32_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_int32_float32_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_int32_float32_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_int32_float64_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_int32_float64_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_int32_float64_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_int32_float64_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_int32_int32_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_int32_int32_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_int32_int32_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_clamp_type_promotion_cuda_int32_int32_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_comparison_ops_with_type_promotion_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_complex_assertraises_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_complex_half_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_complex_promotion_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_complex_scalar_mult_tensor_promotion_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_computation_ignores_out_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_create_bool_tensors_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_cuda_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_cuda_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_cuda_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_cuda_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_cuda_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_cuda_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_inplace_cuda_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_inplace_cuda_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_inplace_cuda_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_inplace_cuda_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_inplace_cuda_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_inplace_cuda_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_inplace_cuda_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_inplace_cuda_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_out_cuda_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_out_cuda_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_out_cuda_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_out_cuda_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_out_cuda_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_out_cuda_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_out_cuda_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_div_promotion_out_cuda_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_float_promotion_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_from_issue_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_half_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_indexing_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_indexing_fail_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_inplace_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_int_promotion_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_int_to_float_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_integer_addcdiv_deprecated_cuda_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_integer_addcdiv_deprecated_cuda_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_integer_addcdiv_deprecated_cuda_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_integer_addcdiv_deprecated_cuda_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_integer_addcdiv_deprecated_cuda_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_lt_with_type_promotion_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_many_promotions_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_mixed_type_backward_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_non_promoting_ops_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_bool_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_bool_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_bool_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_bool_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_bool_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_bool_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_bool_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_bool_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_bool_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_bool_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_bool_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex128_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex128_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex128_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex128_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex128_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex128_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex128_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex128_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex128_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex128_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex128_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex64_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex64_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex64_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex64_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex64_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex64_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex64_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex64_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex64_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex64_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_complex64_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float16_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float16_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float16_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float16_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float16_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float16_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float16_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float16_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float16_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float16_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float16_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float32_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float32_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float32_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float32_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float32_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float32_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float32_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float32_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float32_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float32_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float32_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float64_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float64_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float64_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float64_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float64_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float64_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float64_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float64_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float64_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float64_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_float64_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int16_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int16_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int16_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int16_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int16_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int16_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int16_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int16_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int16_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int16_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int16_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int32_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int32_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int32_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int32_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int32_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int32_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int32_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int32_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int32_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int32_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int32_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int64_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int64_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int64_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int64_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int64_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int64_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int64_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int64_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int64_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int64_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int64_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int8_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int8_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int8_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int8_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int8_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int8_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int8_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int8_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int8_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int8_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_int8_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_uint8_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_uint8_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_uint8_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_uint8_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_uint8_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_uint8_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_uint8_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_uint8_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_uint8_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_uint8_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_numpy_array_binary_ufunc_promotion_cuda_uint8_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_promote_self_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_promote_types_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bfloat16_bfloat16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bfloat16_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bfloat16_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bfloat16_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bfloat16_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bfloat16_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bfloat16_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bfloat16_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bfloat16_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bfloat16_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bfloat16_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bfloat16_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bool_bfloat16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bool_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bool_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bool_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bool_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bool_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bool_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bool_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bool_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bool_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bool_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_bool_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex128_bfloat16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex128_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex128_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex128_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex128_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex128_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex128_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex128_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex128_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex128_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex128_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex128_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex64_bfloat16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex64_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex64_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex64_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex64_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex64_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex64_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex64_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex64_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex64_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex64_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_complex64_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float16_bfloat16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float16_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float16_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float16_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float16_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float16_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float16_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float16_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float16_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float16_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float16_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float16_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float32_bfloat16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float32_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float32_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float32_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float32_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float32_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float32_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float32_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float32_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float32_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float32_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float32_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float64_bfloat16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float64_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float64_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float64_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float64_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float64_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float64_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float64_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float64_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float64_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float64_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_float64_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int16_bfloat16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int16_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int16_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int16_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int16_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int16_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int16_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int16_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int16_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int16_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int16_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int16_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int32_bfloat16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int32_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int32_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int32_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int32_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int32_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int32_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int32_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int32_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int32_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int32_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int32_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int64_bfloat16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int64_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int64_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int64_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int64_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int64_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int64_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int64_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int64_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int64_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int64_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int64_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int8_bfloat16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int8_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int8_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int8_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int8_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int8_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int8_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int8_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int8_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int8_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int8_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_int8_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_uint8_bfloat16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_uint8_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_uint8_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_uint8_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_uint8_float16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_uint8_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_uint8_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_uint8_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_uint8_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_uint8_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_uint8_int8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_cuda_uint8_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_result_type_tensor_vs_scalar_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_sparse_add_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_sparse_div_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_sparse_div_promotion_cuda_bool, test/test_type_promotion.py::TestTypePromotionCUDA::test_sparse_div_promotion_cuda_int16, test/test_type_promotion.py::TestTypePromotionCUDA::test_sparse_div_promotion_cuda_int32, test/test_type_promotion.py::TestTypePromotionCUDA::test_sparse_div_promotion_cuda_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_sparse_div_promotion_cuda_uint8, test/test_type_promotion.py::TestTypePromotionCUDA::test_sparse_mul_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_sparse_sub_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_ternary_out_promotion_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_transpose_cuda, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_complex128_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_complex128_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_complex128_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_complex128_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_complex128_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_complex64_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_complex64_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_complex64_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_complex64_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_complex64_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_float32_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_float32_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_float32_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_float32_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_float32_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_float64_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_float64_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_float64_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_float64_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_float64_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_int64_complex128, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_int64_complex64, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_int64_float32, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_int64_float64, test/test_type_promotion.py::TestTypePromotionCUDA::test_unary_op_out_casting_cuda_int64_int64, test/test_type_promotion.py::TestTypePromotionCUDA::test_unsigned_cuda 2025-12-04T17:16:22.3616504Z 2025-12-04T17:16:22.3616875Z Finished test_type_promotion 1/1 ... [2025-12-04 17:16:22.321278][28966.011665632], took 0.29min 2025-12-04T17:16:22.3771799Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_type_promotion/test_type_promotion-bc48b0383dc65f97.xml 2025-12-04T17:16:22.4702824Z Running lazy/test_reuse_ir 1/1 ... [2025-12-04 17:16:22.469938][28966.160328643] 2025-12-04T17:16:22.4703376Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T17:16:22.4706358Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'lazy/test_reuse_ir.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 17:16:22.470400] 2025-12-04T17:16:28.1431153Z 2025-12-04T17:16:28.1432397Z lazy/test_reuse_ir 1/1 was successful, full logs can be found in artifacts with path test/test-reports/lazy.test_reuse_ir_1.1_bce662c061b92625_.log 2025-12-04T17:16:28.1434405Z Running 4 items in this shard: test/lazy/test_reuse_ir.py::TestLazyReuseIr::testAdd, test/lazy/test_reuse_ir.py::TestLazyReuseIr::testAddSub, test/lazy/test_reuse_ir.py::TestLazyReuseIr::testAddSubFallback, test/lazy/test_reuse_ir.py::TestLazyReuseIr::testBatchNorm 2025-12-04T17:16:28.1435756Z 2025-12-04T17:16:28.1436080Z Finished lazy/test_reuse_ir 1/1 ... [2025-12-04 17:16:28.142931][28971.833324063], took 0.09min 2025-12-04T17:16:28.1986277Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/lazy.test_reuse_ir/lazy.test_reuse_ir-27aa6d3f8959224d.xml 2025-12-04T17:16:28.2544541Z Running test_functional_autograd_benchmark 1/1 ... [2025-12-04 17:16:28.254162][28971.944552201] 2025-12-04T17:16:28.2545339Z SCRIBE_GRAPHQL_ACCESS_TOKEN is set 2025-12-04T17:16:28.2548394Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_functional_autograd_benchmark.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-12-04 17:16:28.254601] 2025-12-04T17:17:21.7942846Z 2025-12-04T17:17:21.7944008Z test_functional_autograd_benchmark 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_functional_autograd_benchmark_1.1_3090448ccb207ef1_.log 2025-12-04T17:17:21.7946244Z Running 2 items in this shard: test/test_functional_autograd_benchmark.py::TestFunctionalAutogradBenchmark::test_fast_tasks, test/test_functional_autograd_benchmark.py::TestFunctionalAutogradBenchmark::test_slow_tasks 2025-12-04T17:17:21.7947388Z 2025-12-04T17:17:21.7947802Z Finished test_functional_autograd_benchmark 1/1 ... [2025-12-04 17:17:21.794135][29025.484527549], took 0.89min 2025-12-04T17:17:21.8503492Z Parsing testcases for test report: /var/lib/jenkins/workspace/test/test-reports/python-pytest/test_functional_autograd_benchmark/test_functional_autograd_benchmark-8f47749aefc55c89.xml 2025-12-04T17:17:28.9175892Z Running test batch 'tests to run' cost 27150.38 seconds 2025-12-04T17:17:28.9194157Z Emitting td_test_failure_stats_v2 2025-12-04T17:17:28.9199749Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764868648_1c7216eed13511f0ade40242ac110002 2025-12-04T17:17:29.0360692Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764868648_1c7216eed13511f0ade40242ac110002 2025-12-04T17:17:29.0375565Z Emitting td_test_failure_stats_v2 2025-12-04T17:17:29.0378048Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764868649_1c8417a4d13511f0ade40242ac110002 2025-12-04T17:17:29.0717003Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764868649_1c8417a4d13511f0ade40242ac110002 2025-12-04T17:17:29.0731882Z Emitting td_test_failure_stats_v2 2025-12-04T17:17:29.0734395Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764868649_1c8987f2d13511f0ade40242ac110002 2025-12-04T17:17:29.1077065Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764868649_1c8987f2d13511f0ade40242ac110002 2025-12-04T17:17:29.1092412Z Emitting td_test_failure_stats_v2 2025-12-04T17:17:29.1094900Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764868649_1c8f0830d13511f0ade40242ac110002 2025-12-04T17:17:29.2013126Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764868649_1c8f0830d13511f0ade40242ac110002 2025-12-04T17:17:29.2029386Z Emitting td_test_failure_stats_v2 2025-12-04T17:17:29.2032251Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764868649_1c9d5598d13511f0ade40242ac110002 2025-12-04T17:17:29.2379490Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764868649_1c9d5598d13511f0ade40242ac110002 2025-12-04T17:17:29.2394966Z Emitting td_test_failure_stats_v2 2025-12-04T17:17:29.2397434Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764868649_1ca2e832d13511f0ade40242ac110002 2025-12-04T17:17:29.2727140Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764868649_1ca2e832d13511f0ade40242ac110002 2025-12-04T17:17:29.2742721Z Emitting td_test_failure_stats_v2 2025-12-04T17:17:29.2745080Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764868649_1ca83670d13511f0ade40242ac110002 2025-12-04T17:17:29.3069176Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764868649_1ca83670d13511f0ade40242ac110002 2025-12-04T17:17:29.3084675Z Emitting td_test_failure_stats_v2 2025-12-04T17:17:29.3087035Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764868649_1cad6df2d13511f0ade40242ac110002 2025-12-04T17:17:29.3606344Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764868649_1cad6df2d13511f0ade40242ac110002 2025-12-04T17:17:29.3622528Z Emitting td_test_failure_stats_v2 2025-12-04T17:17:29.3625572Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764868649_1cb5a396d13511f0ade40242ac110002 2025-12-04T17:17:29.3972199Z Done! Finish writing document to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764868649_1cb5a396d13511f0ade40242ac110002 2025-12-04T17:17:29.3973832Z inductor/test_aot_inductor 3/6 failed! 2025-12-04T17:17:29.3974240Z inductor/test_torchinductor 2/2 failed! 2025-12-04T17:17:29.3974700Z inductor/test_torchinductor_dynamic_shapes 4/5 failed! 2025-12-04T17:17:29.3975189Z inductor/test_cuda_select_algorithm 5/5 failed! 2025-12-04T17:17:29.3975634Z inductor/test_decompose_mem_bound_mm 1/1 failed! 2025-12-04T17:17:29.3976060Z inductor/test_online_softmax 1/1 failed! 2025-12-04T17:17:29.3976468Z inductor/test_mix_order_reduction 2/2 failed! 2025-12-04T17:17:29.3976897Z test_matmul_cuda 1/1 failed! 2025-12-04T17:17:29.3977233Z functorch/test_vmap 1/1 failed! 2025-12-04T17:17:30.2297245Z 2025-12-04T17:17:30.2298213Z real 452m38.983s 2025-12-04T17:17:30.2298541Z user 433m21.821s 2025-12-04T17:17:30.2298805Z sys 68m57.060s 2025-12-04T17:17:30.2299058Z + sccache_epilogue 2025-12-04T17:17:30.2299392Z + echo '::group::Sccache Compilation Log' 2025-12-04T17:17:30.2300195Z ##[group]Sccache Compilation Log 2025-12-04T17:17:30.2300613Z + echo '=================== sccache compilation log ===================' 2025-12-04T17:17:30.2301087Z =================== sccache compilation log =================== 2025-12-04T17:17:30.2301827Z + python /var/lib/jenkins/workspace/.ci/pytorch/print_sccache_log.py /var/lib/jenkins/sccache_error.log 2025-12-04T17:17:30.2454225Z + echo '=========== If your build fails, please take a look at the log above for possible reasons ===========' 2025-12-04T17:17:30.2455039Z =========== If your build fails, please take a look at the log above for possible reasons =========== 2025-12-04T17:17:30.2455604Z + sccache --show-stats 2025-12-04T17:17:30.2491141Z Compile requests 5294 2025-12-04T17:17:30.2491900Z Compile requests executed 597 2025-12-04T17:17:30.2492316Z Cache hits 305 2025-12-04T17:17:30.2492707Z Cache hits (C/C++) 305 2025-12-04T17:17:30.2493366Z Cache misses 271 2025-12-04T17:17:30.2493823Z Cache misses (C/C++) 271 2025-12-04T17:17:30.2494172Z Cache hits rate 52.95 % 2025-12-04T17:17:30.2494548Z Cache hits rate (C/C++) 52.95 % 2025-12-04T17:17:30.2494911Z Cache timeouts 0 2025-12-04T17:17:30.2495451Z Cache read errors 0 2025-12-04T17:17:30.2496376Z Forced recaches 0 2025-12-04T17:17:30.2496741Z Cache write errors 0 2025-12-04T17:17:30.2497144Z Cache errors 0 2025-12-04T17:17:30.2497502Z Compilations 271 2025-12-04T17:17:30.2497861Z Compilation failures 21 2025-12-04T17:17:30.2498234Z Non-cacheable compilations 0 2025-12-04T17:17:30.2498603Z Non-cacheable calls 309 2025-12-04T17:17:30.2498969Z Non-compilation calls 4388 2025-12-04T17:17:30.2499344Z Unsupported compiler calls 0 2025-12-04T17:17:30.2499709Z Average cache write 0.056 s 2025-12-04T17:17:30.2500085Z Average compiler 7.326 s 2025-12-04T17:17:30.2500462Z Average cache read hit 0.027 s 2025-12-04T17:17:30.2500934Z Failed distributed compilations 0 2025-12-04T17:17:30.2501188Z 2025-12-04T17:17:30.2501299Z Non-cacheable reasons: 2025-12-04T17:17:30.2501614Z unknown source language 244 2025-12-04T17:17:30.2501974Z -E 65 2025-12-04T17:17:30.2502217Z 2025-12-04T17:17:30.2502485Z Cache location s3, name: ossci-compiler-cache-circleci-v2, prefix: / 2025-12-04T17:17:30.2503010Z Version (client) 0.10.0 2025-12-04T17:17:30.2503361Z + sccache --stop-server 2025-12-04T17:17:30.2517687Z Stopping sccache server... 2025-12-04T17:17:30.2521010Z Compile requests 5294 2025-12-04T17:17:30.2521447Z Compile requests executed 597 2025-12-04T17:17:30.2521815Z Cache hits 305 2025-12-04T17:17:30.2522158Z Cache hits (C/C++) 305 2025-12-04T17:17:30.2522509Z Cache misses 271 2025-12-04T17:17:30.2522857Z Cache misses (C/C++) 271 2025-12-04T17:17:30.2523216Z Cache hits rate 52.95 % 2025-12-04T17:17:30.2523587Z Cache hits rate (C/C++) 52.95 % 2025-12-04T17:17:30.2523955Z Cache timeouts 0 2025-12-04T17:17:30.2524320Z Cache read errors 0 2025-12-04T17:17:30.2524721Z Forced recaches 0 2025-12-04T17:17:30.2525073Z Cache write errors 0 2025-12-04T17:17:30.2525422Z Cache errors 0 2025-12-04T17:17:30.2525759Z Compilations 271 2025-12-04T17:17:30.2526117Z Compilation failures 21 2025-12-04T17:17:30.2526585Z Non-cacheable compilations 0 2025-12-04T17:17:30.2526943Z Non-cacheable calls 309 2025-12-04T17:17:30.2527307Z Non-compilation calls 4388 2025-12-04T17:17:30.2527679Z Unsupported compiler calls 0 2025-12-04T17:17:30.2528046Z Average cache write 0.056 s 2025-12-04T17:17:30.2528426Z Average compiler 7.326 s 2025-12-04T17:17:30.2528813Z Average cache read hit 0.027 s 2025-12-04T17:17:30.2529197Z Failed distributed compilations 0 2025-12-04T17:17:30.2529450Z 2025-12-04T17:17:30.2529578Z Non-cacheable reasons: 2025-12-04T17:17:30.2529888Z unknown source language 244 2025-12-04T17:17:30.2530243Z -E 65 2025-12-04T17:17:30.2530476Z 2025-12-04T17:17:30.2530747Z Cache location s3, name: ossci-compiler-cache-circleci-v2, prefix: / 2025-12-04T17:17:30.2531276Z Version (client) 0.10.0 2025-12-04T17:17:30.2531641Z + echo ::endgroup:: 2025-12-04T17:17:30.2532185Z ##[endgroup] 2025-12-04T17:17:30.2532440Z + cleanup_workspace 2025-12-04T17:17:30.2533009Z + echo 'sudo may print the following warning message that can be ignored. The chown command will still run.' 2025-12-04T17:17:30.2534542Z sudo may print the following warning message that can be ignored. The chown command will still run. 2025-12-04T17:17:30.2535666Z + echo ' sudo: setrlimit(RLIMIT_STACK): Operation not permitted' 2025-12-04T17:17:30.2536199Z sudo: setrlimit(RLIMIT_STACK): Operation not permitted 2025-12-04T17:17:30.2536982Z + echo 'For more details refer to https://github.com/sudo-project/sudo/issues/42' 2025-12-04T17:17:30.2537707Z For more details refer to https://github.com/sudo-project/sudo/issues/42 2025-12-04T17:17:30.2538340Z + sudo chown -R 1000 /var/lib/jenkins/workspace 2025-12-04T17:17:31.0570573Z ##[error]Process completed with exit code 1. 2025-12-04T17:17:31.0654915Z Prepare all required actions 2025-12-04T17:17:31.0655420Z Getting action download info 2025-12-04T17:17:31.2326665Z ##[group]Run ./.github/actions/pytest-cache-upload 2025-12-04T17:17:31.2327070Z with: 2025-12-04T17:17:31.2327330Z cache_dir: .pytest_cache 2025-12-04T17:17:31.2327626Z shard: 3 2025-12-04T17:17:31.2327904Z sha: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T17:17:31.2328309Z test_config: legacy_nvidia_driver 2025-12-04T17:17:31.2328901Z job_identifier: periodic_linux-jammy-cuda12.4-py3.10-gcc11 2025-12-04T17:17:31.2329339Z env: 2025-12-04T17:17:31.2329564Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:31.2329872Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:31.2330239Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:31.2330877Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:31.2331464Z ##[endgroup] 2025-12-04T17:17:31.2371461Z ##[group]Run nick-fields/retry@v3.0.0 2025-12-04T17:17:31.2371901Z with: 2025-12-04T17:17:31.2372132Z shell: bash 2025-12-04T17:17:31.2372401Z timeout_minutes: 5 2025-12-04T17:17:31.2372685Z max_attempts: 5 2025-12-04T17:17:31.2372948Z retry_wait_seconds: 30 2025-12-04T17:17:31.2373337Z command: set -eu python3 -m pip install boto3==1.35.42 2025-12-04T17:17:31.2373784Z polling_interval_seconds: 1 2025-12-04T17:17:31.2374109Z warning_on_retry: true 2025-12-04T17:17:31.2374396Z continue_on_error: false 2025-12-04T17:17:31.2374695Z env: 2025-12-04T17:17:31.2374932Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:31.2375220Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:31.2375580Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:31.2376235Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:31.2376805Z ##[endgroup] 2025-12-04T17:17:31.9213672Z Defaulting to user installation because normal site-packages is not writeable 2025-12-04T17:17:33.2397993Z Collecting boto3==1.35.42 2025-12-04T17:17:33.2570606Z Downloading boto3-1.35.42-py3-none-any.whl (139 kB) 2025-12-04T17:17:33.3537071Z Collecting s3transfer<0.11.0,>=0.10.0 2025-12-04T17:17:33.3575607Z Downloading s3transfer-0.10.4-py3-none-any.whl (83 kB) 2025-12-04T17:17:33.3636015Z Requirement already satisfied: jmespath<2.0.0,>=0.7.1 in /usr/lib/python3.9/site-packages (from boto3==1.35.42) (0.10.0) 2025-12-04T17:17:34.7778076Z Collecting botocore<1.36.0,>=1.35.42 2025-12-04T17:17:34.7820212Z Downloading botocore-1.35.99-py3-none-any.whl (13.3 MB) 2025-12-04T17:17:34.9622689Z Requirement already satisfied: urllib3<1.27,>=1.25.4 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.42->boto3==1.35.42) (1.25.10) 2025-12-04T17:17:34.9629554Z Requirement already satisfied: python-dateutil<3.0.0,>=2.1 in /usr/lib/python3.9/site-packages (from botocore<1.36.0,>=1.35.42->boto3==1.35.42) (2.8.1) 2025-12-04T17:17:35.2508722Z Requirement already satisfied: six>=1.5 in /usr/lib/python3.9/site-packages (from python-dateutil<3.0.0,>=2.1->botocore<1.36.0,>=1.35.42->boto3==1.35.42) (1.15.0) 2025-12-04T17:17:35.3555669Z Installing collected packages: botocore, s3transfer, boto3 2025-12-04T17:17:36.0013511Z Successfully installed boto3-1.35.42 botocore-1.35.99 s3transfer-0.10.4 2025-12-04T17:17:36.3359990Z Command completed after 1 attempt(s). 2025-12-04T17:17:36.3454651Z ##[group]Run python3 .github/scripts/pytest_cache.py \ 2025-12-04T17:17:36.3455180Z python3 .github/scripts/pytest_cache.py \ 2025-12-04T17:17:36.3455614Z  --upload \ 2025-12-04T17:17:36.3455968Z  --cache_dir "$GITHUB_WORKSPACE/$CACHE_DIR" \ 2025-12-04T17:17:36.3456404Z  --pr_identifier "$GITHUB_REF" \ 2025-12-04T17:17:36.3456812Z  --job_identifier "$JOB_IDENTIFIER" \ 2025-12-04T17:17:36.3457317Z  --sha "$SHA" \ 2025-12-04T17:17:36.3457646Z  --test_config "$TEST_CONFIG" \ 2025-12-04T17:17:36.3458008Z  --shard "$SHARD" \ 2025-12-04T17:17:36.3458562Z  --repo "$REPO" \ 2025-12-04T17:17:36.3458903Z  --temp_dir "$RUNNER_TEMP" \ 2025-12-04T17:17:36.3469527Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T17:17:36.3469969Z env: 2025-12-04T17:17:36.3470226Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:36.3470523Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:36.3470889Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:36.3471545Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:36.3472258Z CACHE_DIR: .pytest_cache 2025-12-04T17:17:36.3472655Z JOB_IDENTIFIER: periodic_linux-jammy-cuda12.4-py3.10-gcc11 2025-12-04T17:17:36.3473150Z SHA: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T17:17:36.3473554Z TEST_CONFIG: legacy_nvidia_driver 2025-12-04T17:17:36.3473873Z SHARD: 3 2025-12-04T17:17:36.3474124Z REPO: pytorch/pytorch 2025-12-04T17:17:36.3474411Z ##[endgroup] 2025-12-04T17:17:36.8725620Z PR identifier for `refs/heads/main` is `96e092540d6b3c4076e3d2bc6f1f9013` 2025-12-04T17:17:36.8728051Z Uploading cache with args Namespace(upload=True, download=False, cache_dir='/home/ec2-user/actions-runner/_work/pytorch/pytorch/.pytest_cache', pr_identifier='refs/heads/main', job_identifier='periodic_linux-jammy-cuda12.4-py3.10-gcc11', sha='ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32', test_config='legacy_nvidia_driver', shard='3', repo='pytorch/pytorch', temp_dir='/home/ec2-user/actions-runner/_work/_temp', bucket=None) 2025-12-04T17:17:36.8730452Z Zipping /home/ec2-user/actions-runner/_work/pytorch/pytorch/.pytest_cache 2025-12-04T17:17:36.8732017Z to /home/ec2-user/actions-runner/_work/_temp/zip-upload/pytest_cache/pytorch/pytorch/96e092540d6b3c4076e3d2bc6f1f9013/periodic_linux-jammy-cuda12_4-py3_10-gcc11/ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32/legacy_nvidia_driver/3 2025-12-04T17:17:36.8734482Z Uploading /home/ec2-user/actions-runner/_work/_temp/zip-upload/pytest_cache/pytorch/pytorch/96e092540d6b3c4076e3d2bc6f1f9013/periodic_linux-jammy-cuda12_4-py3_10-gcc11/ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32/legacy_nvidia_driver/3.zip 2025-12-04T17:17:36.8737097Z to s3://gha-artifacts/pytest_cache/pytorch/pytorch/96e092540d6b3c4076e3d2bc6f1f9013/periodic_linux-jammy-cuda12_4-py3_10-gcc11/ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32/legacy_nvidia_driver/3.zip 2025-12-04T17:17:36.9348840Z ##[group]Run cat test/**/*_toprint.log || true 2025-12-04T17:17:36.9349303Z cat test/**/*_toprint.log || true 2025-12-04T17:17:36.9359775Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T17:17:36.9360229Z env: 2025-12-04T17:17:36.9360483Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:36.9360783Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:36.9361147Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:36.9361803Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:36.9362398Z ##[endgroup] 2025-12-04T17:17:36.9465970Z cat: 'test/**/*_toprint.log': No such file or directory 2025-12-04T17:17:36.9497907Z ##[group]Run kill "$MONITOR_SCRIPT_PID" 2025-12-04T17:17:36.9498343Z kill "$MONITOR_SCRIPT_PID" 2025-12-04T17:17:36.9504788Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T17:17:36.9505227Z env: 2025-12-04T17:17:36.9505472Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:36.9505770Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:36.9506137Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:36.9506787Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:36.9507391Z MONITOR_SCRIPT_PID: 68777 2025-12-04T17:17:36.9507697Z ##[endgroup] 2025-12-04T17:17:36.9533455Z /home/ec2-user/actions-runner/_work/_temp/f12f97c1-8730-43d9-a5ac-4f9996ed3ed3.sh: line 1: kill: (68777) - No such process 2025-12-04T17:17:36.9535978Z ##[error]Process completed with exit code 1. 2025-12-04T17:17:36.9683764Z Prepare all required actions 2025-12-04T17:17:36.9684278Z Getting action download info 2025-12-04T17:17:37.0966224Z Download action repository 'seemethere/upload-artifact-s3@v5' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-12-04T17:17:37.3359326Z Download action repository 'actions/upload-artifact@v4' (SHA:ea165f8d65b6e75b540449e92b4886f43607fa02) 2025-12-04T17:17:37.7962753Z ##[group]Run ./.github/actions/upload-test-artifacts 2025-12-04T17:17:37.7963175Z with: 2025-12-04T17:17:37.7963660Z file-suffix: test-legacy_nvidia_driver-3-5-linux.g4dn.4xlarge.nvidia.gpu_57119749282 2025-12-04T17:17:37.7964361Z s3-bucket: gha-artifacts 2025-12-04T17:17:37.7964652Z env: 2025-12-04T17:17:37.7964903Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:37.7965213Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:37.7965568Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:37.7966227Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:37.7966863Z ##[endgroup] 2025-12-04T17:17:37.8013398Z ##[group]Run # Remove any previous test jsons if they exist 2025-12-04T17:17:37.8013941Z # Remove any previous test jsons if they exist 2025-12-04T17:17:37.8014374Z rm -f test-jsons-*.zip 2025-12-04T17:17:37.8014869Z zip -r "test-jsons-${FILE_SUFFIX}.zip" test/test-reports -i '*.json' 2025-12-04T17:17:37.8021965Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T17:17:37.8022390Z env: 2025-12-04T17:17:37.8022631Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:37.8022936Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:37.8023302Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:37.8023933Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:37.8024739Z FILE_SUFFIX: test-legacy_nvidia_driver-3-5-linux.g4dn.4xlarge.nvidia.gpu_57119749282 2025-12-04T17:17:37.8025309Z ##[endgroup] 2025-12-04T17:17:37.8284778Z adding: test/test-reports/td_exclusions-0dfcec7bd8b6f400f36f.json (deflated 82%) 2025-12-04T17:17:37.8289707Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-cf40a4f46675676c.json (deflated 91%) 2025-12-04T17:17:37.8299222Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-3767af6d1470a6b5.json (deflated 95%) 2025-12-04T17:17:37.8308064Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-bf1966c65cc0ce25.json (deflated 95%) 2025-12-04T17:17:37.8313639Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-4d0cbc1f4fa8e643.json (deflated 95%) 2025-12-04T17:17:37.8331677Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-6ce11cbc918a017c.json (deflated 94%) 2025-12-04T17:17:37.8333047Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-c850c8e21fb6d04c.json (deflated 85%) 2025-12-04T17:17:37.8334396Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-17e608243d035d18.json (deflated 85%) 2025-12-04T17:17:37.8335727Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-7d6e32f4c213ad66.json (deflated 90%) 2025-12-04T17:17:37.8343123Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-e7bcb2cd86408aa6.json (deflated 97%) 2025-12-04T17:17:37.8350432Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-54ad0f704656a459.json (deflated 97%) 2025-12-04T17:17:37.8352825Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-fb514581ec21877e.json (deflated 88%) 2025-12-04T17:17:37.8354546Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-1a3279675b79d710.json (deflated 88%) 2025-12-04T17:17:37.8356505Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-bfb3d1c8c3468bae.json (deflated 88%) 2025-12-04T17:17:37.8376003Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-aa152592c1ebda5c.json (deflated 92%) 2025-12-04T17:17:37.8396629Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-535feae09e5a3ebd.json (deflated 94%) 2025-12-04T17:17:37.8399228Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-b527a05b10cde953.json (deflated 88%) 2025-12-04T17:17:37.8401177Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-0142110b0a563d3e.json (deflated 88%) 2025-12-04T17:17:37.8408612Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-92fbe8f0c3defdbe.json (deflated 92%) 2025-12-04T17:17:37.8441044Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_codegen_dynamic_shapes/inductor.test_torchinductor_codegen_dynamic_shapes-3b673269684e2f76.json (deflated 94%) 2025-12-04T17:17:37.8447330Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-a02012f2395cf6ef.json (deflated 96%) 2025-12-04T17:17:37.8452992Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-96bec2e922d5155c.json (deflated 96%) 2025-12-04T17:17:37.8458923Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-ef6350a9f8bb75e3.json (deflated 95%) 2025-12-04T17:17:37.8542564Z adding: test/test-reports/python-pytest/dynamo.test_logging/dynamo.test_logging-77311c728804d279.json (deflated 94%) 2025-12-04T17:17:37.8609871Z adding: test/test-reports/python-pytest/dynamo.test_repros/dynamo.test_repros-9b6b927930117494.json (deflated 92%) 2025-12-04T17:17:37.8611622Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-97143c683457795e.json (deflated 86%) 2025-12-04T17:17:37.8613311Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-c6ae8965f2e7a602.json (deflated 85%) 2025-12-04T17:17:37.8615043Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-2f238f5f180b8489.json (deflated 85%) 2025-12-04T17:17:37.8616516Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-df27debc7fb08425.json (deflated 85%) 2025-12-04T17:17:37.8618098Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-39f974c40a0e5d07.json (deflated 85%) 2025-12-04T17:17:37.8619605Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-4c5f8d97ccaac9c3.json (deflated 85%) 2025-12-04T17:17:37.8621098Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-fda94603e2b70838.json (deflated 86%) 2025-12-04T17:17:37.8622567Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-ace483b73b471725.json (deflated 85%) 2025-12-04T17:17:37.8624037Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-841659f3603156bc.json (deflated 85%) 2025-12-04T17:17:37.8625511Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-e4e48c14b07c6744.json (deflated 85%) 2025-12-04T17:17:37.8627153Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-887ea8bf31a36206.json (deflated 85%) 2025-12-04T17:17:37.8628634Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-d486c0fac95f9182.json (deflated 85%) 2025-12-04T17:17:37.8630093Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-73bc989177a97c92.json (deflated 86%) 2025-12-04T17:17:37.8631575Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-9368d7516886ec65.json (deflated 85%) 2025-12-04T17:17:37.8633053Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-921cae3f6acd6a84.json (deflated 85%) 2025-12-04T17:17:37.8634603Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-8d346e15c82776ef.json (deflated 86%) 2025-12-04T17:17:37.8636101Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-1b686c821c7b9c77.json (deflated 85%) 2025-12-04T17:17:37.8637569Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-21dcaad0846c92ef.json (deflated 85%) 2025-12-04T17:17:37.8639066Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-633a2407c7cc9faa.json (deflated 85%) 2025-12-04T17:17:37.8640549Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-2a4f8a9a7e80b95a.json (deflated 85%) 2025-12-04T17:17:37.8642037Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-ab0a7cd9bbef732b.json (deflated 85%) 2025-12-04T17:17:37.8643516Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-81245131849bbecb.json (deflated 86%) 2025-12-04T17:17:37.8645010Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-4cf813d975e62c57.json (deflated 85%) 2025-12-04T17:17:37.8646501Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-f694c40280fe3f31.json (deflated 85%) 2025-12-04T17:17:37.8648026Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-a6d9b5e9da7bb417.json (deflated 85%) 2025-12-04T17:17:37.8649504Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-c31955170ca46735.json (deflated 85%) 2025-12-04T17:17:37.8650972Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-1b66a9993689b300.json (deflated 85%) 2025-12-04T17:17:37.8652438Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-79106310b2e4ab0d.json (deflated 86%) 2025-12-04T17:17:37.8653935Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-235ce3baacba0fb6.json (deflated 85%) 2025-12-04T17:17:37.8655420Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-00aa73936a934045.json (deflated 85%) 2025-12-04T17:17:37.8656940Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-333e44976f8ee863.json (deflated 85%) 2025-12-04T17:17:37.8658429Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-acd69551823a4dec.json (deflated 85%) 2025-12-04T17:17:37.8659921Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-c7f6019ad8b4bab5.json (deflated 85%) 2025-12-04T17:17:37.8661482Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-cdc355d159a50f1a.json (stored 0%) 2025-12-04T17:17:37.8662888Z adding: test/test-reports/python-pytest/inductor.test_deterministic/inductor.test_deterministic-142f090bc785acb6.json (deflated 69%) 2025-12-04T17:17:37.8664232Z adding: test/test-reports/python-pytest/inductor.test_deterministic/inductor.test_deterministic-2940c67a6de1d3d8.json (deflated 76%) 2025-12-04T17:17:37.8665577Z adding: test/test-reports/python-pytest/inductor.test_native_matmul/inductor.test_native_matmul-5845c7d4a4d90a90.json (deflated 74%) 2025-12-04T17:17:37.9058185Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-819d138cb6fe4c8f.json (deflated 87%) 2025-12-04T17:17:37.9102978Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-3d7372eabf0c20ef.json (deflated 87%) 2025-12-04T17:17:37.9147707Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-e6eefab2abf634a5.json (deflated 87%) 2025-12-04T17:17:37.9193488Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-fd020a6d441a5738.json (deflated 87%) 2025-12-04T17:17:37.9239647Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-b7f552c9fcec58a4.json (deflated 87%) 2025-12-04T17:17:37.9285555Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-ba19ebfa7960f6c4.json (deflated 87%) 2025-12-04T17:17:37.9631190Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-6aac4defd9d3a3cc.json (deflated 87%) 2025-12-04T17:17:37.9675743Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-041aad81ad32b1c4.json (deflated 87%) 2025-12-04T17:17:37.9720708Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-043eb8ddb5d84fd4.json (deflated 87%) 2025-12-04T17:17:37.9765480Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-555ea3dc15f2a8d0.json (deflated 87%) 2025-12-04T17:17:37.9810455Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-01d96e43057cf9b8.json (deflated 87%) 2025-12-04T17:17:37.9855345Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-563749739dc7a57a.json (deflated 87%) 2025-12-04T17:17:37.9920388Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-d89a888c67de20be.json (deflated 87%) 2025-12-04T17:17:37.9921808Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a12b7a446d46a036.json (deflated 84%) 2025-12-04T17:17:37.9923166Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-c354a873b70ad9a4.json (deflated 83%) 2025-12-04T17:17:37.9924515Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-72b2d7d92c2f3d43.json (deflated 83%) 2025-12-04T17:17:37.9925861Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-c8076b5dd44238e4.json (deflated 83%) 2025-12-04T17:17:37.9927198Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-d2d17ca8c68c28e7.json (deflated 83%) 2025-12-04T17:17:37.9928534Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-b56c2dd706b17a20.json (deflated 83%) 2025-12-04T17:17:37.9929879Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-fa3c7a9046cc8737.json (deflated 83%) 2025-12-04T17:17:37.9931424Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3da5ad44ee5d80af.json (deflated 83%) 2025-12-04T17:17:37.9932788Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2586fe9efcb2d9c1.json (deflated 83%) 2025-12-04T17:17:37.9934127Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-4d9d69bfeda71662.json (deflated 83%) 2025-12-04T17:17:37.9935469Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0045cf67af4f7e36.json (deflated 83%) 2025-12-04T17:17:37.9936927Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-38fbcefce9021f2a.json (deflated 83%) 2025-12-04T17:17:37.9938266Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0625eb2c44495ce0.json (deflated 83%) 2025-12-04T17:17:37.9939619Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-ea417fd0de1ef597.json (deflated 83%) 2025-12-04T17:17:37.9940951Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-e58bcfb06df4ce4d.json (deflated 83%) 2025-12-04T17:17:37.9942302Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0c1484308d51b00a.json (deflated 83%) 2025-12-04T17:17:37.9943637Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-9682f9212923ea0d.json (deflated 83%) 2025-12-04T17:17:37.9944975Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-ef1f71bad0a01b52.json (deflated 83%) 2025-12-04T17:17:37.9946304Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3d18fa53804224f4.json (deflated 85%) 2025-12-04T17:17:37.9947642Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2256fbb35262b330.json (deflated 82%) 2025-12-04T17:17:37.9948984Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a993cfee9c5b5572.json (deflated 82%) 2025-12-04T17:17:37.9950387Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-4097f4a492ce8bb6.json (deflated 84%) 2025-12-04T17:17:37.9951711Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-9df84d07ff6a448b.json (deflated 82%) 2025-12-04T17:17:37.9953055Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-b20a3959da2014da.json (deflated 82%) 2025-12-04T17:17:37.9954397Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-89205029d9c04efd.json (deflated 83%) 2025-12-04T17:17:37.9955742Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-476b174d4b08ad56.json (deflated 83%) 2025-12-04T17:17:37.9957066Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-99551b4e3863e77a.json (deflated 83%) 2025-12-04T17:17:37.9958405Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-498b08ff107b8914.json (deflated 82%) 2025-12-04T17:17:37.9959740Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0e343f505ff50c95.json (deflated 82%) 2025-12-04T17:17:37.9961075Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-486e74c9978d3a7f.json (deflated 82%) 2025-12-04T17:17:37.9962420Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a0676d6303da3615.json (deflated 82%) 2025-12-04T17:17:37.9963849Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-e27b1d19486b6954.json (deflated 82%) 2025-12-04T17:17:37.9965206Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3f1ba8060f3abc48.json (deflated 82%) 2025-12-04T17:17:37.9966554Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0209a84b5ca28602.json (deflated 83%) 2025-12-04T17:17:37.9967894Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3d838b8cb24fd196.json (deflated 83%) 2025-12-04T17:17:37.9969216Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-6a6a2484271011a1.json (deflated 83%) 2025-12-04T17:17:37.9970613Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-9356c0308b0c66f7.json (deflated 82%) 2025-12-04T17:17:37.9971946Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-51758daa3a3dc599.json (deflated 82%) 2025-12-04T17:17:37.9973282Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a0d42e3d123a44e8.json (deflated 82%) 2025-12-04T17:17:37.9974625Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-20d10a20187a2cb9.json (deflated 84%) 2025-12-04T17:17:37.9975977Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-5f3be1bf80a1ce40.json (deflated 81%) 2025-12-04T17:17:37.9977394Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-baeb09a4e62b9d77.json (deflated 81%) 2025-12-04T17:17:37.9978732Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2cdd67c45b570140.json (deflated 84%) 2025-12-04T17:17:37.9980074Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2423c3b275573349.json (deflated 82%) 2025-12-04T17:17:37.9981419Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-66e69110dfb99f31.json (deflated 82%) 2025-12-04T17:17:37.9982761Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-4e57fa3cbc487129.json (deflated 84%) 2025-12-04T17:17:37.9984135Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-7331926cdb107464.json (deflated 82%) 2025-12-04T17:17:37.9985476Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-7d66ea27b5ac2c65.json (deflated 82%) 2025-12-04T17:17:37.9986839Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-356c1a001f2d20d8.json (deflated 86%) 2025-12-04T17:17:37.9988184Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-426f4e961e373829.json (deflated 83%) 2025-12-04T17:17:37.9989526Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-1a1453b5e01c9120.json (deflated 83%) 2025-12-04T17:17:37.9990849Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-09c678f485469fcb.json (deflated 50%) 2025-12-04T17:17:37.9992245Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-fb2ec7b0e0c81fea.json (deflated 88%) 2025-12-04T17:17:37.9993683Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6ae6fdb299843870.json (deflated 82%) 2025-12-04T17:17:37.9995111Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2362ba98ed8b8900.json (deflated 82%) 2025-12-04T17:17:37.9996683Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-52334a8157857cd5.json (deflated 82%) 2025-12-04T17:17:37.9998228Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0b589120e3bef198.json (deflated 82%) 2025-12-04T17:17:37.9999660Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f9457e8d87a8d0e5.json (deflated 82%) 2025-12-04T17:17:38.0001084Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e4f574242b541525.json (deflated 82%) 2025-12-04T17:17:38.0002497Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-19baa3ae62072b2f.json (deflated 82%) 2025-12-04T17:17:38.0003973Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-debf0ef50304edda.json (deflated 82%) 2025-12-04T17:17:38.0005405Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a0af7423e85581ef.json (deflated 82%) 2025-12-04T17:17:38.0006838Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6e37efe6d0d7b5f7.json (deflated 82%) 2025-12-04T17:17:38.0008277Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-64b9c8b908c45f51.json (deflated 82%) 2025-12-04T17:17:38.0009693Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3343e9ff31c5255a.json (deflated 93%) 2025-12-04T17:17:38.0011125Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ae334ed847078f28.json (deflated 86%) 2025-12-04T17:17:38.0012557Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0eab24e125a2cf7d.json (deflated 86%) 2025-12-04T17:17:38.0013992Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6060ecdf9a2577f9.json (deflated 84%) 2025-12-04T17:17:38.0015437Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0eb661cc20714db1.json (deflated 83%) 2025-12-04T17:17:38.0016970Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-5774b4beb84c0975.json (deflated 83%) 2025-12-04T17:17:38.0018404Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-74651aae35dd9844.json (deflated 83%) 2025-12-04T17:17:38.0019838Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-7a5a8b9ba5d60164.json (deflated 83%) 2025-12-04T17:17:38.0021266Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-26faac4e438f6204.json (deflated 83%) 2025-12-04T17:17:38.0022679Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-aa20002b977e7ab4.json (deflated 89%) 2025-12-04T17:17:38.0024106Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6316dfd7aad65a93.json (deflated 83%) 2025-12-04T17:17:38.0025534Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d1f50caa5a44c0ee.json (deflated 83%) 2025-12-04T17:17:38.0026971Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-8edf4f7e001e74f9.json (deflated 85%) 2025-12-04T17:17:38.0028385Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-defc51812af7e0b0.json (deflated 83%) 2025-12-04T17:17:38.0029816Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0bc167e05b941aff.json (deflated 83%) 2025-12-04T17:17:38.0031320Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-98b78c8ed19e0643.json (deflated 83%) 2025-12-04T17:17:38.0032747Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-cc69c969909d9707.json (deflated 83%) 2025-12-04T17:17:38.0034176Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9977f3a62a4ac5d4.json (deflated 83%) 2025-12-04T17:17:38.0035596Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ec39337b11cd8a4d.json (deflated 86%) 2025-12-04T17:17:38.0037031Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f9d493fa40deaa3c.json (deflated 86%) 2025-12-04T17:17:38.0038492Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-492f33e438c88cbb.json (deflated 86%) 2025-12-04T17:17:38.0039930Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-cfe26d7f4e2f142c.json (deflated 83%) 2025-12-04T17:17:38.0041342Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-85dd835c0392b89d.json (deflated 83%) 2025-12-04T17:17:38.0042767Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d745daffa2211db3.json (deflated 83%) 2025-12-04T17:17:38.0044192Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-29d661e244421a0d.json (deflated 86%) 2025-12-04T17:17:38.0045614Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6a080fc2c141366f.json (deflated 83%) 2025-12-04T17:17:38.0047031Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b727a3d5cb43a100.json (deflated 83%) 2025-12-04T17:17:38.0048450Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9820049c2fc9b56d.json (deflated 88%) 2025-12-04T17:17:38.0049876Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-bcfbf334a5fb1b19.json (deflated 83%) 2025-12-04T17:17:38.0051347Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-55cd5c19b34efcf0.json (deflated 83%) 2025-12-04T17:17:38.0052776Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3f17acc66e50cb98.json (deflated 86%) 2025-12-04T17:17:38.0054198Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a7cd72d0cd16fd7c.json (deflated 86%) 2025-12-04T17:17:38.0055653Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-45e1674e77992a96.json (deflated 86%) 2025-12-04T17:17:38.0057135Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-384ea3a6b9159dd5.json (deflated 92%) 2025-12-04T17:17:38.0058569Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-40502088a31692e3.json (deflated 83%) 2025-12-04T17:17:38.0059997Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-1a1d3c182ee91d42.json (deflated 83%) 2025-12-04T17:17:38.0061406Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9bd3fb708dd83987.json (deflated 88%) 2025-12-04T17:17:38.0062836Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-991a5e87e1946535.json (deflated 86%) 2025-12-04T17:17:38.0064340Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-61bd448d788d4193.json (deflated 86%) 2025-12-04T17:17:38.0065759Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0c87f6ca176b74e4.json (deflated 89%) 2025-12-04T17:17:38.0067170Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-00981048468ad3fe.json (deflated 83%) 2025-12-04T17:17:38.0068593Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-7f9686778c56ad66.json (deflated 83%) 2025-12-04T17:17:38.0070015Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dc76d780af365e97.json (deflated 83%) 2025-12-04T17:17:38.0071482Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3d712c264776fdc2.json (deflated 83%) 2025-12-04T17:17:38.0072910Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4da37e7aafa09519.json (deflated 83%) 2025-12-04T17:17:38.0074327Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4f9e658df3b9ce6d.json (deflated 86%) 2025-12-04T17:17:38.0075764Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3d80a9d76d3be932.json (deflated 82%) 2025-12-04T17:17:38.0077192Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ede94f48d428153a.json (deflated 82%) 2025-12-04T17:17:38.0078627Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6ae027693c5dbbf4.json (deflated 82%) 2025-12-04T17:17:38.0080054Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-48db12bdca4a53d6.json (deflated 82%) 2025-12-04T17:17:38.0081489Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0c861fbac71335e0.json (deflated 82%) 2025-12-04T17:17:38.0082925Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ffbb324af5af0672.json (deflated 82%) 2025-12-04T17:17:38.0084391Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-042dc549bdd77a45.json (deflated 82%) 2025-12-04T17:17:38.0085799Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0942d088a0709320.json (deflated 82%) 2025-12-04T17:17:38.0087224Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c3fbd4a060fe6236.json (deflated 93%) 2025-12-04T17:17:38.0088659Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f308a9dbba7be53a.json (deflated 86%) 2025-12-04T17:17:38.0090101Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-bd2ede9c3a5cd1d1.json (deflated 86%) 2025-12-04T17:17:38.0091530Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d27c0d840d6fe564.json (deflated 83%) 2025-12-04T17:17:38.0092942Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2f600ca197676d4b.json (deflated 83%) 2025-12-04T17:17:38.0094375Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ecad5133d6245a6a.json (deflated 83%) 2025-12-04T17:17:38.0095802Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d38b619605c8f606.json (deflated 86%) 2025-12-04T17:17:38.0097652Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ca04900b7cf26510.json (deflated 86%) 2025-12-04T17:17:38.0099234Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-8a1dc4f258524121.json (deflated 86%) 2025-12-04T17:17:38.0100671Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a0112c740677bec4.json (deflated 83%) 2025-12-04T17:17:38.0102100Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-67764f47d32c222e.json (deflated 83%) 2025-12-04T17:17:38.0103529Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c39fdfd87cf64578.json (deflated 83%) 2025-12-04T17:17:38.0104992Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-48a46382e3216972.json (deflated 83%) 2025-12-04T17:17:38.0106418Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d40d195c41f12813.json (deflated 83%) 2025-12-04T17:17:38.0107848Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-35b201d0b4c8de0a.json (deflated 83%) 2025-12-04T17:17:38.0109311Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e10bae22111bfdec.json (deflated 84%) 2025-12-04T17:17:38.0110737Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-5822c2fc7d137525.json (deflated 83%) 2025-12-04T17:17:38.0112177Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a6ba20f02ccefade.json (deflated 83%) 2025-12-04T17:17:38.0113601Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dc63b1e627261305.json (deflated 83%) 2025-12-04T17:17:38.0115039Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9dda8854f56bbc5e.json (deflated 83%) 2025-12-04T17:17:38.0116475Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d35f802df30bd39d.json (deflated 83%) 2025-12-04T17:17:38.0117966Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d1473b4b18944240.json (deflated 84%) 2025-12-04T17:17:38.0119384Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c84649193ffdba8a.json (deflated 83%) 2025-12-04T17:17:38.0120817Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b127809f84c30c37.json (deflated 83%) 2025-12-04T17:17:38.0122251Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-1a7c4cad159549d9.json (deflated 84%) 2025-12-04T17:17:38.0123685Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2f388a73f046ed87.json (deflated 83%) 2025-12-04T17:17:38.0125103Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d68d80f1ba61c89f.json (deflated 83%) 2025-12-04T17:17:38.0126542Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b8a30a3f316b7f0b.json (deflated 86%) 2025-12-04T17:17:38.0127975Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d3917b026aff3e64.json (deflated 86%) 2025-12-04T17:17:38.0129396Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c0c657ac1df627f9.json (deflated 86%) 2025-12-04T17:17:38.0130827Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-175aa275ea059250.json (deflated 86%) 2025-12-04T17:17:38.0132307Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-5ce17ea59d8fd893.json (deflated 83%) 2025-12-04T17:17:38.0133738Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4874d0c887fabf8b.json (deflated 83%) 2025-12-04T17:17:38.0135187Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a28d30f3a161bea7.json (deflated 83%) 2025-12-04T17:17:38.0136621Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-23cf59059171b3d3.json (deflated 83%) 2025-12-04T17:17:38.0138132Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-14e4b4194810001e.json (deflated 83%) 2025-12-04T17:17:38.0139563Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4b58d6ae0af928ee.json (deflated 87%) 2025-12-04T17:17:38.0140999Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9605f50d5f418e61.json (deflated 83%) 2025-12-04T17:17:38.0142427Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2668b63c8efdb7a5.json (deflated 83%) 2025-12-04T17:17:38.0143839Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-8d527629568e40ed.json (deflated 86%) 2025-12-04T17:17:38.0145258Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a273900f3029c317.json (deflated 83%) 2025-12-04T17:17:38.0146685Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dcefced86a4c35d8.json (deflated 83%) 2025-12-04T17:17:38.0148122Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e15b12282035b26d.json (deflated 86%) 2025-12-04T17:17:38.0149554Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e08a7bc39929d142.json (deflated 83%) 2025-12-04T17:17:38.0150973Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-13b16b357405c710.json (deflated 83%) 2025-12-04T17:17:38.0152439Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9275f55235cebf56.json (deflated 86%) 2025-12-04T17:17:38.0153853Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a0875a42ac68f671.json (deflated 86%) 2025-12-04T17:17:38.0155282Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-56c85a19db7b6fad.json (deflated 86%) 2025-12-04T17:17:38.0156697Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b5990bb8b6e78b98.json (deflated 83%) 2025-12-04T17:17:38.0158118Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9e1ed788d5899650.json (deflated 83%) 2025-12-04T17:17:38.0159544Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e97cc37d8f662932.json (deflated 83%) 2025-12-04T17:17:38.0160982Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d35e9b6eebdaac7b.json (deflated 86%) 2025-12-04T17:17:38.0162401Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f6ccc2d46fa5c9ea.json (deflated 86%) 2025-12-04T17:17:38.0163845Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dac88d9039f141ad.json (deflated 86%) 2025-12-04T17:17:38.0165345Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-1d81442a0ab3908c.json (deflated 89%) 2025-12-04T17:17:38.0166769Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-773b425f2e921837.json (deflated 83%) 2025-12-04T17:17:38.0168194Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-09fb0d74c7a00548.json (deflated 83%) 2025-12-04T17:17:38.0169609Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2f000bd6c93ab369.json (deflated 83%) 2025-12-04T17:17:38.0171036Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-64628aab5ce6aeb9.json (deflated 83%) 2025-12-04T17:17:38.0172512Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9fedf51ad965eb7d.json (deflated 83%) 2025-12-04T17:17:38.0173950Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c5a3e4c540d200e7.json (deflated 86%) 2025-12-04T17:17:38.0175373Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ab8ed79c5f08a6c3.json (deflated 83%) 2025-12-04T17:17:38.0176799Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-43cb3996049bee78.json (deflated 83%) 2025-12-04T17:17:38.0178287Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d3a2df1a150e9085.json (deflated 83%) 2025-12-04T17:17:38.0179726Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2b2d696e555e6fc9.json (deflated 83%) 2025-12-04T17:17:38.0181162Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-449e4fae7bbefa8d.json (deflated 83%) 2025-12-04T17:17:38.0182580Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4631de08a3dd09e1.json (deflated 86%) 2025-12-04T17:17:38.0184010Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-7cb60f50c6534a73.json (deflated 83%) 2025-12-04T17:17:38.0185519Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-bbe75c180c40946f.json (deflated 83%) 2025-12-04T17:17:38.0186955Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-19b98f6e1bce5088.json (deflated 89%) 2025-12-04T17:17:38.0188388Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-fd7621eb24f73b8c.json (deflated 83%) 2025-12-04T17:17:38.0189801Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e11a9d6da1493a65.json (deflated 83%) 2025-12-04T17:17:38.0191239Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d245bc4e7ed2a74d.json (deflated 87%) 2025-12-04T17:17:38.0192497Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-d7ddbde7cf40ab13.json (deflated 98%) 2025-12-04T17:17:38.0193576Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-63a23e8fc1ef8137.json (deflated 85%) 2025-12-04T17:17:38.0194633Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-11f9c4926783dfef.json (deflated 85%) 2025-12-04T17:17:38.0195694Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-737659f69f9145f2.json (deflated 85%) 2025-12-04T17:17:38.0196924Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2099d82b2931fb99.json (deflated 85%) 2025-12-04T17:17:38.0197981Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-5f0933bf702563bb.json (deflated 85%) 2025-12-04T17:17:38.0199154Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-ed1b6e1c95269b63.json (deflated 85%) 2025-12-04T17:17:38.0200226Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-ccb90d6b004c5139.json (deflated 85%) 2025-12-04T17:17:38.0201294Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a3b2bb4f2a94fb32.json (deflated 85%) 2025-12-04T17:17:38.0202362Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-581c94c41301ad68.json (deflated 85%) 2025-12-04T17:17:38.0203409Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-809b8dfa72183a04.json (deflated 85%) 2025-12-04T17:17:38.0204525Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-59f3dc780871f4a4.json (deflated 85%) 2025-12-04T17:17:38.0205594Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2648ef692c69d7ff.json (deflated 91%) 2025-12-04T17:17:38.0206667Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-1e4f023d2e76e67f.json (deflated 85%) 2025-12-04T17:17:38.0207726Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-88a75603bd167dcb.json (deflated 85%) 2025-12-04T17:17:38.0208791Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-58d14c6ca3ad4514.json (deflated 85%) 2025-12-04T17:17:38.0209853Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-4358034553dc4d55.json (deflated 85%) 2025-12-04T17:17:38.0210904Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-e604dacd305a4e36.json (deflated 85%) 2025-12-04T17:17:38.0211975Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-29c18a70d3fdf25d.json (deflated 85%) 2025-12-04T17:17:38.0213042Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2a489f89b00449c3.json (deflated 85%) 2025-12-04T17:17:38.0214109Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-1e3f7f3e1ddc719e.json (deflated 85%) 2025-12-04T17:17:38.0215164Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-0a2928546dd5ddd2.json (deflated 85%) 2025-12-04T17:17:38.0216298Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-36be7a929b5f8cea.json (deflated 85%) 2025-12-04T17:17:38.0217427Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-db4ab888748348ef.json (deflated 85%) 2025-12-04T17:17:38.0218496Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a61e284ca72dd15d.json (deflated 96%) 2025-12-04T17:17:38.0219557Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-8da3a511cea25a62.json (deflated 85%) 2025-12-04T17:17:38.0220645Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-aaebada1c5676813.json (deflated 85%) 2025-12-04T17:17:38.0221730Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-c9e9611bbdf13a2b.json (deflated 85%) 2025-12-04T17:17:38.0222797Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-963810090dd7c38b.json (deflated 85%) 2025-12-04T17:17:38.0223857Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-202f1cfa5f6f2e59.json (deflated 85%) 2025-12-04T17:17:38.0224934Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a5f1ba0194e0f361.json (deflated 85%) 2025-12-04T17:17:38.0225995Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-784ae621331a35c7.json (deflated 85%) 2025-12-04T17:17:38.0227057Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2c9da68c717b465c.json (deflated 85%) 2025-12-04T17:17:38.0228118Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a06aff42aa9d7210.json (deflated 85%) 2025-12-04T17:17:38.0229266Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-66f0187c8b5c56f0.json (deflated 85%) 2025-12-04T17:17:38.0230341Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a93a155b666a377d.json (deflated 85%) 2025-12-04T17:17:38.0231410Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-f6900cb9a4d4e17a.json (deflated 91%) 2025-12-04T17:17:38.0232471Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-b4de48ce2713ebfd.json (deflated 85%) 2025-12-04T17:17:38.0233543Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-9d9a1cc8efaa0b76.json (deflated 85%) 2025-12-04T17:17:38.0234624Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-f2c6d1655dbd05d5.json (deflated 85%) 2025-12-04T17:17:38.0235728Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-30cf0cd6fabb5398.json (deflated 85%) 2025-12-04T17:17:38.0236781Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-29e9b8d6f2f6950a.json (deflated 85%) 2025-12-04T17:17:38.0237857Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-0ebf80986aa75a9d.json (deflated 85%) 2025-12-04T17:17:38.0238926Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-ed818918883fb5a6.json (deflated 85%) 2025-12-04T17:17:38.0239995Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-688ec0aa50c9a8c6.json (deflated 85%) 2025-12-04T17:17:38.0241047Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-6fda2bf2ae878b46.json (deflated 85%) 2025-12-04T17:17:38.0242116Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-0857a7a11662fc06.json (deflated 85%) 2025-12-04T17:17:38.0243189Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a3e03d3f7a7bbb81.json (deflated 85%) 2025-12-04T17:17:38.0244266Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-829800aaf6d2511a.json (deflated 99%) 2025-12-04T17:17:38.0245325Z adding: test/test-reports/python-pytest/test_dataloader/test_dataloader-f02fdb9181a8d8ee.json (deflated 90%) 2025-12-04T17:17:38.0246352Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-2c8b2ff0efe02fa9.json (deflated 95%) 2025-12-04T17:17:38.0247371Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-0c1ee1f6151b681e.json (deflated 95%) 2025-12-04T17:17:38.0248354Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-68b4836d5d5fa3ca.json (deflated 95%) 2025-12-04T17:17:38.0397568Z adding: test/test-reports/python-pytest/test_meta/test_meta-029b8ed96c60bfb7.json (deflated 97%) 2025-12-04T17:17:38.0411472Z adding: test/test-reports/python-pytest/test_ops_jit/test_ops_jit-9f733effbdfc0369.json (deflated 95%) 2025-12-04T17:17:38.0432603Z adding: test/test-reports/python-pytest/test_nestedtensor/test_nestedtensor-8e0ff602fd37df16.json (deflated 97%) 2025-12-04T17:17:38.0434878Z adding: test/test-reports/python-pytest/test_public_bindings/test_public_bindings-6fed1053bfc25f05.json (deflated 83%) 2025-12-04T17:17:38.0523748Z adding: test/test-reports/python-pytest/test_ops/test_ops-b6bd03fdf7fbcee0.json (deflated 96%) 2025-12-04T17:17:38.0600650Z adding: test/test-reports/python-pytest/test_ops/test_ops-e04c41c7bbbdce7a.json (deflated 96%) 2025-12-04T17:17:38.0679534Z adding: test/test-reports/python-pytest/test_ops/test_ops-6f4cd93b78224a58.json (deflated 96%) 2025-12-04T17:17:38.0721223Z adding: test/test-reports/python-pytest/functorch.test_ops/functorch.test_ops-37d651ead30b362d.json (deflated 95%) 2025-12-04T17:17:38.0754970Z adding: test/test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-2355f4f75c6e0706.json (deflated 95%) 2025-12-04T17:17:38.0764056Z adding: test/test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-db01917ff3ce307b.json (deflated 96%) 2025-12-04T17:17:38.0773094Z adding: test/test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-12ecde9ddbb651c9.json (deflated 96%) 2025-12-04T17:17:38.0815536Z adding: test/test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-94056324a12f0026.json (deflated 96%) 2025-12-04T17:17:38.0816719Z adding: test/test-reports/python-pytest/dynamo.test_after_aot/dynamo.test_after_aot-9dfa8d2ed594e793.json (deflated 59%) 2025-12-04T17:17:38.0819540Z adding: test/test-reports/python-pytest/inductor.test_snode_runtime/inductor.test_snode_runtime-804cfaf2d58e1050.json (deflated 93%) 2025-12-04T17:17:38.0927281Z adding: test/test-reports/python-pytest/inductor.test_minifier/inductor.test_minifier-e48754fb8b6df415.json (deflated 96%) 2025-12-04T17:17:38.1001023Z adding: test/test-reports/python-pytest/inductor.test_compiled_autograd/inductor.test_compiled_autograd-d055bdf97c26d8bd.json (deflated 93%) 2025-12-04T17:17:38.1002464Z adding: test/test-reports/python-pytest/inductor.test_layout_optim/inductor.test_layout_optim-50eed008ca1555c3.json (stored 0%) 2025-12-04T17:17:38.1006962Z adding: test/test-reports/python-pytest/dynamo.test_unspec/dynamo.test_unspec-4bfba7972407dae5.json (deflated 86%) 2025-12-04T17:17:38.1008141Z adding: test/test-reports/python-pytest/inductor.test_mmdecomp/inductor.test_mmdecomp-6ec8d10db2d708da.json (deflated 92%) 2025-12-04T17:17:38.1018055Z adding: test/test-reports/python-pytest/dynamo.test_ctx_manager/dynamo.test_ctx_manager-2753bd6e949ccdb3.json (deflated 93%) 2025-12-04T17:17:38.1037175Z adding: test/test-reports/python-pytest/dynamo.test_exc/dynamo.test_exc-e06b4c2ef4a1702a.json (deflated 95%) 2025-12-04T17:17:38.1092968Z adding: test/test-reports/python-pytest/dynamo.test_misc/dynamo.test_misc-43b95b63c3862211.json (deflated 90%) 2025-12-04T17:17:38.1102645Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor_arrayref/inductor.test_aot_inductor_arrayref-d7936498b94cbe56.json (deflated 94%) 2025-12-04T17:17:38.1104683Z adding: test/test-reports/python-pytest/inductor.test_provenance_tracing/inductor.test_provenance_tracing-e41a43dd8ef6019c.json (deflated 88%) 2025-12-04T17:17:38.1106086Z adding: test/test-reports/python-pytest/dynamo.test_buffers_override/dynamo.test_buffers_override-01fc103efe1995d6.json (deflated 70%) 2025-12-04T17:17:38.1107607Z adding: test/test-reports/python-pytest/inductor.test_inplacing_pass/inductor.test_inplacing_pass-d670df374fd1e952.json (deflated 90%) 2025-12-04T17:17:38.1110529Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor_custom_ops/inductor.test_aot_inductor_custom_ops-a0ad7eb32926bcd8.json (deflated 94%) 2025-12-04T17:17:38.1112001Z adding: test/test-reports/python-pytest/inductor.test_split_cat_fx_passes/inductor.test_split_cat_fx_passes-4aa43e0fa2d59cfb.json (deflated 86%) 2025-12-04T17:17:38.1113327Z adding: test/test-reports/python-pytest/inductor.test_profiler/inductor.test_profiler-cdccc94d406caaec.json (deflated 80%) 2025-12-04T17:17:38.1114634Z adding: test/test-reports/python-pytest/inductor.test_memory_planning/inductor.test_memory_planning-480d70b98c6f91e2.json (deflated 73%) 2025-12-04T17:17:38.1115997Z adding: test/test-reports/python-pytest/inductor.test_mem_estimation/inductor.test_mem_estimation-f5f67f53fc542f3b.json (deflated 77%) 2025-12-04T17:17:38.1117209Z adding: test/test-reports/python-pytest/dynamo.test_view/dynamo.test_view-5c939a16d5367651.json (deflated 84%) 2025-12-04T17:17:38.1118399Z adding: test/test-reports/python-pytest/inductor.test_cutlass_evt/inductor.test_cutlass_evt-a51e1bccbd1b02fd.json (deflated 89%) 2025-12-04T17:17:38.1119664Z adding: test/test-reports/python-pytest/dynamo.test_reconstruct/dynamo.test_reconstruct-7d75ef23ae79ec60.json (deflated 87%) 2025-12-04T17:17:38.1121354Z adding: test/test-reports/python-pytest/dynamo.test_aot_autograd/dynamo.test_aot_autograd-89fcf170216dd13a.json (deflated 88%) 2025-12-04T17:17:38.1171846Z adding: test/test-reports/python-pytest/export.test_cpp_serdes/export.test_cpp_serdes-1f0305d948167935.json (deflated 96%) 2025-12-04T17:17:38.1173356Z adding: test/test-reports/python-pytest/inductor.test_block_analysis/inductor.test_block_analysis-53978d8777987ceb.json (deflated 89%) 2025-12-04T17:17:38.1176702Z adding: test/test-reports/python-pytest/dynamo.test_subgraphs/dynamo.test_subgraphs-e09a516058cd0e01.json (deflated 94%) 2025-12-04T17:17:38.1178004Z adding: test/test-reports/python-pytest/dynamo.test_pre_dispatch/dynamo.test_pre_dispatch-4a36f2c5e912fcfb.json (deflated 79%) 2025-12-04T17:17:38.1179391Z adding: test/test-reports/python-pytest/inductor.test_custom_post_grad_passes/inductor.test_custom_post_grad_passes-e88ba17552693664.json (deflated 87%) 2025-12-04T17:17:38.1180763Z adding: test/test-reports/python-pytest/dynamo.test_fx_annotate/dynamo.test_fx_annotate-ea1b8128153e57c8.json (deflated 81%) 2025-12-04T17:17:38.1181973Z adding: test/test-reports/python-pytest/dynamo.test_pgo/dynamo.test_pgo-b9bc1426bb4bf646.json (deflated 81%) 2025-12-04T17:17:38.1183145Z adding: test/test-reports/python-pytest/export.test_export_opinfo/export.test_export_opinfo-d8b8b1860a5b7382.json (deflated 87%) 2025-12-04T17:17:38.1201371Z adding: test/test-reports/python-pytest/inductor.test_control_flow/inductor.test_control_flow-12b6e2c7b80b8695.json (deflated 97%) 2025-12-04T17:17:38.1203243Z adding: test/test-reports/python-pytest/dynamo.test_compile/dynamo.test_compile-1298df40b08e7720.json (deflated 84%) 2025-12-04T17:17:38.1207068Z adding: test/test-reports/python-pytest/dynamo.test_nested_graph_breaks/dynamo.test_nested_graph_breaks-3ce3fc0c984a2ae0.json (deflated 91%) 2025-12-04T17:17:38.1208479Z adding: test/test-reports/python-pytest/inductor.test_needs_exact_strides/inductor.test_needs_exact_strides-afd07b0a4c4dd4e2.json (deflated 70%) 2025-12-04T17:17:38.1209957Z adding: test/test-reports/python-pytest/inductor.test_split_cat_fx_aten_passes/inductor.test_split_cat_fx_aten_passes-d90ed49c774dd94e.json (deflated 82%) 2025-12-04T17:17:38.1211275Z adding: test/test-reports/python-pytest/dynamo.test_resume/dynamo.test_resume-1b040f63515910f4.json (deflated 42%) 2025-12-04T17:17:38.1212602Z adding: test/test-reports/python-pytest/dynamo.test_backward_higher_order_ops/dynamo.test_backward_higher_order_ops-8f7cdfdf9264a3b0.json (deflated 77%) 2025-12-04T17:17:38.1214088Z adding: test/test-reports/python-pytest/inductor.test_custom_partitioner_fn/inductor.test_custom_partitioner_fn-44a32a2607e301e4.json (deflated 50%) 2025-12-04T17:17:38.1215531Z adding: test/test-reports/python-pytest/dynamo.test_debug_utils/dynamo.test_debug_utils-d82e5bff075a86e1.json (deflated 76%) 2025-12-04T17:17:38.1216724Z adding: test/test-reports/python-pytest/dynamo.test_base_hop/dynamo.test_base_hop-6ccb089675742efe.json (deflated 79%) 2025-12-04T17:17:38.1217938Z adding: test/test-reports/python-pytest/dynamo.test_package/dynamo.test_package-9643d20377a48274.json (deflated 93%) 2025-12-04T17:17:38.1223611Z adding: test/test-reports/python-pytest/dynamo.test_aot_autograd_cache/dynamo.test_aot_autograd_cache-2e1ac30e1e69e405.json (deflated 94%) 2025-12-04T17:17:38.1224893Z adding: test/test-reports/python-pytest/dynamo.test_comptime/dynamo.test_comptime-61e4f48b41f3f41b.json (deflated 83%) 2025-12-04T17:17:38.1227677Z adding: test/test-reports/python-pytest/test_sort_and_select/test_sort_and_select-ab12b6ad99352fe2.json (deflated 95%) 2025-12-04T17:17:38.1228891Z adding: test/test-reports/python-pytest/functorch.test_rearrange/functorch.test_rearrange-e13c2240713d23f9.json (deflated 88%) 2025-12-04T17:17:38.1230148Z adding: test/test-reports/python-pytest/functorch.test_parsing/functorch.test_parsing-a56fd01f0949f839.json (deflated 88%) 2025-12-04T17:17:38.1232504Z adding: test/test-reports/python-pytest/profiler.test_profiler/profiler.test_profiler-ac51a5ffaf89322c.json (deflated 89%) 2025-12-04T17:17:38.1233766Z adding: test/test-reports/python-pytest/torch_np.test_binary_ufuncs/torch_np.test_binary_ufuncs-e1b3ae8f2439c11d.json (deflated 96%) 2025-12-04T17:17:38.1235046Z adding: test/test-reports/python-pytest/torch_np.test_unary_ufuncs/torch_np.test_unary_ufuncs-0ead42044090ba01.json (deflated 96%) 2025-12-04T17:17:38.1236364Z adding: test/test-reports/python-pytest/test_utils_filelock/test_utils_filelock-bc82f2b31ad6d0a9.json (deflated 62%) 2025-12-04T17:17:38.1237532Z adding: test/test-reports/python-pytest/test_extension_utils/test_extension_utils-1a92a953d2096623.json (deflated 64%) 2025-12-04T17:17:38.1238898Z adding: test/test-reports/python-pytest/test_rename_privateuse1_to_existing_device/test_rename_privateuse1_to_existing_device-a35ba0709779b2ea.json (deflated 44%) 2025-12-04T17:17:38.1240293Z adding: test/test-reports/python-pytest/nn.attention.test_fa4/nn.attention.test_fa4-7e669f8dd97f47a6.json (deflated 98%) 2025-12-04T17:17:38.1243797Z adding: test/test-reports/python-pytest/typing.test_python_operators/typing.test_python_operators-6bf858ca1197a53f.json (deflated 98%) 2025-12-04T17:17:38.1248694Z adding: test/test-reports/python-pytest/test_functionalization/test_functionalization-44af17f38403c26c.json (deflated 92%) 2025-12-04T17:17:38.1249923Z adding: test/test-reports/python-pytest/profiler.test_kineto/profiler.test_kineto-1d87d2a72eae2747.json (deflated 37%) 2025-12-04T17:17:38.1251076Z adding: test/test-reports/python-pytest/test_module_tracker/test_module_tracker-fac08a90d6f586d0.json (deflated 59%) 2025-12-04T17:17:38.1252487Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_scalarinherit/torch_np.numpy_tests.core.test_scalarinherit-9bae58f1bf1858ad.json (deflated 81%) 2025-12-04T17:17:38.1253931Z adding: test/test-reports/python-pytest/test_tensorexpr_pybind/test_tensorexpr_pybind-ea4c0e89818d6b23.json (deflated 92%) 2025-12-04T17:17:38.1270410Z adding: test/test-reports/python-pytest/test_fx_experimental/test_fx_experimental-ab6703fb0b80faae.json (deflated 98%) 2025-12-04T17:17:38.1271519Z adding: test/test-reports/python-pytest/test_fx_passes/test_fx_passes-268f4eef0e7a6c57.json (deflated 96%) 2025-12-04T17:17:38.1272786Z adding: test/test-reports/python-pytest/functorch.test_logging/functorch.test_logging-94ce26f72ae6dfc2.json (deflated 81%) 2025-12-04T17:17:38.1275112Z adding: test/test-reports/python-pytest/test_namedtensor/test_namedtensor-9c123d78058198ac.json (deflated 93%) 2025-12-04T17:17:38.1277294Z adding: test/test-reports/python-pytest/test_tensorexpr/test_tensorexpr-0ea46f0a35c352a2.json (deflated 91%) 2025-12-04T17:17:38.1280040Z adding: test/test-reports/python-pytest/functorch.test_minifier/functorch.test_minifier-320e05a9203d92d3.json (deflated 92%) 2025-12-04T17:17:38.1282386Z adding: test/test-reports/python-pytest/higher_order_ops.test_invoke_quant/higher_order_ops.test_invoke_quant-98f369a4901365e9.json (deflated 93%) 2025-12-04T17:17:38.1289290Z adding: test/test-reports/python-pytest/torch_np.test_basic/torch_np.test_basic-f66c371882977b0d.json (deflated 98%) 2025-12-04T17:17:38.1294930Z adding: test/test-reports/python-pytest/test_jiterator/test_jiterator-faac0f7ad3ce18ff.json (deflated 98%) 2025-12-04T17:17:38.1296547Z adding: test/test-reports/python-pytest/test_native_functions/test_native_functions-ed1f170746938cdb.json (deflated 92%) 2025-12-04T17:17:38.1297727Z adding: test/test-reports/python-pytest/test_typing/test_typing-029ceb92f90c5fb6.json (deflated 90%) 2025-12-04T17:17:38.1298896Z adding: test/test-reports/python-pytest/lazy.test_functionalization/lazy.test_functionalization-3b38bfe5c8489bba.json (deflated 63%) 2025-12-04T17:17:38.1300148Z adding: test/test-reports/python-pytest/torch_np.test_random/torch_np.test_random-92b710d703a1f97f.json (deflated 96%) 2025-12-04T17:17:38.1301373Z adding: test/test-reports/python-pytest/nn.test_multihead_attention/nn.test_multihead_attention-227b95684fd4827a.json (deflated 86%) 2025-12-04T17:17:38.1305707Z adding: test/test-reports/python-pytest/test_legacy_vmap/test_legacy_vmap-1f56d97bf135e06f.json (deflated 93%) 2025-12-04T17:17:38.1306804Z adding: test/test-reports/python-pytest/lazy.test_bindings/lazy.test_bindings-48affd3dc097f9a0.json (deflated 35%) 2025-12-04T17:17:38.1438152Z adding: test/test-reports/python-pytest/test_utils/test_utils-cb6667809c56b359.json (deflated 99%) 2025-12-04T17:17:38.1440375Z adding: test/test-reports/python-pytest/test_pytree/test_pytree-b5c87da58bd7db26.json (deflated 96%) 2025-12-04T17:17:38.1441524Z adding: test/test-reports/python-pytest/test_namedtuple_return_api/test_namedtuple_return_api-9fcfafd59257c871.json (deflated 76%) 2025-12-04T17:17:38.1442871Z adding: test/test-reports/python-pytest/profiler.test_record_function/profiler.test_record_function-5b633e88461adafa.json (deflated 85%) 2025-12-04T17:17:38.1444209Z adding: test/test-reports/python-pytest/test_compile_benchmark_util/test_compile_benchmark_util-df68b65dcac6dfa0.json (deflated 87%) 2025-12-04T17:17:38.1445600Z adding: test/test-reports/python-pytest/test_set_default_mobile_cpu_allocator/test_set_default_mobile_cpu_allocator-e4da27e112b70d28.json (deflated 66%) 2025-12-04T17:17:38.1450812Z adding: test/test-reports/python-pytest/test_fake_tensor/test_fake_tensor-03d6e27aa07cc9f8.json (deflated 94%) 2025-12-04T17:17:38.1452012Z adding: test/test-reports/python-pytest/higher_order_ops.test_print/higher_order_ops.test_print-b7576a32603564cc.json (deflated 88%) 2025-12-04T17:17:38.1453236Z adding: test/test-reports/python-pytest/test_per_overload_api/test_per_overload_api-e690ac2b4ba1f976.json (deflated 75%) 2025-12-04T17:17:38.1454571Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_einsum/torch_np.numpy_tests.core.test_einsum-21e2123caec3dd16.json (deflated 93%) 2025-12-04T17:17:38.1455917Z adding: test/test-reports/python-pytest/test_multiprocessing/test_multiprocessing-66d2144d604c4acf.json (deflated 89%) 2025-12-04T17:17:38.1523608Z adding: test/test-reports/python-pytest/test_modules/test_modules-5eb6bc8f38dae4a5.json (deflated 97%) 2025-12-04T17:17:38.1536791Z adding: test/test-reports/python-pytest/complex_tensor.test_complex_tensor/complex_tensor.test_complex_tensor-00d0e28f118bdf92.json (deflated 97%) 2025-12-04T17:17:38.1539499Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_indexing/torch_np.numpy_tests.core.test_indexing-3b590877a50ad13d.json (deflated 93%) 2025-12-04T17:17:38.1540771Z adding: test/test-reports/python-pytest/test_futures/test_futures-8abfe477c1d087d4.json (deflated 92%) 2025-12-04T17:17:38.1552453Z adding: test/test-reports/python-pytest/test_tensor_creation_ops/test_tensor_creation_ops-c328fea3f872cecd.json (deflated 96%) 2025-12-04T17:17:38.1559026Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_scalarmath/torch_np.numpy_tests.core.test_scalarmath-a9b7ea367b3cdb41.json (deflated 97%) 2025-12-04T17:17:38.1564526Z adding: test/test-reports/python-pytest/test_serialization/test_serialization-43b1fedba2a2b650.json (deflated 94%) 2025-12-04T17:17:38.1566380Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_dlpack/torch_np.numpy_tests.core.test_dlpack-350d972e2ff20e05.json (deflated 97%) 2025-12-04T17:17:38.1567792Z adding: test/test-reports/python-pytest/test_multiprocessing_spawn/test_multiprocessing_spawn-b04898e1c1418acf.json (deflated 92%) 2025-12-04T17:17:38.1569089Z adding: test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-ee0ad83af64551d9.json (deflated 39%) 2025-12-04T17:17:38.1570350Z adding: test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-445675b8d8f15e5e.json (deflated 38%) 2025-12-04T17:17:38.1571604Z adding: test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-d82ae01551c43307.json (deflated 38%) 2025-12-04T17:17:38.1572833Z adding: test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-4e71d02e98da078f.json (deflated 39%) 2025-12-04T17:17:38.1574096Z adding: test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-233c3fc007b40619.json (deflated 38%) 2025-12-04T17:17:38.1575344Z adding: test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-a0f26f1a84f3180c.json (deflated 38%) 2025-12-04T17:17:38.1576706Z adding: test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-4a2a7cd181a00ba9.json (deflated 36%) 2025-12-04T17:17:38.1578006Z adding: test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-24fabc8827e58618.json (deflated 38%) 2025-12-04T17:17:38.1579261Z adding: test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-f87021e84032e697.json (deflated 36%) 2025-12-04T17:17:38.1580480Z adding: test/test-reports/python-pytest/test_mobile_optimizer/test_mobile_optimizer-8da56fb35697f793.json (deflated 83%) 2025-12-04T17:17:38.1581722Z adding: test/test-reports/python-pytest/torch_np.test_function_base/torch_np.test_function_base-fc4ff98d96269d3c.json (deflated 62%) 2025-12-04T17:17:38.1583008Z adding: test/test-reports/python-pytest/test_type_promotion/test_type_promotion-bc48b0383dc65f97.json (deflated 98%) 2025-12-04T17:17:38.1584142Z adding: test/test-reports/python-pytest/lazy.test_reuse_ir/lazy.test_reuse_ir-27aa6d3f8959224d.json (deflated 78%) 2025-12-04T17:17:38.1585430Z adding: test/test-reports/python-pytest/test_functional_autograd_benchmark/test_functional_autograd_benchmark-8f47749aefc55c89.json (deflated 63%) 2025-12-04T17:17:38.1614617Z ##[group]Run # Remove any previous test reports if they exist 2025-12-04T17:17:38.1615187Z # Remove any previous test reports if they exist 2025-12-04T17:17:38.1615640Z rm -f test-reports-*.zip 2025-12-04T17:17:38.1616189Z zip -r "test-reports-${FILE_SUFFIX}.zip" test/test-reports -i '*.xml' -i '*.csv' 2025-12-04T17:17:38.1623307Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T17:17:38.1623759Z env: 2025-12-04T17:17:38.1623998Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:38.1624306Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:38.1624671Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:38.1625310Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:38.1626127Z FILE_SUFFIX: test-legacy_nvidia_driver-3-5-linux.g4dn.4xlarge.nvidia.gpu_57119749282 2025-12-04T17:17:38.1626708Z ##[endgroup] 2025-12-04T17:17:38.1789487Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-cf40a4f46675676c.xml (deflated 89%) 2025-12-04T17:17:38.1800669Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-3767af6d1470a6b5.xml (deflated 93%) 2025-12-04T17:17:38.1811546Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-bf1966c65cc0ce25.xml (deflated 93%) 2025-12-04T17:17:38.1816514Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-4d0cbc1f4fa8e643.xml (deflated 93%) 2025-12-04T17:17:38.1832943Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-6ce11cbc918a017c.xml (deflated 93%) 2025-12-04T17:17:38.1834312Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-c850c8e21fb6d04c.xml (deflated 84%) 2025-12-04T17:17:38.1835660Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-17e608243d035d18.xml (deflated 84%) 2025-12-04T17:17:38.1837003Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-7d6e32f4c213ad66.xml (deflated 88%) 2025-12-04T17:17:38.1846547Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-e7bcb2cd86408aa6.xml (deflated 94%) 2025-12-04T17:17:38.1856404Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-54ad0f704656a459.xml (deflated 94%) 2025-12-04T17:17:38.1858818Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-fb514581ec21877e.xml (deflated 87%) 2025-12-04T17:17:38.1860668Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-1a3279675b79d710.xml (deflated 87%) 2025-12-04T17:17:38.1862308Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-bfb3d1c8c3468bae.xml (deflated 87%) 2025-12-04T17:17:38.1880170Z adding: test/test-reports/python-pytest/inductor.test_torchinductor/inductor.test_torchinductor-aa152592c1ebda5c.xml (deflated 91%) 2025-12-04T17:17:38.1898446Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-535feae09e5a3ebd.xml (deflated 93%) 2025-12-04T17:17:38.1900191Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-b527a05b10cde953.xml (deflated 87%) 2025-12-04T17:17:38.1901945Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-0142110b0a563d3e.xml (deflated 87%) 2025-12-04T17:17:38.1908643Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_dynamic_shapes/inductor.test_torchinductor_dynamic_shapes-92fbe8f0c3defdbe.xml (deflated 90%) 2025-12-04T17:17:38.1937230Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_codegen_dynamic_shapes/inductor.test_torchinductor_codegen_dynamic_shapes-3b673269684e2f76.xml (deflated 93%) 2025-12-04T17:17:38.1942047Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-a02012f2395cf6ef.xml (deflated 93%) 2025-12-04T17:17:38.1946502Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-96bec2e922d5155c.xml (deflated 93%) 2025-12-04T17:17:38.1950993Z adding: test/test-reports/python-pytest/inductor.test_torchinductor_opinfo/inductor.test_torchinductor_opinfo-ef6350a9f8bb75e3.xml (deflated 92%) 2025-12-04T17:17:38.2032273Z adding: test/test-reports/python-pytest/dynamo.test_logging/dynamo.test_logging-77311c728804d279.xml (deflated 94%) 2025-12-04T17:17:38.2097409Z adding: test/test-reports/python-pytest/dynamo.test_repros/dynamo.test_repros-9b6b927930117494.xml (deflated 92%) 2025-12-04T17:17:38.2099032Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-97143c683457795e.xml (deflated 85%) 2025-12-04T17:17:38.2100680Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-c6ae8965f2e7a602.xml (deflated 84%) 2025-12-04T17:17:38.2102146Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-2f238f5f180b8489.xml (deflated 84%) 2025-12-04T17:17:38.2103616Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-df27debc7fb08425.xml (deflated 85%) 2025-12-04T17:17:38.2105064Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-39f974c40a0e5d07.xml (deflated 84%) 2025-12-04T17:17:38.2106543Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-4c5f8d97ccaac9c3.xml (deflated 84%) 2025-12-04T17:17:38.2108009Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-fda94603e2b70838.xml (deflated 85%) 2025-12-04T17:17:38.2109473Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-ace483b73b471725.xml (deflated 84%) 2025-12-04T17:17:38.2110920Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-841659f3603156bc.xml (deflated 84%) 2025-12-04T17:17:38.2112382Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-e4e48c14b07c6744.xml (deflated 85%) 2025-12-04T17:17:38.2113956Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-887ea8bf31a36206.xml (deflated 84%) 2025-12-04T17:17:38.2115415Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-d486c0fac95f9182.xml (deflated 84%) 2025-12-04T17:17:38.2116874Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-73bc989177a97c92.xml (deflated 85%) 2025-12-04T17:17:38.2118317Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-9368d7516886ec65.xml (deflated 84%) 2025-12-04T17:17:38.2119778Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-921cae3f6acd6a84.xml (deflated 84%) 2025-12-04T17:17:38.2121301Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-8d346e15c82776ef.xml (deflated 85%) 2025-12-04T17:17:38.2122781Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-1b686c821c7b9c77.xml (deflated 84%) 2025-12-04T17:17:38.2124248Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-21dcaad0846c92ef.xml (deflated 84%) 2025-12-04T17:17:38.2125699Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-633a2407c7cc9faa.xml (deflated 85%) 2025-12-04T17:17:38.2127174Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-2a4f8a9a7e80b95a.xml (deflated 84%) 2025-12-04T17:17:38.2128648Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-ab0a7cd9bbef732b.xml (deflated 84%) 2025-12-04T17:17:38.2130124Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-81245131849bbecb.xml (deflated 85%) 2025-12-04T17:17:38.2131571Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-4cf813d975e62c57.xml (deflated 84%) 2025-12-04T17:17:38.2133032Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-f694c40280fe3f31.xml (deflated 84%) 2025-12-04T17:17:38.2134534Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-a6d9b5e9da7bb417.xml (deflated 85%) 2025-12-04T17:17:38.2136001Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-c31955170ca46735.xml (deflated 84%) 2025-12-04T17:17:38.2137546Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-1b66a9993689b300.xml (deflated 84%) 2025-12-04T17:17:38.2139007Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-79106310b2e4ab0d.xml (deflated 85%) 2025-12-04T17:17:38.2140486Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-235ce3baacba0fb6.xml (deflated 84%) 2025-12-04T17:17:38.2141962Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-00aa73936a934045.xml (deflated 84%) 2025-12-04T17:17:38.2143424Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-333e44976f8ee863.xml (deflated 85%) 2025-12-04T17:17:38.2144877Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-acd69551823a4dec.xml (deflated 84%) 2025-12-04T17:17:38.2146351Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-c7f6019ad8b4bab5.xml (deflated 84%) 2025-12-04T17:17:38.2147916Z adding: test/test-reports/python-pytest/inductor.test_cuda_select_algorithm/inductor.test_cuda_select_algorithm-cdc355d159a50f1a.xml (deflated 28%) 2025-12-04T17:17:38.2149330Z adding: test/test-reports/python-pytest/inductor.test_deterministic/inductor.test_deterministic-142f090bc785acb6.xml (deflated 63%) 2025-12-04T17:17:38.2150673Z adding: test/test-reports/python-pytest/inductor.test_deterministic/inductor.test_deterministic-2940c67a6de1d3d8.xml (deflated 71%) 2025-12-04T17:17:38.2151995Z adding: test/test-reports/python-pytest/inductor.test_native_matmul/inductor.test_native_matmul-5845c7d4a4d90a90.xml (deflated 71%) 2025-12-04T17:17:38.2544536Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-819d138cb6fe4c8f.xml (deflated 87%) 2025-12-04T17:17:38.2588947Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-3d7372eabf0c20ef.xml (deflated 87%) 2025-12-04T17:17:38.2633781Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-e6eefab2abf634a5.xml (deflated 87%) 2025-12-04T17:17:38.2679389Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-fd020a6d441a5738.xml (deflated 87%) 2025-12-04T17:17:38.2725375Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-b7f552c9fcec58a4.xml (deflated 87%) 2025-12-04T17:17:38.2771086Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-ba19ebfa7960f6c4.xml (deflated 87%) 2025-12-04T17:17:38.3115294Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-6aac4defd9d3a3cc.xml (deflated 87%) 2025-12-04T17:17:38.3159913Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-041aad81ad32b1c4.xml (deflated 87%) 2025-12-04T17:17:38.3204897Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-043eb8ddb5d84fd4.xml (deflated 87%) 2025-12-04T17:17:38.3249360Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-555ea3dc15f2a8d0.xml (deflated 87%) 2025-12-04T17:17:38.3294028Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-01d96e43057cf9b8.xml (deflated 87%) 2025-12-04T17:17:38.3339011Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-563749739dc7a57a.xml (deflated 87%) 2025-12-04T17:17:38.3403947Z adding: test/test-reports/python-pytest/inductor.test_decompose_mem_bound_mm/inductor.test_decompose_mem_bound_mm-d89a888c67de20be.xml (deflated 87%) 2025-12-04T17:17:38.3405399Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a12b7a446d46a036.xml (deflated 83%) 2025-12-04T17:17:38.3406736Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-c354a873b70ad9a4.xml (deflated 82%) 2025-12-04T17:17:38.3408069Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-72b2d7d92c2f3d43.xml (deflated 82%) 2025-12-04T17:17:38.3409403Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-c8076b5dd44238e4.xml (deflated 82%) 2025-12-04T17:17:38.3410727Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-d2d17ca8c68c28e7.xml (deflated 82%) 2025-12-04T17:17:38.3412067Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-b56c2dd706b17a20.xml (deflated 82%) 2025-12-04T17:17:38.3413382Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-fa3c7a9046cc8737.xml (deflated 82%) 2025-12-04T17:17:38.3414924Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3da5ad44ee5d80af.xml (deflated 82%) 2025-12-04T17:17:38.3416260Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2586fe9efcb2d9c1.xml (deflated 82%) 2025-12-04T17:17:38.3417674Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-4d9d69bfeda71662.xml (deflated 82%) 2025-12-04T17:17:38.3418988Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0045cf67af4f7e36.xml (deflated 82%) 2025-12-04T17:17:38.3420312Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-38fbcefce9021f2a.xml (deflated 82%) 2025-12-04T17:17:38.3421720Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0625eb2c44495ce0.xml (deflated 82%) 2025-12-04T17:17:38.3423053Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-ea417fd0de1ef597.xml (deflated 82%) 2025-12-04T17:17:38.3424375Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-e58bcfb06df4ce4d.xml (deflated 82%) 2025-12-04T17:17:38.3425698Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0c1484308d51b00a.xml (deflated 82%) 2025-12-04T17:17:38.3427014Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-9682f9212923ea0d.xml (deflated 82%) 2025-12-04T17:17:38.3428341Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-ef1f71bad0a01b52.xml (deflated 82%) 2025-12-04T17:17:38.3429673Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3d18fa53804224f4.xml (deflated 83%) 2025-12-04T17:17:38.3430993Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2256fbb35262b330.xml (deflated 81%) 2025-12-04T17:17:38.3432320Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a993cfee9c5b5572.xml (deflated 81%) 2025-12-04T17:17:38.3433711Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-4097f4a492ce8bb6.xml (deflated 83%) 2025-12-04T17:17:38.3435039Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-9df84d07ff6a448b.xml (deflated 81%) 2025-12-04T17:17:38.3436358Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-b20a3959da2014da.xml (deflated 81%) 2025-12-04T17:17:38.3437698Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-89205029d9c04efd.xml (deflated 82%) 2025-12-04T17:17:38.3439031Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-476b174d4b08ad56.xml (deflated 82%) 2025-12-04T17:17:38.3440362Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-99551b4e3863e77a.xml (deflated 82%) 2025-12-04T17:17:38.3441673Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-498b08ff107b8914.xml (deflated 81%) 2025-12-04T17:17:38.3442997Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0e343f505ff50c95.xml (deflated 81%) 2025-12-04T17:17:38.3444329Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-486e74c9978d3a7f.xml (deflated 81%) 2025-12-04T17:17:38.3445651Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a0676d6303da3615.xml (deflated 81%) 2025-12-04T17:17:38.3446978Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-e27b1d19486b6954.xml (deflated 81%) 2025-12-04T17:17:38.3448351Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3f1ba8060f3abc48.xml (deflated 81%) 2025-12-04T17:17:38.3449669Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-0209a84b5ca28602.xml (deflated 82%) 2025-12-04T17:17:38.3450994Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-3d838b8cb24fd196.xml (deflated 82%) 2025-12-04T17:17:38.3452311Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-6a6a2484271011a1.xml (deflated 82%) 2025-12-04T17:17:38.3453650Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-9356c0308b0c66f7.xml (deflated 81%) 2025-12-04T17:17:38.3454978Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-51758daa3a3dc599.xml (deflated 81%) 2025-12-04T17:17:38.3456311Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-a0d42e3d123a44e8.xml (deflated 81%) 2025-12-04T17:17:38.3457707Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-20d10a20187a2cb9.xml (deflated 82%) 2025-12-04T17:17:38.3459022Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-5f3be1bf80a1ce40.xml (deflated 80%) 2025-12-04T17:17:38.3460360Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-baeb09a4e62b9d77.xml (deflated 80%) 2025-12-04T17:17:38.3461691Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2cdd67c45b570140.xml (deflated 83%) 2025-12-04T17:17:38.3463028Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-2423c3b275573349.xml (deflated 81%) 2025-12-04T17:17:38.3464354Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-66e69110dfb99f31.xml (deflated 81%) 2025-12-04T17:17:38.3465663Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-4e57fa3cbc487129.xml (deflated 83%) 2025-12-04T17:17:38.3467021Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-7331926cdb107464.xml (deflated 81%) 2025-12-04T17:17:38.3468331Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-7d66ea27b5ac2c65.xml (deflated 81%) 2025-12-04T17:17:38.3469659Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-356c1a001f2d20d8.xml (deflated 85%) 2025-12-04T17:17:38.3470970Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-426f4e961e373829.xml (deflated 82%) 2025-12-04T17:17:38.3472286Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-1a1453b5e01c9120.xml (deflated 82%) 2025-12-04T17:17:38.3473611Z adding: test/test-reports/python-pytest/inductor.test_online_softmax/inductor.test_online_softmax-09c678f485469fcb.xml (deflated 48%) 2025-12-04T17:17:38.3474981Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-fb2ec7b0e0c81fea.xml (deflated 86%) 2025-12-04T17:17:38.3476392Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6ae6fdb299843870.xml (deflated 81%) 2025-12-04T17:17:38.3477812Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2362ba98ed8b8900.xml (deflated 81%) 2025-12-04T17:17:38.3479227Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-52334a8157857cd5.xml (deflated 81%) 2025-12-04T17:17:38.3480711Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0b589120e3bef198.xml (deflated 81%) 2025-12-04T17:17:38.3482137Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f9457e8d87a8d0e5.xml (deflated 81%) 2025-12-04T17:17:38.3483544Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e4f574242b541525.xml (deflated 81%) 2025-12-04T17:17:38.3484954Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-19baa3ae62072b2f.xml (deflated 81%) 2025-12-04T17:17:38.3486374Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-debf0ef50304edda.xml (deflated 81%) 2025-12-04T17:17:38.3487830Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a0af7423e85581ef.xml (deflated 81%) 2025-12-04T17:17:38.3489238Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6e37efe6d0d7b5f7.xml (deflated 81%) 2025-12-04T17:17:38.3490674Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-64b9c8b908c45f51.xml (deflated 81%) 2025-12-04T17:17:38.3492094Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3343e9ff31c5255a.xml (deflated 93%) 2025-12-04T17:17:38.3493506Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ae334ed847078f28.xml (deflated 85%) 2025-12-04T17:17:38.3494922Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0eab24e125a2cf7d.xml (deflated 85%) 2025-12-04T17:17:38.3496492Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6060ecdf9a2577f9.xml (deflated 83%) 2025-12-04T17:17:38.3497987Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0eb661cc20714db1.xml (deflated 82%) 2025-12-04T17:17:38.3499409Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-5774b4beb84c0975.xml (deflated 82%) 2025-12-04T17:17:38.3500907Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-74651aae35dd9844.xml (deflated 82%) 2025-12-04T17:17:38.3502315Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-7a5a8b9ba5d60164.xml (deflated 82%) 2025-12-04T17:17:38.3503740Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-26faac4e438f6204.xml (deflated 82%) 2025-12-04T17:17:38.3505157Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-aa20002b977e7ab4.xml (deflated 88%) 2025-12-04T17:17:38.3506585Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6316dfd7aad65a93.xml (deflated 82%) 2025-12-04T17:17:38.3508006Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d1f50caa5a44c0ee.xml (deflated 82%) 2025-12-04T17:17:38.3509422Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-8edf4f7e001e74f9.xml (deflated 84%) 2025-12-04T17:17:38.3510851Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-defc51812af7e0b0.xml (deflated 82%) 2025-12-04T17:17:38.3512283Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0bc167e05b941aff.xml (deflated 82%) 2025-12-04T17:17:38.3513702Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-98b78c8ed19e0643.xml (deflated 82%) 2025-12-04T17:17:38.3515230Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-cc69c969909d9707.xml (deflated 82%) 2025-12-04T17:17:38.3516652Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9977f3a62a4ac5d4.xml (deflated 82%) 2025-12-04T17:17:38.3518077Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ec39337b11cd8a4d.xml (deflated 85%) 2025-12-04T17:17:38.3519509Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f9d493fa40deaa3c.xml (deflated 85%) 2025-12-04T17:17:38.3520974Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-492f33e438c88cbb.xml (deflated 85%) 2025-12-04T17:17:38.3522397Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-cfe26d7f4e2f142c.xml (deflated 82%) 2025-12-04T17:17:38.3523818Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-85dd835c0392b89d.xml (deflated 82%) 2025-12-04T17:17:38.3525245Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d745daffa2211db3.xml (deflated 82%) 2025-12-04T17:17:38.3526705Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-29d661e244421a0d.xml (deflated 84%) 2025-12-04T17:17:38.3528109Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6a080fc2c141366f.xml (deflated 82%) 2025-12-04T17:17:38.3529533Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b727a3d5cb43a100.xml (deflated 82%) 2025-12-04T17:17:38.3530943Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9820049c2fc9b56d.xml (deflated 86%) 2025-12-04T17:17:38.3532353Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-bcfbf334a5fb1b19.xml (deflated 82%) 2025-12-04T17:17:38.3533803Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-55cd5c19b34efcf0.xml (deflated 82%) 2025-12-04T17:17:38.3535207Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3f17acc66e50cb98.xml (deflated 85%) 2025-12-04T17:17:38.3536632Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a7cd72d0cd16fd7c.xml (deflated 85%) 2025-12-04T17:17:38.3538120Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-45e1674e77992a96.xml (deflated 85%) 2025-12-04T17:17:38.3539541Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-384ea3a6b9159dd5.xml (deflated 90%) 2025-12-04T17:17:38.3540938Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-40502088a31692e3.xml (deflated 82%) 2025-12-04T17:17:38.3542367Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-1a1d3c182ee91d42.xml (deflated 82%) 2025-12-04T17:17:38.3543785Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9bd3fb708dd83987.xml (deflated 87%) 2025-12-04T17:17:38.3545209Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-991a5e87e1946535.xml (deflated 85%) 2025-12-04T17:17:38.3546612Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-61bd448d788d4193.xml (deflated 85%) 2025-12-04T17:17:38.3548093Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0c87f6ca176b74e4.xml (deflated 87%) 2025-12-04T17:17:38.3549514Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-00981048468ad3fe.xml (deflated 82%) 2025-12-04T17:17:38.3550933Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-7f9686778c56ad66.xml (deflated 82%) 2025-12-04T17:17:38.3552345Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dc76d780af365e97.xml (deflated 82%) 2025-12-04T17:17:38.3553777Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3d712c264776fdc2.xml (deflated 82%) 2025-12-04T17:17:38.3555182Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4da37e7aafa09519.xml (deflated 82%) 2025-12-04T17:17:38.3556605Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4f9e658df3b9ce6d.xml (deflated 84%) 2025-12-04T17:17:38.3558023Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-3d80a9d76d3be932.xml (deflated 81%) 2025-12-04T17:17:38.3559430Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ede94f48d428153a.xml (deflated 81%) 2025-12-04T17:17:38.3560845Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-6ae027693c5dbbf4.xml (deflated 81%) 2025-12-04T17:17:38.3562260Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-48db12bdca4a53d6.xml (deflated 81%) 2025-12-04T17:17:38.3563672Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0c861fbac71335e0.xml (deflated 81%) 2025-12-04T17:17:38.3565080Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ffbb324af5af0672.xml (deflated 81%) 2025-12-04T17:17:38.3566500Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-042dc549bdd77a45.xml (deflated 81%) 2025-12-04T17:17:38.3567940Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-0942d088a0709320.xml (deflated 81%) 2025-12-04T17:17:38.3569355Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c3fbd4a060fe6236.xml (deflated 92%) 2025-12-04T17:17:38.3570780Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f308a9dbba7be53a.xml (deflated 85%) 2025-12-04T17:17:38.3572183Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-bd2ede9c3a5cd1d1.xml (deflated 85%) 2025-12-04T17:17:38.3573634Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d27c0d840d6fe564.xml (deflated 82%) 2025-12-04T17:17:38.3575050Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2f600ca197676d4b.xml (deflated 82%) 2025-12-04T17:17:38.3576454Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ecad5133d6245a6a.xml (deflated 82%) 2025-12-04T17:17:38.3577950Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d38b619605c8f606.xml (deflated 85%) 2025-12-04T17:17:38.3579347Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ca04900b7cf26510.xml (deflated 85%) 2025-12-04T17:17:38.3580828Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-8a1dc4f258524121.xml (deflated 85%) 2025-12-04T17:17:38.3582238Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a0112c740677bec4.xml (deflated 82%) 2025-12-04T17:17:38.3583656Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-67764f47d32c222e.xml (deflated 82%) 2025-12-04T17:17:38.3585055Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c39fdfd87cf64578.xml (deflated 82%) 2025-12-04T17:17:38.3586465Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-48a46382e3216972.xml (deflated 82%) 2025-12-04T17:17:38.3587952Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d40d195c41f12813.xml (deflated 82%) 2025-12-04T17:17:38.3589372Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-35b201d0b4c8de0a.xml (deflated 82%) 2025-12-04T17:17:38.3590792Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e10bae22111bfdec.xml (deflated 83%) 2025-12-04T17:17:38.3592200Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-5822c2fc7d137525.xml (deflated 82%) 2025-12-04T17:17:38.3593621Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a6ba20f02ccefade.xml (deflated 82%) 2025-12-04T17:17:38.3595039Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dc63b1e627261305.xml (deflated 82%) 2025-12-04T17:17:38.3596574Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9dda8854f56bbc5e.xml (deflated 82%) 2025-12-04T17:17:38.3597983Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d35f802df30bd39d.xml (deflated 82%) 2025-12-04T17:17:38.3599398Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d1473b4b18944240.xml (deflated 83%) 2025-12-04T17:17:38.3600873Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c84649193ffdba8a.xml (deflated 82%) 2025-12-04T17:17:38.3602281Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b127809f84c30c37.xml (deflated 82%) 2025-12-04T17:17:38.3603680Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-1a7c4cad159549d9.xml (deflated 83%) 2025-12-04T17:17:38.3605096Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2f388a73f046ed87.xml (deflated 82%) 2025-12-04T17:17:38.3606519Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d68d80f1ba61c89f.xml (deflated 82%) 2025-12-04T17:17:38.3607940Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b8a30a3f316b7f0b.xml (deflated 85%) 2025-12-04T17:17:38.3609364Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d3917b026aff3e64.xml (deflated 85%) 2025-12-04T17:17:38.3610767Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c0c657ac1df627f9.xml (deflated 85%) 2025-12-04T17:17:38.3612185Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-175aa275ea059250.xml (deflated 84%) 2025-12-04T17:17:38.3613606Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-5ce17ea59d8fd893.xml (deflated 82%) 2025-12-04T17:17:38.3615113Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4874d0c887fabf8b.xml (deflated 82%) 2025-12-04T17:17:38.3616636Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a28d30f3a161bea7.xml (deflated 82%) 2025-12-04T17:17:38.3642592Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-23cf59059171b3d3.xml (deflated 82%) 2025-12-04T17:17:38.3644055Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-14e4b4194810001e.xml (deflated 82%) 2025-12-04T17:17:38.3645579Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4b58d6ae0af928ee.xml (deflated 85%) 2025-12-04T17:17:38.3646984Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9605f50d5f418e61.xml (deflated 82%) 2025-12-04T17:17:38.3648407Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2668b63c8efdb7a5.xml (deflated 82%) 2025-12-04T17:17:38.3649818Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-8d527629568e40ed.xml (deflated 84%) 2025-12-04T17:17:38.3651231Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a273900f3029c317.xml (deflated 82%) 2025-12-04T17:17:38.3652641Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dcefced86a4c35d8.xml (deflated 82%) 2025-12-04T17:17:38.3654061Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e15b12282035b26d.xml (deflated 85%) 2025-12-04T17:17:38.3655464Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e08a7bc39929d142.xml (deflated 82%) 2025-12-04T17:17:38.3656937Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-13b16b357405c710.xml (deflated 82%) 2025-12-04T17:17:38.3658342Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9275f55235cebf56.xml (deflated 85%) 2025-12-04T17:17:38.3659783Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-a0875a42ac68f671.xml (deflated 85%) 2025-12-04T17:17:38.3661204Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-56c85a19db7b6fad.xml (deflated 85%) 2025-12-04T17:17:38.3662626Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-b5990bb8b6e78b98.xml (deflated 82%) 2025-12-04T17:17:38.3664095Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9e1ed788d5899650.xml (deflated 82%) 2025-12-04T17:17:38.3665506Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e97cc37d8f662932.xml (deflated 82%) 2025-12-04T17:17:38.3666913Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d35e9b6eebdaac7b.xml (deflated 85%) 2025-12-04T17:17:38.3668345Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-f6ccc2d46fa5c9ea.xml (deflated 85%) 2025-12-04T17:17:38.3669766Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-dac88d9039f141ad.xml (deflated 85%) 2025-12-04T17:17:38.3671176Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-1d81442a0ab3908c.xml (deflated 87%) 2025-12-04T17:17:38.3672656Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-773b425f2e921837.xml (deflated 82%) 2025-12-04T17:17:38.3674056Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-09fb0d74c7a00548.xml (deflated 82%) 2025-12-04T17:17:38.3675467Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2f000bd6c93ab369.xml (deflated 82%) 2025-12-04T17:17:38.3676879Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-64628aab5ce6aeb9.xml (deflated 82%) 2025-12-04T17:17:38.3678292Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-9fedf51ad965eb7d.xml (deflated 82%) 2025-12-04T17:17:38.3679726Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-c5a3e4c540d200e7.xml (deflated 85%) 2025-12-04T17:17:38.3681153Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-ab8ed79c5f08a6c3.xml (deflated 82%) 2025-12-04T17:17:38.3682569Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-43cb3996049bee78.xml (deflated 82%) 2025-12-04T17:17:38.3683984Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d3a2df1a150e9085.xml (deflated 82%) 2025-12-04T17:17:38.3685390Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-2b2d696e555e6fc9.xml (deflated 82%) 2025-12-04T17:17:38.3686803Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-449e4fae7bbefa8d.xml (deflated 82%) 2025-12-04T17:17:38.3688222Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-4631de08a3dd09e1.xml (deflated 84%) 2025-12-04T17:17:38.3689636Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-7cb60f50c6534a73.xml (deflated 82%) 2025-12-04T17:17:38.3691035Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-bbe75c180c40946f.xml (deflated 82%) 2025-12-04T17:17:38.3692476Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-19b98f6e1bce5088.xml (deflated 87%) 2025-12-04T17:17:38.3693881Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-fd7621eb24f73b8c.xml (deflated 82%) 2025-12-04T17:17:38.3695282Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-e11a9d6da1493a65.xml (deflated 82%) 2025-12-04T17:17:38.3696959Z adding: test/test-reports/python-pytest/inductor.test_mix_order_reduction/inductor.test_mix_order_reduction-d245bc4e7ed2a74d.xml (deflated 83%) 2025-12-04T17:17:38.3698221Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-d7ddbde7cf40ab13.xml (deflated 97%) 2025-12-04T17:17:38.3699286Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-63a23e8fc1ef8137.xml (deflated 84%) 2025-12-04T17:17:38.3700325Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-11f9c4926783dfef.xml (deflated 84%) 2025-12-04T17:17:38.3701375Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-737659f69f9145f2.xml (deflated 84%) 2025-12-04T17:17:38.3702423Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2099d82b2931fb99.xml (deflated 84%) 2025-12-04T17:17:38.3703477Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-5f0933bf702563bb.xml (deflated 84%) 2025-12-04T17:17:38.3704516Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-ed1b6e1c95269b63.xml (deflated 84%) 2025-12-04T17:17:38.3705680Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-ccb90d6b004c5139.xml (deflated 84%) 2025-12-04T17:17:38.3706733Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a3b2bb4f2a94fb32.xml (deflated 84%) 2025-12-04T17:17:38.3707787Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-581c94c41301ad68.xml (deflated 84%) 2025-12-04T17:17:38.3708825Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-809b8dfa72183a04.xml (deflated 84%) 2025-12-04T17:17:38.3709871Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-59f3dc780871f4a4.xml (deflated 84%) 2025-12-04T17:17:38.3710971Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2648ef692c69d7ff.xml (deflated 89%) 2025-12-04T17:17:38.3712018Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-1e4f023d2e76e67f.xml (deflated 84%) 2025-12-04T17:17:38.3713056Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-88a75603bd167dcb.xml (deflated 84%) 2025-12-04T17:17:38.3714098Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-58d14c6ca3ad4514.xml (deflated 84%) 2025-12-04T17:17:38.3715146Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-4358034553dc4d55.xml (deflated 84%) 2025-12-04T17:17:38.3716196Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-e604dacd305a4e36.xml (deflated 84%) 2025-12-04T17:17:38.3717240Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-29c18a70d3fdf25d.xml (deflated 84%) 2025-12-04T17:17:38.3718286Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2a489f89b00449c3.xml (deflated 84%) 2025-12-04T17:17:38.3719338Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-1e3f7f3e1ddc719e.xml (deflated 84%) 2025-12-04T17:17:38.3720381Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-0a2928546dd5ddd2.xml (deflated 84%) 2025-12-04T17:17:38.3721431Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-36be7a929b5f8cea.xml (deflated 84%) 2025-12-04T17:17:38.3722487Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-db4ab888748348ef.xml (deflated 84%) 2025-12-04T17:17:38.3723588Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a61e284ca72dd15d.xml (deflated 95%) 2025-12-04T17:17:38.3724645Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-8da3a511cea25a62.xml (deflated 84%) 2025-12-04T17:17:38.3725689Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-aaebada1c5676813.xml (deflated 84%) 2025-12-04T17:17:38.3726745Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-c9e9611bbdf13a2b.xml (deflated 84%) 2025-12-04T17:17:38.3727798Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-963810090dd7c38b.xml (deflated 84%) 2025-12-04T17:17:38.3728850Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-202f1cfa5f6f2e59.xml (deflated 84%) 2025-12-04T17:17:38.3729889Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a5f1ba0194e0f361.xml (deflated 84%) 2025-12-04T17:17:38.3730944Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-784ae621331a35c7.xml (deflated 84%) 2025-12-04T17:17:38.3731988Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-2c9da68c717b465c.xml (deflated 84%) 2025-12-04T17:17:38.3733045Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a06aff42aa9d7210.xml (deflated 84%) 2025-12-04T17:17:38.3734087Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-66f0187c8b5c56f0.xml (deflated 84%) 2025-12-04T17:17:38.3735138Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a93a155b666a377d.xml (deflated 84%) 2025-12-04T17:17:38.3736238Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-f6900cb9a4d4e17a.xml (deflated 89%) 2025-12-04T17:17:38.3737360Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-b4de48ce2713ebfd.xml (deflated 84%) 2025-12-04T17:17:38.3738407Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-9d9a1cc8efaa0b76.xml (deflated 84%) 2025-12-04T17:17:38.3739459Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-f2c6d1655dbd05d5.xml (deflated 84%) 2025-12-04T17:17:38.3740511Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-30cf0cd6fabb5398.xml (deflated 84%) 2025-12-04T17:17:38.3741590Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-29e9b8d6f2f6950a.xml (deflated 84%) 2025-12-04T17:17:38.3742645Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-0ebf80986aa75a9d.xml (deflated 84%) 2025-12-04T17:17:38.3743697Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-ed818918883fb5a6.xml (deflated 84%) 2025-12-04T17:17:38.3744750Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-688ec0aa50c9a8c6.xml (deflated 84%) 2025-12-04T17:17:38.3745790Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-6fda2bf2ae878b46.xml (deflated 84%) 2025-12-04T17:17:38.3746844Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-0857a7a11662fc06.xml (deflated 84%) 2025-12-04T17:17:38.3747903Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-a3e03d3f7a7bbb81.xml (deflated 84%) 2025-12-04T17:17:38.3748961Z adding: test/test-reports/python-pytest/test_matmul_cuda/test_matmul_cuda-829800aaf6d2511a.xml (deflated 98%) 2025-12-04T17:17:38.3750006Z adding: test/test-reports/python-pytest/test_dataloader/test_dataloader-f02fdb9181a8d8ee.xml (deflated 82%) 2025-12-04T17:17:38.3751034Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-2c8b2ff0efe02fa9.xml (deflated 91%) 2025-12-04T17:17:38.3752010Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-0c1ee1f6151b681e.xml (deflated 91%) 2025-12-04T17:17:38.3752968Z adding: test/test-reports/python-pytest/test_decomp/test_decomp-68b4836d5d5fa3ca.xml (deflated 91%) 2025-12-04T17:17:38.3839101Z adding: test/test-reports/python-pytest/test_meta/test_meta-029b8ed96c60bfb7.xml (deflated 96%) 2025-12-04T17:17:38.3849738Z adding: test/test-reports/python-pytest/test_ops_jit/test_ops_jit-9f733effbdfc0369.xml (deflated 92%) 2025-12-04T17:17:38.3868254Z adding: test/test-reports/python-pytest/test_nestedtensor/test_nestedtensor-8e0ff602fd37df16.xml (deflated 97%) 2025-12-04T17:17:38.3870447Z adding: test/test-reports/python-pytest/test_public_bindings/test_public_bindings-6fed1053bfc25f05.xml (deflated 82%) 2025-12-04T17:17:38.3944813Z adding: test/test-reports/python-pytest/test_ops/test_ops-b6bd03fdf7fbcee0.xml (deflated 95%) 2025-12-04T17:17:38.4007782Z adding: test/test-reports/python-pytest/test_ops/test_ops-e04c41c7bbbdce7a.xml (deflated 94%) 2025-12-04T17:17:38.4072626Z adding: test/test-reports/python-pytest/test_ops/test_ops-6f4cd93b78224a58.xml (deflated 94%) 2025-12-04T17:17:38.4107585Z adding: test/test-reports/python-pytest/functorch.test_ops/functorch.test_ops-37d651ead30b362d.xml (deflated 93%) 2025-12-04T17:17:38.4135624Z adding: test/test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-2355f4f75c6e0706.xml (deflated 93%) 2025-12-04T17:17:38.4147579Z adding: test/test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-db01917ff3ce307b.xml (deflated 94%) 2025-12-04T17:17:38.4159482Z adding: test/test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-12ecde9ddbb651c9.xml (deflated 94%) 2025-12-04T17:17:38.4194711Z adding: test/test-reports/python-pytest/functorch.test_vmap/functorch.test_vmap-94056324a12f0026.xml (deflated 94%) 2025-12-04T17:17:38.4195869Z adding: test/test-reports/python-pytest/dynamo.test_after_aot/dynamo.test_after_aot-9dfa8d2ed594e793.xml (deflated 52%) 2025-12-04T17:17:38.4199330Z adding: test/test-reports/python-pytest/inductor.test_snode_runtime/inductor.test_snode_runtime-804cfaf2d58e1050.xml (deflated 92%) 2025-12-04T17:17:38.4307794Z adding: test/test-reports/python-pytest/inductor.test_minifier/inductor.test_minifier-e48754fb8b6df415.xml (deflated 96%) 2025-12-04T17:17:38.4374461Z adding: test/test-reports/python-pytest/inductor.test_compiled_autograd/inductor.test_compiled_autograd-d055bdf97c26d8bd.xml (deflated 92%) 2025-12-04T17:17:38.4375815Z adding: test/test-reports/python-pytest/inductor.test_layout_optim/inductor.test_layout_optim-50eed008ca1555c3.xml (deflated 28%) 2025-12-04T17:17:38.4379851Z adding: test/test-reports/python-pytest/dynamo.test_unspec/dynamo.test_unspec-4bfba7972407dae5.xml (deflated 83%) 2025-12-04T17:17:38.4381029Z adding: test/test-reports/python-pytest/inductor.test_mmdecomp/inductor.test_mmdecomp-6ec8d10db2d708da.xml (deflated 85%) 2025-12-04T17:17:38.4390193Z adding: test/test-reports/python-pytest/dynamo.test_ctx_manager/dynamo.test_ctx_manager-2753bd6e949ccdb3.xml (deflated 91%) 2025-12-04T17:17:38.4409201Z adding: test/test-reports/python-pytest/dynamo.test_exc/dynamo.test_exc-e06b4c2ef4a1702a.xml (deflated 95%) 2025-12-04T17:17:38.4461434Z adding: test/test-reports/python-pytest/dynamo.test_misc/dynamo.test_misc-43b95b63c3862211.xml (deflated 88%) 2025-12-04T17:17:38.4469961Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor_arrayref/inductor.test_aot_inductor_arrayref-d7936498b94cbe56.xml (deflated 93%) 2025-12-04T17:17:38.4471854Z adding: test/test-reports/python-pytest/inductor.test_provenance_tracing/inductor.test_provenance_tracing-e41a43dd8ef6019c.xml (deflated 86%) 2025-12-04T17:17:38.4473226Z adding: test/test-reports/python-pytest/dynamo.test_buffers_override/dynamo.test_buffers_override-01fc103efe1995d6.xml (deflated 64%) 2025-12-04T17:17:38.4474670Z adding: test/test-reports/python-pytest/inductor.test_inplacing_pass/inductor.test_inplacing_pass-d670df374fd1e952.xml (deflated 89%) 2025-12-04T17:17:38.4477355Z adding: test/test-reports/python-pytest/inductor.test_aot_inductor_custom_ops/inductor.test_aot_inductor_custom_ops-a0ad7eb32926bcd8.xml (deflated 93%) 2025-12-04T17:17:38.4478916Z adding: test/test-reports/python-pytest/inductor.test_split_cat_fx_passes/inductor.test_split_cat_fx_passes-4aa43e0fa2d59cfb.xml (deflated 78%) 2025-12-04T17:17:38.4480240Z adding: test/test-reports/python-pytest/inductor.test_profiler/inductor.test_profiler-cdccc94d406caaec.xml (deflated 77%) 2025-12-04T17:17:38.4481524Z adding: test/test-reports/python-pytest/inductor.test_memory_planning/inductor.test_memory_planning-480d70b98c6f91e2.xml (deflated 69%) 2025-12-04T17:17:38.4482878Z adding: test/test-reports/python-pytest/inductor.test_mem_estimation/inductor.test_mem_estimation-f5f67f53fc542f3b.xml (deflated 61%) 2025-12-04T17:17:38.4484073Z adding: test/test-reports/python-pytest/dynamo.test_view/dynamo.test_view-5c939a16d5367651.xml (deflated 78%) 2025-12-04T17:17:38.4485252Z adding: test/test-reports/python-pytest/inductor.test_cutlass_evt/inductor.test_cutlass_evt-a51e1bccbd1b02fd.xml (deflated 84%) 2025-12-04T17:17:38.4486502Z adding: test/test-reports/python-pytest/dynamo.test_reconstruct/dynamo.test_reconstruct-7d75ef23ae79ec60.xml (deflated 83%) 2025-12-04T17:17:38.4487747Z adding: test/test-reports/python-pytest/dynamo.test_aot_autograd/dynamo.test_aot_autograd-89fcf170216dd13a.xml (deflated 85%) 2025-12-04T17:17:38.4533784Z adding: test/test-reports/python-pytest/export.test_cpp_serdes/export.test_cpp_serdes-1f0305d948167935.xml (deflated 96%) 2025-12-04T17:17:38.4535041Z adding: test/test-reports/python-pytest/inductor.test_block_analysis/inductor.test_block_analysis-53978d8777987ceb.xml (deflated 80%) 2025-12-04T17:17:38.4538643Z adding: test/test-reports/python-pytest/dynamo.test_subgraphs/dynamo.test_subgraphs-e09a516058cd0e01.xml (deflated 93%) 2025-12-04T17:17:38.4539980Z adding: test/test-reports/python-pytest/dynamo.test_pre_dispatch/dynamo.test_pre_dispatch-4a36f2c5e912fcfb.xml (deflated 76%) 2025-12-04T17:17:38.4541349Z adding: test/test-reports/python-pytest/inductor.test_custom_post_grad_passes/inductor.test_custom_post_grad_passes-e88ba17552693664.xml (deflated 82%) 2025-12-04T17:17:38.4542704Z adding: test/test-reports/python-pytest/dynamo.test_fx_annotate/dynamo.test_fx_annotate-ea1b8128153e57c8.xml (deflated 79%) 2025-12-04T17:17:38.4543837Z adding: test/test-reports/python-pytest/dynamo.test_pgo/dynamo.test_pgo-b9bc1426bb4bf646.xml (deflated 74%) 2025-12-04T17:17:38.4544996Z adding: test/test-reports/python-pytest/export.test_export_opinfo/export.test_export_opinfo-d8b8b1860a5b7382.xml (deflated 77%) 2025-12-04T17:17:38.4561925Z adding: test/test-reports/python-pytest/inductor.test_control_flow/inductor.test_control_flow-12b6e2c7b80b8695.xml (deflated 97%) 2025-12-04T17:17:38.4563741Z adding: test/test-reports/python-pytest/dynamo.test_compile/dynamo.test_compile-1298df40b08e7720.xml (deflated 81%) 2025-12-04T17:17:38.4567379Z adding: test/test-reports/python-pytest/dynamo.test_nested_graph_breaks/dynamo.test_nested_graph_breaks-3ce3fc0c984a2ae0.xml (deflated 90%) 2025-12-04T17:17:38.4568784Z adding: test/test-reports/python-pytest/inductor.test_needs_exact_strides/inductor.test_needs_exact_strides-afd07b0a4c4dd4e2.xml (deflated 64%) 2025-12-04T17:17:38.4570235Z adding: test/test-reports/python-pytest/inductor.test_split_cat_fx_aten_passes/inductor.test_split_cat_fx_aten_passes-d90ed49c774dd94e.xml (deflated 68%) 2025-12-04T17:17:38.4571536Z adding: test/test-reports/python-pytest/dynamo.test_resume/dynamo.test_resume-1b040f63515910f4.xml (deflated 42%) 2025-12-04T17:17:38.4572847Z adding: test/test-reports/python-pytest/dynamo.test_backward_higher_order_ops/dynamo.test_backward_higher_order_ops-8f7cdfdf9264a3b0.xml (deflated 70%) 2025-12-04T17:17:38.4574334Z adding: test/test-reports/python-pytest/inductor.test_custom_partitioner_fn/inductor.test_custom_partitioner_fn-44a32a2607e301e4.xml (deflated 49%) 2025-12-04T17:17:38.4575672Z adding: test/test-reports/python-pytest/dynamo.test_debug_utils/dynamo.test_debug_utils-d82e5bff075a86e1.xml (deflated 62%) 2025-12-04T17:17:38.4576897Z adding: test/test-reports/python-pytest/dynamo.test_base_hop/dynamo.test_base_hop-6ccb089675742efe.xml (deflated 74%) 2025-12-04T17:17:38.4578112Z adding: test/test-reports/python-pytest/dynamo.test_package/dynamo.test_package-9643d20377a48274.xml (deflated 91%) 2025-12-04T17:17:38.4582950Z adding: test/test-reports/python-pytest/dynamo.test_aot_autograd_cache/dynamo.test_aot_autograd_cache-2e1ac30e1e69e405.xml (deflated 93%) 2025-12-04T17:17:38.4584530Z adding: test/test-reports/python-pytest/dynamo.test_comptime/dynamo.test_comptime-61e4f48b41f3f41b.xml (deflated 80%) 2025-12-04T17:17:38.4586267Z adding: test/test-reports/python-pytest/test_sort_and_select/test_sort_and_select-ab12b6ad99352fe2.xml (deflated 91%) 2025-12-04T17:17:38.4587481Z adding: test/test-reports/python-pytest/functorch.test_rearrange/functorch.test_rearrange-e13c2240713d23f9.xml (deflated 77%) 2025-12-04T17:17:38.4588738Z adding: test/test-reports/python-pytest/functorch.test_parsing/functorch.test_parsing-a56fd01f0949f839.xml (deflated 77%) 2025-12-04T17:17:38.4590552Z adding: test/test-reports/python-pytest/profiler.test_profiler/profiler.test_profiler-ac51a5ffaf89322c.xml (deflated 83%) 2025-12-04T17:17:38.4591822Z adding: test/test-reports/python-pytest/torch_np.test_binary_ufuncs/torch_np.test_binary_ufuncs-e1b3ae8f2439c11d.xml (deflated 91%) 2025-12-04T17:17:38.4593094Z adding: test/test-reports/python-pytest/torch_np.test_unary_ufuncs/torch_np.test_unary_ufuncs-0ead42044090ba01.xml (deflated 91%) 2025-12-04T17:17:38.4594295Z adding: test/test-reports/python-pytest/test_utils_filelock/test_utils_filelock-bc82f2b31ad6d0a9.xml (deflated 49%) 2025-12-04T17:17:38.4595429Z adding: test/test-reports/python-pytest/test_extension_utils/test_extension_utils-1a92a953d2096623.xml (deflated 52%) 2025-12-04T17:17:38.4597127Z adding: test/test-reports/python-pytest/test_rename_privateuse1_to_existing_device/test_rename_privateuse1_to_existing_device-a35ba0709779b2ea.xml (deflated 38%) 2025-12-04T17:17:38.4598518Z adding: test/test-reports/python-pytest/nn.attention.test_fa4/nn.attention.test_fa4-7e669f8dd97f47a6.xml (deflated 97%) 2025-12-04T17:17:38.4601113Z adding: test/test-reports/python-pytest/typing.test_python_operators/typing.test_python_operators-6bf858ca1197a53f.xml (deflated 95%) 2025-12-04T17:17:38.4605140Z adding: test/test-reports/python-pytest/test_functionalization/test_functionalization-44af17f38403c26c.xml (deflated 87%) 2025-12-04T17:17:38.4606338Z adding: test/test-reports/python-pytest/profiler.test_kineto/profiler.test_kineto-1d87d2a72eae2747.xml (deflated 37%) 2025-12-04T17:17:38.4607558Z adding: test/test-reports/python-pytest/test_module_tracker/test_module_tracker-fac08a90d6f586d0.xml (deflated 49%) 2025-12-04T17:17:38.4608954Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_scalarinherit/torch_np.numpy_tests.core.test_scalarinherit-9bae58f1bf1858ad.xml (deflated 76%) 2025-12-04T17:17:38.4610368Z adding: test/test-reports/python-pytest/test_tensorexpr_pybind/test_tensorexpr_pybind-ea4c0e89818d6b23.xml (deflated 87%) 2025-12-04T17:17:38.4625158Z adding: test/test-reports/python-pytest/test_fx_experimental/test_fx_experimental-ab6703fb0b80faae.xml (deflated 97%) 2025-12-04T17:17:38.4626233Z adding: test/test-reports/python-pytest/test_fx_passes/test_fx_passes-268f4eef0e7a6c57.xml (deflated 92%) 2025-12-04T17:17:38.4627370Z adding: test/test-reports/python-pytest/functorch.test_logging/functorch.test_logging-94ce26f72ae6dfc2.xml (deflated 81%) 2025-12-04T17:17:38.4629149Z adding: test/test-reports/python-pytest/test_namedtensor/test_namedtensor-9c123d78058198ac.xml (deflated 88%) 2025-12-04T17:17:38.4631035Z adding: test/test-reports/python-pytest/test_tensorexpr/test_tensorexpr-0ea46f0a35c352a2.xml (deflated 84%) 2025-12-04T17:17:38.4633567Z adding: test/test-reports/python-pytest/functorch.test_minifier/functorch.test_minifier-320e05a9203d92d3.xml (deflated 92%) 2025-12-04T17:17:38.4635758Z adding: test/test-reports/python-pytest/higher_order_ops.test_invoke_quant/higher_order_ops.test_invoke_quant-98f369a4901365e9.xml (deflated 92%) 2025-12-04T17:17:38.4641021Z adding: test/test-reports/python-pytest/torch_np.test_basic/torch_np.test_basic-f66c371882977b0d.xml (deflated 96%) 2025-12-04T17:17:38.4644866Z adding: test/test-reports/python-pytest/test_jiterator/test_jiterator-faac0f7ad3ce18ff.xml (deflated 96%) 2025-12-04T17:17:38.4646211Z adding: test/test-reports/python-pytest/test_native_functions/test_native_functions-ed1f170746938cdb.xml (deflated 91%) 2025-12-04T17:17:38.4647276Z adding: test/test-reports/python-pytest/test_typing/test_typing-029ceb92f90c5fb6.xml (deflated 81%) 2025-12-04T17:17:38.4648436Z adding: test/test-reports/python-pytest/lazy.test_functionalization/lazy.test_functionalization-3b38bfe5c8489bba.xml (deflated 50%) 2025-12-04T17:17:38.4649681Z adding: test/test-reports/python-pytest/torch_np.test_random/torch_np.test_random-92b710d703a1f97f.xml (deflated 92%) 2025-12-04T17:17:38.4650898Z adding: test/test-reports/python-pytest/nn.test_multihead_attention/nn.test_multihead_attention-227b95684fd4827a.xml (deflated 78%) 2025-12-04T17:17:38.4654605Z adding: test/test-reports/python-pytest/test_legacy_vmap/test_legacy_vmap-1f56d97bf135e06f.xml (deflated 91%) 2025-12-04T17:17:38.4655716Z adding: test/test-reports/python-pytest/lazy.test_bindings/lazy.test_bindings-48affd3dc097f9a0.xml (deflated 37%) 2025-12-04T17:17:38.4754869Z adding: test/test-reports/python-pytest/test_utils/test_utils-cb6667809c56b359.xml (deflated 98%) 2025-12-04T17:17:38.4756280Z adding: test/test-reports/python-pytest/test_pytree/test_pytree-b5c87da58bd7db26.xml (deflated 92%) 2025-12-04T17:17:38.4757424Z adding: test/test-reports/python-pytest/test_namedtuple_return_api/test_namedtuple_return_api-9fcfafd59257c871.xml (deflated 73%) 2025-12-04T17:17:38.4758904Z adding: test/test-reports/python-pytest/profiler.test_record_function/profiler.test_record_function-5b633e88461adafa.xml (deflated 74%) 2025-12-04T17:17:38.4760248Z adding: test/test-reports/python-pytest/test_compile_benchmark_util/test_compile_benchmark_util-df68b65dcac6dfa0.xml (deflated 86%) 2025-12-04T17:17:38.4761643Z adding: test/test-reports/python-pytest/test_set_default_mobile_cpu_allocator/test_set_default_mobile_cpu_allocator-e4da27e112b70d28.xml (deflated 53%) 2025-12-04T17:17:38.4765702Z adding: test/test-reports/python-pytest/test_fake_tensor/test_fake_tensor-03d6e27aa07cc9f8.xml (deflated 90%) 2025-12-04T17:17:38.4766894Z adding: test/test-reports/python-pytest/higher_order_ops.test_print/higher_order_ops.test_print-b7576a32603564cc.xml (deflated 77%) 2025-12-04T17:17:38.4768188Z adding: test/test-reports/python-pytest/test_per_overload_api/test_per_overload_api-e690ac2b4ba1f976.xml (deflated 60%) 2025-12-04T17:17:38.4769518Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_einsum/torch_np.numpy_tests.core.test_einsum-21e2123caec3dd16.xml (deflated 88%) 2025-12-04T17:17:38.4770842Z adding: test/test-reports/python-pytest/test_multiprocessing/test_multiprocessing-66d2144d604c4acf.xml (deflated 81%) 2025-12-04T17:17:38.4822521Z adding: test/test-reports/python-pytest/test_modules/test_modules-5eb6bc8f38dae4a5.xml (deflated 95%) 2025-12-04T17:17:38.4830563Z adding: test/test-reports/python-pytest/complex_tensor.test_complex_tensor/complex_tensor.test_complex_tensor-00d0e28f118bdf92.xml (deflated 96%) 2025-12-04T17:17:38.4832586Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_indexing/torch_np.numpy_tests.core.test_indexing-3b590877a50ad13d.xml (deflated 89%) 2025-12-04T17:17:38.4833851Z adding: test/test-reports/python-pytest/test_futures/test_futures-8abfe477c1d087d4.xml (deflated 83%) 2025-12-04T17:17:38.4842719Z adding: test/test-reports/python-pytest/test_tensor_creation_ops/test_tensor_creation_ops-c328fea3f872cecd.xml (deflated 94%) 2025-12-04T17:17:38.4847474Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_scalarmath/torch_np.numpy_tests.core.test_scalarmath-a9b7ea367b3cdb41.xml (deflated 96%) 2025-12-04T17:17:38.4851853Z adding: test/test-reports/python-pytest/test_serialization/test_serialization-43b1fedba2a2b650.xml (deflated 90%) 2025-12-04T17:17:38.4853520Z adding: test/test-reports/python-pytest/torch_np.numpy_tests.core.test_dlpack/torch_np.numpy_tests.core.test_dlpack-350d972e2ff20e05.xml (deflated 96%) 2025-12-04T17:17:38.4854917Z adding: test/test-reports/python-pytest/test_multiprocessing_spawn/test_multiprocessing_spawn-b04898e1c1418acf.xml (deflated 86%) 2025-12-04T17:17:38.4856194Z adding: test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-ee0ad83af64551d9.xml (deflated 36%) 2025-12-04T17:17:38.4857494Z adding: test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-445675b8d8f15e5e.xml (deflated 36%) 2025-12-04T17:17:38.4858735Z adding: test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-d82ae01551c43307.xml (deflated 37%) 2025-12-04T17:17:38.4859974Z adding: test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-4e71d02e98da078f.xml (deflated 36%) 2025-12-04T17:17:38.4861204Z adding: test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-233c3fc007b40619.xml (deflated 36%) 2025-12-04T17:17:38.4862447Z adding: test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-a0f26f1a84f3180c.xml (deflated 37%) 2025-12-04T17:17:38.4863694Z adding: test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-4a2a7cd181a00ba9.xml (deflated 34%) 2025-12-04T17:17:38.4864937Z adding: test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-24fabc8827e58618.xml (deflated 36%) 2025-12-04T17:17:38.4866257Z adding: test/test-reports/python-pytest/test_cuda_nvml_based_avail/test_cuda_nvml_based_avail-f87021e84032e697.xml (deflated 34%) 2025-12-04T17:17:38.4867473Z adding: test/test-reports/python-pytest/test_mobile_optimizer/test_mobile_optimizer-8da56fb35697f793.xml (deflated 80%) 2025-12-04T17:17:38.4868705Z adding: test/test-reports/python-pytest/torch_np.test_function_base/torch_np.test_function_base-fc4ff98d96269d3c.xml (deflated 48%) 2025-12-04T17:17:38.4869928Z adding: test/test-reports/python-pytest/test_type_promotion/test_type_promotion-bc48b0383dc65f97.xml (deflated 96%) 2025-12-04T17:17:38.4871031Z adding: test/test-reports/python-pytest/lazy.test_reuse_ir/lazy.test_reuse_ir-27aa6d3f8959224d.xml (deflated 61%) 2025-12-04T17:17:38.4872359Z adding: test/test-reports/python-pytest/test_functional_autograd_benchmark/test_functional_autograd_benchmark-8f47749aefc55c89.xml (deflated 54%) 2025-12-04T17:17:38.4901677Z ##[group]Run # Remove any previous usage logs if they exist 2025-12-04T17:17:38.4902219Z # Remove any previous usage logs if they exist 2025-12-04T17:17:38.4902732Z rm -f logs-*.zip 2025-12-04T17:17:38.4903132Z zip "logs-${FILE_SUFFIX}.zip" 'usage_log.txt' || true 2025-12-04T17:17:38.4903715Z zip -r "logs-${FILE_SUFFIX}.zip" test/test-reports -i '*.log' || true 2025-12-04T17:17:38.4910612Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T17:17:38.4911058Z env: 2025-12-04T17:17:38.4911294Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:38.4911601Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:38.4911963Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:38.4912609Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:38.4913420Z FILE_SUFFIX: test-legacy_nvidia_driver-3-5-linux.g4dn.4xlarge.nvidia.gpu_57119749282 2025-12-04T17:17:38.4913997Z ##[endgroup] 2025-12-04T17:17:38.4987965Z adding: usage_log.txt (deflated 58%) 2025-12-04T17:17:38.5081206Z adding: test/test-reports/inductor.test_aot_inductor_3.6_01eb6db542665098_.log (deflated 93%) 2025-12-04T17:17:38.5120238Z adding: test/test-reports/inductor.test_torchinductor_2.2_198ec367be5c71a3_.log (deflated 93%) 2025-12-04T17:17:38.5122008Z adding: test/test-reports/torch_np.numpy_tests.core.test_einsum_1.1_62576103f884ccec_.log (deflated 85%) 2025-12-04T17:17:38.5143112Z adding: test/test-reports/inductor.test_torchinductor_dynamic_shapes_4.5_2869e792aa617cfa_.log (deflated 92%) 2025-12-04T17:17:38.5144087Z adding: test/test-reports/higher_order_ops.test_print_1.1_da02828dbc3b3692_.log (deflated 73%) 2025-12-04T17:17:38.5163167Z adding: test/test-reports/inductor.test_torchinductor_codegen_dynamic_shapes_4.4_0fcd01f998dd3903_.log (deflated 92%) 2025-12-04T17:17:38.5171654Z adding: test/test-reports/inductor.test_torchinductor_opinfo_5.17_b8d59ea0e8d50306_.log (deflated 90%) 2025-12-04T17:17:38.5173188Z adding: test/test-reports/test_multiprocessing_1.1_9528eb91822da171_.log (deflated 80%) 2025-12-04T17:17:38.5183551Z adding: test/test-reports/inductor.test_torchinductor_opinfo_10.17_664b3f49cadd1793_.log (deflated 93%) 2025-12-04T17:17:38.5264612Z adding: test/test-reports/test_modules_1.1_8b9aed1ec5d9ad87_.log (deflated 95%) 2025-12-04T17:17:38.5272648Z adding: test/test-reports/inductor.test_torchinductor_opinfo_15.17_68b5d6354fd5686b_.log (deflated 90%) 2025-12-04T17:17:38.5274530Z adding: test/test-reports/dynamo.test_logging_1.1_65e572a46b19d5f7_.log (deflated 87%) 2025-12-04T17:17:38.5285955Z adding: test/test-reports/dynamo.test_repros_1.1_3353f5f590b0aac4_.log (deflated 85%) 2025-12-04T17:17:38.5326452Z adding: test/test-reports/inductor.test_cuda_select_algorithm_5.5_387c3291169f376b_.log (deflated 97%) 2025-12-04T17:17:38.5327410Z adding: test/test-reports/inductor.test_deterministic_2.8_649a40a09099c603_.log (deflated 60%) 2025-12-04T17:17:38.5328309Z adding: test/test-reports/inductor.test_deterministic_7.8_cf2a089c308daa96_.log (deflated 66%) 2025-12-04T17:17:38.5329372Z adding: test/test-reports/inductor.test_native_matmul_2.2_564ae2b75fe7723b_.log (deflated 56%) 2025-12-04T17:17:38.5345262Z adding: test/test-reports/complex_tensor.test_complex_tensor_1.1_4feb438836f89d3d_.log (deflated 95%) 2025-12-04T17:17:38.6458621Z adding: test/test-reports/inductor.test_decompose_mem_bound_mm_1.1_7b3d736361a2b2dc_.log (deflated 87%) 2025-12-04T17:17:38.6487231Z adding: test/test-reports/inductor.test_online_softmax_1.1_d592e7340b8c6a36_.log (deflated 98%) 2025-12-04T17:17:38.6579077Z adding: test/test-reports/inductor.test_mix_order_reduction_2.2_7b65ea7a55b42f88_.log (deflated 98%) 2025-12-04T17:17:38.6644662Z adding: test/test-reports/test_matmul_cuda_1.1_e0f80f5015105a39_.log (deflated 97%) 2025-12-04T17:17:38.6648583Z adding: test/test-reports/test_dataloader_2.2_8685276eb73ca04c_.log (deflated 85%) 2025-12-04T17:17:38.6664593Z adding: test/test-reports/test_decomp_5.17_3ed363c2fcab4a0f_.log (deflated 89%) 2025-12-04T17:17:38.6680142Z adding: test/test-reports/test_decomp_10.17_cae88f2962bfc8d3_.log (deflated 89%) 2025-12-04T17:17:38.6695810Z adding: test/test-reports/test_decomp_15.17_7513624a35207a1d_.log (deflated 89%) 2025-12-04T17:17:38.6903287Z adding: test/test-reports/test_meta_1.5_23737c7a1f6b460a_.log (deflated 93%) 2025-12-04T17:17:38.6919088Z adding: test/test-reports/test_ops_jit_1.2_893406403e6f8fa5_.log (deflated 91%) 2025-12-04T17:17:38.6934193Z adding: test/test-reports/test_nestedtensor_2.4_5194ad5857dee29b_.log (deflated 91%) 2025-12-04T17:17:38.6935365Z adding: test/test-reports/test_public_bindings_1.1_c7580692556b690c_.log (deflated 59%) 2025-12-04T17:17:38.7025844Z adding: test/test-reports/test_ops_1.11_b5934d9d9b184ff9_.log (deflated 91%) 2025-12-04T17:17:38.7112663Z adding: test/test-reports/test_ops_6.11_ab7dea6c6996f1c3_.log (deflated 91%) 2025-12-04T17:17:38.7198045Z adding: test/test-reports/test_ops_11.11_c9baa1a26aa25252_.log (deflated 91%) 2025-12-04T17:17:38.7237295Z adding: test/test-reports/functorch.test_ops_3.7_580309c6de0de560_.log (deflated 92%) 2025-12-04T17:17:38.7312339Z adding: test/test-reports/functorch.test_vmap_1.1_151d84486b0c8e0c_.log (deflated 94%) 2025-12-04T17:17:38.7313186Z adding: test/test-reports/dynamo.test_after_aot_1.1_c57cf0a68fd0c5b4_.log (deflated 54%) 2025-12-04T17:17:38.7314241Z adding: test/test-reports/inductor.test_snode_runtime_1.1_55d09bf84d516763_.log (deflated 79%) 2025-12-04T17:17:38.7320629Z adding: test/test-reports/inductor.test_minifier_1.1_9cc263edd080c3f5_.log (deflated 93%) 2025-12-04T17:17:38.7354757Z adding: test/test-reports/inductor.test_compiled_autograd_1.1_bff035ff4424e9fc_.log (deflated 91%) 2025-12-04T17:17:38.7355665Z adding: test/test-reports/inductor.test_layout_optim_1.1_12d2e0f1c8d2c698_.log (deflated 50%) 2025-12-04T17:17:38.7357285Z adding: test/test-reports/dynamo.test_unspec_1.1_9e5ffa0a28ec4624_.log (deflated 80%) 2025-12-04T17:17:38.7358705Z adding: test/test-reports/inductor.test_mmdecomp_1.1_8ac94a9aa95bcda4_.log (deflated 81%) 2025-12-04T17:17:38.7362348Z adding: test/test-reports/dynamo.test_ctx_manager_1.1_7cf597ef36c79024_.log (deflated 87%) 2025-12-04T17:17:38.7363358Z adding: test/test-reports/dynamo.test_exc_1.1_d00a7ab51481eb46_.log (deflated 72%) 2025-12-04T17:17:38.7384431Z adding: test/test-reports/dynamo.test_misc_1.1_cef59efd3c8e72ba_.log (deflated 85%) 2025-12-04T17:17:38.7391850Z adding: test/test-reports/inductor.test_aot_inductor_arrayref_2.2_d8fa4d4f4ba9cf83_.log (deflated 91%) 2025-12-04T17:17:38.7392755Z adding: test/test-reports/inductor.test_halide_1.1_8d6772cbd633567a_.log (deflated 7%) 2025-12-04T17:17:38.7393568Z adding: test/test-reports/inductor.test_xpu_basic_1.1_8048a1f844b679fa_.log (deflated 7%) 2025-12-04T17:17:38.7394478Z adding: test/test-reports/inductor.test_provenance_tracing_1.1_fa80dd40f7fc5a69_.log (deflated 79%) 2025-12-04T17:17:38.7395406Z adding: test/test-reports/dynamo.test_buffers_override_1.1_f8bb7c2759bc3846_.log (deflated 58%) 2025-12-04T17:17:38.7396661Z adding: test/test-reports/inductor.test_inplacing_pass_1.1_1204e772d28433be_.log (deflated 80%) 2025-12-04T17:17:38.7399007Z adding: test/test-reports/torch_np.numpy_tests.core.test_indexing_1.1_33940d3a3f5ad817_.log (deflated 86%) 2025-12-04T17:17:38.7400984Z adding: test/test-reports/inductor.test_aot_inductor_custom_ops_1.1_663d4eb0571b35c8_.log (deflated 88%) 2025-12-04T17:17:38.7401961Z adding: test/test-reports/inductor.test_split_cat_fx_passes_1.1_64f7c65379e18b89_.log (deflated 74%) 2025-12-04T17:17:38.7402871Z adding: test/test-reports/inductor.test_profiler_1.1_eb004b5dbdc54b3f_.log (deflated 67%) 2025-12-04T17:17:38.7403771Z adding: test/test-reports/inductor.test_memory_planning_1.1_7d8eac35fee71fb8_.log (deflated 59%) 2025-12-04T17:17:38.7404748Z adding: test/test-reports/inductor.test_mem_estimation_1.1_db7eba60d0c86675_.log (deflated 63%) 2025-12-04T17:17:38.7405597Z adding: test/test-reports/dynamo.test_view_1.1_285b149edbf19f8b_.log (deflated 65%) 2025-12-04T17:17:38.7406461Z adding: test/test-reports/inductor.test_cutlass_evt_1.1_5894d03fbbc62b34_.log (deflated 69%) 2025-12-04T17:17:38.7407345Z adding: test/test-reports/dynamo.test_reconstruct_1.1_abcfd742020bd6ec_.log (deflated 77%) 2025-12-04T17:17:38.7410233Z adding: test/test-reports/dynamo.test_aot_autograd_1.1_a815d5749913f0e4_.log (deflated 90%) 2025-12-04T17:17:38.7427799Z adding: test/test-reports/export.test_cpp_serdes_1.1_9da4c12e7e041500_.log (deflated 89%) 2025-12-04T17:17:38.7428889Z adding: test/test-reports/inductor.test_block_analysis_1.1_b3c20d306fe594c2_.log (deflated 75%) 2025-12-04T17:17:38.7430368Z adding: test/test-reports/dynamo.test_subgraphs_1.1_e1155ecf654700c6_.log (deflated 82%) 2025-12-04T17:17:38.7431227Z adding: test/test-reports/dynamo.test_pre_dispatch_1.1_99fd2ec956edaa5c_.log (deflated 59%) 2025-12-04T17:17:38.7432999Z adding: test/test-reports/test_futures_1.1_2500c18a32eb03f9_.log (deflated 86%) 2025-12-04T17:17:38.7433880Z adding: test/test-reports/inductor.test_custom_post_grad_passes_1.1_14ce4ccc6fc3bcc1_.log (deflated 71%) 2025-12-04T17:17:38.7434822Z adding: test/test-reports/dynamo.test_fx_annotate_1.1_a80c3def106c395c_.log (deflated 66%) 2025-12-04T17:17:38.7435642Z adding: test/test-reports/dynamo.test_pgo_1.1_c150ddd3d0d86d52_.log (deflated 65%) 2025-12-04T17:17:38.7436792Z adding: test/test-reports/export.test_export_opinfo_1.1_ca68eeb5bfbf3cab_.log (deflated 75%) 2025-12-04T17:17:38.7940209Z adding: test/test-reports/inductor.test_control_flow_2.4_11abfa8a57b42ba2_.log (deflated 96%) 2025-12-04T17:17:38.7941055Z adding: test/test-reports/dynamo.test_compile_1.1_b494fc7b48d70de4_.log (deflated 75%) 2025-12-04T17:17:38.7942203Z adding: test/test-reports/dynamo.test_nested_graph_breaks_1.1_06faf157c67ea009_.log (deflated 81%) 2025-12-04T17:17:38.7943171Z adding: test/test-reports/inductor.test_needs_exact_strides_1.1_fcd620a2b11e818d_.log (deflated 58%) 2025-12-04T17:17:38.7957682Z adding: test/test-reports/test_tensor_creation_ops_1.1_8ce9396a5d7487a9_.log (deflated 93%) 2025-12-04T17:17:38.7958625Z adding: test/test-reports/inductor.test_split_cat_fx_aten_passes_1.1_c6d6a90bc763f053_.log (deflated 76%) 2025-12-04T17:17:38.7959503Z adding: test/test-reports/dynamo.test_resume_1.1_a3d627742d7240d5_.log (deflated 50%) 2025-12-04T17:17:38.7965354Z adding: test/test-reports/test_serialization_1.1_39980f3b0bf1286b_.log (deflated 90%) 2025-12-04T17:17:38.7966263Z adding: test/test-reports/dynamo.test_backward_higher_order_ops_1.1_c04481e0c8fd9912_.log (deflated 72%) 2025-12-04T17:17:38.7967249Z adding: test/test-reports/inductor.test_custom_partitioner_fn_1.1_2eb6199e2d992792_.log (deflated 55%) 2025-12-04T17:17:38.7968145Z adding: test/test-reports/dynamo.test_debug_utils_1.1_3242effe45123c25_.log (deflated 62%) 2025-12-04T17:17:38.7968977Z adding: test/test-reports/dynamo.test_base_hop_1.1_81fd091fbb10b281_.log (deflated 70%) 2025-12-04T17:17:38.7971029Z adding: test/test-reports/dynamo.test_package_1.1_bbe95160f77383a1_.log (deflated 87%) 2025-12-04T17:17:38.7974996Z adding: test/test-reports/dynamo.test_aot_autograd_cache_1.1_f0d2e6b1214663ab_.log (deflated 89%) 2025-12-04T17:17:38.7975880Z adding: test/test-reports/inductor.test_mps_basic_1.1_979b26eef19a62b7_.log (deflated 7%) 2025-12-04T17:17:38.7976702Z adding: test/test-reports/dynamo.test_comptime_1.1_a63436f12def3f22_.log (deflated 71%) 2025-12-04T17:17:38.7979170Z adding: test/test-reports/test_sort_and_select_1.1_5068bfed31d9fabf_.log (deflated 89%) 2025-12-04T17:17:38.7980216Z adding: test/test-reports/functorch.test_rearrange_1.1_dc24a0c907a22605_.log (deflated 71%) 2025-12-04T17:17:38.7981084Z adding: test/test-reports/functorch.test_parsing_1.1_0964d504887b4f27_.log (deflated 73%) 2025-12-04T17:17:38.7985122Z adding: test/test-reports/profiler.test_profiler_1.1_f03b04cece9b91aa_.log (deflated 82%) 2025-12-04T17:17:38.7986225Z adding: test/test-reports/torch_np.test_binary_ufuncs_1.1_3f5612601f960528_.log (deflated 85%) 2025-12-04T17:17:38.7987457Z adding: test/test-reports/torch_np.test_unary_ufuncs_1.1_2c643a870062efc9_.log (deflated 85%) 2025-12-04T17:17:38.7988296Z adding: test/test-reports/test_utils_filelock_1.1_4b9e876ba6fc808a_.log (deflated 54%) 2025-12-04T17:17:38.7989098Z adding: test/test-reports/test_extension_utils_1.1_d74913d9452f40a0_.log (deflated 57%) 2025-12-04T17:17:38.7994857Z adding: test/test-reports/torch_np.numpy_tests.core.test_scalarmath_1.1_7a9b3c7e6c8b9fbf_.log (deflated 93%) 2025-12-04T17:17:38.7995895Z adding: test/test-reports/test_rename_privateuse1_to_existing_device_1.1_066a744b408e3e3f_.log (deflated 56%) 2025-12-04T17:17:38.7998351Z adding: test/test-reports/nn.attention.test_fa4_1.1_c563975b7201948d_.log (deflated 94%) 2025-12-04T17:17:38.8006905Z adding: test/test-reports/typing.test_python_operators_1.1_7abeab1ea263e7df_.log (deflated 93%) 2025-12-04T17:17:38.8009739Z adding: test/test-reports/test_functionalization_1.1_e1689373b3a112cf_.log (deflated 89%) 2025-12-04T17:17:38.8010578Z adding: test/test-reports/profiler.test_kineto_1.1_5a67f12c59b2d264_.log (deflated 51%) 2025-12-04T17:17:38.8011403Z adding: test/test-reports/test_module_tracker_1.1_ac47ee0370be4ffc_.log (deflated 58%) 2025-12-04T17:17:38.8012206Z adding: test/test-reports/test_per_overload_api_1.1_c9215af73fca3909_.log (deflated 59%) 2025-12-04T17:17:38.8013226Z adding: test/test-reports/torch_np.numpy_tests.core.test_scalarinherit_1.1_98f4b4754a63741f_.log (deflated 63%) 2025-12-04T17:17:38.8014183Z adding: test/test-reports/test_tensorexpr_pybind_1.1_7dab8d3742d21018_.log (deflated 78%) 2025-12-04T17:17:38.8032547Z adding: test/test-reports/test_fx_experimental_1.1_6141111e1c7aeffb_.log (deflated 94%) 2025-12-04T17:17:38.8034244Z adding: test/test-reports/test_fx_passes_1.1_01e7f0194a1e970b_.log (deflated 88%) 2025-12-04T17:17:38.8035061Z adding: test/test-reports/functorch.test_logging_1.1_ac84cd84fa23725e_.log (deflated 52%) 2025-12-04T17:17:38.8037475Z adding: test/test-reports/test_namedtensor_1.1_b9881adc6407f0b9_.log (deflated 84%) 2025-12-04T17:17:38.8039683Z adding: test/test-reports/test_tensorexpr_1.1_d8551b64b93513ad_.log (deflated 83%) 2025-12-04T17:17:38.8040615Z adding: test/test-reports/functorch.test_minifier_1.1_a1ba5cac68d3f419_.log (deflated 63%) 2025-12-04T17:17:38.8041531Z adding: test/test-reports/higher_order_ops.test_invoke_quant_1.1_1d7f8dcadfd047a9_.log (deflated 76%) 2025-12-04T17:17:38.8050688Z adding: test/test-reports/torch_np.test_basic_1.1_83f086e013fa5452_.log (deflated 93%) 2025-12-04T17:17:38.8058416Z adding: test/test-reports/test_jiterator_1.1_8988ac20057f1363_.log (deflated 94%) 2025-12-04T17:17:38.8059421Z adding: test/test-reports/test_native_functions_1.1_be6b38187a1d4fbd_.log (deflated 74%) 2025-12-04T17:17:38.8060318Z adding: test/test-reports/test_typing_1.1_d8e70b3c726a1ce6_.log (deflated 74%) 2025-12-04T17:17:38.8061145Z adding: test/test-reports/lazy.test_functionalization_1.1_2e9665a4a8fbdab9_.log (deflated 57%) 2025-12-04T17:17:38.8062343Z adding: test/test-reports/torch_np.test_random_1.1_5e1f3614f612e456_.log (deflated 87%) 2025-12-04T17:17:38.8063720Z adding: test/test-reports/nn.test_multihead_attention_1.1_aa528fb509f77678_.log (deflated 83%) 2025-12-04T17:17:38.8067123Z adding: test/test-reports/test_legacy_vmap_1.1_88ccad856a60ef1f_.log (deflated 87%) 2025-12-04T17:17:38.8067909Z adding: test/test-reports/lazy.test_bindings_1.1_b260f38330c47014_.log (deflated 50%) 2025-12-04T17:17:38.8193912Z adding: test/test-reports/test_utils_1.1_f544fe8617f67377_.log (deflated 96%) 2025-12-04T17:17:38.8196713Z adding: test/test-reports/test_pytree_1.1_52fd485987cd9dff_.log (deflated 87%) 2025-12-04T17:17:38.8197517Z adding: test/test-reports/test_namedtuple_return_api_1.1_ed47513e0b44d1fb_.log (deflated 60%) 2025-12-04T17:17:38.8198567Z adding: test/test-reports/profiler.test_record_function_1.1_ca949195f6b456e2_.log (deflated 70%) 2025-12-04T17:17:38.8199468Z adding: test/test-reports/test_compile_benchmark_util_1.1_fb69b29a06036afc_.log (deflated 53%) 2025-12-04T17:17:38.8200388Z adding: test/test-reports/test_set_default_mobile_cpu_allocator_1.1_a92884857a29fe5f_.log (deflated 59%) 2025-12-04T17:17:38.8208848Z adding: test/test-reports/test_fake_tensor_1.1_51e2ac03aee772c5_.log (deflated 90%) 2025-12-04T17:17:38.8210552Z adding: test/test-reports/torch_np.numpy_tests.core.test_dlpack_1.1_be350d36e188bf08_.log (deflated 90%) 2025-12-04T17:17:38.8211876Z adding: test/test-reports/test_multiprocessing_spawn_1.1_a0b8d9d26431815a_.log (deflated 82%) 2025-12-04T17:17:38.8213320Z adding: test/test-reports/test_cuda_nvml_based_avail_1.1_c66f584abc6fe991_.log (deflated 92%) 2025-12-04T17:17:38.8214382Z adding: test/test-reports/test_mobile_optimizer_1.1_f99b91a9a53c0fc9_.log (deflated 67%) 2025-12-04T17:17:38.8215237Z adding: test/test-reports/torch_np.test_function_base_1.1_2a583a88d944c488_.log (deflated 55%) 2025-12-04T17:17:38.8226190Z adding: test/test-reports/test_type_promotion_1.1_fda622c3de5d55b3_.log (deflated 94%) 2025-12-04T17:17:38.8226986Z adding: test/test-reports/lazy.test_reuse_ir_1.1_bce662c061b92625_.log (deflated 59%) 2025-12-04T17:17:38.8228526Z adding: test/test-reports/test_functional_autograd_benchmark_1.1_3090448ccb207ef1_.log (deflated 87%) 2025-12-04T17:17:38.8271742Z ##[group]Run # Remove any previous debugging artifacts if they exist 2025-12-04T17:17:38.8272988Z # Remove any previous debugging artifacts if they exist 2025-12-04T17:17:38.8273813Z rm -f debug-*.zip 2025-12-04T17:17:38.8274371Z if [ -d 'test/debug' ]; then 2025-12-04T17:17:38.8275062Z  zip -r "debug-${FILE_SUFFIX}.zip" test/debug 2025-12-04T17:17:38.8275533Z fi 2025-12-04T17:17:38.8282316Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T17:17:38.8282759Z env: 2025-12-04T17:17:38.8282995Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:38.8283304Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:38.8283665Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:38.8284308Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:38.8285127Z FILE_SUFFIX: test-legacy_nvidia_driver-3-5-linux.g4dn.4xlarge.nvidia.gpu_57119749282 2025-12-04T17:17:38.8285700Z ##[endgroup] 2025-12-04T17:17:38.8378380Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T17:17:38.8378792Z with: 2025-12-04T17:17:38.8379049Z s3-bucket: gha-artifacts 2025-12-04T17:17:38.8379402Z s3-prefix: pytorch/pytorch/19922826259/1/artifact 2025-12-04T17:17:38.8379817Z retention-days: 14 2025-12-04T17:17:38.8380108Z if-no-files-found: warn 2025-12-04T17:17:38.8380426Z path: test-jsons-*.zip 2025-12-04T17:17:38.8380711Z name: artifact 2025-12-04T17:17:38.8380972Z region: us-east-1 2025-12-04T17:17:38.8381238Z env: 2025-12-04T17:17:38.8381464Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:38.8381766Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:38.8382133Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:38.8382775Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:38.8383354Z ##[endgroup] 2025-12-04T17:17:39.2336276Z NOTE: s3-prefix specified, ignoring name parameter 2025-12-04T17:17:39.2336889Z With the provided path, there will be 1 file uploaded 2025-12-04T17:17:39.2337424Z Uploading to s3 prefix: pytorch/pytorch/19922826259/1/artifact 2025-12-04T17:17:39.2391489Z Starting upload of test-jsons-test-legacy_nvidia_driver-3-5-linux.g4dn.4xlarge.nvidia.gpu_57119749282.zip 2025-12-04T17:17:39.5283078Z Finished upload of test-jsons-test-legacy_nvidia_driver-3-5-linux.g4dn.4xlarge.nvidia.gpu_57119749282.zip 2025-12-04T17:17:39.5495399Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T17:17:39.5495779Z with: 2025-12-04T17:17:39.5496438Z s3-bucket: gha-artifacts 2025-12-04T17:17:39.5496814Z s3-prefix: pytorch/pytorch/19922826259/1/artifact 2025-12-04T17:17:39.5497304Z retention-days: 14 2025-12-04T17:17:39.5497586Z if-no-files-found: error 2025-12-04T17:17:39.5497907Z path: test-reports-*.zip 2025-12-04T17:17:39.5498216Z name: artifact 2025-12-04T17:17:39.5498484Z region: us-east-1 2025-12-04T17:17:39.5498742Z env: 2025-12-04T17:17:39.5498983Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:39.5499274Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:39.5499637Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:39.5500297Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:39.5500862Z ##[endgroup] 2025-12-04T17:17:39.9643134Z NOTE: s3-prefix specified, ignoring name parameter 2025-12-04T17:17:39.9643919Z With the provided path, there will be 1 file uploaded 2025-12-04T17:17:39.9644646Z Uploading to s3 prefix: pytorch/pytorch/19922826259/1/artifact 2025-12-04T17:17:39.9699432Z Starting upload of test-reports-test-legacy_nvidia_driver-3-5-linux.g4dn.4xlarge.nvidia.gpu_57119749282.zip 2025-12-04T17:17:40.1313191Z Finished upload of test-reports-test-legacy_nvidia_driver-3-5-linux.g4dn.4xlarge.nvidia.gpu_57119749282.zip 2025-12-04T17:17:40.1530343Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T17:17:40.1530876Z with: 2025-12-04T17:17:40.1531260Z s3-bucket: gha-artifacts 2025-12-04T17:17:40.1531707Z s3-prefix: pytorch/pytorch/19922826259/1/artifact 2025-12-04T17:17:40.1532219Z retention-days: 14 2025-12-04T17:17:40.1532734Z if-no-files-found: ignore 2025-12-04T17:17:40.1533178Z path: logs-*.zip 2025-12-04T17:17:40.1533502Z name: artifact 2025-12-04T17:17:40.1533889Z region: us-east-1 2025-12-04T17:17:40.1534320Z env: 2025-12-04T17:17:40.1534622Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:40.1535063Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:40.1535557Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:40.1536288Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:40.1537105Z ##[endgroup] 2025-12-04T17:17:40.5312040Z NOTE: s3-prefix specified, ignoring name parameter 2025-12-04T17:17:40.5312637Z With the provided path, there will be 1 file uploaded 2025-12-04T17:17:40.5313196Z Uploading to s3 prefix: pytorch/pytorch/19922826259/1/artifact 2025-12-04T17:17:40.5367143Z Starting upload of logs-test-legacy_nvidia_driver-3-5-linux.g4dn.4xlarge.nvidia.gpu_57119749282.zip 2025-12-04T17:17:40.7504324Z Finished upload of logs-test-legacy_nvidia_driver-3-5-linux.g4dn.4xlarge.nvidia.gpu_57119749282.zip 2025-12-04T17:17:40.7716584Z ##[group]Run seemethere/upload-artifact-s3@v5 2025-12-04T17:17:40.7716974Z with: 2025-12-04T17:17:40.7717209Z s3-bucket: gha-artifacts 2025-12-04T17:17:40.7717582Z s3-prefix: pytorch/pytorch/19922826259/1/artifact 2025-12-04T17:17:40.7717988Z retention-days: 14 2025-12-04T17:17:40.7718266Z if-no-files-found: ignore 2025-12-04T17:17:40.7718588Z path: debug-*.zip 2025-12-04T17:17:40.7718852Z name: artifact 2025-12-04T17:17:40.7719096Z region: us-east-1 2025-12-04T17:17:40.7719350Z env: 2025-12-04T17:17:40.7719591Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:40.7719884Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:40.7720249Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:40.7721085Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:40.7721670Z ##[endgroup] 2025-12-04T17:17:41.1421561Z No files were found with the provided path: debug-*.zip. No artifacts will be uploaded. 2025-12-04T17:17:41.1638697Z ##[group]Run # shellcheck disable=SC2156 2025-12-04T17:17:41.1639385Z # shellcheck disable=SC2156 2025-12-04T17:17:41.1640477Z find . -iname "core.[1-9]*" -exec docker exec "${DOCKER_CONTAINER_ID}" sh -c "gdb python {} -ex 'bt' -ex 'q'" \; 2025-12-04T17:17:41.1650735Z shell: /usr/bin/bash -e {0} 2025-12-04T17:17:41.1651239Z env: 2025-12-04T17:17:41.1651773Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:41.1652331Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:41.1652944Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:41.1654049Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:41.1654875Z ##[endgroup] 2025-12-04T17:17:41.5475261Z ##[group]Run seemethere/upload-artifact-s3@baba72d0712b404f646cebe0730933554ebce96a 2025-12-04T17:17:41.5475825Z with: 2025-12-04T17:17:41.5476242Z name: coredumps-legacy_nvidia_driver-3-5-linux.g4dn.4xlarge.nvidia.gpu 2025-12-04T17:17:41.5476774Z retention-days: 14 2025-12-04T17:17:41.5477075Z if-no-files-found: ignore 2025-12-04T17:17:41.5477376Z path: ./**/core.[1-9]* 2025-12-04T17:17:41.5477675Z s3-bucket: gha-artifacts 2025-12-04T17:17:41.5477983Z region: us-east-1 2025-12-04T17:17:41.5478232Z env: 2025-12-04T17:17:41.5478476Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:41.5478779Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:41.5479138Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:41.5479794Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:41.5480405Z ##[endgroup] 2025-12-04T17:17:51.0015265Z No files were found with the provided path: ./**/core.[1-9]*. No artifacts will be uploaded. 2025-12-04T17:17:51.0331931Z Prepare all required actions 2025-12-04T17:17:51.0332387Z Getting action download info 2025-12-04T17:17:51.1582524Z Download action repository 'actions/setup-python@v6' (SHA:83679a892e2d95755f2dac6acb0bfd1e9ac5d548) 2025-12-04T17:17:51.4714325Z ##[group]Run ./.github/actions/upload-utilization-stats 2025-12-04T17:17:51.4714741Z with: 2025-12-04T17:17:51.4714988Z job_id: 57119749282 2025-12-04T17:17:51.4715709Z job_name: linux-jammy-cuda12.4-py3.10-gcc11 / test (legacy_nvidia_driver, 3, 5, linux.g4dn.4xlarge.nvidia.gpu, mem_leak_check, unstable) 2025-12-04T17:17:51.4716508Z workflow_name: periodic 2025-12-04T17:17:51.4716806Z workflow_run_id: 19922826259 2025-12-04T17:17:51.4717136Z workflow_attempt: 1 2025-12-04T17:17:51.4717407Z env: 2025-12-04T17:17:51.4717634Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:51.4717937Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:51.4718296Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:51.4718993Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:51.4719561Z ##[endgroup] 2025-12-04T17:17:51.4762261Z ##[group]Run actions/setup-python@v6 2025-12-04T17:17:51.4762616Z with: 2025-12-04T17:17:51.4762864Z python-version: 3.10 2025-12-04T17:17:51.4763154Z check-latest: false 2025-12-04T17:17:51.4763545Z token: *** 2025-12-04T17:17:51.4763814Z update-environment: true 2025-12-04T17:17:51.4764135Z allow-prereleases: false 2025-12-04T17:17:51.4764431Z freethreaded: false 2025-12-04T17:17:51.4764703Z env: 2025-12-04T17:17:51.4764941Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:51.4765236Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:51.4765604Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:51.4766262Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:51.4766833Z ##[endgroup] 2025-12-04T17:17:51.9637865Z ##[group]Installed versions 2025-12-04T17:17:51.9650271Z Version 3.10 was not found in the local cache 2025-12-04T17:17:51.9963182Z (node:326118) [DEP0040] DeprecationWarning: The `punycode` module is deprecated. Please use a userland alternative instead. 2025-12-04T17:17:51.9964909Z (Use `node --trace-deprecation ...` to show where the warning was created) 2025-12-04T17:17:52.3597647Z ##[error]The version '3.10' with architecture 'x64' was not found for this operating system. The list of all available versions can be found here: https://raw.githubusercontent.com/actions/python-versions/main/versions-manifest.json 2025-12-04T17:17:52.3785204Z ##[group]Run pytorch/test-infra/.github/actions/teardown-linux@main 2025-12-04T17:17:52.3785707Z with: 2025-12-04T17:17:52.3786030Z env: 2025-12-04T17:17:52.3786259Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:52.3786571Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:52.3786940Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:52.3787580Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:52.3788163Z ##[endgroup] 2025-12-04T17:17:52.3805946Z ##[group]Run set -eou pipefail 2025-12-04T17:17:52.3806562Z set -eou pipefail 2025-12-04T17:17:52.3806871Z  2025-12-04T17:17:52.3807301Z echo "Holding runner for 2 hours until all ssh sessions have logged out" 2025-12-04T17:17:52.3807857Z for _ in $(seq 1440); do 2025-12-04T17:17:52.3808230Z  # Break if no ssh session exists anymore 2025-12-04T17:17:52.3808634Z  if [ "$(who)" = "" ]; then 2025-12-04T17:17:52.3809007Z  break 2025-12-04T17:17:52.3809278Z  fi 2025-12-04T17:17:52.3809531Z  echo "." 2025-12-04T17:17:52.3809814Z  sleep 5 2025-12-04T17:17:52.3810071Z done 2025-12-04T17:17:52.3816705Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T17:17:52.3817227Z env: 2025-12-04T17:17:52.3817460Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:52.3817765Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:52.3818123Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:52.3818763Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:52.3819346Z ##[endgroup] 2025-12-04T17:17:52.3848588Z Holding runner for 2 hours until all ssh sessions have logged out 2025-12-04T17:17:52.3928717Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T17:17:52.3929385Z # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T17:17:52.3929902Z # shellcheck disable=SC2046 2025-12-04T17:17:52.3930292Z docker stop $(docker ps -q) || true 2025-12-04T17:17:52.3930692Z # Prune all of the docker images 2025-12-04T17:17:52.3931079Z docker system prune -af 2025-12-04T17:17:52.3937616Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T17:17:52.3938040Z env: 2025-12-04T17:17:52.3938284Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:17:52.3938595Z HAS_NVIDIA_GPU: true 2025-12-04T17:17:52.3938942Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:17:52.3939590Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:17:52.3940172Z ##[endgroup] 2025-12-04T17:18:03.9204270Z 0b946b67a9e0 2025-12-04T17:18:08.3497377Z Deleted Containers: 2025-12-04T17:18:08.3497892Z 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:18:08.3498293Z 2025-12-04T17:18:16.0992467Z Deleted Images: 2025-12-04T17:18:16.0992930Z untagged: public.ecr.aws/docker/library/python:3.13 2025-12-04T17:18:16.0993787Z untagged: public.ecr.aws/docker/library/python@sha256:3f986299a7b8b44b0d8cf9bda2b22361ce5c3058ef5d7cb17fb7452506680ab0 2025-12-04T17:18:16.0994801Z deleted: sha256:44438aecfedf7b6086fce506dae0db5ba7fc0027f9b743f1a75a6b5cbc7de70a 2025-12-04T17:18:16.0995589Z deleted: sha256:6f09a1f5d8a107c2532fbd116e75116cb75fa77b1a7d72d3bdf1ac12de152acd 2025-12-04T17:18:16.0996570Z deleted: sha256:fe5f3ac0be086125eb1e3cd10cc33e8e426f4e079381f7ce5a987b626e99fa67 2025-12-04T17:18:16.0997336Z deleted: sha256:79dd2061a22cf919cfc4f1f02704bfda09afadb017265e670ee54441d296c06c 2025-12-04T17:18:16.0998122Z deleted: sha256:9447ad402aafdbee17e999b0ec84ad89c2646dbebf054d469d4f8bee77f66212 2025-12-04T17:18:16.0998890Z deleted: sha256:7a4909f3c1975be52292f53107495ee1b41c17494918767ccedf1cf1688ae318 2025-12-04T17:18:16.0999631Z deleted: sha256:3474923d97f1f498237650a7d51bd4aea37d5e6b9d8a778777920584af5dd560 2025-12-04T17:18:16.1000671Z deleted: sha256:683afd1773444401a9cbd24842ee5d9154a11abb4fab63ddea5c03df788597ee 2025-12-04T17:18:16.1001896Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-cuda12.4-cudnn9-py3-gcc11-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T17:18:16.1003567Z untagged: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image@sha256:ae30f11a5b50741bd652aa0c94ad89ef791c4e50157eff642748620825cf7940 2025-12-04T17:18:16.1004642Z deleted: sha256:5465aa79632b68f6240c23f0d0b021df4d0fd595333b61a40d36a0cf73656024 2025-12-04T17:18:16.1005380Z deleted: sha256:f57a578c46f36a858c2be92210a89558688ee36b619af78c698952c0e3ef05ad 2025-12-04T17:18:16.1006149Z deleted: sha256:ce0698bd1efc811ccead0ecdad944b4839bf17bff387495b58e64cf8db0e210c 2025-12-04T17:18:16.1006927Z deleted: sha256:f0ee66f328fa98c40f336c64fee9a4b42e51a793cceea7f81932068bdc7bd315 2025-12-04T17:18:16.1007684Z deleted: sha256:ea24b30a25c161bd4bd564bfd90c36d88674a1aa59ef3e65647e926c76685be0 2025-12-04T17:18:16.1008451Z deleted: sha256:15bc0847ce5e60cc1a9b36d25283dc5648fb45e04aa9a8dec984af3c193e2f0b 2025-12-04T17:18:16.1009210Z deleted: sha256:3639aa26691090ef45641c75bffcb2e3f427f5e282abc93d607de4433bf90488 2025-12-04T17:18:16.1009959Z deleted: sha256:86258272ba477934c917d08b21e0da6000c268b60f5a9ae907038e7bf3236532 2025-12-04T17:18:16.1010702Z deleted: sha256:ba8e0040c98ddbf87acbc3ae6575b2933c09421ac7094a96e027d1fc9356fbb6 2025-12-04T17:18:16.1011475Z deleted: sha256:ca0176fc0de6cc059c4dbfc313434b5dea2c90dc24f2dc3a1061b941c7b3e6ca 2025-12-04T17:18:16.1012238Z deleted: sha256:cc6a480ab9e6091c6c206bc9b340611b3863258975e835769bd8f2a38b5d8c13 2025-12-04T17:18:16.1012987Z deleted: sha256:8465c24f0b284d8589ea191edeb80d1da07e4a59dfcfdcfa153bdf3d5d678d3e 2025-12-04T17:18:16.1013843Z deleted: sha256:b93bfbd3b55899c606fb98c5edbd21fd63114862a4f5a5b67c7aa63fc9ada9a3 2025-12-04T17:18:16.1014606Z deleted: sha256:6b7582e3ce445d82e9d2ae7769502119c39c1edbf5fe11c195615db8da846931 2025-12-04T17:18:16.1015362Z deleted: sha256:9d79615a9d9ae67110cc9da697933492b385b1e4708d30c2211625bea5d42f27 2025-12-04T17:18:16.1016097Z deleted: sha256:7132c6db5e7d5692786167dfb22dea62d8203dc7837b2d1de435c6e5c85e906e 2025-12-04T17:18:16.1016848Z deleted: sha256:d61bc13a0957d633ff633186c6cbdf48da1c551991d814281262e58709e225a8 2025-12-04T17:18:16.1017678Z deleted: sha256:0c348bbc3988acd329b3e42de4d2c73d5dc4942618716ca312d389d4f704f4bb 2025-12-04T17:18:16.1018432Z deleted: sha256:28d30dd15686ab6819c2f03388c9999bbdaef35e8756817297d795e00dd623fc 2025-12-04T17:18:16.1019170Z deleted: sha256:0a57608df6cffb31a0b24f2537b4dfe7a55bbe6ea02216703cc3172062ab9d75 2025-12-04T17:18:16.1019934Z deleted: sha256:43d23f49f4d70a54b4aff6f4f10d5c5a3d75b100abbbf281ad510177cc80cd99 2025-12-04T17:18:16.1020700Z deleted: sha256:f9e33c2e4c7b8e7179fba052da4d7c4acdc8287f253c95328ae04055755f88a4 2025-12-04T17:18:16.1021454Z deleted: sha256:cfce0930cf33c7136fc92511b9bcad570958363b55e9e0c82e9b8ebc29301356 2025-12-04T17:18:16.1022212Z deleted: sha256:9a709ae20528f500f51271ad2ce6a3d7196fe814a28ae73881901ecef9748c2a 2025-12-04T17:18:16.1022960Z deleted: sha256:68a1d16e9392be6fe939a58c5f941a0919408b5852e52cb04027b0b8777e2b0e 2025-12-04T17:18:16.1023714Z deleted: sha256:042a0022b3eea78f54015f4cf2888bcfa3b91deb0b08830a33c2814b93285dd9 2025-12-04T17:18:16.1024463Z deleted: sha256:a7ba703ff0aa305a608f3b4afd89c2ecd0d1244b127629145a2e691490abb271 2025-12-04T17:18:16.1025238Z deleted: sha256:be44f5fbae55066faba60eebf7065a082abf517ab8f2ebf8ece69e74d45def07 2025-12-04T17:18:16.1026009Z deleted: sha256:a01f1b0d88a8936d648f78787f56579bdb6617edf4620d0410ab6b118351bbb2 2025-12-04T17:18:16.1026752Z deleted: sha256:dc93f45553adafb5c6e7473711c833996f6884dab2da708ffc76b5cf65b8db9d 2025-12-04T17:18:16.1027518Z deleted: sha256:ffdba9ecb5890a9cb23368d781ff5484270b7f13c6d5629feca3512b58b9a0ac 2025-12-04T17:18:16.1028269Z deleted: sha256:268a91c420865628895871795b524436f5cc4403aa53d71f457db21bf42dd530 2025-12-04T17:18:16.1029009Z deleted: sha256:72450bfd97986ccc53d8fa76252130b464fdb3c5fd8e688546e8c3ce0b9d4394 2025-12-04T17:18:16.1029865Z deleted: sha256:63954235d3be0420af6ad2dae2b24849e3eee1edb10cf86d29137c3e19621f47 2025-12-04T17:18:16.1030629Z deleted: sha256:1c4e2d3e68e8a166d1965962077fe194ea00cad2ee636399c0c17ba5a94bdb9c 2025-12-04T17:18:16.1031441Z deleted: sha256:361cacbab7154a0cb62486f57d75b112feedbcc751a7d8f7bb02ec7a61b1fe0d 2025-12-04T17:18:16.1032196Z deleted: sha256:e653f6af92265f4300717bd617aab954cfbf049d4be32e890e57c2e8135be7f9 2025-12-04T17:18:16.1032958Z deleted: sha256:bfffeb2974ffc58c0669724812f701df860257ac3d047a7315a100beb0ea0507 2025-12-04T17:18:16.1033727Z deleted: sha256:6ae48d8efc75420f721058928fe8b1ccf48aa1bdc92de539b1f0db9248a41fcf 2025-12-04T17:18:16.1034474Z deleted: sha256:535c7026785a690366fc69ecbc9a81f1b58a46f63c782620591c1297406a2731 2025-12-04T17:18:16.1035207Z deleted: sha256:8462076c3cc8db6030f38e1137bfbef1aad85404ed4231285c1e06cd414d3e57 2025-12-04T17:18:16.1035961Z deleted: sha256:fe340d63ccb66e5b395b7900c1002a513e4afd7f610e9df5e7262c4f71e93bef 2025-12-04T17:18:16.1036711Z deleted: sha256:b61085386114396fe42144a4aa739b2a0b45f0c30a083462a2ea7b9b675c02aa 2025-12-04T17:18:16.1037459Z deleted: sha256:7772f25c05bcd5ede631d287b826aa108db67c773e377db98ffa73b0917f3629 2025-12-04T17:18:16.1038204Z deleted: sha256:3ea8a43d8193d05ecd6aa473b523a3569e11ae691eed9e6ffd693f23b0106035 2025-12-04T17:18:16.1038956Z deleted: sha256:34647b4087d29cf48a18668bb935a95fc8b2dac3522c2581397f0f27227047fd 2025-12-04T17:18:16.1039705Z deleted: sha256:b6a169f1ab01281c16562ad43b462a1a47a33be8d3cfae0a117ffa5c47d0b532 2025-12-04T17:18:16.1040448Z deleted: sha256:664173a33cd21248a2d73d2eba7887602e36fbc96002d991eb0bd0a2d574ac88 2025-12-04T17:18:16.1041203Z deleted: sha256:d67fdfe94c9a0228f17991cd3e958e36da96d4d597b46773cb7eed98c489f947 2025-12-04T17:18:16.1041956Z deleted: sha256:f2be0722250908742f067756b56ed3fa169daa2f1c8201a7ed4335b2fed2cae5 2025-12-04T17:18:16.1042747Z deleted: sha256:8614db257d8dc9e0f0ee8398a4a4d3c061b2797d6017daaf0696dd7f87633b3e 2025-12-04T17:18:16.1043494Z deleted: sha256:23ee0908a1bf254f1d4dd0591cc0c6801571b4d93950b6fd4fee57ca7e361da0 2025-12-04T17:18:16.1044255Z deleted: sha256:f627a99df4c0f370bd7fc8ea6be7695d8027f988aed52b65233cbcf78b01989b 2025-12-04T17:18:16.1045009Z deleted: sha256:d5e92389b59d4134cdb96113af964186602e98c392e76a8f26d4ea6e54056ccc 2025-12-04T17:18:16.1045759Z deleted: sha256:cbfccf44b9dc670c109634fbf19c2bfff2a3d5243bfa351c851d9fad3f1acfc2 2025-12-04T17:18:16.1046524Z deleted: sha256:1242535e81ad4bd713910a6c5e1b38375b12ed1bcd1b48419813a5ef28a5c84c 2025-12-04T17:18:16.1047304Z deleted: sha256:10b1394079cfe756a1ad9aa9aa3a2995bd5e46ef1e18029eb9eae0398f6d4e88 2025-12-04T17:18:16.1048054Z deleted: sha256:1d32da9a5f10e10c4a97a839151a1943d4db18494e8080bea91a6c9784fde067 2025-12-04T17:18:16.1048798Z deleted: sha256:af2fd59653ebd685a032ef800f8227c0d7b9b0e5ef397b30d4301e001c943e8b 2025-12-04T17:18:16.1049695Z deleted: sha256:c48d351980e3bd24d533ae55d1acc6a27911dffcbb03b2ae552d7ccc3e4cd74f 2025-12-04T17:18:16.1050458Z deleted: sha256:e663afac609b1b6c812ab45265c27d870b92c9fc6849939f0b8635da83cbfb53 2025-12-04T17:18:16.1051217Z deleted: sha256:f79dc17668331d4214ef24000d5c54a0bb2ba70f152d8523f571e2b76a303f4f 2025-12-04T17:18:16.1051965Z deleted: sha256:00de9606a6cd2a2dfb4ceffcb076474d027a1f6273894677090aee7478035865 2025-12-04T17:18:16.1052729Z deleted: sha256:cf35fe1d0317253b75ee17c12783c2561faebf9bf2c59c07ad4712c053246586 2025-12-04T17:18:16.1053480Z deleted: sha256:06622801490739d9db884c23c05a31a1ee86c41e888b34c3ccef23d37f2bdbb5 2025-12-04T17:18:16.1054238Z deleted: sha256:df5dafcaee865ddfb66e22075c63769836e01a627d6fe46658b6f4b4a25318d3 2025-12-04T17:18:16.1054995Z deleted: sha256:7949ae5c4df921feb0e2cd7bac1e402e1ab9135e758fa41cd567880b354b40bc 2025-12-04T17:18:16.1055765Z deleted: sha256:9f19148d820adb1d6e86d0ce68e21fbcedafa7c7ec6c45c9004fa3a607096923 2025-12-04T17:18:16.1056545Z deleted: sha256:1d37d963e85ce22ffaab56a1cf35b3411f34f9432dc5e49ebbdf6f30816cdfa8 2025-12-04T17:18:16.1057370Z deleted: sha256:bac6d91e3830e51e96879deaa3e6d0d39da076fa802ebda68f81bdf7ef8342d5 2025-12-04T17:18:16.1058231Z deleted: sha256:ffd496b07151c90e7ddd68a81a36471f51a544187982db5e34621358e1b29681 2025-12-04T17:18:16.1058991Z deleted: sha256:890b2042bdb9e22a614cea1be88366cd3ae15159bf78ac510b9daa6f802493a6 2025-12-04T17:18:16.1059799Z deleted: sha256:ddd9a57b20a8b45ae0e8e350ec266d50a1b9e9a7ff4921470eb38f004d50eb20 2025-12-04T17:18:16.1060553Z deleted: sha256:2f4f91684b8221bc5cbc3f14c7e00bb693854027a1a6de5ad6bdcd000bb579f2 2025-12-04T17:18:16.1061326Z deleted: sha256:9c01ec5e73233284a0f9bb42de59696a1fa61caacacdf63d04df5ebd73895d77 2025-12-04T17:18:16.1062084Z deleted: sha256:f6153a90f0f5316b03f1464826325a1578231b89b3c1f1c83cc7cebdd41cee2a 2025-12-04T17:18:16.1062825Z deleted: sha256:4e89cd2181813af7fd2219923bae493e33111d8b4ebd76f257b7fb26744fda28 2025-12-04T17:18:16.1063768Z deleted: sha256:a0b77eb4054db8f2ea2ec957b3941b4aeee14b59e94a99a1521f90d6e41faf0e 2025-12-04T17:18:16.1064573Z deleted: sha256:1a1b2848f15aa5114f5a67e3705439512880bf1a7a6436cc67760c59b5f10c46 2025-12-04T17:18:16.1065315Z deleted: sha256:004fc01362840c164664c18580e479546fa0b7f9599487558f80190aec30e2b5 2025-12-04T17:18:16.1066047Z deleted: sha256:35f36e20799f0a0dead81bc3701732e43489264e6bee9fcb789b376a99e17e78 2025-12-04T17:18:16.1066829Z deleted: sha256:1207fd2ede86015c3f105620cb491e8199d2060a4a87490de358286d0ae52e4e 2025-12-04T17:18:16.1067594Z deleted: sha256:02dccb85ee744d1fbb819c6da618b2c52a3e4affc89e407f79b875e7b3bbb7df 2025-12-04T17:18:16.1068359Z deleted: sha256:d22e6ff9c3ac9dabbcc6052e1459f8dc4ebd19bd057bd0688615d6cc3ebb5cf0 2025-12-04T17:18:16.1069133Z deleted: sha256:73974f74b436f39a2fdb6461b1e3f7c3e41c73325776fa71d16b942a5b4a365b 2025-12-04T17:18:16.1069596Z 2025-12-04T17:18:16.1069735Z Total reclaimed space: 36.13GB 2025-12-04T17:18:16.1105032Z ##[group]Run set +e 2025-12-04T17:18:16.1105428Z set +e 2025-12-04T17:18:16.1105783Z set -x 2025-12-04T17:18:16.1106037Z  2025-12-04T17:18:16.1106282Z nvidia-smi 2025-12-04T17:18:16.1106809Z # NB: Surprisingly, nvidia-smi command returns successfully with return code 0 even in 2025-12-04T17:18:16.1107619Z # the case where the driver has already crashed as it still can get the driver version 2025-12-04T17:18:16.1108410Z # and some basic information like the bus ID. However, the rest of the information 2025-12-04T17:18:16.1109025Z # would be missing (ERR!), for example: 2025-12-04T17:18:16.1109389Z # 2025-12-04T17:18:16.1109800Z # +-----------------------------------------------------------------------------+ 2025-12-04T17:18:16.1110429Z # | NVIDIA-SMI 525.89.02 Driver Version: 525.89.02 CUDA Version: 12.0 | 2025-12-04T17:18:16.1111083Z # |-------------------------------+----------------------+----------------------+ 2025-12-04T17:18:16.1111718Z # | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T17:18:16.1112404Z # | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | 2025-12-04T17:18:16.1112974Z # | | | MIG M. | 2025-12-04T17:18:16.1113396Z # |===============================+======================+======================| 2025-12-04T17:18:16.1113891Z # | 0 ERR! Off | 00000000:00:1E.0 Off | ERR! | 2025-12-04T17:18:16.1114460Z # |ERR! ERR! ERR! ERR! / ERR! | 4184MiB / 23028MiB | ERR! Default | 2025-12-04T17:18:16.1114975Z # | | | ERR! | 2025-12-04T17:18:16.1115465Z # +-------------------------------+----------------------+----------------------+ 2025-12-04T17:18:16.1115913Z # 2025-12-04T17:18:16.1116266Z # +-----------------------------------------------------------------------------+ 2025-12-04T17:18:16.1116804Z # | Processes: | 2025-12-04T17:18:16.1117352Z # | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T17:18:16.1117879Z # | ID ID Usage | 2025-12-04T17:18:16.1118395Z # |=============================================================================| 2025-12-04T17:18:16.1118892Z # +-----------------------------------------------------------------------------+ 2025-12-04T17:18:16.1119350Z # 2025-12-04T17:18:16.1119807Z # This should be reported as a failure instead as it will guarantee to fail when 2025-12-04T17:18:16.1120402Z # Docker tries to run with --gpus all 2025-12-04T17:18:16.1120777Z # 2025-12-04T17:18:16.1121202Z # So, the correct check here is to query one of the missing piece of info like 2025-12-04T17:18:16.1121816Z # GPU name, so that the command can fail accordingly 2025-12-04T17:18:16.1122393Z nvidia-smi --query-gpu=gpu_name --format=csv,noheader --id=0 2025-12-04T17:18:16.1122905Z NVIDIA_SMI_STATUS=$? 2025-12-04T17:18:16.1123223Z  2025-12-04T17:18:16.1123725Z # These are acceptable return code from nvidia-smi as copied from setup-nvidia GitHub action 2025-12-04T17:18:16.1124514Z if [ "$NVIDIA_SMI_STATUS" -ne 0 ] && [ "$NVIDIA_SMI_STATUS" -ne 14 ]; then 2025-12-04T17:18:16.1125216Z  echo "NVIDIA driver installation has failed, shutting down the runner..." 2025-12-04T17:18:16.1125824Z  .github/scripts/stop_runner_service.sh 2025-12-04T17:18:16.1126196Z fi 2025-12-04T17:18:16.1126435Z  2025-12-04T17:18:16.1127002Z # For runner with multiple GPUs, we also want to confirm that the number of GPUs are the 2025-12-04T17:18:16.1127771Z # power of 2, i.e. 1, 2, 4, or 8. This is to avoid flaky test issue when one GPU fails 2025-12-04T17:18:16.1128402Z # https://github.com/pytorch/test-infra/issues/4000 2025-12-04T17:18:16.1128917Z GPU_COUNT=$(nvidia-smi --list-gpus | wc -l) 2025-12-04T17:18:16.1129336Z NVIDIA_SMI_STATUS=$? 2025-12-04T17:18:16.1129635Z  2025-12-04T17:18:16.1130144Z # These are acceptable return code from nvidia-smi as copied from setup-nvidia GitHub action 2025-12-04T17:18:16.1130907Z if [ "$NVIDIA_SMI_STATUS" -ne 0 ] && [ "$NVIDIA_SMI_STATUS" -ne 14 ]; then 2025-12-04T17:18:16.1131581Z  echo "NVIDIA driver installation has failed, shutting down the runner..." 2025-12-04T17:18:16.1132176Z  .github/scripts/stop_runner_service.sh 2025-12-04T17:18:16.1132553Z fi 2025-12-04T17:18:16.1132792Z  2025-12-04T17:18:16.1133061Z # Check the GPU count to be a power of 2 2025-12-04T17:18:16.1133719Z if [ "$GPU_COUNT" -le 8 ] && [ "$GPU_COUNT" -ne 1 ] && [ "$GPU_COUNT" -ne 2 ] && [ "$GPU_COUNT" -ne 4 ] && [ "$GPU_COUNT" -ne 8 ]; then 2025-12-04T17:18:16.1134611Z  echo "NVIDIA driver detects $GPU_COUNT GPUs. The runner has a broken GPU, shutting it down..." 2025-12-04T17:18:16.1135286Z  .github/scripts/stop_runner_service.sh 2025-12-04T17:18:16.1135662Z fi 2025-12-04T17:18:16.1146134Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T17:18:16.1146575Z env: 2025-12-04T17:18:16.1146825Z GIT_DEFAULT_BRANCH: main 2025-12-04T17:18:16.1147117Z HAS_NVIDIA_GPU: true 2025-12-04T17:18:16.1147477Z GPU_FLAG: --gpus all -e NVIDIA_DRIVER_CAPABILITIES=all 2025-12-04T17:18:16.1148127Z DOCKER_CONTAINER_ID: 0b946b67a9e0a55bf84a305b3a85e33c7dd885b0b95ae28802376c23a12269aa 2025-12-04T17:18:16.1148710Z ##[endgroup] 2025-12-04T17:18:16.1177879Z + nvidia-smi 2025-12-04T17:18:16.1383319Z Thu Dec 4 17:18:16 2025 2025-12-04T17:18:16.1383795Z +-----------------------------------------------------------------------------+ 2025-12-04T17:18:16.1384401Z | NVIDIA-SMI 525.105.17 Driver Version: 525.105.17 CUDA Version: 12.0 | 2025-12-04T17:18:16.1384972Z |-------------------------------+----------------------+----------------------+ 2025-12-04T17:18:16.1385565Z | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | 2025-12-04T17:18:16.1386335Z | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | 2025-12-04T17:18:16.1386869Z | | | MIG M. | 2025-12-04T17:18:16.1387263Z |===============================+======================+======================| 2025-12-04T17:18:16.1549173Z | 0 Tesla T4 On | 00000000:00:1E.0 Off | 0 | 2025-12-04T17:18:16.1549723Z | N/A 24C P8 14W / 70W | 2MiB / 15360MiB | 0% Default | 2025-12-04T17:18:16.1550173Z | | | N/A | 2025-12-04T17:18:16.1550646Z +-------------------------------+----------------------+----------------------+ 2025-12-04T17:18:16.1551114Z 2025-12-04T17:18:16.1551570Z +-----------------------------------------------------------------------------+ 2025-12-04T17:18:16.1552058Z | Processes: | 2025-12-04T17:18:16.1552582Z | GPU GI CI PID Type Process name GPU Memory | 2025-12-04T17:18:16.1553064Z | ID ID Usage | 2025-12-04T17:18:16.1553475Z |=============================================================================| 2025-12-04T17:18:16.1554497Z | No running processes found | 2025-12-04T17:18:16.1555072Z +-----------------------------------------------------------------------------+ 2025-12-04T17:18:16.2407247Z + nvidia-smi --query-gpu=gpu_name --format=csv,noheader --id=0 2025-12-04T17:18:16.2583494Z Tesla T4 2025-12-04T17:18:16.2615985Z + NVIDIA_SMI_STATUS=0 2025-12-04T17:18:16.2616715Z + '[' 0 -ne 0 ']' 2025-12-04T17:18:16.2622825Z ++ nvidia-smi --list-gpus 2025-12-04T17:18:16.2623967Z ++ wc -l 2025-12-04T17:18:16.2835028Z + GPU_COUNT=1 2025-12-04T17:18:16.2835374Z + NVIDIA_SMI_STATUS=0 2025-12-04T17:18:16.2835694Z + '[' 0 -ne 0 ']' 2025-12-04T17:18:16.2835938Z + '[' 1 -le 8 ']' 2025-12-04T17:18:16.2836190Z + '[' 1 -ne 1 ']' 2025-12-04T17:18:16.2911467Z Post job cleanup. 2025-12-04T17:18:16.2997567Z Post job cleanup. 2025-12-04T17:18:16.3065928Z Post job cleanup. 2025-12-04T17:18:16.4229943Z [command]/usr/bin/git version 2025-12-04T17:18:16.4289709Z git version 2.50.1 2025-12-04T17:18:16.4329790Z Copying '/home/ec2-user/.gitconfig' to '/home/ec2-user/actions-runner/_work/_temp/d5f2a111-2a10-4525-93de-d6a441d84dfe/.gitconfig' 2025-12-04T17:18:16.4340060Z Temporarily overriding HOME='/home/ec2-user/actions-runner/_work/_temp/d5f2a111-2a10-4525-93de-d6a441d84dfe' before making global git config changes 2025-12-04T17:18:16.4341205Z Adding repository directory to the temporary git global config as a safe directory 2025-12-04T17:18:16.4345362Z [command]/usr/bin/git config --global --add safe.directory /home/ec2-user/actions-runner/_work/pytorch/pytorch 2025-12-04T17:18:16.4387170Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-12-04T17:18:16.4430792Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-12-04T17:18:16.4784328Z Entering 'android/libs/fbjni' 2025-12-04T17:18:16.4849399Z Entering 'third_party/FP16' 2025-12-04T17:18:16.4920636Z Entering 'third_party/FXdiv' 2025-12-04T17:18:16.4984202Z Entering 'third_party/NNPACK' 2025-12-04T17:18:16.5049632Z Entering 'third_party/NVTX' 2025-12-04T17:18:16.5115944Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T17:18:16.5181689Z Entering 'third_party/XNNPACK' 2025-12-04T17:18:16.5263925Z Entering 'third_party/aiter' 2025-12-04T17:18:16.5329472Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T17:18:16.5407433Z Entering 'third_party/benchmark' 2025-12-04T17:18:16.5471399Z Entering 'third_party/composable_kernel' 2025-12-04T17:18:16.5545963Z Entering 'third_party/cpp-httplib' 2025-12-04T17:18:16.5610350Z Entering 'third_party/cpuinfo' 2025-12-04T17:18:16.5675217Z Entering 'third_party/cudnn_frontend' 2025-12-04T17:18:16.5739689Z Entering 'third_party/cutlass' 2025-12-04T17:18:16.5814311Z Entering 'third_party/fbgemm' 2025-12-04T17:18:16.5880110Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T17:18:16.5948785Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T17:18:16.6020882Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T17:18:16.6084675Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T17:18:16.6157703Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T17:18:16.6221709Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T17:18:16.6284903Z Entering 'third_party/fbgemm/external/json' 2025-12-04T17:18:16.6351499Z Entering 'third_party/flash-attention' 2025-12-04T17:18:16.6415217Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T17:18:16.6485435Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T17:18:16.6557852Z Entering 'third_party/flatbuffers' 2025-12-04T17:18:16.6626856Z Entering 'third_party/fmt' 2025-12-04T17:18:16.6690196Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T17:18:16.6756339Z Entering 'third_party/gloo' 2025-12-04T17:18:16.6819634Z Entering 'third_party/googletest' 2025-12-04T17:18:16.6882753Z Entering 'third_party/ideep' 2025-12-04T17:18:16.6944558Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T17:18:16.7016629Z Entering 'third_party/ittapi' 2025-12-04T17:18:16.7080275Z Entering 'third_party/kineto' 2025-12-04T17:18:16.7143163Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T17:18:16.7207033Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T17:18:16.7271882Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T17:18:16.7335073Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T17:18:16.7398809Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T17:18:16.7460452Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T17:18:16.7525860Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T17:18:16.7591126Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T17:18:16.7655143Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T17:18:16.7719407Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T17:18:16.7782789Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T17:18:16.7848936Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T17:18:16.7915701Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T17:18:16.7984483Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T17:18:16.8052414Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T17:18:16.8116817Z Entering 'third_party/kleidiai' 2025-12-04T17:18:16.8185063Z Entering 'third_party/mimalloc' 2025-12-04T17:18:16.8249585Z Entering 'third_party/nlohmann' 2025-12-04T17:18:16.8315106Z Entering 'third_party/onnx' 2025-12-04T17:18:16.8400265Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T17:18:16.8467436Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T17:18:16.8535064Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T17:18:16.8598999Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T17:18:16.8662060Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T17:18:16.8722936Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T17:18:16.9145154Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T17:18:16.9209627Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T17:18:16.9274474Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T17:18:16.9336670Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T17:18:16.9402133Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T17:18:16.9467055Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T17:18:16.9554084Z Entering 'third_party/pocketfft' 2025-12-04T17:18:16.9618403Z Entering 'third_party/protobuf' 2025-12-04T17:18:16.9685876Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T17:18:16.9747502Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T17:18:16.9814419Z Entering 'third_party/psimd' 2025-12-04T17:18:16.9877091Z Entering 'third_party/pthreadpool' 2025-12-04T17:18:16.9941477Z Entering 'third_party/pybind11' 2025-12-04T17:18:17.0007335Z Entering 'third_party/python-peachpy' 2025-12-04T17:18:17.0070323Z Entering 'third_party/sleef' 2025-12-04T17:18:17.0134103Z Entering 'third_party/tensorpipe' 2025-12-04T17:18:17.0198621Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T17:18:17.0260728Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T17:18:17.0323069Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T17:18:17.0386260Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T17:18:17.0448773Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T17:18:17.0540119Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-12-04T17:18:17.0564190Z http.https://github.com/.extraheader 2025-12-04T17:18:17.0575058Z [command]/usr/bin/git config --local --unset-all http.https://github.com/.extraheader 2025-12-04T17:18:17.0608136Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-12-04T17:18:17.0946238Z Entering 'android/libs/fbjni' 2025-12-04T17:18:17.0990548Z http.https://github.com/.extraheader 2025-12-04T17:18:17.1029858Z Entering 'third_party/FP16' 2025-12-04T17:18:17.1073588Z http.https://github.com/.extraheader 2025-12-04T17:18:17.1113327Z Entering 'third_party/FXdiv' 2025-12-04T17:18:17.1156313Z http.https://github.com/.extraheader 2025-12-04T17:18:17.1195656Z Entering 'third_party/NNPACK' 2025-12-04T17:18:17.1239098Z http.https://github.com/.extraheader 2025-12-04T17:18:17.1277025Z Entering 'third_party/NVTX' 2025-12-04T17:18:17.1321589Z http.https://github.com/.extraheader 2025-12-04T17:18:17.1362415Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T17:18:17.1406277Z http.https://github.com/.extraheader 2025-12-04T17:18:17.1445018Z Entering 'third_party/XNNPACK' 2025-12-04T17:18:17.1488756Z http.https://github.com/.extraheader 2025-12-04T17:18:17.1544603Z Entering 'third_party/aiter' 2025-12-04T17:18:17.1588012Z http.https://github.com/.extraheader 2025-12-04T17:18:17.1627818Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T17:18:17.1670315Z http.https://github.com/.extraheader 2025-12-04T17:18:17.1722527Z Entering 'third_party/benchmark' 2025-12-04T17:18:17.1765904Z http.https://github.com/.extraheader 2025-12-04T17:18:17.1805380Z Entering 'third_party/composable_kernel' 2025-12-04T17:18:17.1848174Z http.https://github.com/.extraheader 2025-12-04T17:18:17.1897187Z Entering 'third_party/cpp-httplib' 2025-12-04T17:18:17.1940112Z http.https://github.com/.extraheader 2025-12-04T17:18:17.1979475Z Entering 'third_party/cpuinfo' 2025-12-04T17:18:17.2023854Z http.https://github.com/.extraheader 2025-12-04T17:18:17.2062537Z Entering 'third_party/cudnn_frontend' 2025-12-04T17:18:17.2107391Z http.https://github.com/.extraheader 2025-12-04T17:18:17.2145040Z Entering 'third_party/cutlass' 2025-12-04T17:18:17.2187724Z http.https://github.com/.extraheader 2025-12-04T17:18:17.2240476Z Entering 'third_party/fbgemm' 2025-12-04T17:18:17.2284281Z http.https://github.com/.extraheader 2025-12-04T17:18:17.2326408Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T17:18:17.2369139Z http.https://github.com/.extraheader 2025-12-04T17:18:17.2407593Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T17:18:17.2449490Z http.https://github.com/.extraheader 2025-12-04T17:18:17.2498433Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T17:18:17.2542819Z http.https://github.com/.extraheader 2025-12-04T17:18:17.2581499Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T17:18:17.2625347Z http.https://github.com/.extraheader 2025-12-04T17:18:17.2673556Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T17:18:17.2715942Z http.https://github.com/.extraheader 2025-12-04T17:18:17.2753311Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T17:18:17.2803115Z http.https://github.com/.extraheader 2025-12-04T17:18:17.2841886Z Entering 'third_party/fbgemm/external/json' 2025-12-04T17:18:17.2884240Z http.https://github.com/.extraheader 2025-12-04T17:18:17.2925739Z Entering 'third_party/flash-attention' 2025-12-04T17:18:17.2970294Z http.https://github.com/.extraheader 2025-12-04T17:18:17.3010054Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T17:18:17.3055239Z http.https://github.com/.extraheader 2025-12-04T17:18:17.3102269Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T17:18:17.3143929Z http.https://github.com/.extraheader 2025-12-04T17:18:17.3193439Z Entering 'third_party/flatbuffers' 2025-12-04T17:18:17.3238609Z http.https://github.com/.extraheader 2025-12-04T17:18:17.3279295Z Entering 'third_party/fmt' 2025-12-04T17:18:17.3324156Z http.https://github.com/.extraheader 2025-12-04T17:18:17.3364014Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T17:18:17.3408327Z http.https://github.com/.extraheader 2025-12-04T17:18:17.3446773Z Entering 'third_party/gloo' 2025-12-04T17:18:17.3490383Z http.https://github.com/.extraheader 2025-12-04T17:18:17.3529239Z Entering 'third_party/googletest' 2025-12-04T17:18:17.3572792Z http.https://github.com/.extraheader 2025-12-04T17:18:17.3612729Z Entering 'third_party/ideep' 2025-12-04T17:18:17.3654658Z http.https://github.com/.extraheader 2025-12-04T17:18:17.3691670Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T17:18:17.3734648Z http.https://github.com/.extraheader 2025-12-04T17:18:17.3782074Z Entering 'third_party/ittapi' 2025-12-04T17:18:17.3828356Z http.https://github.com/.extraheader 2025-12-04T17:18:17.3867014Z Entering 'third_party/kineto' 2025-12-04T17:18:17.3911025Z http.https://github.com/.extraheader 2025-12-04T17:18:17.3949076Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T17:18:17.3992051Z http.https://github.com/.extraheader 2025-12-04T17:18:17.4031029Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T17:18:17.4073780Z http.https://github.com/.extraheader 2025-12-04T17:18:17.4114468Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T17:18:17.4158296Z http.https://github.com/.extraheader 2025-12-04T17:18:17.4197798Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T17:18:17.4238473Z http.https://github.com/.extraheader 2025-12-04T17:18:17.4276297Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T17:18:17.4319054Z http.https://github.com/.extraheader 2025-12-04T17:18:17.4355927Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T17:18:17.4400535Z http.https://github.com/.extraheader 2025-12-04T17:18:17.4444609Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T17:18:17.4488922Z http.https://github.com/.extraheader 2025-12-04T17:18:17.4527379Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T17:18:17.4570117Z http.https://github.com/.extraheader 2025-12-04T17:18:17.4609482Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T17:18:17.4652844Z http.https://github.com/.extraheader 2025-12-04T17:18:17.4692906Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T17:18:17.4736226Z http.https://github.com/.extraheader 2025-12-04T17:18:17.4774095Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T17:18:17.4817922Z http.https://github.com/.extraheader 2025-12-04T17:18:17.4855854Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T17:18:17.4902247Z http.https://github.com/.extraheader 2025-12-04T17:18:17.4947566Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T17:18:17.4991885Z http.https://github.com/.extraheader 2025-12-04T17:18:17.5036479Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T17:18:17.5081220Z http.https://github.com/.extraheader 2025-12-04T17:18:17.5119487Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T17:18:17.5162121Z http.https://github.com/.extraheader 2025-12-04T17:18:17.5203946Z Entering 'third_party/kleidiai' 2025-12-04T17:18:17.5248199Z http.https://github.com/.extraheader 2025-12-04T17:18:17.5287220Z Entering 'third_party/mimalloc' 2025-12-04T17:18:17.5331524Z http.https://github.com/.extraheader 2025-12-04T17:18:17.5369896Z Entering 'third_party/nlohmann' 2025-12-04T17:18:17.5414699Z http.https://github.com/.extraheader 2025-12-04T17:18:17.5453951Z Entering 'third_party/onnx' 2025-12-04T17:18:17.5499155Z http.https://github.com/.extraheader 2025-12-04T17:18:17.5556687Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T17:18:17.5599333Z http.https://github.com/.extraheader 2025-12-04T17:18:17.5640719Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T17:18:17.5684027Z http.https://github.com/.extraheader 2025-12-04T17:18:17.5724549Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T17:18:17.5766381Z http.https://github.com/.extraheader 2025-12-04T17:18:17.5804996Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T17:18:17.5846539Z http.https://github.com/.extraheader 2025-12-04T17:18:17.5884938Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T17:18:17.5927825Z http.https://github.com/.extraheader 2025-12-04T17:18:17.5964748Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T17:18:17.6008370Z http.https://github.com/.extraheader 2025-12-04T17:18:17.6046907Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T17:18:17.6088343Z http.https://github.com/.extraheader 2025-12-04T17:18:17.6126794Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T17:18:17.6169732Z http.https://github.com/.extraheader 2025-12-04T17:18:17.6208502Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T17:18:17.6251701Z http.https://github.com/.extraheader 2025-12-04T17:18:17.6287803Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T17:18:17.6331625Z http.https://github.com/.extraheader 2025-12-04T17:18:17.6372011Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T17:18:17.6414818Z http.https://github.com/.extraheader 2025-12-04T17:18:17.6454143Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T17:18:17.6497526Z http.https://github.com/.extraheader 2025-12-04T17:18:17.6560436Z Entering 'third_party/pocketfft' 2025-12-04T17:18:17.6606928Z http.https://github.com/.extraheader 2025-12-04T17:18:17.6645265Z Entering 'third_party/protobuf' 2025-12-04T17:18:17.6688497Z http.https://github.com/.extraheader 2025-12-04T17:18:17.6731134Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T17:18:17.6773312Z http.https://github.com/.extraheader 2025-12-04T17:18:17.6811502Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T17:18:17.6853606Z http.https://github.com/.extraheader 2025-12-04T17:18:17.6894666Z Entering 'third_party/psimd' 2025-12-04T17:18:17.6939608Z http.https://github.com/.extraheader 2025-12-04T17:18:17.6977235Z Entering 'third_party/pthreadpool' 2025-12-04T17:18:17.7021070Z http.https://github.com/.extraheader 2025-12-04T17:18:17.7059061Z Entering 'third_party/pybind11' 2025-12-04T17:18:17.7102622Z http.https://github.com/.extraheader 2025-12-04T17:18:17.7141728Z Entering 'third_party/python-peachpy' 2025-12-04T17:18:17.7186457Z http.https://github.com/.extraheader 2025-12-04T17:18:17.7224353Z Entering 'third_party/sleef' 2025-12-04T17:18:17.7267103Z http.https://github.com/.extraheader 2025-12-04T17:18:17.7306056Z Entering 'third_party/tensorpipe' 2025-12-04T17:18:17.7348660Z http.https://github.com/.extraheader 2025-12-04T17:18:17.7386674Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T17:18:17.7429798Z http.https://github.com/.extraheader 2025-12-04T17:18:17.7467349Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T17:18:17.7510557Z http.https://github.com/.extraheader 2025-12-04T17:18:17.7548718Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T17:18:17.7590287Z http.https://github.com/.extraheader 2025-12-04T17:18:17.7628934Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T17:18:17.7675701Z http.https://github.com/.extraheader 2025-12-04T17:18:17.7712597Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T17:18:17.7755066Z http.https://github.com/.extraheader 2025-12-04T17:18:17.7823787Z [command]/usr/bin/git config --local --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:17.7868058Z [command]/usr/bin/git submodule foreach --recursive git config --local --show-origin --name-only --get-regexp remote.origin.url 2025-12-04T17:18:17.8215795Z Entering 'android/libs/fbjni' 2025-12-04T17:18:17.8244166Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-12-04T17:18:17.8263746Z Entering 'third_party/FP16' 2025-12-04T17:18:17.8292983Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-12-04T17:18:17.8312183Z Entering 'third_party/FXdiv' 2025-12-04T17:18:17.8342436Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-12-04T17:18:17.8367434Z Entering 'third_party/NNPACK' 2025-12-04T17:18:17.8398481Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-12-04T17:18:17.8417285Z Entering 'third_party/NVTX' 2025-12-04T17:18:17.8446665Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-12-04T17:18:17.8467895Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T17:18:17.8497585Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-12-04T17:18:17.8516274Z Entering 'third_party/XNNPACK' 2025-12-04T17:18:17.8545476Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-12-04T17:18:17.8582251Z Entering 'third_party/aiter' 2025-12-04T17:18:17.8612517Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-12-04T17:18:17.8632454Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T17:18:17.8661435Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-12-04T17:18:17.8689864Z Entering 'third_party/benchmark' 2025-12-04T17:18:17.8719656Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-12-04T17:18:17.8738760Z Entering 'third_party/composable_kernel' 2025-12-04T17:18:17.8767419Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-12-04T17:18:17.8797282Z Entering 'third_party/cpp-httplib' 2025-12-04T17:18:17.8825947Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-12-04T17:18:17.8848089Z Entering 'third_party/cpuinfo' 2025-12-04T17:18:17.8878523Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-12-04T17:18:17.8898001Z Entering 'third_party/cudnn_frontend' 2025-12-04T17:18:17.8926095Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-12-04T17:18:17.8945632Z Entering 'third_party/cutlass' 2025-12-04T17:18:17.8974754Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-12-04T17:18:17.9004828Z Entering 'third_party/fbgemm' 2025-12-04T17:18:17.9033760Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-12-04T17:18:17.9054933Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T17:18:17.9083288Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-12-04T17:18:17.9101808Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T17:18:17.9131725Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-12-04T17:18:17.9158988Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T17:18:17.9187692Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-12-04T17:18:17.9205854Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T17:18:17.9234302Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-12-04T17:18:17.9261490Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T17:18:17.9288726Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-12-04T17:18:17.9306847Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T17:18:17.9335679Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-12-04T17:18:17.9353214Z Entering 'third_party/fbgemm/external/json' 2025-12-04T17:18:17.9382233Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-12-04T17:18:17.9404195Z Entering 'third_party/flash-attention' 2025-12-04T17:18:17.9434782Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-12-04T17:18:17.9454062Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T17:18:17.9482648Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-12-04T17:18:17.9509001Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T17:18:17.9537069Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-12-04T17:18:17.9566499Z Entering 'third_party/flatbuffers' 2025-12-04T17:18:17.9595288Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-12-04T17:18:17.9618741Z Entering 'third_party/fmt' 2025-12-04T17:18:17.9647718Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-12-04T17:18:17.9667184Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T17:18:17.9697416Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-12-04T17:18:17.9716483Z Entering 'third_party/gloo' 2025-12-04T17:18:17.9745439Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-12-04T17:18:17.9765750Z Entering 'third_party/googletest' 2025-12-04T17:18:17.9795383Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-12-04T17:18:17.9814521Z Entering 'third_party/ideep' 2025-12-04T17:18:17.9843580Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-12-04T17:18:17.9862008Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T17:18:17.9890438Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-12-04T17:18:17.9918654Z Entering 'third_party/ittapi' 2025-12-04T17:18:17.9947968Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-12-04T17:18:17.9966853Z Entering 'third_party/kineto' 2025-12-04T17:18:17.9996325Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-12-04T17:18:18.0014858Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T17:18:18.0043428Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-12-04T17:18:18.0061382Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T17:18:18.0090570Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-12-04T17:18:18.0110524Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T17:18:18.0139344Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-12-04T17:18:18.0158568Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T17:18:18.0187233Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-12-04T17:18:18.0206073Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T17:18:18.0234979Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-12-04T17:18:18.0251970Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T17:18:18.0281557Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-12-04T17:18:18.0302223Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T17:18:18.0330560Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-12-04T17:18:18.0349372Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T17:18:18.0378598Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-12-04T17:18:18.0397063Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T17:18:18.0425917Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-12-04T17:18:18.0445295Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T17:18:18.0474400Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-12-04T17:18:18.0492467Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T17:18:18.0522221Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T17:18:18.0539719Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T17:18:18.0568045Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T17:18:18.0588589Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T17:18:18.0618181Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T17:18:18.0641025Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T17:18:18.0671290Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-12-04T17:18:18.0688836Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T17:18:18.0719274Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-12-04T17:18:18.0740321Z Entering 'third_party/kleidiai' 2025-12-04T17:18:18.0769984Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-12-04T17:18:18.0789709Z Entering 'third_party/mimalloc' 2025-12-04T17:18:18.0819132Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-12-04T17:18:18.0838192Z Entering 'third_party/nlohmann' 2025-12-04T17:18:18.0867814Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-12-04T17:18:18.0889674Z Entering 'third_party/onnx' 2025-12-04T17:18:18.0918684Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-12-04T17:18:18.0958904Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T17:18:18.0988739Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-12-04T17:18:18.1009969Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T17:18:18.1039085Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-12-04T17:18:18.1059072Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T17:18:18.1086516Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-12-04T17:18:18.1105841Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T17:18:18.1133922Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-12-04T17:18:18.1152569Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T17:18:18.1180332Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-12-04T17:18:18.1198747Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T17:18:18.1226532Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-12-04T17:18:18.1245925Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T17:18:18.1274409Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-12-04T17:18:18.1291876Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T17:18:18.1320416Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-12-04T17:18:18.1337567Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T17:18:18.1366221Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T17:18:18.1383323Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T17:18:18.1412430Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T17:18:18.1432254Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T17:18:18.1461853Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T17:18:18.1482204Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T17:18:18.1510966Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-12-04T17:18:18.1551913Z Entering 'third_party/pocketfft' 2025-12-04T17:18:18.1581768Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-12-04T17:18:18.1601297Z Entering 'third_party/protobuf' 2025-12-04T17:18:18.1630318Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-12-04T17:18:18.1652202Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T17:18:18.1680509Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-12-04T17:18:18.1700094Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T17:18:18.1727658Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-12-04T17:18:18.1747968Z Entering 'third_party/psimd' 2025-12-04T17:18:18.1777164Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-12-04T17:18:18.1795573Z Entering 'third_party/pthreadpool' 2025-12-04T17:18:18.1825686Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-12-04T17:18:18.1844070Z Entering 'third_party/pybind11' 2025-12-04T17:18:18.1872674Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-12-04T17:18:18.1891610Z Entering 'third_party/python-peachpy' 2025-12-04T17:18:18.1924241Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-12-04T17:18:18.1942851Z Entering 'third_party/sleef' 2025-12-04T17:18:18.1971803Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-12-04T17:18:18.1991294Z Entering 'third_party/tensorpipe' 2025-12-04T17:18:18.2020172Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-12-04T17:18:18.2037995Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T17:18:18.2069702Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-12-04T17:18:18.2087206Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T17:18:18.2116172Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-12-04T17:18:18.2133955Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T17:18:18.2162462Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-12-04T17:18:18.2181366Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T17:18:18.2210133Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-12-04T17:18:18.2226919Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T17:18:18.2255369Z file:/home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-12-04T17:18:18.2296330Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2327462Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2355073Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2382200Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2408248Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2435844Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2463870Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2489119Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2515975Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2542144Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2568589Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2596940Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2623790Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2649599Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2676260Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2703031Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2729528Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2756173Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2781891Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2808233Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2835625Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2863017Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2890966Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2916998Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2943084Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2967706Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.2992603Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3018166Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3046122Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3072245Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3098755Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3125318Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3151665Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3178847Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3204727Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3231322Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3257369Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3283646Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3310063Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3338256Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3364735Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3390771Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3417850Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3444153Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3470542Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3496770Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3534157Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3561630Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3587732Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3613927Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3639218Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3664646Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3689647Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3715051Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3741498Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3766698Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3792043Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3817330Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3843294Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3869171Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3894193Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3919975Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3946376Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3971927Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.3997868Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.4024573Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.4050682Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.4076784Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.4108738Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.4137425Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.4165662Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.4191446Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.4220978Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.4246894Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.4272499Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.4298748Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.4324418Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.4350491Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.4375675Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.4402332Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.4428206Z [command]/usr/bin/git config --file /home/ec2-user/actions-runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T17:18:18.4556710Z A job completed hook has been configured by the self-hosted runner administrator 2025-12-04T17:18:18.4579766Z ##[group]Run '/home/ec2-user/runner-scripts/after_job.sh' 2025-12-04T17:18:18.4585987Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T17:18:18.4586426Z ##[endgroup] 2025-12-04T17:18:26.9175302Z Cleaning up orphan processes